2023
DOI: 10.48550/arxiv.2302.07248
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generation Probabilities Are Not Enough: Exploring the Effectiveness of Uncertainty Highlighting in AI-Powered Code Completions

Abstract: Large-scale generative models enabled the development of AI-powered code completion tools to assist programmers in writing code. However, much like other AI-powered tools, AI-powered code completions are not always accurate, potentially introducing bugs or even security vulnerabilities into code if not properly detected and corrected by a human programmer. One technique that has been proposed and implemented to help programmers identify potential errors is to highlight uncertain tokens. However, there have bee… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 33 publications
0
5
0
Order By: Relevance
“…What is a useful notion of uncertainty for LLMs? While LLMs have a notion of uncertainty baked into them -the likelihood that the model would generate a specific token given its preceding or surrounding context (Bengio et al, 2003), what we have referred to in past work as the generation probability (Vasconcelos et al, 2023)-whether this notion would be useful to different stakeholders is questionable. In particular, this notion may not line up with people's intuition about what it means for the model to be uncertain.…”
Section: Communicating Uncertaintymentioning
confidence: 99%
See 1 more Smart Citation
“…What is a useful notion of uncertainty for LLMs? While LLMs have a notion of uncertainty baked into them -the likelihood that the model would generate a specific token given its preceding or surrounding context (Bengio et al, 2003), what we have referred to in past work as the generation probability (Vasconcelos et al, 2023)-whether this notion would be useful to different stakeholders is questionable. In particular, this notion may not line up with people's intuition about what it means for the model to be uncertain.…”
Section: Communicating Uncertaintymentioning
confidence: 99%
“…Carefully selecting a notion of uncertainty to convey to stakeholders matters because the particular notion used impacts their behavior and trust. In our recent work with collaborators (Vasconcelos et al, 2023) provided in (e.g., its precision and modality), and what the effect is (e.g., on trust or behaviors), as well as taking into consideration the characteristics of the receiver (Van Der Bles et al, 2019). For example, in our study on uncertainty in the context of code completion tools (Vasconcelos et al, 2023), by soliciting participants' feedback on different uncertainty communication design choices, we found that programmers prefer uncertainty about granular or meaningful blocks to guide them to make token-level changes and prefer less precise communication (as opposed to exact quantification) for easy processing-both ultimately supporting their goal of producing correct code efficiently.…”
Section: Communicating Uncertaintymentioning
confidence: 99%
“…What is a useful notion of uncertainty for LLMs? While LLMs have a notion of uncertainty baked in them-the likelihood that the model would generate a specific token given its preceding or surrounding context [17], what we have referred to in past work as the generation probability [178]whether this notion would be useful to different stakeholders is questionable. In particular, this notion may not line up with people's intuition about what it means for the model to be uncertain.…”
Section: Communicating Uncertaintymentioning
confidence: 99%
“…Carefully selecting a notion of uncertainty to convey to stakeholders matters because the particular notion used impacts their behavior and trust. In our recent work with collaborators [178], we explored the effectiveness of displaying two alternative notions of uncertainty to programmers interacting with an LLM-powered code completion tool. In a mixed-methods study with 30 programmers, we compared three conditions: providing a code completion alone, highlighting those tokens with the lowest likelihood of being generated by the underlying LLM (i.e., lowest generation probability), and highlighting tokens with the highest predicted likelihood of being edited by a programmer according to a separate "edit model" trained on logged data from past programmer interactions.…”
Section: Communicating Uncertaintymentioning
confidence: 99%
See 1 more Smart Citation