You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
After reviewing your source code, I noticed that when you generate your prompt using the script “make_prompt.py,” you only include the function name, the preceding context (located within the same file as the function to be completed), and the following context (also within the same file). However, I see that you’ve already gathered additional information such as the dependencies (“intra_class”, “intra_file”, and “cross_file”) in a separate “data.jsonl” file. My question is: Why don’t you include this dependency information in the prompt to give the LLM hints about which external elements might be needed to complete the code? Even if you prefer not to directly provide the LLM with the “ground truth” dependencies, you could implement a retrieval module to bring in relevant information from other parts of the project.
From what I can tell, your approach focuses primarily on local context (within the same file) without considering dependencies across the repository. This makes it seem like there isn’t a significant difference between your so-called “Repository-Level Code Completion” and the more traditional “Function-Level Code Completion/Generation.” I would appreciate it if you could clarify this distinction. Thanks in advance.
lambdaofgod, EvensXia, YaLing-Luo0, lihaoling, pan2013e and 2 more