-
-
Notifications
You must be signed in to change notification settings - Fork 78
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Send multiple files into the chat #206
Comments
- starting with @context_file (issue: #206)
Take a look at this branch of my fork. I adapted the work done in this PR: #174 basically, I felt the PR went too far and overcomplicated itself with by trying to be able to reference a specific function in a code file. A truly amazing feat they accomplished it, but definitely not something I personally need. I took the basic template of their I also added lastly I added an The other thing I did different was make it so that the the imported text always appears IN the message where the command is present AND at the top of the message. so it's always necessary to say 'in the context above', rather than 'in the following context`, although I'm not sure if there are any studies that show which strategy leads to better results, I personally think it generally makes no difference to the LLM where the context is, so long as it is included so it can be referenced. And the current strategy leads to simpler code. hope this helps! (I will try to keep my fork continually updated with the main repo, but since it's the main branch I use, I may update it with other features occasionally as well, since I'm lazy and don't feel like keeping the different features on separate branches and then merging them into a single production branch) edit: oh and I forgot to mention that I made it so that each |
I should also note that this pr is adding 'macro' support, which looks like is going to be an extendable feature where we can define new macros easily, and one of the first macros being developed for the system is |
At least with programming, the file being discussed with the LLM and/or other related files might immediately have some changes applied to it as the result of the discussion. This means, if you're not saving and storing the referenced file away in a separate location (like the 'artifacts' directory discussed some time ago), attempts to recreate the exact state of the previous messages using the current state of the files will likely fail. The files could have changed drastically since the time the old message was issued, the file might have been moved, or even deleted altogether. |
- starting with @context_file (issue: #206)
@UreshiiPanda Just use GpChatPaste multiple times. Configure a good hotkey for it or set it up to select a file from Telescope, e.g. |
Looks like an awesome plugin, but after reading through the ReadMe it wasn't apparent to me how I can send whole files into the chat for the chat bot to reference when answering questions. Is this not a feature of gp.nvim?
The text was updated successfully, but these errors were encountered: