# gpt-contextfiles ** currently in development, if you'll like to contribute or provide any feedback check out the [link](https://github.com/Iheuzio/gpt-contextfiles/issues) ** I was annoyed with copying responses into chatgpt and other LLMs for debugging my code across files, so I decided to make an extension that will do that. You simply right click each file you want to pass through, check or uncheck the checkbox, then enter your question and pass along the response over the api to your LLM. This extension uses the openai api, there are many models avaliable: https://openai.com/pricing > However, being orientated with managing files this project defaults to the 16k context with GPT-3.5-turbo-16k If you wish to change the model, you must change the model in the extension.js file https://platform.openai.com/docs/models/gpt-3-5 # Installation Add your api key to `OPENAI_API_KEY` for your windows/linux environment variable (tested with system variable) # Features Clear -> Clears the files currently available Submit -> Submits the query to the api Refresh -> refreshes the window so that all new files will be available for that session. - Right click to add files to the context window - Click on the extension addon to open the context window, refresh to update the files to check. - Select the files uses checkboxes - After submit is pressed, wait until the question disappears, this means the query is processed by openai and was fully sent - Click `API Response` to view your query # Examples Demo of how to use the extension: ![](./images/demo-program.gif) # How it works We can select two files we want to pass through, however we can uncheck one of them for later debugging and enter our question: ``` What does this do? c:\dev\test\gpt-contextfiles-test\program.js: \``` window.alert("Hello World!") \``` ``` Functions based on the principle of files passed into it ``` Selected Files: [x] c:\dev\test\gpt-contextfiles-test\program.js [ ] c:\dev\test\gpt-contextfiles-test\program2.js ```