-
-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEAT] Code Interpreter for higher quality response and output, while summarizing the files and documents #856
Comments
Im having the same issue. Ive offloaded the embedding to ollama using nomic-embed-text to see if that was the issue since I saw the loacal embedding has this.embeddingMaxChunkLength = 1_00; but am having the same results. I have also pushed the LLM to LM studio to see the logs in the API call and noticed the context is VERY limited. Limited enough to make the LLM hallucinate quite often. |
Yes we need to improve this quality and less hallucinate. |
Try to jam the BERT tokenizer over the existing one and see what you get.
You will want to In some pdfs, im able to get better context. still testing on my side |
I should add: I'm using
as the embedder in |
Still chugging along on this. Things I've learned
Best Results So Far
|
Step by step instructions would be much appreciated 🙏
|
Update. V2. Results have been sub par by using different embedding models. The setup I am currently running is
I have been messing with the chunking and been getting better success though. I am going to try and mess with different chunking and text splitting methods and overlaps. I read that a parent child method works pretty good and will give that a try at some point. I believe chunking occurs at the
Even by just changing the overlap manually to 100 or so I feel like I get better results. Also tied this with a 400 and 40 just to see what it would do and it was performing alright. This wouldnt be great for lots of hits though because it would clutter the context in the LLM and theres a hard coded max that will be hit. |
@tylerfeldstein fyi, Related issue! #490 We can prioritize this so you can mess with it more easily. Are you using Docker, Desktop, or local dev? |
ACK. I'll jump over to that one. |
How are you running AnythingLLM?
AnythingLLM desktop app
What happened?
I was trying to chat with my documents. It was a basic document with the employee data with name and IDs and it wasn't able to generate the high-quality response. Most of the time it was giving me random data, and the output was not matching with the already tools that are available� for this.
I have tested.
1.
Claude-2.1
.2.
Gemini Pro
.3.
Local Models.
And I'm not trying to promote my product, but I compared with already available tools that are code interpreters that can generate the code, and it can analyze and all the files in our local system,
Link Code-Interpreter this is my tool i tested with same models like
Claude 2.1
and I got better results with this and that was more accurate, and other tools that are already available that are called code interpreters.Are there known steps to reproduce?
You can try with the very basic file and you can ask about the data and it will try to generate like table but the data will not be accurate all the time even though if you use the same models and different software available by code interpreters.
The text was updated successfully, but these errors were encountered: