Replies: 3 comments 1 reply
-
in my case , when running with this config the answer is very low comparing to however please consider the answer reflects the ingested documents' content |
Beta Was this translation helpful? Give feedback.
-
Yes, I also found GPT4All gpt-j models etc. are quite poor, sometimes even natively giving no output. I recommend better than Vicuna, instead use WizardLM uncensored models from TheBloke for CPU and his -HF models, and use our prompt_type that matches in h2oGPT: https://github.com/h2oai/h2ogpt#windows-1011 |
Beta Was this translation helpful? Give feedback.
-
Why is that? Why can it not read over all documents? Can i do something about that? |
Beta Was this translation helpful? Give feedback.
-
I keep testing the privateGPT for several weeks with different versions, I can say that privateGPT's accuracy is very low.
It looks like it can only read the last document, and mostly it cannot get the correct answer.
Recently I watch youtube and found a localGPT project, which is similar to privateGPT.
(I can only use CPU to run the projects, so... I can only keep going on privateGPT )
The main point is, I copy localGPT's ingest.py, use hkunlp/instructor-xl as the embedding model to generate chomo database files.
(which requires a very long time to generate that database file...)
Then I use WizardLM-30B-Uncensored.ggmlv3.q4_0 as the LlamaCpp model.
(which takes a very long time to get the answer in qa...)
It gets acceptable answers. I can say that is a great improved.
So I suspect the accuracy problem is coming from the fastest embedding model all-MiniLM-L12-v2 and all-MiniLM-L6-v2.
Because I have tested it with WizardLM-30B-Uncensored.ggmlv3.q4_0, the accuracy is still very low...
Beta Was this translation helpful? Give feedback.
All reactions