Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Replace Falcon 🦅 model with Llama V2 🦙 for offline chat #352

Merged
merged 11 commits into from
Jul 28, 2023

Conversation

sabaimran
Copy link
Member

@sabaimran sabaimran commented Jul 27, 2023

Incoming

  • Rather than release the first local LLM with Falcon, use Llama with the existing GPT4All setup.
  • Rework the extract_questions flow. While Falcon's output wasn't high enough quality to actually use this functionality, Llama generally provides good enough responses (at least able to pass ~50% of the benchmark tests). That being said, query times might be too long to justify using this.
  • Rename all references from Falcon -> Llama
  • Update relevant unit tests to match Llama capability

Closes #347

@sabaimran sabaimran requested a review from debanjum July 27, 2023 20:31
Copy link
Member

@debanjum debanjum left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice!! Feels like Llama v2 may just make offline chat useful enough 🤞🏾🏕️

tests/test_gpt4all_chat_actors.py Show resolved Hide resolved
tests/test_gpt4all_chat_actors.py Outdated Show resolved Hide resolved
src/khoj/routers/api.py Outdated Show resolved Hide resolved
src/khoj/routers/helpers.py Outdated Show resolved Hide resolved
- Change llama-specific naming in chat_model methods to be general offline
- Fix reasoning of assertion failure in one of the gpt4all actor tests
@sabaimran sabaimran merged commit 124d97c into master Jul 28, 2023
3 of 4 checks passed
@sabaimran sabaimran deleted the features/replace-falcon-with-llamav2 branch July 28, 2023 03:51
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Use llamav2 instead of falcon as open source model for Khoj
2 participants