Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create an easy way for users to share statistics and execution environment #517

Open
nickanderson opened this issue Oct 26, 2023 · 1 comment
Labels
upgrade New feature or request

Comments

@nickanderson
Copy link

nickanderson commented Oct 26, 2023

From: https://discord.com/channels/1112065956647284756/1112066421577482262/1166885403798798427

I would like to have a way that made it easy to provide feedback about the performance of local llm responses.

I think it could be useful for both the developers and users to be able to easily see information about the performance and quality of the local llm in use.

Information that come to mind:

  • OS & version
  • Relevant hardware info
    • CPU
      • cores
      • model
    • Memory
    • GPU model
    • Vram
      • available
      • in use by khoj
      • free
  • chat streaming response time
  • quality of response (if user rated, i think there is another issue that is specifically this)
  • Time to do RAG
  • Time to decode user input
  • Time to emit output to user (I think chat streaming time logged currently is combination of input decode time and output, not sure if possible to get separate timeing
@sabaimran
Copy link
Collaborator

Interesting note! These are definitely relevant performance metrics, and thank you for collating this information. We have the /help endpoint in chat which would output some of this, but could definitely be more detailed.

@sabaimran sabaimran added the upgrade New feature or request label Nov 3, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
upgrade New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants