{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"cortex","owner":"janhq","isFork":false,"description":"Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan","allTopics":["ai","cuda","llama","accelerated","inference-engine","openai-api","llm","stable-diffusion","llms","llamacpp","llama2","gguf","tensorrt-llm"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":4,"issueCount":76,"starsCount":1679,"forksCount":85,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,42,46,22,20,5,6,22,28,111,62,23,47,33,14,32,35,50,40,76,30,22,18,17,2,0,2,6,20,23,13,2,0,2,38,40,49],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T05:27:39.823Z"}},{"type":"Public","name":"cortex.llamacpp","owner":"janhq","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":2,"issueCount":5,"starsCount":3,"forksCount":1,"license":"GNU Affero General Public License v3.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,5,5,8,13,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T01:17:53.294Z"}},{"type":"Public","name":"cortex.tensorrt-llm","owner":"janhq","isFork":true,"description":"Nitro is an C++ inference server on top of TensorRT-LLM. OpenAI-compatible API. Run blazing fast inference on Nvidia GPUs. Used in Jan","allTopics":["jan","tensorrt","llm","tensorrt-llm"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":8,"starsCount":28,"forksCount":747,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,5,0,6,8,1,1,3,5,1,1,4,1,0,1,3,2,1,1,6,1,1,3,3,28,30,1,1,1,1,1,1,1,2,3,2,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T01:05:08.277Z"}},{"type":"Public","name":"cortex.python","owner":"janhq","isFork":false,"description":"C++ code that run Python embedding","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"GNU Affero General Public License v3.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,2,4,9,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T13:36:05.468Z"}},{"type":"Public","name":"infinity","owner":"janhq","isFork":true,"description":"The AI-native database built for LLM applications, providing incredibly fast vector and full-text search ","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":152,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-19T06:45:53.239Z"}},{"type":"Public","name":"llama.cpp-avx-vnni","owner":"janhq","isFork":true,"description":"Port of Facebook's LLaMA model in C/C++","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":8464,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-19T07:04:40.274Z"}},{"type":"Public","name":"TensorRT","owner":"janhq","isFork":true,"description":"NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1989,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-03T03:08:17.003Z"}}],"repositoryCount":7,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}