{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"modelscope","owner":"modelscope","isFork":false,"description":"ModelScope: bring the notion of Model-as-a-Service to life.","topicNames":["nlp","science","cv","speech","multi-modal","python","machine-learning","deep-learning"],"topicsNotShown":0,"allTopics":["nlp","science","cv","speech","multi-modal","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":124,"starsCount":6202,"forksCount":653,"license":"Apache License 2.0","participation":[9,8,4,9,32,18,42,19,55,36,18,17,25,44,30,15,23,49,1,20,23,18,7,11,10,1,7,9,4,3,12,7,6,9,4,1,1,0,11,9,12,2,0,3,0,4,4,3,0,0,4,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T07:37:43.980Z"}},{"type":"Public","name":"swift","owner":"modelscope","isFork":false,"description":"ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs","topicNames":["agent","deploy","llama","lora","finetune","peft","multimodal","sft","dpo","pre-training"],"topicsNotShown":9,"allTopics":["agent","deploy","llama","lora","finetune","peft","multimodal","sft","dpo","pre-training","awq","llm","modelscope","llava","qwen","galore","unsloth","llama3","pissa"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":111,"starsCount":1557,"forksCount":162,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,9,18,3,3,3,8,8,16,11,1,6,5,1,11,5,8,11,5,14,11,19,7,16,25,9,14,20,28,0,13,16,18,33,15,17,11,13,28,22,19,20,22,24],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T08:04:49.181Z"}},{"type":"Public","name":"data-juicer","owner":"modelscope","isFork":false,"description":"A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!","topicNames":["nlp","data-science","opendata","data-visualization","pytorch","dataset","chinese","data-analysis","llama","gpt"],"topicsNotShown":10,"allTopics":["nlp","data-science","opendata","data-visualization","pytorch","dataset","chinese","data-analysis","llama","gpt","multi-modal","pre-training","sora","streamlit","gpt-4","large-language-models","llm","llms","instruction-tuning","llava"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":3,"starsCount":1566,"forksCount":99,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,10,5,0,2,0,2,1,5,1,0,2,4,5,2,1,9,6,4,2,2,5,6,4,4,4,4,3,1,0,5,5,4,16,12,6,0,4,7,5,1,4,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T07:50:49.127Z"}},{"type":"Public","name":"agentscope","owner":"modelscope","isFork":false,"description":"Start building LLM-empowered multi-agent applications in an easier way.","topicNames":["agent","chatbot","multi-agent","multi-modal","distributed-agents","gpt-4","large-language-models","llm","llm-agent","llama3"],"topicsNotShown":1,"allTopics":["agent","chatbot","multi-agent","multi-modal","distributed-agents","gpt-4","large-language-models","llm","llm-agent","llama3","gpt-4o"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":11,"starsCount":2437,"forksCount":152,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,9,2,3,0,3,10,5,8,16,15,12,7,8,7,9,5,15,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T06:26:07.339Z"}},{"type":"Public","name":"modelscope-agent","owner":"modelscope","isFork":false,"description":"ModelScope-Agent: An agent framework connecting models in ModelScope with the world","topicNames":["agent","multi-agents","gpts","llm","qwen","chatglm-4","open-gpts"],"topicsNotShown":0,"allTopics":["agent","multi-agents","gpts","llm","qwen","chatglm-4","open-gpts"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":39,"starsCount":1989,"forksCount":227,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,1,0,0,1,1,3,4,4,59,2,7,3,0,0,0,2,2,7,16,23,38,19,31,15,1,3,1,1,0,2,2,1,0,5,4,9,10,5,1,6,12,4,6,3,11,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T07:53:25.086Z"}},{"type":"Public","name":"eval-scope","owner":"modelscope","isFork":false,"description":"A streamlined and customizable framework for efficient large model evaluation and performance benchmarking","topicNames":["performance","evaluation","llm"],"topicsNotShown":0,"allTopics":["performance","evaluation","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":5,"starsCount":67,"forksCount":9,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,3,28,2,0,4,3,4,19,0,0,0,5,18,2,1,7,2,1,1,1,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T08:56:42.543Z"}},{"type":"Public","name":"scepter","owner":"modelscope","isFork":false,"description":"SCEPTER is an open-source framework used for training, fine-tuning, and inference with generative models.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":254,"forksCount":15,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,2,3,0,3,9,0,10,0,0,1,0,0,0,0,8,2,1,9,3,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-30T01:37:24.623Z"}},{"type":"Public","name":"lite-sora","owner":"modelscope","isFork":false,"description":"An initiative to replicate Sora","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":90,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,7,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-10T08:35:58.543Z"}},{"type":"Public","name":"facechain","owner":"modelscope","isFork":false,"description":"FaceChain is a deep-learning toolchain for generating your Digital-Twin.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":71,"starsCount":8432,"forksCount":774,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,7,81,22,73,34,31,84,5,0,4,4,5,4,8,0,1,8,2,1,5,0,0,3,1,0,0,0,0,0,1,1,3,0,7,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-07T01:40:32.628Z"}},{"type":"Public","name":"richdreamer","owner":"modelscope","isFork":false,"description":"Live Demo:https://modelscope.cn/studios/Damo_XR_Lab/3D_AIGC","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":326,"forksCount":10,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,8,5,2,0,3,14,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-06T13:55:31.760Z"}},{"type":"Public","name":"normal-depth-diffusion","owner":"modelscope","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":106,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,0,6,0,1,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-07T14:39:51.966Z"}},{"type":"Public","name":"FunCodec","owner":"modelscope","isFork":false,"description":"FunCodec is a research-oriented toolkit for audio quantization and downstream applications, such as text-to-speech synthesis, music generation et.al. ","topicNames":["tts","speech-synthesis","codec","speech-to-text","audio-generation","encodec","voicecloning","audio-quantization"],"topicsNotShown":0,"allTopics":["tts","speech-synthesis","codec","speech-to-text","audio-generation","encodec","voicecloning","audio-quantization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":13,"starsCount":292,"forksCount":23,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,0,13,0,0,0,0,0,0,33,0,7,0,11,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-25T11:56:17.823Z"}},{"type":"Public","name":"AdaSeq","owner":"modelscope","isFork":false,"description":"AdaSeq: An All-in-One Library for Developing State-of-the-Art Sequence Understanding Models","topicNames":["natural-language-processing","information-extraction","chinese-nlp","word-segmentation","bert","sequence-labeling","relation-extraction","natural-language-understanding","entity-typing","token-classification"],"topicsNotShown":6,"allTopics":["natural-language-processing","information-extraction","chinese-nlp","word-segmentation","bert","sequence-labeling","relation-extraction","natural-language-understanding","entity-typing","token-classification","multi-modal-ner","nlp","crf","pytorch","named-entity-recognition","ner"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":24,"starsCount":370,"forksCount":33,"license":"Apache License 2.0","participation":[0,0,0,1,4,0,0,0,0,0,0,0,0,0,0,5,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-15T12:08:20.988Z"}},{"type":"Public","name":"motionagent","owner":"modelscope","isFork":false,"description":"MotionAgent is your AI assistent to convert ideas into motion pictures.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":262,"forksCount":28,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,6,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-07T14:44:46.195Z"}},{"type":"Public","name":"AdaDet","owner":"modelscope","isFork":false,"description":"AdaDet: A Development Toolkit for Object Detection based on ModelScope","topicNames":["python","tracking","machine-learning","ocr","computer-vision","deep-learning","domain-specific-models","keypoint-detection","detection-toolkit","detection"],"topicsNotShown":2,"allTopics":["python","tracking","machine-learning","ocr","computer-vision","deep-learning","domain-specific-models","keypoint-detection","detection-toolkit","detection","face-detection","object-detection"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":41,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-17T07:31:31.697Z"}}],"repositoryCount":15,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}