Stars
LlamaIndex is the leading framework for building LLM-powered agents over your data.
A high-throughput and memory-efficient inference and serving engine for LLMs
Production-ready platform for agentic workflow development.
Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
8000 Create Customized Software using Natural Language Idea (through LLM-powered Multi-Agent Collaboration)
Convert documents to structured data effortlessly. Unstructured is open-source ETL solution for transforming complex documents into clean, structured formats for language models. Visit our website …
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen2.5-VL, Qwen2.5-Omni, Qwen2-Audio, Ovis2, InternVL3, Llava, GLM4…
AgentTuning: Enabling Generalized Agent Abilities for LLMs
MS-Agent: Lightweight Framework for Empowering Agents with Autonomous Exploration
An Autonomous LLM Agent for Complex Task Solving
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)
AutoChain: Build lightweight, extensible, and testable LLM Agents
Official inference library for Mistral models
中文Mixtral-8x7B(Chinese-Mixtral-8x7B)
A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI
Start building LLM-empowered multi-agent applications in an easier way.
A platform for building proxies to bypass network restrictions.
ReLE中文大模型能力评测(持续更新):目前已囊括257个大模型,覆盖chatgpt、gpt-4.1、o4-mini、谷歌gemini-2.5、Claude、智谱GLM-Z1、文心一言、qwen-max、百川、讯飞星火、商汤senseChat、minimax等商用模型, 以及DeepSeek-R1-0528、qwq-32b、deepseek-v3、qwen3、llama4、phi-4、glm…
SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
XVERSE-13B: A multilingual large language model developed by XVERSE Technology Inc.
Llama中文社区,实时汇总最新Llama学习资料,构建最好的中文Llama大模型开源生态,完全开源可商用
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
A 13B large language model developed by Baichuan Intelligent Technology
A large-scale 7B pretraining language model developed by BaiChuan-Inc.