Lists (4)
Sort Name ascending (A-Z)
Starred repositories
A fast, clean, responsive Hugo theme.
❤️A clean, elegant but advanced blog theme for Hugo 一个简洁、优雅且高效的 Hugo 主题
A series of math-specific large language models of our Qwen2 series.
Seed1.5-VL, a vision-language foundation model designed to advance general-purpose multimodal understanding and reasoning, achieving state-of-the-art performance on 38 out of 60 public benchmarks.
simpleR1: A Simple Framework for Training R1-like Models
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
ToRA is a series of Tool-integrated Reasoning LLM Agents designed to solve challenging mathematical reasoning problems by interacting with tools [ICLR'24].
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
Search-R1: An Efficient, Scalable RL Training Framework for Reasoning & Search Engine Calling interleaved LLM based on veRL
✨✨Latest Advances on Multimodal Large Language Models
MTEB: Massive Text Embedding Benchmark
Understanding R1-Zero-Like Training: A Critical Perspective
Efficient Triton Kernels for LLM Training
This package contains the original 2012 AlexNet code.
Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen2.5-VL, Qwen2.5-Omni, Qwen2-Audio, Ovis2, InternVL3, Llava, GLM4…
A series of technical report on Slow Thinking with LLM
OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models
verl: Volcano Engine Reinforcement Learning for LLMs
Fully open reproduction of DeepSeek-R1
Democratizing Reinforcement Learning for LLMs
SGLang is a fast serving framework for large language models and vision language models.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
[ICML'24] Magicoder: Empowering Code Generation with OSS-Instruct
This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or rejection sampling fine-tuning.