Stars
pip install nb_log 各种日志handler和自动转化项目的任意print的效果。日志自动彩色炫酷,可点击控制台的日志自动精确跳转到pycharm的文件和行号。文件日志多进程切割安全。在10个最重要方面全方位超过loguru
A powerful tool for creating fine-tuning datasets for LLM
Official repository for DistFlashAttn: Distributed Memory-efficient Attention for Long-context LLMs Training
Qihoo360 / 360-LLaMA-Factory
Forked from hiyouga/LLaMA-Factoryadds Sequence Parallelism into LLaMA-Factory
Unleashing the Power of Reinforcement Learning for Math and Code Reasoners
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
Qwen2.5-Omni is an end-to-end multimodal model by Qwen team at Alibaba Cloud, capable of understanding text, audio, vision, video, and performing real-time speech generation.
The official Python SDK for Model Context Protocol servers and clients
Model Context Protocol Servers
Model Context Protocol(MCP) 编程极速入门
An open-source solution for full parameter fine-tuning of DeepSeek-V3/R1 671B, including complete code and scripts from training to inference, as well as some practical experiences and conclusions.…
Analyze computation-communication overlap in V3/R1.
A bidirectional pipeline parallelism algorithm for computation-communication overlap in V3/R1 training.
EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
🧑🚀 全世界最好的LLM资料总结(视频生成、Agent、辅助编程、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.
Efficient Mixture of Experts for LLM Paper List
DeepEP: an efficient expert-parallel communication library
FlashMLA: Efficient MLA decoding kernels
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
A PyTorch native platform for training generative AI models
Minimalistic large language model 3D-parallelism training
MoBA: Mixture of Block Attention for Long-Context LLMs
Building DeepSeek R1 from Scratch