- Beijing, China
- https://aberhu.github.io/
Stars
Video-R1: Reinforcing Video Reasoning in MLLMs [🔥the first paper to explore R1 for video]
Official repository of ’Visual-RFT: Visual Reinforcement Fine-Tuning’
R1-onevision, a visual language model capable of deep CoT reasoning.
Solve Visual Understanding with Reinforced VLMs
Extend OpenRLHF to support LMM RL training for reproduction of DeepSeek-R1 on multimodal tasks.
A fork to add multimodal model training to open-r1
Fully open reproduction of DeepSeek-R1
A library for advanced large language model reasoning
Scalable RL solution for advanced reasoning of language models
The Next Step Forward in Multimodal LLM Alignment
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & LoRA & vLLM & RFT)
Unified KV Cache Compression Methods for Auto-Regressive Models
A streamlined and customizable framework for efficient large model evaluation and performance benchmarking
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
Open-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks
⚡FlashRAG: A Python Toolkit for Efficient RAG Research (WWW2025 Resource)
Get your documents ready for gen AI
Python tool for converting files and office documents to Markdown.
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, m…
[ECCV 2022] ByteTrack: Multi-Object Tracking by Associating Every Detection Box
A lightweight data processing framework built on DuckDB and 3FS.
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
A bidirectional pipeline parallelism algorithm for computation-communication overlap in V3/R1 training.
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
DeepEP: an efficient expert-parallel communication library
FlashMLA: Efficient MLA decoding kernels
Making large AI models cheaper, faster and more accessible
Efficient Triton Kernels for LLM Training