Highlights
- Pro
Lists (19)
Sort Name ascending (A-Z)
Stars
This is the repository of DEER, a Dynamic Early Exit in Reasoning method for Large Reasoning Language Models.
The official implementation of "Well Begun is Half Done: Low-resource Preference Alignment by Weak-to-Strong Decoding"
[ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data generation pipeline!
[ACL'25] Code for "Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering"
🚀 The fast, Pythonic way to build MCP servers and clients
🔥 How to efficiently and effectively compress the CoTs or directly generate concise CoTs during inference while maintaining the reasoning performance is an important topic!
Code for the paper: "Learning to Reason without External Rewards"
Open-source Multi-agent Poster Generation from Papers
VeriThinker: Learning to Verify Makes Reasoning Model Efficient
Align Anything: Training All-modality Model with Feedback
[NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct
Official implementation for "ALI-Agent: Assessing LLMs'Alignment with Human Values via Agent-based Evaluation"
[ICML 2025] Fourier Position Embedding: Enhancing Attention’s Periodic Extension for Length Generalization
利用AI大模型,一键生成高清短视频 Generate short videos with one click using AI LLM.
LoRI: Reducing Cross-Task Interference in Multi-Task Low-Rank Adaptation
AlphaEdit: Null-Space Constrained Knowledge Editing for Language Models, ICLR 2025 (Outstanding Paper)
Implementing DeepSeek R1's GRPO algorithm from scratch
[NAACL 2025 Main Selected Oral] Repository for the paper: Prompt Compression for Large Language Models: A Survey
Distributed Compiler Based on Triton for Parallel Systems
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton