Lists (1)
Sort Name ascending (A-Z)
Starred repositories
Block Diffusion: Interpolating Between Autoregressive and Diffusion Language Models
Official PyTorch implementation for "Large Language Diffusion Models"
A collection of research papers on low-precision training methods
MiMo: Unlocking the Reasoning Potential of Language Model – From Pretraining to Posttraining
[ICLR 2025] Repository for Show-o series, One Single Transformer to Unify Multimodal Understanding and Generation.
Official Repo for Open-Reasoner-Zero
Pretraining code for a large-scale depth-recurrent language model
Training Large Language Model to Reason in a Continuous Latent Space
Witness the aha moment of VLM with less than $3.
MM-Eureka V0 also called R1-Multimodal-Journey, Latest version is in MM-Eureka
Fully open reproduction of DeepSeek-R1
[CVPR 2025] The First Investigation of CoT Reasoning (RL, TTS, Reflection) in Image Generation
Minimal reproduction of DeepSeek R1-Zero
A debugging and profiling tool that can trace and visualize python code execution
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
[NeurIPS 2024] The official code of "U-DiTs: Downsample Tokens in U-Shaped Diffusion Transformers"
[ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear attention mechanism.
SAM-DiffSR: Structure-Modulated Diffusion Model for Image Super-Resolution
[AAAI 2025] Official PyTorch implementation of "TinySAM: Pushing the Envelope for Efficient Segment Anything Model"
[ICML'24] The official implementation of “Rethinking Optimization and Architecture for Tiny Language Models”