-
University of California, San Diego
- La Jolla, California
-
12:38
(UTC -07:00)
Highlights
- Pro
Stars
(best/better) practices of megatron on veRL and tuning guide
A repo for open research on building large reasoning models
Official implementation for DenseMixer: Improving MoE Post-Training with Precise Router Gradient
The evaluation framework for training-free sparse attention in LLMs
An Awesome List of Reinforcement Learning-based Large Language Agent Works. Collect directly from official code base.
Implementation and datasets for "Training Language Models to Generate Quality Code with Program Analysis Feedback"
Official repository for FLAME-MoE: A Transparent End-to-End Research Platform for Mixture-of-Experts Language Models
Code for our tutorial on Discrete Variational Autoencoders
[TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.
[ACL2025 Oral] Cuckoo: A Series of IE Free Riders Using LLM's Resources to Scale up Themselves.
A bibliography and survey of the papers surrounding o1
The road to hack SysML and become an system expert
A framework for the evaluation of autoregressive code generation language models.
Tools for merging pretrained large language models.
Large Language Model Text Generation Inference
GPU programming related news and material links
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
alibaba / Megatron-LLaMA
Forked from NVIDIA/Megatron-LMBest practice for training LLaMA models in Megatron-LM
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
A curated list of papers and applications on tool learning.
A repo lists papers related to LLM based agent
Train transformer language models with reinforcement learning.
A curated list of awesome resources dedicated to Scaling Laws for LLMs
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming