More
Stars
Awesome LLMs on Device: A Comprehensive Survey
AIInfra(AI 基础设施)指AI系统从底层芯片等硬件,到上层软件栈支持AI大模型训练和推理。
Democratizing Reinforcement Learning for LLMs
Minimal reproduction of DeepSeek R1-Zero
[COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition
SGLang is a fast serving framework for large language models and vision language models.
mina86 / pygtrie
Forked from google/pygtriePython library implementing a trie data structure.
Python library implementing a trie data structure.
A comprehensive library for implementing LLMs, including a unified training pipeline and comprehensive model evaluation.
Summarize existing representative LLMs text datasets.
Curated list of datasets and tools for post-training.
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
A quick guide (especially) for trending instruction finetuning datasets
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
CA-LoRA: Adapting Existing LoRA for Compressed LLMs to Enable Efficient Multi-Tasking on Personal Devices (COLM 2024)
Exploring Efficient Fine-Grained Perception of Multimodal Large Language Models
CoreNet: A library for training deep neural networks
📰 Must-read papers and blogs on Speculative Decoding ⚡️
[TMLR 2024] Efficient Large Language Models: A Survey
Visualizer for neural network, deep learning and machine learning models
Reaching LLaMA2 Performance with 0.1M Dollars
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Code for "AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling"
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".