Stars
Official Code Repository for the paper "Distilling LLM Agent into Small Models with Retrieval and Code Tools"
EvaByte: Efficient Byte-level Language Models at Scale
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
Efficient triton implementation of Native Sparse Attention.
Official Repo for Open-Reasoner-Zero
🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"
A bibliography and survey of the papers surrounding o1
Client Code Examples, Use Cases and Benchmarks for Enterprise h2oGPTe RAG-Based GenAI Platform
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton
Code for Neurips24 paper: QuaRot, an end-to-end 4-bit inference of large language models.
An Open-source Framework for Data-centric, Self-evolving Autonomous Language Agents
[ICCV 2023] VAD: Vectorized Scene Representation for Efficient Autonomous Driving
Official Code for Paper: RecurrentGPT: Interactive Generation of (Arbitrarily) Long Text
Official Code for DragGAN (SIGGRAPH 2023)
https://csstipendrankings.org
An open-source tool-augmented conversational language model from Fudan University
Running large language models on a single GPU for throughput-oriented scenarios.
Fast and memory-efficient exact attention
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
A PyTorch-based framework for Quantum Classical Simulation, Quantum Machine Learning, Quantum Neural Networks, Parameterized Quantum Circuits with support for easy deployments on real quantum compu…
Benchmark for Natural Temporal Distribution Shift (NeurIPS 2022)
A machine learning benchmark of in-the-wild distribution shifts, with data loaders, evaluators, and default models.
About Code release for "Flowformer: Linearizing Transformers with Conservation Flows" (ICML 2022), https://arxiv.org/pdf/2202.06258.pdf
Code for a series of work in LiDAR perception, including SST (CVPR 22), FSD (NeurIPS 22), FSD++ (TPAMI 23), FSDv2, and CTRL (ICCV 23, oral).
Official PyTorch implementation for the following KDD2022 paper: Variational Inference for Training Graph Neural Networks in Low-Data Regime through Joint Structure-Label Estimation
[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation
A scientific and useful toolbox, which contains practical and effective long-tail related tricks with extensive experimental results