-
KAIST AI (OSI LAB)
- Seoul, Korea
-
11:54
(UTC +09:00) - namgyu.com
- https://orcid.org/0000-0002-2445-3026
- @itsnamgyu
Stars
Doing simple retrieval from LLM models at various context lengths to measure accuracy
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
A PyTorch native platform for training generative AI models
XAttention: Block Sparse Attention with Antidiagonal Scoring
Gemma open-weight LLM library, from Google DeepMind
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry lead…
An interactive HTML pretty-printer for machine learning research in IPython notebooks.
Modular, scalable library to train ML models
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Code for paper called Self-Training Elicits Concise Reasoning in Large Language Models
📰 Must-read papers on KV Cache Compression (constantly updating 🤗).
Evaluation of speculative inference over multilingual tasks
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Doing simple retrieval from LLM models at various context lengths to measure accuracy
Deep learning for dummies. All the practical details and useful utilities that go into working with real models.
Official repository for EXAONE built by LG AI Research
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)
Official implementation of "Perturbed-Attention Guidance"
Official repository of "Distort, Distract, Decode: Instruction-Tuned Model Can Refine its Response from Noisy Instructions", ICLR 2024 Spotlight