-
Mila, McGill University
- Montreal, Canada
- sahandrez.github.io
Highlights
- Pro
Lists (3)
Sort Name ascending (A-Z)
Stars
Everything about the SmolLM and SmolVLM family of models
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
DeepEP: an efficient expert-parallel communication library
Minimal reproduction of DeepSeek R1-Zero
Fully open reproduction of DeepSeek-R1
A framework for few-shot evaluation of language models.
Sharing both practical insights and theoretical knowledge about LLM evaluation that we gathered while managing the Open LLM Leaderboard and designing lighteval!
Modeling, training, eval, and inference code for OLMo
Minimalistic large language model 3D-parallelism training
Unsupervised text tokenizer for Neural Network-based text generation.
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
This repository contains LLM (Large language model) interview question asked in top companies like Google, Nvidia , Meta , Microsoft & fortune 500 companies.
[WIP] Resources for AI engineers. Also contains supporting materials for the book AI Engineering (Chip Huyen, 2025)
Solve puzzles. Improve your pytorch.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Persian raw text - حدود ۸۰ گیگابایت متن خام فارسی
Persian (Farsi) Question Answering Dataset (+ Models)
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Reinforcement Learning from Human Feedback (RLHF) with Unpaired Preferences
https://huyenchip.com/ml-interviews-book/
Machine Learning and Computer Vision Engineer - Technical Interview Questions
Curated list of data science interview questions and answers
Test your prompts, agents, and RAGs. Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with comma…
A high-throughput and memory-efficient inference and serving engine for LLMs