Lists (3)
Sort Name ascending (A-Z)
Stars
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
A responsive, good looking with modern design documentation theme for Sphinx, with great supports for many sphinx extensions.
Tile primitives for speedy kernels
🚀 Kick-start your C++! A template for modern C++ projects using CMake, CI, code coverage, clang-format, reproducible dependency management and much more.
AI Assistant running within your browser.
A Easy-to-understand TensorOp Matmul Tutorial
An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.
MLC-LLM fork of oobabooga/text-generation-webui
Vercel and web-llm template to run wasm models directly in the browser.
OpenGVLab / OmniQuant
[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.
FlashInfer: Kernel Library for LLM Serving
pacman100 / mlc-llm
Forked from mlc-ai/mlc-llmEnable everyone to develop, optimize and deploy AI models natively on everyone's devices.
Neovim plugin for interacting with LLM's and building editor integrated prompts.
Run Large Language Models on RK3588 with GPU-acceleration
Structured inference with Llama 2 in your browser
Status column plugin that provides a configurable 'statuscolumn' and click handlers.
A template for modern C++ projects using CMake, Clang-Format, CI, unit testing and more, with support for downstream inclusion.