Stars
a python command-line tool which draws basic graphs in the terminal
Example Code of Calling Python from C++ with PyBind11.
Development repository for the Triton language and compiler
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
Hackable and optimized Transformers building blocks, supporting a composable construction.
Warmup learning rate wrapper for Pytorch Scheduler
Write workflows scripting the GitHub API in JavaScript
KErnel OPerationS, on CPUs and GPUs, with autodiff and without memory overflows
Research and development for optimizing transformers
Nintendo Joy-Con/ProController Key mapper for macOS
Is the attention layer even necessary? (https://arxiv.org/abs/2105.02723)
Pytorch library for fast transformer implementations
Deformable DETR: Deformable Transformers for End-to-End Object Detection.
Pytorch Optimizer for Simulated Annealing
Low Precision Arithmetic Simulation in PyTorch