RetentionLabs
Popular repositories Loading
-
RWKV-LM
RWKV-LM PublicForked from BlinkDL/RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
Python
-
titans-pytorch
titans-pytorch PublicForked from lucidrains/titans-pytorch
Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch
Python
-
-
MLP-Mixer-CIFAR
MLP-Mixer-CIFAR PublicForked from omihub777/MLP-Mixer-CIFAR
PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.
Python
-
ragflow
ragflow PublicForked from infiniflow/ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Python
Repositories
- ragflow Public Forked from infiniflow/ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
retentionlabs/ragflow’s past year of commit activity - RWKV-LM Public Forked from BlinkDL/RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
retentionlabs/RWKV-LM’s past year of commit activity - titans-pytorch Public Forked from lucidrains/titans-pytorch
Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch
retentionlabs/titans-pytorch’s past year of commit activity - MLP-Mixer-CIFAR Public Forked from omihub777/MLP-Mixer-CIFAR
PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.
retentionlabs/MLP-Mixer-CIFAR’s past year of commit activity
Top languages
Loading…
Most used topics
Loading…