8000 GitHub - eramax/LLMs
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

eramax/LLMs

< 8000 button type="button" aria-haspopup="true" aria-expanded="false" tabindex="0" class="prc-Button-ButtonBase-c50BI" data-loading="false" data-size="medium" data-variant="primary" aria-describedby=":R55ab:-loading-announcement" id=":R55ab:">Code
 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LLMs Architecture

Alt Text

This repository contains implementations of LLMs

Implemented Architectures

  1. GPT2
  2. BERT
  3. Llama

📚 Description

This repository is structured as follows:

  1. Data Folder contains small text files for simple training
  2. GPT2 Folder contains training, model, textsampler
  3. BERT Folder contains Bert Model and Bert Modules
  4. Llama Folder contains Llama Model with GQA and RoPE Modules

🚀 Goals

  • Implement and experiment with different model architecture.
  • Develop foundational components for future research in transformers.
  • Provide clean and modular code.

📝 Research Papers Implemented (or To Be Implemented)

  1. GPT2 Paper (https://paperswithcode.com/paper/language-models-are-unsupervised-multitask)
  2. Bert Paper (https://arxiv.org/abs/1810.04805)
  3. Llama Paper (https://arxiv.org/abs/2307.09288)

💡 Future Work

Implement more LLMs architectures.

⭐ Contributions

Feel free to contribute to this repository or suggest improvements.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%
0