8000 GitHub - lparisi/rag-gpt: Langchain RAG with local LLMs
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

lparisi/rag-gpt

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits< 8000 div class="d-sm-none">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LLM RAG Assignment

Running

  1. Install Ollama by following the instructions in the Ollama documentation.

  2. Run Ollama with the following commands ollama run "phi3:latest for the LLM model ollama run "nomic-embed-text"for the embeddings model

  3. Install Chromadb by following the instructions in the Chromadb documentation.

  4. Once Ollama and Chromadb are installed, you can use the following commands to run the project:

    • To populate the database using Ollama:
      python populate_database.py

    • To query the database using Ollama:

      python query_data.py "What are the latest Credit challenges for Amazon currently?"

    • To reset the database
      python populate_database.py --reset


Additional Notes: - Embedding Model: The current embedding model being used is nomic-embed-text. You can utilize the get_embedding_function.py script to select different embedding models such as AWS Bedrock, Hugging Face, or OpenAI. The quality of the embeddings directly impacts the retrieval accuracy.
- Vector Store: Chroma is being utilized as the vector store for local development due to time constraints. However, any other vector store supported by Langchain community can be used.

About

Langchain RAG with local LLMs

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%
0