- Seoul, Korea
- @_the_young
Stars
the AI-native open-source embedding database
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of …
Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search
Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
Open-source vector similarity search for Postgres
Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas…
MCP (Model Context Protocol) server for Weaviate
An official Qdrant Model Context Protocol (MCP) server implementation
🚀 The fast, Pythonic way to build MCP servers and clients
Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, using the browser, and more with your permission every step of the way.
MySQL/MariaDB database management for macOS
An extremely fast Python package and project manager, written in Rust.
This repository contains various advanced techniques for Retrieval-Augmented Generation (RAG) systems.
Simple, unified interface to multiple Generative AI providers
Python SDK for AI agent monitoring, LLM cost tracking, benchmarking, and more. Integrates with most LLMs and agent frameworks including OpenAI Agents SDK, CrewAI, Langchain, Autogen, AG2, and CamelAI
Collection of awesome LLM apps with AI Agents and RAG using OpenAI, Anthropic, Gemini and opensource models.
LLM-powered multiagent persona simulation for imagination enhancement and business insights.
Awesome-LLM: a curated list of Large Language Model
Solar Prompt Cookbook
Langflow is a powerful tool for building and deploying AI-powered agents and workflows.
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting yo…
Stable Diffusion with Core ML on Apple Silicon