You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A high-performance API server that provides OpenAI-compatible endpoints for MLX models. Developed using Python and powered by the FastAPI framework, it provides an efficient, scalable, and user-friendly solution for running MLX-based vision and language models locally with an OpenAI-compatible interface.
All-in-one local AI stack for Mistral-13B and Llama.cpp, with one-step CUDA wheel install, OpenAI-compatible API, and modern web dashboard. Switch between local and cloud chat, run on your own GPU, and deploy instantly—no API keys or paywalls. Designed for easy install, custom builds, and fast remote access. Enjoy!
🧠 Synapse - Supercharge your AI coding assistants with memory, context, MCP tools, and intelligent routing. Works with Cline, Roo, Cursor, and any OpenAI-compatible tool. Just change the API endpoint and get superpowers! ✨
This repository offers a straightforward guide to using GitHub Copilot as an OpenAI-compatible API. With easy setup steps and support for both Node.js and Python servers, you can quickly integrate this functionality into your projects. 🐙✨