8000 GitHub - Predacons/predacons-server: an openai api based server built on predacons to host any torch and hugging face llm model
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

an openai api based server built on predacons to host any torch and hugging face llm model

License

AGPL-3.0 and 2 other licenses found

Licenses found

AGPL-3.0
LICENSE-AGPL
Unknown
LICENSE-COMMERCIAL-PERPETUAL
Unknown
LICENSE-COMMERCIAL-YEARLY
Notifications You must be signed in to change notification settings

Predacons/predacons-server

Repository files navigation

Predacons Server

About

An OpenAI API-based server built on Predacons to host any Torch and Hugging Face LLM model.

Running Locally

To run the application locally, follow these steps:

Prerequisites

  • Docker installed on your system
  • Docker Compose installed on your system

Steps

  1. Clone the repository:

    git clone https://github.com/Predacons/predacons-server.git
  2. Change into the project directory:

    cd predacons-server
  3. Build the Docker image:

    docker-compose build
  4. Start the containers:

    docker-compose up
  5. Access the application at:

    http://localhost:8000

Environment Variables

You can configure environment variables in the env directory. For example, you can add a .env file with the following contents:

API_KEYS=your-api-key

predacons-server

an openai api based server built on predacons to host any torch and hugging face llm model

Features

  • Model Hosting: Easily host and manage PyTorch and Hugging Face models.
  • API Key Authentication: Secure access through API key authentication for both free and paid users.
  • Scalable Architecture: Designed to scale horizontally to handle increasing loads and concurrent requests.
  • Easy Integration: Provides a straightforward REST API for integrating with existing applications or services.

Getting Started

To get started with predacons-server, follow these steps:

  1. Clone the repository to your local machine.
  2. Install the required dependencies using pip install -r requirements.txt.
  3. Set up your environment variables, including your API keys in a .env file.
  4. Run the server using uvicorn main:app --reload.

For detailed instructions and documentation, please refer to our Wiki or Getting Started Guide.

Contribution

Contributions are welcome! If you'd like to contribute, please fork the repository and use a feature branch. Pull requests are warmly welcome.

Support

If you encounter any issues or require assistance, please open an issue on our GitHub repository.

Thank you for considering predacons-server for your AI model hosting needs!

License

This project is licensed under multiple licenses:

  • For free users, the project is licensed under the terms of the GNU Affero General Public License (AGPL). See LICENSE-AGPL for more details.

  • For paid users, there are two options:

Please ensure you understand and comply with the license that applies to you.

About

an openai api based server built on predacons to host any torch and hugging face llm model

Resources

License

AGPL-3.0 and 2 other licenses found

Licenses found

AGPL-3.0
LICENSE-AGPL
Unknown
LICENSE-COMMERCIAL-PERPETUAL
Unknown
LICENSE-COMMERCIAL-YEARLY

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  
0