Easy LLM API allows you to quickly deploy any open-source Large Language Model (LLM) with optimized performance using platforms like Kaggle or Google Colab. This is perfect for projects like hackathons, where you need an instant and efficient LLM deployment.
- Run any open-source LLM easily.
- Supports platforms like Kaggle and Google Colab for seamless execution.
- Provides an API endpoint to interact with the model.
- Uses Ngrok for instant public access.
- Built with FastAPI for high performance.
- CORS enabled for flexible integrations.
Click the button below to open and run the notebook on Kaggle:
Click the button below to run the notebook on Google Colab:
-
Clone this repository:
git clone https://github.com/{YOUR_GITHUB_USERNAME}/easy-llm-api.git cd easy-llm-api
-
Install dependencies:
pip install -r requirements.txt
-
Run the API locally:
python main.py
POST /chat
{
"prompt": "Hello, how are you?",
"context": ""
}
{
"response": "I am fine, how can I assist you?"
}
- Support for additional LLM models.
- Implement model selection via API request.
- Add authentication for API access.
- Optimize inference speed with quantization.
- Provide WebSocket support for real-time chat.
- Add a front-end UI for easier interaction.
- Deploy on cloud platforms (AWS, GCP, Azure).
- Enable multi-user handling for production use.
- Enhance logging and monitoring.
- Improve documentation with example use cases.
This project is licensed under the MIT License.