A transparent proxy service that allows applications to use both Ollama and OpenAI API formats seamlessly with OpenAI-compatible LLM servers like OpenAI, vLLM, LiteLLM, OpenRouter, Ollama, and any ...
API/ ├── docker-compose.yml # Orquestación de servicios ├── .env.example # Variables de entorno de ejemplo ├── api1/ # API básica CRUD │ ├── Dockerfile │ ├── app.py │ ├── database.py │ ├── ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results