Skip to content

Velesio/velesio-aiserver

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

105 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🚀 Velesio AI Server

High-performance, microservice-based AI inference server with Unity integration support.

Deploy on RunPod

✨ Features

  • 🎯 Unity Ready: Seamless integration with Unity!
  • 📈 Scalable: Redis queue-based worker architecture
  • 🐳 Easy Deploy: Docker Compose setup for inference setup, api wrapper, nginx & monitoring
  • 📊 Monitoring: Grafana template for System, GPU & Application observability

⚡ Quick Start

🎮 Unity Integration

Built specifically for Unity developers:

📚 Documentation

📖 Complete Documentation - Full guides, API reference, and examples

🏗️ Architecture

Distributed microservice design for maximum flexibility:

┌─────────────┐    ┌─────────┐    ┌─────────────┐
│    API      │────│  Redis  │────│ GPU Workers │
│  (FastAPI)  │    │ Queue   │    │ (LLM + SD)  │
└─────────────┘    └─────────┘    └─────────────┘
       │                                  │
       │           ┌─────────────┐        │
       └───────────│ Monitoring  │────────┘
                   │(Grafana+Prom)│
                   └─────────────┘
  • API Service: FastAPI with token auth and job queuing
  • GPU Workers: Custom llama.cpp + Stable Diffusion inference engines
  • Redis Queue: Decoupled job processing for scalability
  • Monitoring: Pre-configured Grafana dashboards

📖 Learn more: Architecture Documentation


🔌 Open Source References

Automatic1111 SD Web server LLAMACPP

Questions? Check the Documentation or open an issue!

About

Conveniently self host AI Inference for various Unity Integrations

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors 2

  •  
  •