ScaiInfer — Inference Compute
ScaiInfer provides the GPU-powered compute infrastructure that serves AI model inference. ScaiInfer nodes register with ScaiGrid and handle the actual model execution — loading models into GPU memory, processing requests, and returning results.
What makes ScaiInfer flexible is where it runs. Nodes can operate on ScaiLabs' own infrastructure, on a partner's data center hardware, or on-premises at a customer site. ScaiGrid handles the routing transparently — applications don't know or care which node processes their request.
Key capabilities:
- GPU-optimized inference serving
- Automatic node registration with ScaiGrid
- Health monitoring and capacity reporting
- Support for multiple concurrent models per node
- Hardware flexibility: cloud, data center, or on-premises