Create robust binary AI runtimes

Paramus AI Service Specification 1.1

This specification defines a universal, lightweight interface for AI model deployment, monitoring, and interoperability. Use any language to implement, Python, C++ …

This specification standardizes REST-based endpoints for inference, health, metrics, and citation, enabling a reproducible and testable workflow for AI in science. Models using this design can be deployed as local executables, integrated cloud microservices, or registered modules inside the Paramus Marketplace.

The specification embodies Paramus principles: simplicity, observability, and interoperability. Each service provides self-describing capabilities, versioned responses, and full attribution to scientific and software sources.

It ensures that every Paramus model behaves predictably across all environments: stable system runtimes that cannot block each other.

Example

EndpointPurposeExample Response
/healthService status check{ "status": "ok", "model": "TransPolymer" }
/capabilitiesDeclares supported tasks{ "capabilities": ["embedding","regression"] }
/predictExecutes inference{ "success": true, "results": { "embedding": [...] } }
/metricsReports uptime and latency{ "uptime_s": 3600, "avg_latency_ms": 45.2 }
/citationProvides research attribution"Xu et al., npj Comput. Mater. 9, 64 (2023)"
Scroll to Top