Abstract
Last updated
Last updated
The high-level architecture of InfiniRoute is designed to efficiently route AI inference requests to the most suitable models, optimising for cost and accuracy. The system begins with the Intelligence Consumer, who initiates requests. These requests are managed by the Chat Session Manager, which coordinates user interaction and forwards the requests to the Router. The Router, a central component, determines the best model by consulting the Models Library for cost and endpoint information. Selected models, hosted externally, process the inference and return results.
For data management, session data is stored in the Chat Session Store and passed to the AVS (EigenLayer) for onchain storage. The Node Operator writes this metadata to the IPFS. Attestors validate the integrity of the metadata, which is then aggregated and written to the L2 Blockchain, with final storage on Ethereum (L1 Blockchain). This architecture ensures seamless request handling, optimal model selection, and robust, transparent metadata management using blockchain technology.