Ungate InfiniRoute AVS
  • 🔥Introduction
    • Overview
    • Scope
  • 📓ARCHITECTURE
    • Abstract
    • Keys Components
  • 🧠Inference Engine
    • Abstract
    • User Interface Layer
    • Coordination Layer
    • Data Layer
  • ⚓AVS
    • Abstract
    • Task Consensus
    • Consensus Engine
    • Rewards and Penalties
  • 🏪Models Marketplace
    • Abstract
    • Externally Hosted Models
    • Benchmarking Engine
    • Marketplace Economy
  • 👩‍💼Operators
    • MainNet
      • Registration
      • Installation
      • Quickstart
    • TestNet
      • Installation
      • Registration
      • Quickstart
  • 🎁Summary
    • Immediate Next Steps
    • Building the First SubNet
    • Future Vision
    • Conclusion
Powered by GitBook
On this page
  1. Inference Engine

Abstract

PreviousKeys ComponentsNextUser Interface Layer

Last updated 10 months ago

The Inference Engine is the core component of the InfiniRoute system, designed to dynamically route AI inference requests to the most suitable models. It optimises for both cost and accuracy by leveraging real-time performance benchmarks and data from the Models Library.

Upon receiving a prompt from the Chat Session Manager, the Inference Engine evaluates available models based on predefined criteria and selects the best-fit model for execution. This process ensures efficient utilisation of resources, delivering high-quality outputs while managing operational costs.

The Inference Engine's dynamic routing capabilities allow it to continuously adapt to changing performance metrics, ensuring optimal results for each request. By automating model selection and execution, the Inference Engine enhances overall system efficiency, scalability, and effectiveness, making it a critical element in the InfiniRoute architecture.

🧠
Inference Engine