The Usage page on the NextAI platform provides detailed insights into your usage of our services, helping you understand and optimize your AI model deployments. This document outlines the types of metrics available and how to interpret them for better resource management and planning.

Overview

NextAI’s Usage page is designed to offer comprehensive visibility into your interactions with our platform, focusing on three primary areas:

  1. User Usage Metrics
    • API Usage
    • Tokens Usage
  2. Deployed Models Metrics
    Each section provides valuable data over a specified date range, allowing for a granular analysis of your activities.

User Usage Metrics

Deployed Models Metrics

This area offers insights into the usage and performance of your deployed AI models within NextAI. Before accessing the metrics, you need to select the specific cluster where your models are deployed.

Cluster Selection

Since models can be deployed across different clusters for scalability and redundancy, selecting a cluster allows for focused analysis of model performance and resource utilization in that specific environment. Once a cluster is selected, you can view various metrics related to the models deployed in that cluster, such as the number of inference requests, processing times, and resource consumption.

Conclusion

By regularly monitoring these metrics, you can gain insights into your usage patterns, identify potential bottlenecks, and make informed decisions about scaling, optimizing, and budgeting for your AI projects. If you have any questions or need further assistance with interpreting your metrics, our support team is here to help.