AI-based System Service Level Indicator (SLI) Measure

From GM-RKB
Jump to navigation Jump to search

An AI-based System Service Level Indicator (SLI) Measure is a domain-specific SLI (SLI) that quantifies the quality and effectiveness of an AI system.

  • Context:
    • It enables organizations to monitor, benchmark, and enhance the reliability, accuracy, and operational efficiency of their AI-driven services, helping to maintain user trust and satisfaction.
    • It can range from being an Offline AI SLI Measure (evaluated during the training phase using historical data) and online measures (assessed in real-time operational environments with live data).
    • It can range from being a Proxy AI SLI Measure (indirect indicators of system performance, such as click-through rates) and direct real-world impact measures (actual outcomes, such as revenue generated or customer satisfaction scores).
    • It can help identify potential issues or areas for improvement in AI systems, such as data quality problems, model overfitting, or performance degradation over time.
    • It can inform decisions about when to retrain or update AI models, ensuring they remain accurate and relevant as data patterns and business needs evolve.
    • ...
  • Example(s):
    • Model Accuracy Measure, such as:
      • a Classification Accuracy that calculates the percentage of correct predictions made by an AI model, e.g., an image recognition system correctly identifying objects in 95% of cases,
      • a Regression Accuracy, such as Mean Squared Error (MSE) for regression models that measures the average of the squares of the errors—that is, the average squared difference between the estimated values and the actual value, e.g., a stock price prediction model with an MSE of 0.02, comprehensive view of model performance,
    • Model Performance Monitoring, such as:
      • Latency Measurement which records the time taken by an AI model to make a prediction, e.g., a fraud detection system flagging suspicious transactions within 100 milliseconds,
      • Throughput Measurement which evaluates the number of requests an AI model can handle per unit of time, e.g., a chatbot capable of processing 1,000 user queries per minute,
      • Resource Utilization Monitoring which tracks the computational resources consumed by an AI model, helping to optimize efficiency and costs,
    • Data Quality Metrics, such as:
      • Completeness Check which assesses whether all necessary data fields in input datasets are populated, e.g., ensuring customer records have no missing values,
      • Consistency Check which ensures that all data follows the same formats and adheres to the same rules, e.g., validating that date fields use a consistent YYYY-MM-DD format,
      • Accuracy Check which verifies that data values are correct and up-to-date, e.g., confirming customer addresses against a trusted external database,
    • Model Fairness Evaluation, such as:
      • Demographic Parity which compares the proportion of positive outcomes across different groups to detect potential bias, e.g., ensuring a loan approval model grants loans to male and female applicants at similar rates,
      • Equal Opportunity which assesses whether a model provides equal true positive rates for different groups, e.g., checking that a hiring AI selects qualified candidates from all ethnic backgrounds at comparable levels,
    • ...
  • Counter-Example(s):
  • See: Predictive Quality Analytics, Machine Learning Model Monitoring, Data Governance, Performance Indicator, Quality of Service (QoS), Responsible AI.


References


NOTOC