Monitoring

Athina Monitoring

Advanced Monitoring & Analytics. Your production environment will thank you.

Watch Demo Video → (opens in a new tab)

Athina Monitor assists developers in several key areas:

  • Visibility: By logging prompt-response pairs using our SDK, you get complete visibility into your LLM touchpoints, allowing you to trace through and debug your retrievals and generations.
  • Usage Analytics: Athina will keep track of usage metrics like response time, cost, token usage, feedback, and more, regardless of which LLM you are using.
  • Query Topic Classification: Automatically classify user queries into topics to get detailed insights into popular subjects and AI performance per topic.
  • Granular Segmentation: You can segment your usage and performance metrics based on different metadata properties such as customer_id, prompt_slug, language_model_id, topic, and more to slice and dice your metrics.
  • Data Exports: Export your inferences to CSV or JSON for external analysis.

Here are some examples of how Athina can help you understand your LLM behavior:

💡
  • For queries related to Refunds, retrieval accuracy is only 55%.

  • 80% of user feedback is negative for customer ID nike-california-244

  • Your model's answer relevance is only 37% for prompt customer_support/v2.5

  • Your avg. response time is 6.7s for gpt-4, and 2.1s for gpt-3.5-turbo

  • You are spending an average of $381.44 per day on OpenAI inferences