Get Started

Overview

Nora provides end-to-end visibility into modern LLM applications by automatically tracking LLM calls, agent workflows, and tool executions.

Modern AI systems are no longer simple prompt–response interactions. They involve multi-step workflows, agent coordination, routing decisions, tool usage, and streaming responses.

Nora makes these executions inspectable and debuggable by capturing not only what happened, but why it happened — without invasive or manual instrumentation.

Key Features

  • Automatic tracking of OpenAI, Anthropic, and Google Gemini API calls

  • First-class decision tracking (tool selection, routing, RAG choices)

  • Full async/await and streaming support

  • Trace groups for organizing multi-step logical workflows

  • Custom function tracking via lightweight decorators

  • Session- and user-level trace correlation

  • Automatic tool and function calling detection

  • Token usage, latency, and error capture

  • Multiple span types for precise operation classification

Built for Production AI Systems

Nora is designed for real-world production environments:

  • Low-overhead batching and async flushing

  • Safe for FastAPI, background tasks, and streaming servers

  • Works with synchronous, async, and parallel execution models

  • Captures partial failures and errors as structured traces

This makes Nora suitable not only for experimentation, but for monitoring real user traffic at scale.

Supported Providers

  • OpenAI: GPT-5, GPT-4, GPT-3.5 (synchronous, asynchronous, streaming)

  • Anthropic: Claude Opus, Sonnet (synchronous, asynchronous, streaming)

  • Google Gemini: Gemini Pro, Flash (synchronous, asynchronous, streaming)

Was this page helpful?