Setup
To capture the complete picture—including both the pipeline structure and the underlying token costs—we highly recommend initializing both the LlamaIndex integration and your specific LLM provider (e.g., OpenAI, Anthropic, Google GenAI).Usage
Once initialized, all LlamaIndex operations are captured automatically:What Gets Captured
| Attribute | Description |
|---|---|
| Document chunking | Each SentenceSplitter call as a span |
| Embedding calls | Each batch embedding with text inputs and vector outputs |
| Retrieval | Top-k retrieval with query and retrieved documents |
| LLM synthesis | Response generation with context and final answer |
| Tokens & Cost | Input/output tokens and calculated cost per LLM call |
| Latency | Duration of each pipeline stage |