-
Notifications
You must be signed in to change notification settings - Fork 2
Update tracing documentation with new features #92
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -22,7 +22,7 @@ | |
|
|
||
| - **Observability / AI spans / request logs**: We capture standard OpenTelemetry traces and spans for LLM calls and related operations. | ||
| - **Agent runs / tools / function calls**: These appear as nested spans in the trace tree, with inputs/outputs when available. | ||
| - **Prompt/Completion pairs**: Extracted from common keys (`openinference.*`, `ai.prompt` / `ai.response`, `gen_ai.*`) so they can be turned into testcases and scored. | ||
|
|
||
| --- | ||
|
|
||
|
|
@@ -106,10 +106,40 @@ | |
|
|
||
| - **Time ranges**: 30m, 24h, 3d, 7d, 30d, All. | ||
| - **Project scope**: toggle between Current project and All projects. | ||
| - **SearchText**: full‑text across span/resource attributes (including prompt/response fields). | ||
| - **SearchText**: full‑text across span name, span kind, span ID, and all span attributes (including nested values in prompt/response fields). | ||
| - **Match previews**: quick context snippets with deep links to traces. | ||
| - **Cursor pagination**: efficient browsing with shareable URLs. | ||
|
|
||
| ### Conversation view | ||
|
|
||
| The **Conversation** tab displays AI interactions in a familiar chat-like format, making it easy to follow the flow of prompts and responses. This view: | ||
|
|
||
| - Extracts messages from `gen_ai` spans and renders them as user/assistant bubbles | ||
| - Shows system prompts in a distinct format | ||
| - Displays tool calls and their results inline | ||
| - Supports Claude Code traces with agent identification and rich tool rendering | ||
| - Automatically deduplicates messages across spans | ||
|
|
||
| Use the Conversation view to quickly understand what happened in a trace without navigating the span tree. | ||
|
|
||
| ### User annotations | ||
|
|
||
| Add human feedback directly to traces and individual spans using the **Annotations** feature. Annotations help you: | ||
|
|
||
| - **Rate interactions**: Give thumbs up/down ratings to mark good or problematic responses | ||
| - **Add comments**: Document observations, issues, or suggestions for improvement | ||
| - **Track feedback by span**: Annotate specific spans to pinpoint exactly where issues occurred | ||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The introduction and 'how-to' guide mention that annotations can be applied to both traces and spans, but this benefit focuses only on spans. To improve consistency and better reflect the feature's capabilities, consider rephrasing this to cover both trace and span level annotations. |
||
|
|
||
| To add an annotation: | ||
| 1. Open a trace and select a span (or view the trace overview) | ||
| 2. Expand the **Annotations** section | ||
| 3. Click **Add Annotation** | ||
| 4. Optionally select a thumbs up or thumbs down rating | ||
| 5. Add a comment describing your feedback | ||
| 6. Click **Submit** | ||
|
|
||
| Annotations are visible to your team and persist with the trace. Spans with annotations show a feedback indicator in the span tree for easy identification. | ||
|
|
||
| --- | ||
|
|
||
| ## Trace Grouping | ||
|
|
@@ -173,12 +203,12 @@ | |
|
|
||
| --- | ||
|
|
||
| ## Turn traces to testcases | ||
|
|
||
| Live traffic exposes edge-cases synthetic datasets miss. From any span that contains prompt/response attributes click **Create Testcase** and Scorecard will: | ||
|
|
||
| 1. Extract `openinference.*`, `ai.prompt` / `ai.response`, or `gen_ai.*` fields. | ||
| 2. Save the pair into a chosen **Testset**. | ||
| 3. Make it immediately available for offline evaluation runs. | ||
|
|
||
| Read more in [Trace to Testcase](/features/trace-to-testcase). | ||
|
|
@@ -209,9 +239,9 @@ | |
| - **CrewAI** – Multi-agent collaboration | ||
| - **Haystack** – Search and question-answering pipelines | ||
| - **LangChain** – Chains, agents, and tool calls | ||
| - **Langflow** – Visual workflow builder | ||
| - **LangGraph** – Multi-step workflows and state machines | ||
| - **LiteLLM** – Unified interface for 100+ LLMs | ||
| - **LlamaIndex** – RAG pipelines and document retrieval | ||
| - **[OpenAI Agents SDK](https://github.com/openai/openai-agents-python?tab=readme-ov-file#tracing)** – Assistants API and function calling | ||
| - **[Vercel AI SDK](https://ai-sdk.dev/providers/observability/scorecard)** – Full-stack AI applications | ||
|
|
@@ -229,11 +259,11 @@ | |
| - Cohere | ||
| - Google Gemini | ||
| - Google Vertex AI | ||
| - Groq | ||
| - HuggingFace | ||
| - IBM Watsonx AI | ||
| - Mistral AI | ||
| - Ollama | ||
| - OpenAI | ||
| - Replicate | ||
| - Together AI | ||
|
|
@@ -242,11 +272,11 @@ | |
| ### Vector Databases | ||
| - Chroma | ||
| - LanceDB | ||
| - Marqo | ||
| - Milvus | ||
| - Pinecone | ||
| - Qdrant | ||
| - Weaviate | ||
|
|
||
| <Info> | ||
| For the complete list of supported integrations, see the [OpenLLMetry repository](https://github.com/traceloop/openllmetry). All integrations are built on OpenTelemetry standards and maintained by the community. | ||
|
|
@@ -269,7 +299,7 @@ | |
| - **Debugging slow/failed requests with full span context** | ||
| - **Auditing prompts/completions for compliance** | ||
| - **Attributing token cost and latency to services/cohorts** | ||
| - **Building evaluation datasets from real traffic (Trace to Testcase)** | ||
|
|
||
| --- | ||
|
|
||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For improved clarity and consistency with the phrasing in step 4 of the 'how-to' list ('thumbs up or thumbs down'), you could make the terminology for ratings more explicit here.