| date | 2025-10-10 |
|---|---|
| id | mastra-observability-with-opentelemetry |
| title | Mastra Observability with SigNoz |
| description | Observability and monitoring for Mastra performance and usage with OpenTelemetry instrumentation to send traces to SigNoz |
This guide walks you through setting up observability and monitoring for Mastra using OpenTelemetry and exporting traces to SigNoz. With this integration, you can observe model performance, capture request/response details, giving you real-time visibility into latency, error rates, and usage trends for your Mastra applications.
Instrumenting Mastra in your AI applications with telemetry ensures full observability across your AI workflows, making it easier to debug issues, optimize performance, and understand user interactions.
- A SigNoz Cloud account with an active ingestion key
- Internet access to send telemetry data to SigNoz Cloud
- Node.js (v20.0 or higher)
- An API key from a supported Model Provider
You can get the full Mastra installation instructions here.
Step 1: Start the CLI Wizard
Run the following command to start the interactive setup:
```bash npx create-mastra@latest ``` ```bash yarn dlx create-mastra@latest ``` ```bash pnpm create mastra@latest ``` ```bash bun create mastra@latest ```Step 2: Add Your API key
Add your API key to the .env file:
OPENAI_API_KEY=<your-api-key>
This example uses OpenAI. Each LLM provider uses a unique name. See Model Capabilities for more information.
Step 1: Create an agent in your project.
For example, create a file src/mastra/agents/chefAgent.ts:
import { openai } from "@ai-sdk/openai";
import { Agent } from "@mastra/core/agent";
export const chefAgent = new Agent({
name: "chef-agent",
instructions:
"You are Michel, a practical and experienced home chef" +
"You help people cook with whatever ingredients they have available.",
model: openai("gpt-4o-mini"),
});Step 2: Register the Agent with Mastra
In your src/mastra/index.ts file, register the agent with telemetry enabled:
import { Mastra } from "@mastra/core";
import { chefAgent } from "./agents/chefAgent";
import { OtelExporter } from "@mastra/otel-exporter";
export const mastra = new Mastra({
agents: { chefAgent },
observability: {
configs: {
otel: {
serviceName: '<service-name>',
exporters: [
new OtelExporter({
provider: {
signoz: {
apiKey: process.env.SIGNOZ_INGESTION_KEY,
region: 'us', // 'us' | 'eu' | 'in',
}
},
})
],
},
},
},
});<service_name>is the name of your serviceSIGNOZ_INGESTION_KEY→ Your SigNoz ingestion key. Set the value as an environment variable.region→ appropriate region
Step 3: Run Mastra Dev Server and Interact with Agent
npm run devYou should see some output similar to this:
Visit the playground url, and start chatting with the registered agent. All your interactions should be traced and sent to SigNoz under the traces tab.
Your Mastra interactions should now automatically emit traces.
You should be able to view traces in Signoz Cloud under the traces tab:
When you click on a trace in SigNoz, you'll see a detailed view of the trace, including all associated spans, along with their events and attributes.
You can also check out our custom Mastra dashboard here which provides specialized visualizations for monitoring your Masrta usage in applications. The dashboard includes pre-built charts specifically tailored for LLM usage, along with import instructions to get started quickly.