Immersive Press Releases

Get latest news about us.
Immersive Fusion Open-Sources Trace Generator That Exposes the Gaps in Modern APM

Immersive Fusion Open-Sources Trace Generator That Exposes the Gaps in Modern APM

Dan Kowalski - 2026-03-08

Immersive Fusion proudly announces the open-source release of tracegen, a single-binary distributed trace generator that produces production-realistic OpenTelemetry data at a scale and complexity that pushes APM platforms to their visualization limits. Available today under the Apache 2.0 license, tracegen is both a practical testing tool and an invitation to rethink what observability should look like.

The Problem: APM Evolved Along the Wrong Axis

The APM industry optimized for data collection and query performance, not for how engineers actually experience their systems. The result: platforms that ingest enormous volumes of telemetry and render it as flat panels, waterfall charts, and paginated tables. Engineers pay per host, per GB, per seat, and what they get back is a 2D view of a fundamentally multi-dimensional problem.

Your application is a living distributed system: dozens of services, hundreds of pods, thousands of traces flowing through message queues, caches, databases, and now LLM inference chains. The data is there. The visualization has not kept up.

Meanwhile, the rise of AI workloads created an entirely new observability gap. LLM calls, agent loops, vector searches, and token budgets produce telemetry that most platforms surface in separate views rather than as part of a unified system topology. Specialized LLM observability tools (Langfuse, LangSmith, Helicone) emerged to fill this gap before a standard existed. The OpenTelemetry GenAI semantic conventions are now working to unify this space, but remain experimental, and traditional APM pipelines were not designed with LLM-specific telemetry in mind. The result: engineers still need multiple tools to observe different parts of the same system.

Immersive Fusion built tracegen to make these gaps visible.

28 services. 59 pods. 40 flows. One binary.

What tracegen Generates

tracegen simulates a complete e-commerce platform with AI capabilities, streaming real OTLP data to any backend:

  • 28 microservices across API gateway, authentication, order processing, payment, inventory, cart, product catalog, search, caching, notifications, email, shipping, tax, scheduler, user management, analytics, fraud detection, recommendations, configuration, and a web frontend, plus 8 dedicated AI/LLM services (LLM gateway, embeddings, vector DB, agent orchestration, content safety, model registry, feature store, data pipeline)
  • 59 pod instances distributed across 5 simulated AKS VMSS nodes with realistic Kubernetes resource attributes
  • 40 scenario flows covering checkout, payment processing, search, authentication, inventory sync, and 12 AI agentic patterns including RAG, multi-step agents, and conversational commerce
  • Realistic stack traces emulating .NET ecosystem libraries (Npgsql, StackExchange.Redis, Stripe SDK, OpenAI SDK)
  • Database, cache, and messaging spans with proper semantic conventions for PostgreSQL, Redis, RabbitMQ, and Kafka

One command. Zero containers. Point it at any OTLP-compatible endpoint, cloud or local:

tracegen -apikey YOUR_KEY -endpoint otlp.iapm.app:443
tracegen -apikey local -endpoint localhost:4317 -insecure

AI Agentic Tracing: Unified, Not Separated

While some APM vendors have begun adding AI observability features (Datadog LLM Observability, New Relic AI Monitoring), these typically appear as dedicated views alongside, rather than spatially integrated with, the distributed system topology. tracegen generates both traditional and AI traces as part of the same topology, because that is how they run in production.

All 12 AI scenarios emit spans following the OpenTelemetry GenAI semantic conventions (experimental, v0.x as of March 2026), using span naming conventions compatible with Microsoft Semantic Kernel and Agent Framework:

  • RAG pipelines: Embedding generation, vector search against Qdrant, LLM reranking
  • Multi-step agents: Plan/act/reflect loops with tool calls and iteration tracking
  • Content moderation: Parallel safety classifiers with guardrail enforcement
  • Conversational commerce: Multi-turn sessions with growing context windows

Each LLM span carries proper attributes: gen_ai.usage.input_tokens, gen_ai.usage.output_tokens, gen_ai.request.model, gen_ai.response.finish_reasons, and agent-specific fields like gen_ai.agent.id and gen_ai.tool.name.

In Immersive APM, these AI traces appear alongside traditional distributed traces in the same 3D environment. Services are rendered as spatial objects with edges encoding latency and throughput. AI agent loops, LLM inference chains, and vector searches are visible in context with the API calls and database queries that surround them. No tab switching. No separate dashboards. One unified view of the entire system.

AI agents and microservices, visualized together in real time

Chaos Engineering That Shows the Blast Radius

tracegen includes failure injection modes with a tunable error rate (0-10 scale):

  • Lost messages: Configurable message loss per queue hop (default 5%) simulating unreliable consumers
  • Dead consumers: Producers send messages that are never consumed (also available via -no-consumers flag)
  • Retry storms: Payment retries 3x with exponential backoff
  • Timeout cascades: Circuit breakers serving stale cache data
  • Saga compensation: 4-way parallel rollback after mid-transaction failure
  • LLM rate limits: OpenAI 429 errors with token budget details
  • Token budget exceeded: Agent exceeds iteration token limit mid-conversation
  • Hallucinated tool calls: Agents request tools that do not exist
  • Content filter blocks: Safety classifiers reject generated content

These are the failure modes that cause 3 AM pages. The question is not whether your APM platform detects them. The question is whether it shows you the cascade: which pods are affected, which downstream services are degrading, and where the blast radius ends. In Immersive APM, failure propagation is rendered spatially, with affected nodes, edges, and pods highlighted in the same 3D view where you see healthy traffic. tracegen generates the chaos so you can evaluate the difference.

Trace Generator Comparison (as of March 2026)

Capability tracegen telemetrygen Astronomy Shop HotROD
Single binary Yes Yes 15+ containers 4 containers
Services 28 1 ~22 4
Pod instances 59 0 1/svc 0
Scenario flows 40 0 ~10 1
AI agentic scenarios 12 No No No
OTel GenAI conventions Yes No No No
Tunable chaos (0-10) Yes No Fixed No
Startup time <1s <1s 3-5 min 30s

This table compares open-source trace generators, not APM platforms. Sources: telemetrygen, Astronomy Shop (OTel Demo), HotROD. Each tool is valuable for its intended scope. tracegen fills a different gap: generating the volume, variety, and failure complexity needed to stress-test how platforms visualize distributed systems.

Configurable Throughput

Aggressiveness levels from 1-10 control generation rate:

  • Level 1 (whisper): ~2 traces/second for careful evaluation
  • Level 5 (steady): ~7 traces/second for realistic production load
  • Level 10 (full blast): ~350 traces/second for pipeline stress testing (measured on a 4-core workstation with 8 GB RAM; results vary by hardware and network conditions)

Scenario filtering flags (-ai-only, -no-ai-backends, -no-consumers) let teams isolate specific workload types.

tracegen -apikey $KEY -level 10 -errors 10

Compatible With Every OTLP Backend

tracegen speaks OTLP over gRPC. Use it to evaluate any OpenTelemetry-compatible platform, including ours:

  • Immersive APM (3D real-time visualization)
  • Jaeger
  • Grafana Tempo
  • Honeycomb
  • New Relic
  • Datadog
  • Splunk Observability
  • Elastic APM
  • Any OpenTelemetry Collector

We built tracegen to be platform-agnostic because we believe the best way to demonstrate Immersive APM's spatial visualization is to let engineers compare it directly against whatever they use today.

Open Source. Apache 2.0. No Strings.

tracegen is available on GitHub under the Apache 2.0 license. Built in Go, it compiles to a single binary for Linux, macOS, and Windows (amd64 and arm64). Download the latest release, run it, fork it, contribute.

See What Observability Should Look Like

"We built tracegen because we needed the world to see what happens when you stop treating observability as a spreadsheet and start treating it as an experience. The 3D visualization is not a gimmick. When you render services as spatial objects, cluster pods by node, encode latency as edge thickness, and animate trace flow in real time, patterns emerge that flat dashboards structurally cannot show. That is the gap we are closing." - Dobri Kowalski, Founder, Immersive Fusion

28 services. 59 pods. 40 flows. AI agents. Chaos. All rendered in immersive 3D. This is what APM should look like.

Enter the World of Your Application®

About Immersive Fusion

Immersive Fusion LLC builds Immersive APM (IAPM), an application performance monitoring platform that renders distributed systems in real-time 3D. Founded on the belief that engineers deserve to see their applications the way those applications actually run, Immersive Fusion combines spatial visualization, AI-guided diagnostics, and full OpenTelemetry support to deliver observability that goes beyond dashboards.

Media Contact: Dobri Kowalski, dkowalski@immersivefusion.com

Download tracegen on GitHub. Free, open source, no signup required.

Start Free. Immersive. AI-guided. Full-stack observability. Enter the World of Your Application®.

Dan Kowalski

Father, technology aficionado, gamer, Gridmaster

About Immersive Fusion

Immersive Fusion (immersivefusion.com) is pioneering the next generation of observability by merging spatial computing and AI to make complex systems intuitive, interactive, and intelligent. As the creators of IAPM, we deliver solutions that combine web, 3D/VR, and AI technologies, empowering teams to visualize and troubleshoot their applications in entirely new ways. This approach enables rapid root-cause analysis, reduces downtime, and drives higher productivity—transforming observability from static dashboards into an immersive, intelligent experience. Learn more about or join Immersive Fusion on LinkedIn, Mastodon, X, YouTube, Facebook, Instagram, GitHub, Discord.

The Better Way to Monitor and Manage Your Software

Streamlined Setup

Simple integration

Cloud-native and open source friendly

Rapid Root Cause Analysis

Intuitive tooling

Find answers in a single glance. Know the health of your application

AI Powered

AI Assistant by your side

Unlock the power of AI for assistance and resolution

Intuitive Solutions

Conventional and Immersive

Expert tools for every user:
DevOps, SRE, Infra, Education

The Better Way to Monitor and Manage Your Software

A fusion of real-time data, immersive diagnostics, and AI Assistant that accelerate resolution.

Start Free