Skip to Main Content
Talk Intermediate

Why LLM Observability is the New Frontier in AI Safety

Rejected
Session Description

Large Language Models (LLMs) are no longer just research toys, they're in production, making decisions, influencing users, and shaping outcomes at scale. But here's the hard truth: you can't trust what you can't see. LLMs can fail silently, hallucinate with confidence, and degrade without warning. "Trust but verify" isn't optional it's the new baseline for responsible AI.

This session cuts to the core of the issue: traditional observability wasn't built for non-deterministic, black-box systems like LLMs. We'll explore why LLM observability is fundamentally different and how we can build the tools to meet that challenge.

Open source is leading the charge toward AI transparency. If you're deploying LLMs into Kubernetes-native stacks or building safety-critical AI services, this talk will give you a roadmap to make your models observable, measurable, and verifiable just like the rest of your infrastructure.

Key Takeaways

A key part of the solution lies in OpenTelemetry, the open-source standard for instrumenting distributed systems. By extending its tracing and metrics pipelines to capture LLM-specific signals like prompt/response payloads, latency, cost, token usage, and model confidence, we can build visibility into how these models behave across real-world workflows.

We'll explore:

  • How OpenTelemetry enables structured tracing of LLM chains, across micro-services and vector stores.

  • What to log and trace for LLMs prompt/response deltas, temperature drift, user feedback, error modes, and more.

  • How to pair OpenTelemetry with Prometheus, Grafana, and CNCF-native tools to create actionable observability dashboards.

  • Real-world patterns for monitoring prompt injections, data leakage, and performance regressions in production LLM apps.

References

Session Categories

Technology architecture
Which track are you applying for?
Main track

Speakers

Swapnil Kulkarni Observability | Kloudfuse

Swapnil is a passionate technologist and a cloud native leader, with 16+ years of experience in cloud computing and enterprise software architecture helping customers achieve their goals and solve their challenges.

Swapnil is SME for all public cloud offerings AWS, GCP, Azure & IBM Cloud, Kubernetes, and Containerization ecosystem, and has worked with multiple product teams for security, compliance, and operational excellence of the cloud SaaS platform. Swapnil has contributed to open source projects like OpenStack, Docker, and Kubernetes, and received the OpenStack Community Contributor Award. Swapnil believes in enabling and empowering organizations to adopt cloud-native technologies and practices, and to foster a culture of collaboration, innovation, and excellence.

Swapnil is currently leading Customer Success & Solution Architecture at Kloudfuse, a cloud-native platform that enables businesses to observe infrastructure & applications on any cloud platform

Swapnil Kulkarni
https://www.linkedin.com/in/meswapnilk/

Reviews

Reviewer #1 Approved

The proposal is detailed, and it discusses solving a timely problem using FOSS tools and standards. It is, however, a niche topic and might not be broadly relevant to the IndiaFOSS audience.

Reviewer #2 Not Sure

Thank you for submitting your proposal for IndiaFOSS 2025. Your submission was well-received and progressed to our final review stages.

Unfortunately, due to the high volume of excellent proposals this year, we were unable to select your talk for the final program. We appreciate the effort you put into your submission and encourage you to apply again for future events.

Reviewer #3 Rejected