Redesigning LLM Observability for Developers

Founding Product Designer. Helped raise 3M in Seed and reach product market fit. Led end-to-end design of Helicone's core platform, reducing time-to-insight from minutes to seconds.

April 23, 2024
/ User ResearchProduct DesignGrowthContent Strategy
Screenshot of Helicone's redesigned request page showing a clean interface with request and response data

Challenge

Designing for Developers = Designing for Productivity

I joined Helicone (YC W23) as a founding designer to reimagine the core experience of LLM observability from scratch. As the company grew, we needed to redesign the core platform to support enterprise clients.

Role & Timeline

Founding Product Designer
March 2024 - Present

Contributions

User research, workflow redesign, and content strategy.

Team

Justin Torre (CEO), Cole Gottdank (CTO), Kavin Desi Valli (Engineer)

Product & GTM Impact

After the redesign, we saw a consistent 10-20% MoM growth in signups, a 3.6% increase in free to paid conversion, and doubled organic traffic in one week through content strategy.

Team feedback and testimonials about the product design process

Philosophy

Prioritize, Ship Fast, Ship Often

We anchored our design strategy around three core principles to accelerate user value and business outcomes:

Improve Core Features

Double down on the UX of the most valued features.

Remove Biggest Friction

Make it easy to get started and shorten time-to-insight.

Establish GTM Channels

Create content, YouTube partnerships and engineering-as-marketing.

Before we go on...

The rest of the case study covers wonderful research, creative problem solving, and architectural redesigns of the core platform. But if you're in a time crunch, feel free to:

Request Redesign

Having Strong Opinions About the Core Information We Show

Comprehensive view of Helicone's request page showing the main interface with request details and analytics

Problem

As the product grew, the complexity was working against developer productivity rather than supporting it.

The request page is Helicone's most visited page —live in product—where users can see and analyze individual LLM interactions. I focused the redesign on surfacing only the most essential debugging information, cutting noise and aligning with real usage patterns.

A simple audit reveals that the drawer needed to surface the right information without overwhelming users, and look consistent with the rest of the product.

Design audit of the original request drawer showing areas for improvement in layout and information hierarchy

How might we display the most critical data for developers who need quick observability answers?

After speaking to 5 companies with top usages of Helicone, I found that:

90% Use Case

Users spend 90% of their time in the request drawer reading the request and response bodies.

10% Use Case

Users spend 10% of their time tracing back to sessions for context.

Technical Constraint

The request drawer needed to work both as a standalone component and within the request page.

Solution

Defaultly Hide Metadata, Shortcuts to Session/Prompts

By hiding metadata by default and simplifying the drawer, we helped developers cut time-to-insight by 60% and focus immediately on key debugging info - the request and response pair.

Side-by-side comparison of Helicone's request page in dark and light modes, highlighting the interface's adaptability

Edge Cases

While most users log only LLM calls today, the future of AI applications is multimodal — increasingly dependent on tool and vector calls.

LLM calls focus on prompt/response pairs, while tool calls need parameter validation and vector queries require similarity scoring visualization. Thus they required a different type of visualization than LLM calls.

Examples of vector and tool calls in Helicone's interface, showing different visualization patterns for various request types

Interactions

By default, metadata is hidden so developers can focus on request/response bodies, with full details accessible when needed. I designed this progressive disclosure approach to match how developers actually debug: as they scroll through responses, non-essential UI elements fade away, keeping focus on the core data.

Interactive behaviors of the request drawer showing progressive disclosure of metadata and focus states

Accessibility

Since most developers work in dark-themed environments (VS Code, terminals, dashboards), dark mode was essential for Helicone to fit seamlessly into their workflows without jarring visual interruptions that bright interfaces create.

Comparison of Helicone's interface in light and dark modes, demonstrating accessibility considerations

Sessions Redesign

Help Developers Pinpoint Errors in Complex LLM Workflows

With the request page simplified, I redesigned the Sessions view to solve our second biggest user pain point.

High-fidelity mockup of Helicone's sessions view showing the redesigned interface for tracking LLM workflows

User Research

90% of users follow this debugging workflow: pinpoint errors, understand surrounding context, aggregate metrics for deeper LLM application analysis.

User interview insights and feedback about the sessions feature, displayed in a structured format

This revealed that users spend most of their time on steps 1 and 2, which we were well-positioned to optimize. Our interface needed to be opinionated about surfacing key insights upfront, answering critical questions like "What's the average session cost?" and "What's the typical latency?" without forcing developers to dig for answers.

How might we guide developers to the root cause error in their LLM workflows containing many sequential requests?

Before

Initial view of the sessions page showing the original interface design

Detailed view of a session showing the original interface for analyzing LLM interactions

Insights

What questions do developers want to ask their LLM data?

While engineering wanted a list of bug reports and performance improvements, from a product/marketing perspective, we want to allow developers to navigate from a request to Sessions easily in order to upsell on advanced features. In order to create a smooth upsell experience, we needed to ensure developers get the most value out of the Sessions page.

Where did the error happen?

Developers want to be able to pinpoint errors in their LLM workflows.

What's the length of an average session?

Having all the data in one place, Helicone is well-positioned to help answer aggregate questions.

What was the most expensive request?

Developers want to be able to find the bottleneck in their LLM workflows and optimize costs.

Solution

Error Indicator, Aggregate Metrics to Answer Questions Easily

"It's nice to be able to collapse everything and then determine, 'Okay, we think the error is in this part,' and expand that and look there. Sessions really helps us locate where the issue is using the paths."

Founding Engineer at Tusk

Filter Improvement

78% of debugging time was spent not on fixing issues, but on simply finding the right request.

This insight shifted our design from better analytics to better filtering, designed to help developers find the right request faster and ultimately reducing time-to-insight by 60%.

Top-of-Funnel

Leading Developer-First Content Strategy

Mapping out the user journey, I found that developers at different stages of their journey with Helicone need different support. The most pressing need was to increase awareness of the product.

Visual representation of Helicone's user journey from discovery to implementation

Designing a Scalable Content System to Drive Conversion

Helicone was winning word-of-mouth in the dev community, but our traffic did not convert into signups. After auditing 50+ dev-first content sites (e.g. Vercel, Supabase, LangChain), I created a modular content system that allowed us to create consistent, high-quality content that converts.

Capturing Customers at Scale with Engineering-as-Marketing Tools

We built high-utility tools like LLM comparison tool, API cost calculator, open stats and provider status page, turning fun engineering pages into scalable marketing assets that brings in qualified traffic. This approach helped us establish trust with technical audiences, and drive top-of-funnel growth organically.

Impact

Our content strategy helped Helicone 2x traffic MoM, with a 3.9% conversion rate on average, and helped Helicone become the most visible brand in AI search.

Analytics dashboard showing blog statistics and content performance metrics

Key Learnings

Building systems that multiply effort

Rather than just delivering designs, I created branding guidelines, asset libraries, and SEO checklists to help the team succeed.

Empowering the team to share UX insights

Great design isn't solo. Our most impactful features often came from enabling engineers to contribute UX insights, especially around performance.

Great developer experience prioritizes productivity

This meant designing flexible interfaces that adapted to different debugging workflows and sometimes letting developers find their own paths to solutions.

Team collaboration and workflow diagram showing how design and development teams work together at Helicone

Next Steps

While the redesign achieved significant improvements in developer productivity, the next phase focuses on systematic testing and measurement to drive continuous iteration.

  1. Watch session recordings on the request drawer and sessions to validate that users actually spend 90% of time on request/response bodies.
  2. Track average time from landing on request page to finding target error request.
  3. Conduct monthly interviews with 5-8 power users to track workflow evolution.

Curious to learn more? Reach out to me.