Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
74 changes: 52 additions & 22 deletions contents/teams/replay/objectives.mdx
Original file line number Diff line number Diff line change
@@ -1,33 +1,63 @@
## 🚀 Goal 1: Get Hayne rocketing
## 🧠 Goal 1: Nail single-session on-demand summaries

It's great to have Hayne on the team, let's make the time and priority to make it a success!
Ship a reliable, on-demand AI summary for individual sessions. Users should be able to hit a button and get a useful natural-language summary of what happened in a session — key user actions, errors, frustration signals, and outcomes.

## 👀 Goal 2: Make Replay low-maintenance 🔧
- Wire up the rasterizer → AI pipeline end-to-end
- Iterate on summary quality (prompt tuning, frame selection, context enrichment)
- Configurable customer-specific context (let customers provide their own domain context to make summaries more relevant)
- Measure latency and cost per summary, set targets for both

* Clean up legacy features and finish outstanding migrations from old to new
* Move everyone to PostHog recorder
* Remove support for Blobby V1
* Remove LTS feature and related Celery tasks
* Improve fidelity - let's squash the filtering bugs and visual inconsistencies that make Replay seem "rough" to use
* Improve debugging facilities - it can be hard to reproduce issues locally or get our hands on the data we need, let's build the internal tooling needed to solve that
## 🎬 Goal 2: Scale video export pipeline

## ⌚ Goal 3: Replay everywhere!
The rasterizer can render sessions to video — now make it production-ready at scale.

* When a recording doesn't exist, explain why!
* Make the view recording buttons consistent
* Make the view recording buttons session aware
* Out with modals, in with tabs!
* Where else can we link to recordings from within other products?
* External integrations - jam.dev? GitHub? ZenDesk? Linear? FreshDesk? Jira? Trello?
- Harden the Temporal worker for reliability and throughput
- Optimize resource usage (browser concurrency, memory, CPU)
- Explore GPU acceleration for further speedup

## 🏎️ Goal 4: Make Replay compliant
## 🏷️ Goal 3: Build session categorizer

* Let's shred some recordings 🤘
Train a binary classifier that can label sessions as "interesting" or not. Use this model to power a categorizer that groups sessions by properties (rage clicks, errors, conversion flows, etc.) but only surfaces the ones the model flags as interesting.

## 💰 Goal 5: How are our unit economics trending?
- Extend the session summarizer (Goal 1) to also classify sessions as interesting/not — this becomes our training dataset
- Train and evaluate the binary classifier
- Build the categorizer layer on top: group by session properties, filter by interestingness
- Run classifier on a schedule

How have recent infrastructure changes impacted our unit economics? Is our pricing still competitive?
## 🖥️ Goal 4: Build new AI-native frontend for Replay

## 👨‍🔬 Goal 6. Make "show me something interesting" tangible
Reimagine the Replay UI around AI-first workflows. Move away from recency as the primary way to rank recordings — instead, leverage the session categorizer (Goal 3) to surface interesting recordings across categories.
Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

😻
I've heard of this need a lot in all sorts of user interviews.

Note: A major part of this job will be served by the signals inbox, and so in @PostHog/team-signals we'll build an integration with session summarization to surface concretely fixable problems (visual errors and the like, pretty much).

But people will still be watching recordings manually to understand users intimately, so categorization will matter both in the UI and for agents.


Customers want us to highlight interesting moments and sessions, but that means different things to different people. Can we generate a library of what Interesting means to our customers?
- Augment the chronological recording list with category-driven views
- Surface AI-generated summaries and labels directly in the list UI
- Design the UX for exploring sessions by category rather than scrolling by time

## 🔌 Goal 5: Basic MCP for Session Replay

Expose Session Replay as MCP tools so AI agents (Claude Desktop, etc.) can search, retrieve, and summarize sessions programmatically.
Copy link
Copy Markdown
Member

@Twixes Twixes Apr 2, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I believe search and retrieve have already been exposed as MCP tools by @VojtechBartos. And as for "summarize session", it's in fact key for us in @PostHog/team-signals for research of errors - so I'll likely build a targeted version of this (specifically "summarize part of session") 2-4 weeks from now – will certainly need your help guys.

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's in PR, hopefully merged soon 🚀 PostHog/posthog#51757


## 💰 Goal 6: Business model for AI features

Figure out how to price and sustain AI-powered Replay features.

- Map out costs per session for summaries, categorization, and video export
- Identify break-even points at different usage tiers
- Propose a pricing model that works for customers and for us

## 🔒 Goal 7: Map out PII, data retention, and IP concerns

Before we begin using machine learning we need to figure out the compliance, legal, and brand concerns around training future models on recording data.

- What PII risks exist when sending session data to LLMs?
- How do data retention policies interact with AI-generated artifacts?
- What are the IP implications of training on customer recordings?
- What do customers expect and what do we need to communicate?
- What does s-tier consent look like?

## 🗂️ Goal 8: Set up data labelling system for replays

Build the infrastructure for labelling replay data at scale — needed to train and evaluate future models on session data.

- Evaluate labelling tools / build internal tooling
- Design the labelling workflow and quality controls
- Start building a labelled dataset
Loading