<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Netflix TechBlog - Medium]]></title>
        <description><![CDATA[Learn about Netflix’s world class engineering efforts, company culture, product developments and more. - Medium]]></description>
        <link>https://netflixtechblog.com?source=rss----2615bd06b42e---4</link>
        <image>
            <url>https://cdn-images-1.medium.com/proxy/1*TGH72Nnw24QL3iV9IOm4VA.png</url>
            <title>Netflix TechBlog - Medium</title>
            <link>https://netflixtechblog.com?source=rss----2615bd06b42e---4</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sat, 16 May 2026 22:46:33 GMT</lastBuildDate>
        <atom:link href="https://netflixtechblog.com/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Scaling ArchUnit with Nebula ArchRules]]></title>
            <link>https://netflixtechblog.com/scaling-archunit-with-nebula-archrules-b4642c464c5a?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/b4642c464c5a</guid>
            <category><![CDATA[gradle]]></category>
            <category><![CDATA[nebula]]></category>
            <category><![CDATA[java]]></category>
            <category><![CDATA[archunit]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 08 May 2026 15:55:59 GMT</pubDate>
            <atom:updated>2026-05-08T15:56:00.919Z</atom:updated>
            <content:encoded><![CDATA[<p>By <a href="https://github.com/wakingrufus">John Burns</a> and <a href="https://www.linkedin.com/in/emilyyuan03/">Emily Yuan</a></p><h3>Introduction</h3><p>At Netflix, we operate using a <a href="https://netflixtechblog.com/towards-true-continuous-integration-distributed-repositories-and-dependencies-2a2e3108c051">polyrepo</a> strategy with tens of thousands of Java repositories. This means that we need to have ways of sharing common build logic across these repositories. On the <a href="https://sites.google.com/netflix.com/javaplatformnetflix/jvm">JVM Ecosystem team</a> within Java Platform, we build tooling such as the <a href="https://github.com/nebula-plugins">Nebula suite of Gradle plugins</a> to provide standard ways to build projects, keep dependencies up-to-date, and publish artifacts reliably across the Java ecosystem. Our mission also entails providing build-time feedback to the developer when they deviate from the <a href="https://netflixtechblog.com/how-we-build-code-at-netflix-c5d9bd727f15">paved road</a>, or when their code base contains technical debt.</p><h3>Case Study</h3><p>After a Netflix incident relating to a library releasing a backwards-incompatible change, our team was asked to provide some tooling and practices to improve the Java library lifecycle management. This was not a simple case of a library making a reckless breaking change. The code removed had been deprecated for years. Library authors often struggle to know when it is safe to remove deprecated code, or refactor code that is not meant to be used by downstream applications. Fleet-wide migrations, such as upgrading major Spring Boot versions, also involve deprecated code removal. To help with this, we established a suite of API lifecycle annotations:</p><ul><li>@Deprecated from the Java standard library</li><li>@Public A custom annotation to use on APIs meant to be used downstream</li><li>@Experimental A custom annotation for new APIs which may not yet be stable</li><li>All other APIs are assumed to be “internal”</li></ul><p>Library authors can annotate their APIs with these annotations. However, how will they know which downstream projects are using their API incorrectly, based on these?</p><p>As we sought to improve the paved road for JVM-based libraries at Netflix, we needed a good way of identifying this kind of technical debt, not only for the benefit of the Java Platform-provided libraries, but any team delivering shared libraries to the organization. For this, we looked at ArchUnit.</p><p><a href="https://www.archunit.org/">ArchUnit</a> is a popular OSS library (3.5k stars, 84 contributors) used to enforce “architectural” code rules as part of a JUnit suite. It is used internally by Gradle, Spring, and is provided as part of the <a href="https://spring.io/projects/spring-modulith">Spring Modulith</a> platform. The rules engine, which is built directly on top of <a href="https://asm.ow2.io/">ASM</a>, can be used for a wide variety of use cases. It is powerful enough to be a general purpose static analysis tool with the following distinctive features:</p><p>1. Works cross-language (JVM), because it uses ASM/bytecode, not AST parsing.</p><p>2. Exposes a builder API pattern that makes it easy to write rules</p><p>3. Also has a lower level API ideal for writing more complex custom rules.</p><p>The limitation of ArchUnit is that it is designed to be used as part of a JUnit suite in a single repository. The Nebula ArchRules plugins give organizations the ability to share and apply rules across any number of repositories. Rules can be sourced from OSS libraries or private internal libraries. This makes the plugin generally useful for any JVM+Gradle engineering organization.</p><h3>Why ArchUnit?</h3><p>Before we go into how ArchRules works, it is good to understand why we would want to use ArchUnit in this way instead of other static analysis tools.</p><h4>AST vs Bytecode</h4><p>Some tools, such as PMD, process rules against an AST (abstract syntax tree). An AST is a structured representation of source code. This kind of tool will have rules that are syntax dependent. Rules that need to support multiple JVM languages, such as Kotlin or Scala, often need to be rewritten for each language. It also allows code which should be found to be hidden under syntactic sugar not anticipated by the rule author. ArchUnit uses <a href="https://asm.ow2.io/">ASM</a> to analyze actual compiled bytecode, which means it doesn’t matter how that code was produced. What is analyzed is the actual code that will be run.</p><h4>Rule Authorship</h4><p>Tools like PMD and Spotbugs are not optimized for custom rule authorships. Most usage of these tools run built-in provided rules, or add in pre-made third party plugins. Take a look at what a custom rule for PMD might look like:</p><pre>&lt;![CDATA[<br> //AllocationExpression/ClassOrInterfaceType[<br>   @Image=&#39;DateTime&#39; and (<br>       (count(..//Name[@Image=&#39;DateTimeZone.UTC&#39;])&lt;=0)<br>       and<br>       (count(..//Name[@Image=&#39;DateTimeZone.forID&#39;])&lt;=0)<br>    ) or (<br>       (<br>           (count(..//Name[@Image=&#39;DateTimeZone.UTC&#39;])&gt;0)<br>             or<br>           (count(..//Name[@Image=&#39;DateTimeZone.forID&#39;])&gt;0)<br>       ) and (../Arguments/ArgumentList and count(../Arguments/ArgumentList/Expression) = 1)<br>   )<br> ]<br>]]&gt;</pre><p>This rule ensures that DateTimes are not instantiated without an explicit zone. This is a raw string meant to be used within PMD’s xpath parser. There is no IDE guidance on crafting it. To test it, a whole separate PMD process needs to be wired up to interpret the rule and evaluate it against a source file. Let’s see how a similar rule would look with ArchUnit:</p><pre>ArchRuleDefinition.priority(Priority.MEDIUM)<br>.noClasses()<br>.should()<br>.callConstructorWhere(<br>    // constructor does not have a zone arguement<br>    target(doesNot(have(rawParameterTypes(DateTimeZone.class))))<br>   // constructor is for DateTime<br>        .and(targetOwner(assignableTo(DateTime.class)))<br>)</pre><p>This is type-safe Java code with a fluent API. It is also simple to unit test, as ArchUnit has a method to pass a rule object and class references to evaluate the rule against those classes.</p><h4>Class Relations</h4><p>Because ArchUnit processes the entire classpath with ASM, it retains a graph of the class data, allowing rules to easily traverse class relationships and call sites. This allows rules to have much more context about the code it is evaluating.</p><h3>Rules Libraries</h3><p>The first step was to build the ability to write ArchUnit rules which can be shared and published. In order to do this, we have the <a href="https://github.com/nebula-plugins/nebula-archrules-plugin?tab=readme-ov-file#authoring-rules">ArchRules Library Plugin</a>. This plugin adds an additional source set to your Gradle project called archRules. In this source set, you can create a class which implements the ArchRulesService interface. This interface has a single abstract method which returns a Map&lt;String, ArchRule&gt;. The keys of this map are the names of your rules, and the ArchRule is the rule you would like to define using the standard ArchUnit API. Here is an example:</p><pre>public class GuavaRules implements ArchRulesService {<br>  static final ArchRule OPTIONAL = ArchRuleDefinition.priority(Priority.MEDIUM)<br>        .noClasses()<br>        .should()<br>        .dependOnClassesThat()<br>        .haveFullyQualifiedName(&quot;com.google.common.base.Optional&quot;)<br>        .because(&quot;Java Optional is preferred over Guava Optional&quot;);<br><br>    @Override<br>    public Map&lt;String, ArchRule&gt; getRules() {<br>        Map&lt;String, ArchRule&gt; rules = new HashMap&lt;&gt;();<br>        rules.put(&quot;guava optional&quot;, OPTIONAL);<br>        return rules;<br>    }<br>}</pre><p>This code and its dependencies will not be bundled with your main code. It is bundled into a separate Jar with the arch-rules classifier. When publishing, your library will publish this jar as a separate variant with the usage attribute set to arch-rules. This means that in order for downstream projects to use these rules, they must use <a href="https://docs.gradle.org/current/userguide/publishing_gradle_module_metadata.html">Gradle Module Metadata</a> for dependency resolution. There are 2 flavors of rules Libraries: Standalone rules libraries, bundled rule libraries.</p><h4>Standalone Rule Libraries</h4><p>A Standalone Rule library contains no main code: only archRules. These are useful for defining rules for code you don’t own, such as Core Java APIs or OSS libraries. They are also useful for generic rules that can apply to any code, such as “don’t use code marked as @Deprecated”. We maintain a <a href="https://github.com/nebula-plugins/nebula-archrules">collection</a> of OSS Standalone rule libraries which anyone is free to use, and serve as examples of the types of rules you may want to write yourself. However, the real power of ArchRules is in “bundled rule libraries”.</p><h4>Bundled Rule Libraries</h4><p>A bundled rule library is a library with both main and archRules sources. The main source set will contain useful library code, whatever it may be. The archRules will contain rules specific to the usage of that library. For example, rules scoped to that library’s package, or referencing that library’s specific API. Whenever possible, we recommend writing rules in this bundled way. That is because the <a href="https://github.com/nebula-plugins/nebula-archrules-plugin?tab=readme-ov-file#running-rules">ArchRules Runner Plugin</a> will be able to automatically detect these rules and run them in only the source sets that use this library as a dependency. An example of this can be seen in our <a href="https://github.com/nebula-plugins/nebula-test/blob/main/src/archRules/java/com/netflix/nebula/test/archrules/NebulaTestArchRules.java">Nebula Test</a> library.</p><p>In any case, the library plugin will automatically generate a service loader registration entry for your ArchRulesService so that the runner can discover your rules.</p><h3>Running Rules</h3><p>The <a href="https://github.com/nebula-plugins/nebula-archrules-plugin?tab=readme-ov-file#running-rules">ArchRules Runner Plugin</a> allows rules to be evaluated against your code. Standalone rule libraries can be evaluated against all source sets by adding them to the archRules configuration in your build. For example:</p><pre>dependencies {<br>    archRules(&quot;your:rules:1.0.0&quot;)<br>}</pre><p>As mentioned before, bundled rules will be evaluated automatically. To do this, the runner plugin creates a separate configuration for each of your source sets. In each of these configurations, the archRules classpath is combined with the runtimeClasspath with the arch-rules variant selected. This configuration is the classpath used when the ServiceLoader discovers implementations of ArchRulesService. In the following example, we have a Project which uses a test helper library as a testImplementation dependency, and also adds a standalone rules library to the archRules configuration. The test runtime classpath will only contain the implementation jar for the helper library, but the arch rules runtime will contain the archrules jar for the bundled rules and standalone rules. This all happens automatically.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*pWrmMaPCSm3XRIsTyMEpgQ.png" /><figcaption>Gradle configurations used by ArchRules</figcaption></figure><p>Once the rules classpath is determined, the runner plugin will create a Gradle work action to evaluate rules against that specific source set. This action runs with classpath isolation using the *archRuleRuntime configuration. Within this action, a ServiceLoader is used to discover rule definitions. The action ends by writing a binary serialization of rule violations to a file for reporting.</p><p>In a project running rules, you also have the ability to customize rule configurations using the archRules extension. For example, you can override a rule’s priority level:</p><pre>archRules {<br>    ruleClass(&quot;com.netflix.nebula.archrules.deprecation&quot;) {<br>        priority(&quot;HIGH&quot;)<br>    }<br>}</pre><p>Other <a href="https://github.com/nebula-plugins/nebula-archrules-plugin?tab=readme-ov-file#running-rules">customizations</a> include disabling running rules on certain source sets and configuring the failure threshold (i.e., high priority failures will cause the build to fail).</p><h3>Reporting</h3><p>The ArchRules runner plugin has two built-in reports: JSON and console. The json report will collect the output from all source sets within a project and create a single json file with all of the data. The console report also collects the output from all source sets within a project, but it prints to the console an easy to read report, for example:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*BJ2ONrMNHEFsBEks3WMPaQ.png" /><figcaption>Console Report output</figcaption></figure><p>Note that failure details feature a detailed plain English description, along with a pointer to the exact line of code in violation.</p><p>For custom reporting, you can either use the JSON file, or create your own task that reads the binary files. Take a look at the source code for the ArchRules runner plugin’s report tasks for an example of how to do this.</p><h3>Case Study Solution</h3><p>Going back to our original problem, using ArchRules, we were able to deliver a platform for library authors to track the usage of their APIs. They write ArchRules to detect usage of the annotations, scoped to their library’s package, such as:</p><pre>ArchRuleDefinition.priority(Priority.MEDIUM)<br>    .noClasses().that(resideOutsideOfPackage(packageName + &quot;..&quot;))<br>    .should()<br>    .dependOnClassesThat(resideInAPackage(packageName + &quot;..&quot;).and(are(deprecated())))<br>    .orShould().accessTargetWhere(targetOwner(resideInAPackage(packageName + &quot;..&quot;))<br>        .and(target(is(deprecated())).or(targetOwner(is(deprecated())))))<br>    .allowEmptyShould(true)<br>    .because(&quot;Deprecated APIs are subject to removal&quot;);</pre><p>NB: the deprecated() predicate comes from <a href="https://github.com/nebula-plugins/nebula-archrules/blob/main/archrules-common/src/main/java/com/netflix/nebula/archrules/common/CanBeAnnotated.java">nebula-archrules</a>.</p><p>Our internal Nebula standard Gradle wrapper and plugin suite automatically enable the ArchRules runner on every project, and provides a custom reporter which sends the report data to our Internal Developer Portal on every main-branch CI build. This way, library authors can easily see a report of all downstream consumers using their experimental, deprecated, or non-public APIs, giving them confidence to make “breaking” changes, knowing that it will not actually break downstream consumers. If their changes are currently blocked by downstream usage, they can easily see exactly which projects are reporting those usages.</p><h3>OSS Rule Libraries</h3><p>While the most powerful way to use ArchRules is for you to write your own rules, we have built some <a href="https://github.com/nebula-plugins/nebula-archrules">OSS rule libraries</a> that anyone is free to use, or reference as examples.</p><h4>Nullability</h4><p>These rules enforce proper nullability annotation in Java, for example, that every public class is marked with <a href="https://jspecify.dev/">JSpecify</a>’s @NullMarked. It is smart enough to exclude Kotlin code, as Kotlin has built-in nullability.</p><h4>Gradle Plugin Best Practices</h4><p><a href="https://docs.gradle.org/current/userguide/writing_plugins.html">Writing Gradle plugins</a> can be hard, especially since there are many APIs and patterns that should not be used anymore. These rules help enforce current best practices when writing Gradle plugins.</p><h4>Joda / Guava Rules</h4><p>These rule libraries discourage the use of Joda Time and Guava classes (respectively) as these have been superseded by java.time and standard library enhancements.</p><h4>Security Rules</h4><p>These rules help mitigate CVEs by detecting usage of known vulnerable APIs. Ideally, we keep dependencies up to date to mitigate CVEs. But sometimes that is not immediately feasible, and in those cases, a compile time check to ensure the specific vulnerable API is not used is often good enough.</p><h3>Conclusion</h3><p>We are now running 358 (and counting) rules across over 5,000 repositories detecting over nearly 1 million issues. About 1,000 of these issues are for “High” priority rules. Being able to run these rules on this scale allows us to quickly gain insight into our large fleet of microservices, and identify the areas carrying the most critical technical debt. This makes it easier to focus and prioritize our efforts.</p><p>Going forward, we will be exploring how to tie auto-remediation solutions into the ArchRules findings. ArchUnit currently provides very specific and detailed information about failures in reports, which makes a very strong input signal to an auto remediation tool. We will explore deterministic solutions such as <a href="https://docs.openrewrite.org/">OpenRewrite</a> and non-deterministic solutions such as LLMs. Pairing the easy rule authorship and deterministic results of ArchUnit with an auto-remediation tool that can correctly interpret the results to solve the issue at hand will be a very powerful combination.</p><p>We also will investigate how to get ArchRule failure information surfaced in the IDE as inspections.</p><p>If you have questions or feedback about Nebula ArchRules, reach out to us by posting in the #nebula channel on the <a href="http://gradle-community.slack.com">Gradle Community</a> Slack.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=b4642c464c5a" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/scaling-archunit-with-nebula-archrules-b4642c464c5a">Scaling ArchUnit with Nebula ArchRules</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Democratizing Machine Learning at Netflix: Building the Model Lifecycle Graph]]></title>
            <link>https://netflixtechblog.com/democratizing-machine-learning-at-netflix-building-the-model-lifecycle-graph-5cc6d5828bb1?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/5cc6d5828bb1</guid>
            <category><![CDATA[mlops]]></category>
            <category><![CDATA[event-driven-architecture]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[distributed-systems]]></category>
            <category><![CDATA[knowledge-graph]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Mon, 04 May 2026 16:01:02 GMT</pubDate>
            <atom:updated>2026-05-15T18:44:41.436Z</atom:updated>
            <content:encoded><![CDATA[<p><a href="https://www.linkedin.com/in/saishsali/">Saish Sali</a>, <a href="https://www.linkedin.com/in/nipunk/">Nipun Kumar</a>, <a href="https://www.linkedin.com/in/suraelamurugu/">Sura Elamurugu</a></p><h3>Introduction</h3><p>As Netflix has grown, machine learning continues to support our ability to deliver value to members and drive excellence across multiple areas of our business. When Netflix began investing in machine learning over a decade ago, it was primarily focused on a single domain: personalization. Scala was the industry standard, our ML teams were relatively small, and optimizing member engagement was our primary use case. Fast forward to today, and machine learning has become the backbone of Netflix’s business transformation. We now apply ML across various business domains, including:</p><ul><li><strong>Personalization</strong>: Optimizing engagement and helping members discover content they’ll love</li><li><strong>Studio</strong>: Pre and post-production workflows</li><li><strong>Payments</strong><em>: </em>Fraud detection, payment routing, and recurring billing optimization</li><li><strong>Ads</strong>: Our newest domain, requiring real-time decisioning and targeting</li></ul><p>… and a growing number of additional use cases across the company</p><p>Each domain operates with a different tech stack, different business metrics, and a distinct organizational structure. While this diversity is a testament to how machine learning has evolved to drive value across many verticals at Netflix, this growth introduces a new challenge: <strong>enabling cross-pollination of models and data across domains.</strong></p><h3>The Challenge: A Fragmented ML Landscape</h3><p>As our ML investments scaled across these domains, a critical problem emerged: the models produced largely became black boxes. Without any discovery infrastructure, ML practitioners couldn’t easily collaborate or share work across business verticals.</p><p>Consider a concrete example: <a href="https://netflixtechblog.com/mediafm-the-multimodal-ai-foundation-for-media-understanding-at-netflix-e8c28df82e2d">content embeddings</a>. Our Studio teams create sophisticated embeddings that identify scene boundaries, detect visual transitions, and understand content structure. These embeddings were originally built for production workflows.</p><p>But those same embeddings could be incredibly valuable elsewhere. Ads could hypothetically use content embeddings for context matching (ensuring advertisements align with the tone and content of what’s currently playing). Personalization could leverage them for episodic merchandising and recommendations (matching the topic or mood of an episode with a user’s preferred viewing preferences). Yet making this cross-pollination happen is extraordinarily difficult.</p><p>Why? Our ML tools exist in silos, each with its own backend services and user interface. The model registry is unaware of which A/B tests were using its models, and the pipeline orchestrator is unaware of downstream model dependencies. ML practitioners have to traverse multiple systems to answer basic questions about their work. Finding a model requires opening the model registry, understanding its lineage means switching to the pipeline orchestrator, and tracking which A/B tests use that model requires navigating to the experimentation platform. This fragmentation prevents practitioners from answering critical questions:</p><ul><li><strong>Discovery: </strong>What features exist? What data sources are available for generating features for a model?</li><li><strong>Lineage:</strong> Which pipeline is generating data for a specific model? What data sources feed those features?</li><li><strong>Impact:</strong> Which A/B tests are running this model? Which models will break if I change this feature? Who owns each piece of this chain?</li></ul><h3>The Hard Problem: Connecting everything</h3><p>The real challenge wasn’t just building a consolidated UI. We needed to connect the different pieces of infrastructure our ML practitioners were using to perform different parts of the ML lifecycle.</p><p>Our ML ecosystem generates metadata from dozens of sources:</p><ul><li>Pipeline orchestration systems emit execution details, stage dependencies, and data transformations</li><li>Deployed model registry tracks model versions, artifacts, staleness, and deployment history</li><li>Experimentation platform manages A/B tests and their configurations</li><li>Feature store catalog feature definitions and usage</li><li>AI Dataset platform tracks the creation, management, discovery, and loading of datasets.</li><li>Identity platform maintains user, team, and organization metadata</li></ul><p>Each system employs different formats, identifiers, and mental models. The hard technical problem we had to solve was: <strong>How do we collect this heterogeneous metadata, transform it into a unified entity model, and build a connected graph that enables true exploration and collaboration across business domains?</strong></p><h4>The Solution: Metadata Service and the Model Lifecycle Graph</h4><p>Our answer was the Metadata Service (MDS), which builds a Model Lifecycle Graph that indexes and connects ML-related entities across Netflix. MDS is optimized for real-time ingestion of ML metadata (e.g., models, features, pipelines, experiments, datasets) and to answer cross-domain questions such as “Which experiments are running this model?” or “Which models share these features?” It is the foundation that enables discovery, ingesting events from diverse sources, enriching them with context, and materializing relationships across entities.</p><p>Our vision: to make every ML asset at Netflix discoverable, understandable, and reusable by every ML practitioner, regardless of their team or domain.</p><h3>Core Abstractions: The Vocabulary of the System</h3><p>Before diving into the technical implementation, it’s helpful to understand the conceptual model that underpins MDS. This vocabulary enables consistent communication across teams and systems:</p><p><strong>Component:</strong> Any object that is uniquely addressable using an AI Platform’s (AIP) Uniform Resource Identifier (URI). An AIP URI follows the formataip://&lt;componentType&gt;/&lt;platformId&gt;/&lt;resourceId&gt;, ensuring global uniqueness. For example:</p><ul><li>Models: aip://model/registry/ranking-v5</li><li>Users: aip://user/identity/alice</li><li>Pipelines: aip://pipeline/orchestrator/weekly-training</li></ul><p><strong>Entity:</strong> A component within the ML ecosystem, characterized by additional properties such as name, description, creation date, and owners. Entities represent ML-specific assets, such as models, features, and pipelines.</p><p><strong>Entity Type:</strong> A group of entities that share the same data shape. A data shape is a set of property constraints that specify the attributes and relationships an entity must have.</p><p><strong>Domain:</strong> A functional grouping of related entity types that defines the abstract interface for a category of ML assets. For example, the Models domain defines what a Model and Model Instance look like, while the Pipelines domain defines Schedules, Requests, and Executions.</p><p><strong>Provider:</strong> A concrete implementation of a domain, backed by a specific source system. For example, the Models domain is currently backed by our internal model registry. This separation allows MDS to support multiple providers for the same domain. If a new model registry were introduced, it could be added as an additional provider without changing the domain interface.</p><p>We can summarize these concepts with a concrete example:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*RQuXyzwooTZcUZ5rCejOug.png" /></figure><p>This URI-based addressing scheme is crucial as it allows any service to reference any ML asset with a single string, and MDS can resolve that reference back to rich, connected metadata.</p><h3><strong>From Events to Entities to Graph</strong></h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*dXe2XJMjTpZ6o5wwvVxGUA.png" /></figure><p>The journey from raw system events to a queryable graph happens in stages. Let’s walk through each with a concrete example: connecting a model to its A/B tests through relationship inference.</p><h4>1 Event Ingestion</h4><p>MDS integrates with various source systems via Kafka and AWS SNS/SQS, consuming events in real-time. Source systems emit thin events that include an identifier and an event type.</p><p>Example event:</p><pre>{<br>  &quot;event_type&quot;: &quot;model_instance_created&quot;,<br>  &quot;instance_id&quot;: &quot;ranking-model-v5-20XX0101&quot;,<br>  ...<br>}</pre><p>This design keeps producers simple. Source systems only need to announce that a change occurred, without building complete payloads or understanding downstream requirements.</p><p>Each source system has dedicated event handlers in MDS:</p><ul><li><strong>Pipeline Orchestration</strong>: Ingests pipeline execution events, including node definitions, schedules, requests, and job attempts</li><li><strong>Model Registry</strong>: Captures model deployments, configurations, and version updates</li><li><strong>Feature Store</strong>: Tracks feature definitions and their versions</li><li><strong>Experimentation Platform</strong>: Monitors A/B test configurations and allocations</li><li><strong>Datasets:</strong> Tracks ML datasets and their versions</li><li><strong>Identity Platform</strong>: Maintains ownership and team membership information</li></ul><h4>2 Entity Enrichment</h4><p>MDS implements a hydration contract for each event type. When an event arrives, MDS:</p><ol><li>Validates the event schema</li><li>Calls the source system’s API to fetch the complete, current state</li><li>Transforms the response into a normalized entity</li></ol><p>This design has a crucial property: the order of events doesn’t matter. MDS always fetches the latest facts from the source of truth. This pattern decouples the event stream from state consistency. If the event bus drops a message or delivers it out of order, the next event corrects the state. The event stream becomes a notification of change rather than a log of changes.</p><p>This notification of change pattern has a few important tradeoffs. On the plus side, it keeps producers simple, makes us robust to out-of-order or dropped events, and ensures that MDS can always reconcile to the latest state by reading from the source of truth. The tradeoff is that we place additional read load on source systems during hydration and need to be deliberate about rate limiting, caching, and backoff in our enrichment workers so that we don’t overload them.</p><p>For our ranking model example, when the model_instance_created event arrives, MDS calls the Model Registry API: GET /api/v1/instances/ranking-model-v5-20XX0101</p><p>The registry responds with a full descriptor. Example response (key fields only):</p><pre>{<br>  &quot;id&quot;: &quot;ranking-model-v5-20XX0101&quot;,<br>  &quot;pipeline_run_id&quot;: &quot;train-weekly-ranking-20XX0101&quot;,<br>  &quot;owner_emails&quot;: [&quot;alice@netflix.com&quot;],<br>  &quot;labels&quot;: [{&quot;key&quot;: &quot;team&quot;, &quot;value&quot;: &quot;personalization&quot;}],<br>  ...<br>}</pre><h4>3 Data Transformation and Normalization</h4><p>Raw events are heterogeneous and each source system has its own schema and semantics. MDS workers transform these events into a unified entity model with standardized fields.</p><p>Without normalization, downstream consumers would need to understand every source system’s schema. Normalization creates a consistent interface, allowing queries and relationships to work across all entity types. Here is an example.</p><p>Normalized MDS entity:</p><pre>{<br>  &quot;id&quot;: &quot;aip://model/registry/ranking-model-v5-20XX0101&quot;,<br>  &quot;pipeline_run&quot;: &quot;aip://pipeline-run/orchestrator/train-weekly-ranking-20XX0101&quot;,<br>  &quot;entity_type&quot;: &quot;ModelInstance&quot;,<br>  &quot;owners&quot;: [&quot;aip://user/identity/alice&quot;],<br>  &quot;tags&quot;: [{&quot;tag&quot;: &quot;team&quot;, &quot;value&quot;: &quot;personalization&quot;}],<br>  ...<br>}</pre><p>The normalization process standardizes field names and formats. For example, platform-specific IDs become global AIP URIs, owner_emails becomes owners with resolved user URIs, and labels become tags. Foreign keys like pipeline_run_id are transformed into entity references. However, there’s still no reference to which A/B tests are using this model. The Model Registry doesn’t track experiments, and the Experimentation Platform doesn’t track which pipeline produced a given model. This is where knowledge enrichment becomes critical.</p><h4>4 Storage and Indexing</h4><p>Once normalized, entities are persisted to Datomic and immediately indexed in Elasticsearch. This happens synchronously within the event processing flow.</p><p><strong>Datomic for Caching and Relationships</strong><br>Normalized entities are first written to Datomic, which serves as both a local cache and a graph database.</p><p>Why Datomic? Datomic serves as both the system of record for MDS and the working dataset for enrichment processes. Its immutable fact model means we can continuously add relationships without losing the original entity state.</p><p><strong>What we store:</strong></p><ul><li>All entity attributes as facts</li><li>Entity references (foreign keys that may point to entities not yet fully resolved)</li><li>All relationships as reified edges (added by enrichment processes)</li><li>Entity lifecycle state (tracking which entities are fully enriched vs awaiting hydration)</li></ul><p><strong>This enables:</strong></p><ul><li><strong>Complex graph traversals:</strong> Navigate from a model to its features to their data sources in a single query</li><li><strong>Entity relationships:</strong> Join across multiple domains without N+1 query problems</li><li><strong>Flexible schema evolution:</strong> Easy to add new entity types and attributes as the catalog grows</li><li><strong>Progressive enrichment</strong>: Background jobs efficiently identify and process entities requiring additional hydration, enabling gradual graph completion without reprocessing fully enriched entities</li></ul><p>In practice, we use Datomic for relationship-heavy, navigational queries such as:</p><ul><li>Starting from this model instance, show me all upstream datasets and downstream experiments.</li><li>Given this feature, list all consuming models and their owning teams.</li></ul><p>These queries often span multiple hops in the graph and benefit from Datomic’s immutable fact model and efficient joins across entity relationships.</p><p><strong>Elasticsearch for Discovery</strong><br>Immediately after writing to Datomic, entities are indexed in Elasticsearch to power fast, full-text search across the catalog.</p><p><strong>What we index:</strong></p><ul><li>Primary fields: Entity name, description, entity type, owner names</li><li>Relationship metadata: Names of related entities (e.g., a model’s features, pipelines, A/B tests) stored in the related field</li><li>Tags: Domain-specific metadata stored as key-value pairs (e.g., <em>team::personalization, env::production, model.state::released</em>)</li></ul><p><strong>Index structure:</strong></p><ul><li>Single entities index: All entity types (models, features, pipelines, etc.) are indexed in one unified index, differentiated by the entityType field</li><li>Separate owners index: Dedicated index for users and groups to enable cross-entity owner searches</li><li>Relevance boosting: Exact name matches score higher than other relevant matches</li></ul><p><strong>This enables:</strong></p><ul><li>Multi-field text search across entity names, descriptions, tags, and related metadata</li><li>Relevance ranking with boosting (exact name matches score significantly higher)</li><li>Complex filtering by entity type, ownership, tags, and domain-specific attributes (stored as tags)</li><li>Fuzzy matching to handle typos and partial queries</li></ul><p>Elasticsearch powers the entry point into the system: users typically start with a free-text search in the AIP Portal (for a model name, a team, or a domain term), and then switch to graph navigation once they land on an entity page. Indexing happens in near real-time as part of the ingestion and enrichment workflows, so changes are usually visible in the Portal with a short delay that is acceptable for interactive use.</p><h4>5 Knowledge Enrichment and Graph Formation</h4><p>Once entity metadata is persisted in Datomic, scheduled background processes take over to discover and materialize relationships. These enrichment jobs run periodically, scanning for uncached or partially resolved entities (entities that exist only as references without full metadata).</p><p>The enrichment workflow:</p><ul><li><strong>Identify candidates:</strong> Find entities marked as uncached or with unresolved references</li><li><strong>Hydrate relationships:</strong> Query source-of-truth systems to fetch related entity details</li><li><strong>Materialize edges:</strong> Write discovered relationships back to Datomic</li><li><strong>Re-index:</strong> Trigger Elasticsearch indexing for updated entities</li><li><strong>Mark as enriched:</strong> Update entity status to prevent redundant processing</li></ul><p>This asynchronous approach allows MDS to handle the computational cost of graph formation without blocking real-time event ingestion. It also enables retry logic and gradual enrichment as new entities become available.</p><p>Because enrichment is asynchronous, newly discovered relationships may appear with a short delay after the underlying entities are created (typically minutes rather than seconds). We track when each entity was last enriched and surface this timestamp in the AIP Portal, so practitioners can reason about staleness and know when it’s safe to rely on a particular relationship for debugging or impact analysis.</p><p><strong>Why enrich?</strong> Source systems are purpose-built and don’t know about entities in other domains. Enrichment discovers and materializes cross-system relationships that enable powerful lineage and impact queries.</p><h4>Example: Connecting Models to A/B Tests</h4><p>When MDS processes a new model instance, background enrichment jobs discover relationships through multi-hop inference:</p><p><strong>Step 1: Direct link to pipeline</strong></p><p>The model references a pipeline_run_id. An enrichment job hydrates the pipeline and discovers its A/B test associations: GET /api/v1/pipeline-runs/train-weekly-ranking-20XX0101</p><p>Response:</p><pre>{<br>&quot;run_id&quot;: &quot;train-weekly-ranking-20XX0101&quot;, &quot;pipeline&quot;:  &quot;weekly-ranking-trainer&quot;,<br>&quot;ab_test_cells&quot;: [<br>   {&quot;test_id&quot;: &quot;12345&quot;,&quot;cell_number&quot;: 2,&quot;cell_name&quot;: &quot;treatment_ranking_v5&quot;}<br> ]<br> ...<br>}</pre><p><strong>Step 2: Discover A/B test context</strong><br>The enrichment job discovers the pipeline ran for A/B test cell #2 and queries the Experimentation Platform for test details: GET /api/v1/tests/12345</p><pre>{<br> &quot;test_id&quot;: &quot;12345&quot;,<br> &quot;name&quot;: &quot;Ranking Model v5 vs v4&quot;,<br> &quot;status&quot;: &quot;ACTIVE&quot;,<br> &quot;cells&quot;: [{&quot;cell_number&quot;: 1, &quot;name&quot;: &quot;control_ranking_v4&quot;}],<br> ...<br>}</pre><p><strong>Step 3: Infer transitive relationships</strong><br>The enrichment job now has the complete chain:</p><ul><li>Model Instance was produced by Pipeline Run</li><li>Pipeline Run was executed for A/B Test Cell #2</li><li>The A/B Test Cell #2 belongs to A/B Test “Ranking Model v5 vs v4”</li><li>Model Instance now gets associated with this A/B Test</li></ul><p>The job writes the inferred relationship back to Datomic and triggers re-indexing, and materializes these edges in the graph. MDS doesn’t just store what it’s told; it derives new knowledge by <em>walking</em> the graph in the background.</p><p><strong>Why this matters:</strong> Without MDS, answering “Which A/B tests are using this model?” requires:</p><ol><li>Looking up the model in the Model Registry</li><li>Finding which pipeline produced it</li><li>Checking the Pipeline Orchestrator for A/B test tags</li><li>Querying the Experimentation Platform for test details</li></ol><p>With the model lifecycle graph, it’s a single query:</p><pre>query {<br>  model(id: &quot;aip://model/registry/ranking-model-v5-20XX0101&quot;) {<br>    name<br>    owners { name }<br>    currentInstance {<br>      version<br>      pipeline {<br>        name<br>        owners { name }<br>      }<br>      features {<br>        edges {<br>          node {<br>            name<br>            data { edges { node { name } } }<br>          }<br>        }<br>      }<br>      associatedAbTests {<br>        name<br>        cells { number name }<br>      }<br>    }<br>  }<br>}</pre><p>The reverse query also works: “What models are being tested in experiment 12345?”</p><h3>Enabling Exploration, Not Just Search</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/682/1*j8moOl19CHOfIDRvfnPk5A.png" /></figure><p>With the Model Lifecycle Graph in place, we shift from entity search to entity exploration. Discovery isn’t just about finding a model; It’s about traversing relationships:</p><ul><li>Start with a model, explore its features</li><li>From features, navigate to the core data driving them</li><li>From the data, trace back to the pipelines generating it</li><li>From pipelines, see which teams own and depend on them</li><li>From experiments, understand which models are being tested</li></ul><p>For example, imagine an engineer investigating a degraded engagement metric for a personalization model. They might:</p><ol><li>Start with the model instance powering the affected recommendations in the AIP Portal.</li><li>Inspect the model’s features and follow a suspicious feature to its upstream dataset.</li><li>From the dataset page, see that its pipeline recently had failed runs and identify the owning team.</li><li>Confirm which A/B tests are currently running this model instance to understand which members and surfaces are impacted.</li></ol><p>Before MDS and the Model Lifecycle Graph, this required manual checks across multiple tools (model registry, pipeline orchestrator, experiment platform). Now it’s a contiguous journey in a single interface.</p><p>This graph-based exploration answers questions that were previously impossible:</p><ul><li>Lineage queries: What is the complete lineage of this model, from training data to production experiments?</li><li>Impact analysis: Which models will be affected if I change this feature?</li><li>Usage discovery: Which A/B tests are using this model?</li><li>Dependency mapping: What data sources does my pipeline transitively depend on?</li><li>Deprecation planning: Which entities are no longer being used and can be retired?</li></ul><p>Every entity has deep context: its creation time, ownership, update history, and most importantly, its relationships to other entities.</p><p>The Model Lifecycle Graph is surfaced to practitioners through the AIP Portal, a unified interface that provides full-text search across all entity types, detailed entity pages with navigable relationships, and personalized views for teams and individuals.</p><p>A typical interaction in the AIP Portal looks like:</p><ul><li><strong>Search:</strong> Type a model, feature, dataset, or team name into the single search box backed by Elasticsearch.</li><li><strong>Inspect:</strong> Land on an entity page that shows key metadata (description, owners, domains, tags) alongside a relationships panel.</li><li><strong>Explore:</strong> Click through to related entities (upstream datasets, downstream experiments, and sibling model versions) to navigate the Model Lifecycle Graph without leaving the portal.</li></ul><p>When new entity types are introduced into MDS, the portal automatically provides baseline search, entity pages, and relationship navigation, and we can then layer on domain-specific visualizations (such as model deployment history or dataset version timelines) over time.</p><h3>The Road Ahead: Open Challenges</h3><p>Building the ML lifecycle graph is an ongoing journey. Significant challenges remain, and these represent the future opportunities for us:</p><ul><li><strong>Tool Proliferation:</strong> As new ML tools emerge, we need robust integration patterns that scale. How do we design plugin architectures that make adding new sources seamless? If we don’t keep up with new tools, practitioners will be forced back into fragmented views, and the Model Lifecycle Graph will lose coverage and trust.</li><li><strong>Domain-Specific Visualizations:</strong> Different entity types require distinct visualization experiences. Model pages should display deployment history, A/B test associations, and performance metrics. Feature pages should highlight data lineage and consuming models. Pipeline pages must show execution history, dependencies, and schedules. Dataset pages require versioning timelines and downstream consumers. How do we design a flexible UI framework that allows each entity type to have its own tailored experience while maintaining consistent navigation and interaction patterns across the portal? Without rich, domain-specific experiences, the portal risks becoming a generic catalog rather than a tool that ML practitioners rely on in their daily workflows.</li><li><strong>Metadata Quality:</strong> Today, MDS ensures data consistency through source-of-truth hydration and schema validation at ingestion. Background enrichment jobs continuously infer relationships and materialize entities from source systems. However, challenges remain in ensuring completeness and timeliness at scale. When source systems fail to emit events, when ownership information becomes stale, or when entities lack descriptions and contextual metadata, the graph’s utility degrades. How do we build automated validation and enrichment systems to detect metadata anomalies, suggest missing relationships, and maintain quality benchmarks across millions of entities? Poor or stale metadata erodes practitioner trust: if the graph is incomplete or incorrect, teams will revert to ad hoc knowledge and one-off integrations rather than using MDS as their source of truth.</li><li><strong>Advanced Relationship Inference:</strong> Beyond explicit relationships declared in source systems, how do we infer implicit connections? Can we detect that two models serve similar purposes based on shared features? Can we recommend features based on usage patterns from similar pipelines? We are in the early stages of exploring these ideas. Done well, they would turn MDS from a passive catalog into an active recommendation engine for ML assets, accelerating reuse and reducing duplicate work across domains.</li></ul><h3>Acknowledgments</h3><p>This work represents the collective effort of stunning colleagues across the AI Platform organization: <a href="https://www.linkedin.com/in/emma-carney-6a700b17a/">Emma Carney</a>, <a href="https://www.linkedin.com/in/megan-ren-7b78a81a8/">Megan Ren</a>, <a href="https://www.linkedin.com/in/nadeem-ahmad-80000983/">Nadeem Ahmad</a>, <a href="https://www.linkedin.com/in/poleniuk/">Pat Oleniuk</a>, <a href="https://www.linkedin.com/in/prateekagarwal17/">Prateek Agarwal</a>, <a href="https://www.linkedin.com/in/tikhakobyan/">Tigran Hakobyan</a>, <a href="https://www.linkedin.com/in/yinglao-liu-6b48b6126/">Yinglao Liu</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=5cc6d5828bb1" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/democratizing-machine-learning-at-netflix-building-the-model-lifecycle-graph-5cc6d5828bb1">Democratizing Machine Learning at Netflix: Building the Model Lifecycle Graph</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[State of Routing in Model Serving]]></title>
            <link>https://netflixtechblog.com/state-of-routing-in-model-serving-16e22fe18741?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/16e22fe18741</guid>
            <category><![CDATA[ai-platform]]></category>
            <category><![CDATA[distributed-systems]]></category>
            <category><![CDATA[infrastructure]]></category>
            <category><![CDATA[machine-learning]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 01 May 2026 21:03:13 GMT</pubDate>
            <atom:updated>2026-05-01T21:22:25.368Z</atom:updated>
            <content:encoded><![CDATA[<p>By <a href="https://www.linkedin.com/in/nipunk/">Nipun Kumar</a>, <a href="https://www.linkedin.com/in/rajatsshah/">Rajat Shah</a>, <a href="https://www.linkedin.com/in/peterchng/">Peter Chng</a></p><h3>Introduction</h3><p><em>This is the first blog post in a multi-part series that shares technical insights into how our ML model serving infrastructure powers several personalized experiences at scale across various domains (e.g., title recommendations, commerce). In this introductory blog post, we will dive into our domain-independent API abstraction and its traffic routing capabilities that the central ML model serving platform exposes to several domain-specific microservices for model inference. This singular API, or entry point, into the ML model serving platform has significantly increased the speed of innovation for iterating on newer versions of existing ML experiences, as well as enabling completely new product experiences with ML.</em></p><p>Machine Learning use cases powering member experiences on Netflix require rapid iteration and evolution in response to new learnings. The success of our ML model serving infrastructure largely depends on enabling researchers to rapidly experiment with new hypotheses and safely, at scale, release their models into production. Equally important is enabling multiple microservices at Netflix to seamlessly get model inference without exposing the complexities of ML model inference. To achieve this in a uniform and scalable manner, we created a centralized ML serving platform. As of 2025, the platform serves hundreds of model types and versions, netting 1 million requests per second. In this post, we’ll zoom in on a core challenge of any large-scale ML serving system: How to route traffic to the right model instance, on the right cluster shard, for the right user and use case, while preserving a simple abstraction for both client services and model researchers.</p><h3>Background</h3><h3>Models at Netflix</h3><p>To properly frame our discussion, let’s first clarify the distinction between model <em>serving</em> and model <em>inference</em>. At Netflix, the definition of an ML model has historically been somewhat unique. While model <em>inference</em> typically focuses only on an infer(features) -&gt; score capability, models at Netflix act as self-contained workflows that transform inputs to outputs. A “model” encapsulates pre- and post-processing, feature computation logic, and an optional ML-trained component, all packaged in a standard format suitable for use across multiple contexts. We refer to the end-to-end execution of this workflow as model <em>serving</em>. This distinction matters because our routing and API abstractions operate at the level of workflows, not just individual scoring functions.</p><p>A few <em>simplified</em> examples of model serving use cases:</p><p><strong>Use case</strong>: Personalized Continue Watching row on Netflix Homepage</p><ul><li>Input: UserId, Country, Device ID</li><li>Output: Ranked List of movies and shows (aka title): [titleId1, titleId2, titleId3,…]</li></ul><p><strong>Use case</strong>: Payment Fraud Detection</p><ul><li>Input: UserId, Country, Payment Transaction details</li><li>Output: Probability of the transaction being fraudulent</li></ul><p>A typical flow of this serving workflow is depicted below:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/804/1*V0IaBLQSyADbjdnlhRDJdA.png" /></figure><p>To achieve this higher level of abstraction, the model definition contains a list of facts (raw, unprocessed data or observations built as states in different business workflows) that it needs to compute features, and it relies on the model serving platform to supply these facts at serving time by calling several other microservices. Likewise, during offline training, <a href="https://netflixtechblog.com/evolution-of-ml-fact-store-5941d3231762">Netflix’s ML fact store</a> provides snapshots for bulk access to facilitate feature computation.</p><p>The important takeaway from this model definition is that the calling services only need to provide standard request context (such as userId, country, device), and the relevant domain context (such as titles to rank, or payment transaction for fraud detection), and the model can itself compute features and perform inference as part of the execution flow. This common set of request contexts across domains enables them to share a standard API abstraction and standardizes how various client microservices can uniformly integrate with the serving app. Furthermore, clients are shielded from the model selection and execution, allowing the model architecture and data inputs to evolve with minimal client coordination.</p><p>This post focuses on showcasing the technical details to support this design paradigm. We’ll first describe how we implemented this abstraction with Switchboard, a centralized routing service, and then discuss the operational challenges we encountered at scale and how they led us to the Lightbulb architecture.</p><h3>ML Model Serving Platform Principles</h3><p>We envisioned a central model serving platform for all of Netflix’s member-facing ML Model serving needs. This ambitious effort required principled thinking to provide the right level of abstraction for both the researchers and client applications. The following ideas, which are relevant to the topic of this blog post, ensured that the platform acts as an enabler of rapid ML innovation and limits the exposure of ML model iterations to the client apps:</p><ul><li><strong>Model innovation independent of client apps: </strong>There should be only a one-time integration effort by the calling app with the ML serving platform for a new use case. After that, almost all model iterations, including intermediate model A/B experiments, should be mostly opaque to the calling apps. This implies that the platform should handle tasks such as model selection based on a user’s A/B allocation, fetching additional data needed by experimental models, logging for further training or observability, and more. This also benefits the ML researcher, as they only need to coordinate with one platform for model innovation.</li><li><strong>Decouple clients from model sharding: </strong>Models are distributed across multiple serving compute cluster shards, each with its own Virtual IP (VIP) Address. Various factors, such as traffic patterns, SLAs, model architecture, and CPU/Memory availability, affect model-to-cluster mapping, and changes to this mapping result in changes to the VIP address at which a model is reachable. The serving platform should make clients agnostic to such frequent VIP address changes while ensuring high availability.</li><li><strong>Flexible traffic routing rules: </strong>Support flexible mechanisms to introduce new traffic routing rules. This includes supporting traffic routing based on A/B experiments, providing a knob to slowly shift traffic to new models and VIP addresses, and allowing client overrides.</li></ul><h3>Introducing Switchboard</h3><p>Standard out-of-the-box API Gateway solutions (such as AWS API Gateway, a standalone Service Mesh proxy) did not meet all our requirements. In particular, we needed first-class integration with Netflix’s experimentation platform, the ability to expose gRPC endpoints to clients, and the ability to use rich domain-specific context for routing customizations, which generic proxies were not designed to handle. Furthermore, the platform required customizations to model-specific lifecycle stages (shadow mode, canaries, rollbacks) to enable safe rollouts and migrations.</p><p>Hence, we embarked on building a custom service that serves as a flexible proxy layer for all traffic, handling over 1 million requests per second while maintaining high availability and reliability. We named it Switchboard.</p><p>Switchboard serves as the central entry point for the system, <strong>acting as a mandatory interface </strong>for all clients to access the appropriate model based on their context. Its role is to perform context-aware routing and to apply any configured context enrichment to the model inputs.</p><p>Here is a visual representation of the request flow from different clients to different serving clusters:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/935/1*HEV6B_6F5ci3dyoKXyq55A.png" /></figure><h3>Objective Abstraction</h3><p>To support this system design, we introduce the concept of an “Objective”. It’s an Enumeration defined by the serving platform that every request into the system must provide. It has three key purposes:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/703/1*V6bhyHhYQ4W5baVzvygW9g.png" /></figure><p>In short, an <strong>Objective</strong> is the serving platform’s name for a specific business use case (e.g., ContinueWatchingRanking), which decouples clients from concrete models and guides the platform’s routing and model selection decisions.</p><h3>Key Capabilities of Switchboard</h3><p>To summarize, these are the key capabilities of Switchboard:</p><ol><li><strong>Common Client Abstraction: </strong>Switchboard provides a single point of contact for all our clients’ model needs. When clients wish to consume additional models for new ML applications addressing the same business need, there is no new service dependency to introduce or new clients to manage to make requests to the models. From an ML Ops perspective, this also gives us knobs to control client rate limits across model versions and manage central concurrency limits to deal with bad clients.</li><li><strong>Context-Aware Routing:</strong> Switchboard can route a request based on a rich set of contextual features, such as the user’s current device, locale, ranking surface type (e.g., home page vs. search results), or the current A/B test a user is in.</li><li><strong>Dynamic Traffic Splitting:</strong> It enables real-time traffic splitting for canary deployments and experimentation. This allows engineers to safely roll out a new model version to a small, controlled percentage of users before a full launch.</li><li><strong>Model Versioning and Lifecycle Management:</strong> Switchboard inherently manages concurrent request traffic to multiple versions of the same model. This is crucial for:</li></ol><ul><li><strong>Shadow Mode Testing:</strong> Routing production traffic to a new model version without affecting the user experience, enabling performance comparisons.</li><li><strong>Instant Rollback:</strong> Immediate switching of traffic away from a problematic new model version back to a stable one.</li></ul><p>But is this the whole story? Not quite. Introducing this routing layer adds complexity to our model deployment cycles. In addition, we need a mechanism to collect the context-based routing information from the researchers when they choose to deploy model variants.</p><h3>The Glue — Switchboard Rules</h3><p>Given that Objectives serve as the contract between clients and the serving platform, we needed a way for researchers to attach model variants, experiments, and traffic splits to those Objectives without changing client code. This is where Switchboard Rules comes in.</p><p>The primary UX for model researchers to define models associated with an objective in a flexible manner is a JavaScript configuration, which we call <em>Switchboard Rules</em>. It’s used to produce a set of rules (typically a JSON file) that primarily dictate the following things to the serving platform:</p><ol><li>The default model to use for a given Objective</li><li>A/B experiments to configure for a set of Objectives and the corresponding models to load for those experiments</li><li>Customizations to gradually shift traffic to a new model</li></ol><p>Here is an example of an A/B test rule in the context of the Continue Watching row:</p><pre>/**<br>Configuration rule written by a Model Researcher to add an A/B experiment in the Model Serving system.<br>Cell 1: Uses the default, currently productized model<br>Cell 2 and Cell 3: Use different experimental (candidate) models<br>**/<br><br>function defineAB12345Rule() {<br>    const abTestId = 12345;<br><br>    const objectives = Objectives.ContinueWatchingRanking;<br>    const abTestCellToModel = {<br>        1: {name: &quot;netflix-continue-watching-model-default&quot;},<br>        2: {name: &quot;netflix-continue-watching-model-cell-2&quot;},<br>        3: {name: &quot;netflix-continue-watching-model-cell-3&quot;}<br>    };<br><br>    return {<br>        cellToModel: abTestCellToModel,<br>        abTestId: abTestId,<br>        targetObjectives: [objectives],<br>        modelInputType: constants.TITLE_INPUT_TYPE,<br>        modelType: &#39;SCORER&#39;<br>    };<br>}</pre><p>These rules are consumed by both the Switchboard and the Model Serving clusters. Given these rules, the serving platform components can take various actions, some detailed below:</p><p><strong>Control Plane Flow</strong>:</p><ol><li><strong>Assignment:</strong> Produce model-to-cluster shard assignment.</li><li><strong>Validation:</strong> Load all specified models into the Serving Cluster Shard and validate model dependencies to ensure successful execution.</li><li><strong>Mapping:</strong> Provide the model-to-shard VIP address mapping to Switchboard.</li></ol><p><strong>Data Plane Flow</strong>:</p><ol><li><strong>Allocation:</strong> If the request is for Objective=ContinueWatchingRanking, query the <a href="https://netflixtechblog.com/its-all-a-bout-testing-the-netflix-experimentation-platform-4e1ca458c15">Experimentation Platform</a> for the userId’s cell allocation.</li><li><strong>Model Selection:</strong> Use the allocation and A/B test rule to select the appropriate model.</li><li><strong>Request Routing:</strong> Route the request to the serving cluster shard with the selected model and context.</li><li><strong>Model Execution (on the serving host):</strong> Run the model workflow steps and return the response.</li></ol><p>A key highlight of this setup is the decoupling of the experimentation config from the serving platform code. This includes having an independent release cycle for the rules, separate from the code deployments. <a href="https://netflixtechblog.com/how-netflix-microservices-tackle-dataset-pub-sub-4a068adcc9a">Netflix’s Gutenberg</a> system provides an excellent ecosystem that enables a flexible pub-sub architecture, facilitating proper versioning, dynamic loading, easy rollbacks, and more. Both Switchboard and the Serving Cluster Host subscribe to the same Switchboard Rules configuration.</p><p>To prevent race conditions and ensure proper sync of the dynamic Switchboard Rules configuration, the following flow is considered:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/944/1*iSNuD8ZuSC9E2zN-wZdH-A.png" /></figure><h3>Evolving Challenges</h3><p>Switchboard solved the primary problem of improving model iteration and innovation velocity, and provided an excellent ML serving abstraction to over 30 service clients. However, as the system scale increased, a few challenges and problems with this design became apparent:</p><ul><li><strong>Single point of failure: </strong>The presence of Switchboard in the critical request path clearly highlights the risks of shutting down access to all serving hosts in extreme cases, such as unintentional bugs or noisy neighbors sending excessive traffic.</li><li><em>Why this matters: Switchboard became a shared dependency whose failure would degrade or disable multiple ML-powered experiences at Netflix.</em></li><li><strong>Added latency due to additional network hop:</strong> Switchboard in the request path adds between 10–20ms of latency due to serialization-deserialization operations, depending on payload size. Additionally, it further exposes a request to tail latency amplification.</li><li><em>Why this matters: The added latency is unacceptable for some latency-sensitive clients, resulting in end-user impact due to service timeouts.</em></li><li><strong>Reduced Client flexibility</strong>: Switchboard obscures visibility into client request origins from the serving clusters. Consequently, distinguishing data logged for real vs artificial traffic, which is essential for model training, is difficult and requires ongoing customization and increased MLOps overhead.</li><li><em>Why this matters: It makes it harder to do tenant separation and test traffic isolation.</em></li></ul><h3>What Next? — Lightbulb</h3><p>The aforementioned challenges of operating Switchboard at scale forced us to rethink the core implementation while retaining its key features. Our goal was not to throw away Switchboard’s design, but to refactor where and how its responsibilities were executed, keeping the benefits while reducing risk and latency. Particularly:</p><ul><li><em>Common Client Abstraction</em></li><li><em>Decouple clients from model sharding</em></li><li><em>Flexible traffic routing rules</em></li><li><em>Lightweight system client</em></li><li><em>Single place to define model and experimentation config</em></li><li><em>Fast experimentation config propagation</em></li><li><em>Fallback and client-side caching in case of failures</em></li></ul><p>However, we did want to address some of the previous design choices to move forward with:</p><ul><li><strong>Remove the routing service from the direct request path: </strong>Having a single service in the active request path introduces another failure mode and limits fallback flexibility. While routing rules change infrequently, maintaining consistency comes at the cost of increased availability risks.</li><li><strong>Separate model inputs from the request metadata</strong>: In certain cases, the request payload could be quite large. Needing to deserialize and then re-serialize the payload as it flowed through Switchboard to make a routing decision was a significant contributor to latency and increased serving costs.</li><li><strong>Provide better isolation for the routing layer: </strong>Consolidating multiple use cases (tenants) into a single routing cluster poses two main challenges. First, error propagation posed a risk, as a surge of problematic requests from one tenant could cascade errors back to Switchboard, potentially impacting other users. Second, the cluster had to accommodate diverse latency requirements because the requests from different use cases varied significantly in complexity.</li></ul><p>This required some changes in our setup flow: While it largely remained unchanged, however, we created separate components for Routing and Model Selection (Lightbulb):</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/897/1*wtVpe5xJMNEENitvkd1FPQ.png" /></figure><p>We now take the rules for an Objective and break them into distinct sets of configuration:</p><ul><li><strong>Model Serving Configuration</strong>: This allows us to determine which model should be used at request time, along with the required metadata</li><li><strong>Routing Rules</strong>: Given a model we want to serve at request time, this tells us which VIP the request should be routed to.</li></ul><p>The Data Plane changes also reflect this separation, as we now rely on <a href="https://github.com/envoyproxy/envoy">Envoy</a> to take care of the routing details:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/912/1*jbW4NlcnKucGKBi_vNjXbw.png" /></figure><p>Envoy is <a href="https://netflixtechblog.com/zero-configuration-service-mesh-with-on-demand-cluster-discovery-ac6483b52a51">already used</a> for all egress communication between apps at Netflix, and it can route requests to different clusters (VIPs) based on the configurable Routing Rules published from our control plane. However, it lacks the information needed to make routing decisions and the ability to enrich the request body with additional serving parameters required for A/B testing model variants. We introduced Lightbulb to cover this gap:</p><ul><li>Lightbulb consumes the minimal request context, which contains use-case information, and provides the metadata mapping required for routing at the Envoy layer.</li><li>Lightbulb resolves the request context to determine a routingKey configuration along with the <strong>ObjectiveConfig</strong> — this is where we place the model id along with other request-specific configurations required for model execution. This is done to separate the config resolution associated with the request from the placement and routing information needed to reach it on the inference cluster.</li><li>While the routingKey is added to the headers for Envoy proxy to consume, the client adds the ObjectiveConfig parameters to the request itself. This is done to avoid bloating the request headers while passing additional parameters for the model to process the request appropriately.</li><li>The routing of the actual request is performed by the Envoy proxy, which has the metadata to map the routingKey to the actual cluster VIP running the model. Because the routingKey is in a header, this determination can be made with minimal overhead.</li></ul><p>These changes retain the advantages of Switchboard, such as a single integration point, abstraction of model id from use case, context-aware routing, while addressing the challenges we observed over time.</p><h3>Conclusion</h3><p>The evolution from Switchboard to Lightbulb marks a significant architectural refinement in our ML model serving infrastructure. While Switchboard provided the initial abstraction layer critical for rapid innovation, its latency and single-point-of-failure risk posed scaling hurdles. The subsequent adoption of Lightbulb, a decoupled service focused solely on routing metadata, and its integration with Envoy successfully resolved these challenges. This sophisticated new architecture preserves the key benefits — seamless client integration and flexible experimentation — while ensuring reliable, efficient, and scalable delivery of personalized member experiences, positioning us well for future ML growth.</p><p>In future posts in this series, we’ll dive deeper into other aspects of our ML serving platform, including inference and feature fetching, and how they interact with the routing architecture described here.</p><p>Special thanks to <strong>Sura Elamurugu</strong>, <strong>Sri Krishna Vempati</strong>, <strong>Ed Maddox</strong>, and <strong>Sreepathi Prasanna</strong> for their invaluable feedback and partnership in iterating on this idea and bringing this blog post to life.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=16e22fe18741" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/state-of-routing-in-model-serving-16e22fe18741">State of Routing in Model Serving</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Scaling Camera File Processing at Netflix]]></title>
            <link>https://netflixtechblog.com/scaling-camera-file-processing-at-netflix-6dab2b1e80be?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/6dab2b1e80be</guid>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 24 Apr 2026 15:06:01 GMT</pubDate>
            <atom:updated>2026-04-24T15:06:01.452Z</atom:updated>
            <content:encoded><![CDATA[<p><em>Orchestrating Media Workflows Through Strategic Collaboration</em></p><p>Authors: <a href="https://www.linkedin.com/in/ericreinecke/">Eric Reinecke</a>, <a href="https://www.linkedin.com/in/bhanusrikanth/">Bhanu Srikanth</a></p><h3>Introduction to Content Hub’s Media Production Suite</h3><p>At Netflix, we want to provide filmmakers with the tools they need to produce content at a global scale, with quick turnaround and choice from an extraordinary variety of cameras, formats, workflows, and collaborators. Every series or film arrives with its own creative ambitions and technical requirements. To reduce friction and keep productions moving smoothly, we built <a href="https://netflixtechblog.com/globalizing-productions-with-netflixs-media-production-suite-fc3c108c0a22">Netflix’s Media Production Suite (MPS)</a> with the goal of automating repeatable tasks, standardizing key workflows, and giving productions more time to focus on creative collaboration and craftsmanship.</p><p>A critical part of this effort is how we handle image processing and camera metadata across the hundreds of hours and terabytes of camera footage that Netflix productions ingest on a daily basis. Rather than build every component from scratch, we chose to partner where it made sense–especially in areas where the industry already had trusted, battle-tested solutions.</p><p>This article explores how Netflix’s Media Production Suite integrates with FilmLight’s API (FLAPI) as the core studio media processing engine in Netflix’s cloud compute infrastructure, and how that collaboration helps us deliver smarter, more reliable workflows at scale.</p><h3>Why We Built MPS</h3><p>As Netflix’s production slate grew, so did the complexity of file-based workflows. We saw recurring challenges across productions:</p><ul><li>File wrangling sapping time from creative decision-making</li><li>Inconsistent media handling across shows, regions, or vendors</li><li>Difficult to audit manual processes that are prone to human error</li><li>Duplication of effort as teams reinvented similar workflows for each production</li></ul><p>Content Hub Media Production Suite was created to address these pain points. MPS is designed to:</p><ul><li>Bring efficiency, consistency, and quality control to global productions</li><li>Streamline media management and movement from production through post-production</li><li>Reduce time spent on non-creative file management</li><li>Minimize human error while maximizing creative time</li></ul><p>To achieve this, MPS needed a robust, flexible, and trusted way to handle camera-original media and metadata at scale.</p><h3>The Right Tool for the Job</h3><p>From the start, we knew that building a world-class image processing engine in-house is a significant, long-term commitment: one that would require deep, continuous collaboration with camera manufacturers and the wider industry.</p><p>When designing the system, we set out some core requirements:</p><ul><li><strong>Inspect, trim, and transcode original camera files and metadata</strong> for any Netflix production with trusted color science</li><li><strong>Support a wide variety of cameras and recording formats</strong> used worldwide while staying current as new ones are released</li><li><strong>Run well in our paved-path encoding infrastructure,</strong> enabling us to take advantage of proven compute and storage scalability with robust observability</li></ul><p>FilmLight develops Baselight and Daylight, which are commonly used in the industry for color grading, dailies, and transcoding. Their FilmLight API (FLAPI) allows us to use that same media processing engine as a backend API.</p><p>Rather than duplicating that work, we chose to integrate. FilmLight became a trusted technology partner, and FLAPI is now a foundational part of how MPS processes media.</p><h3>The Media Processing Engine</h3><p>MPS is not a single application; it’s an ecosystem of tools and services that support Netflix productions globally. Within that ecosystem, the FilmLight API plays the following key roles.</p><ol><li>Parsing camera metadata on ingest</li></ol><p>Productions upload media to Netflix’s <strong>Content Hub</strong> with <a href="https://theasc.com/society/ascmitc/asc-media-hash-list">ASC MHL</a> (Media Hash List) files to ensure completeness and integrity of initial ingest, but soon after, it’s important to understand the technical characteristics of each piece of media. We call this workflow phase “inspection.”</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*OYBDXSUJ6D0tVXVGO3w5TQ.jpeg" /><figcaption>Footage ingested with MPS is inspected using FLAPI and all metadata is indexed and stored</figcaption></figure><p>At this stage, we:</p><ul><li>Use FLAPI to gather <strong>camera metadata</strong> from the original camera files</li><li>Conform the workflow critical fields to <strong>Netflix’s normalized schema</strong></li><li>Make it <strong>searchable and reusable</strong> for downstream processes</li></ul><p>This metadata is integral to:</p><ul><li>Matching footage based on timing and reel name for automated retrieval</li><li>Debugging (e.g., why a shot looks a certain way after processing)</li><li>Validations and checks across the pipeline</li></ul><p>FLAPI provides consistent, camera-aware insight into footage that may have originated anywhere in the world. Additionally, since we’re able to package FLAPI in a Docker image, we can deploy almost identical code to both cloud and our production compute and storage centers around the world, ensuring a consistent assessment of footage wherever it may exist.</p><p>2. Generating VFX plates and other deliverables</p><p>Visual effects workflows constantly push image processing pipelines to their absolute limits. For MPS to succeed, it must generate images with <strong>accurate</strong> framing, <strong>consistent</strong> color management, and <strong>correct</strong> debayering/decoding parameters — all while maintaining rapid turnaround times.</p><p>To achieve this, we leverage Netflix’s <a href="https://netflixtechblog.com/the-netflix-cosmos-platform-35c14d9351ad">Cosmos</a> compute and storage platform and use open standards to provide predictable and consistent creative control.</p><p>At this phase, we use the FilmLight API to:</p><ul><li><strong>Debayer</strong> original camera files with the correct format-specific decoding parameters</li><li>Crop and de-squeeze images using <strong>Framing Decision Lists (ASC FDL)</strong> to ensure spatial creative decisions are preserved</li><li><strong>Apply ACES Metadata Files (AMF), </strong>providing repeatable color pipelines from dailies through finishing</li><li>Generate <strong>an array of media deliverables</strong> in varied formats</li></ul><p>These processes are automated, repeatable, and auditable. We deliver AMFs alongside the OpenEXRs to ensure recipients know exactly what color transforms are already applied, and which need to be applied to match dailies.</p><p>Because we use FilmLight’s tools on the backend, our workflow specialists can use Baselight on their workstations to manually validate pipeline decisions for productions before the first day of principal photography.</p><h3>The Media Processing Factory in the Cloud</h3><p>Finding an engine that competently processes media in line with open standards is an important part of the equation. To maximize impact, we want to make these tools available to all of the filmmakers we work with. Luckily, we’re no strangers to scaled processing at Netflix, and our <a href="https://netflixtechblog.com/the-netflix-cosmos-platform-35c14d9351ad">Cosmos compute platform</a> was ready for the job!</p><h4>Cloud-first integration</h4><p>The traditional model for this kind of processing in filmmaking has been to invest in beefy computers with large GPUs and high-performance storage arrays to rip through debayering and encoding at breakneck speed. However, constraints in the cloud environment are different.</p><p>Factors that are essential for tools in our runtime environment include that they:</p><ul><li>Are <strong>packageable as Serverless Functions in Linux Docker images</strong> that can be quickly invoked to run a single unit of work and shut down on completion</li><li>Can <strong>run on CPU-only instances</strong> to allow us to take advantage of a wide array of available compute</li><li>Support <strong>headless invocation </strong>via Java, Python, or CLI</li><li><strong>Operate statelessly,</strong> so when things do go wrong, we can simply terminate and re-launch the worker</li></ul><p>Operating within these constraints lets us focus on increasing throughput via parallel encoding rather than focusing on single-instance processing power. We can then target the sweet spot of the cost/performance efficiency curve while still hitting our target turnaround times.</p><p>When tools are API-driven, easily packaged in Linux containers, and don’t require a lot of external state management, Netflix can quickly integrate and deploy them with operational reliability. FilmLight API fit the bill for us. At Netflix, we leverage:</p><ul><li><strong>Java</strong> and <strong>Python</strong> as the primary integration languages</li><li><strong>Ubuntu-based Docker images</strong> with Java and Python code to expose functionality to our workflows</li><li><strong>CPU instances in the cloud and local compute centers</strong> for running inspection, rendering, and trimming jobs</li></ul><p>While FLAPI also supports GPU rendering, CPU instances give us access to a much wider segment of Netflix’s vast encoding compute pool and free up GPU instances for other workloads.</p><p>To use FilmLight API, we bundle it in a package that can be easily installed via a Dockerfile. Then, we built Cosmos Stratum Functions that accept an input clip, output location, and varying parameters such as frame ranges and AMF or FDL files when debayering footage. These functions can be quickly invoked to process a single clip or sub-segment of a clip and shut down again to free up resources.</p><h4>Elastic scaling for production workloads</h4><p>Production workloads are inherently spiky:</p><ul><li>A quiet day on set may mean minimal new footage to inspect.</li><li>A full VFX turnover or pulling trimmed OCF for finishing might require <strong>thousands of parallel renders</strong> in a short time window.</li></ul><p>By deploying FLAPI in the cloud as functions, MPS can:</p><ul><li>Allocate compute on demand and release it when our work queue dies down</li><li>Avoid tying capacity to a fixed pool of local hardware</li><li>Smooth demand across many types of encoding workload in a shared resource pool</li></ul><p>This elasticity lets us swarm pull requests to get them through quickly, then immediately yield resources back to lower priority workloads. Even in peak production periods, we avoid the pain of manually managing render queues and prioritization by avoiding fixed resource allocation. All this means <strong>lightning-fast</strong> turnaround times and <strong>less anxiety</strong> around deadlines for our filmmakers.</p><h3>Designed for Seasoned Pros and Emerging Filmmakers</h3><p>Netflix productions range from highly experienced teams with very specific workflows to newer teams who may be less familiar with potential pitfalls in complex file-based pipelines.</p><p>MPS is designed to support both:</p><ul><li>Industry veterans who need to configure precise, bespoke workflows and trust that underlying image processing will respect those decisions.</li><li>Productions without a color scientist on staff — those who benefit from guardrails and sane defaults that help them avoid common workflow issues (e.g., mismatched color transforms, inconsistent debayering, or incomplete metadata handling).</li></ul><p>The partnership with FilmLight lets Netflix focus on workflow design, orchestration, and production support, while FilmLight focuses on providing competent handling of a wide variety of camera formats with world-class image science!</p><h3>Collaboration and Co-Evolution</h3><p>Netflix aimed to integrate MPS into a wider tool ecosystem by developing a comprehensive solution based on emerging open standards, rather than making MPS a self-contained system. Integrating FLAPI into our system requires more than an API reference–it requires ongoing partnership. FilmLight worked closely with Netflix teams to:</p><ul><li>Align on <strong>feature roadmaps</strong>, particularly around new camera formats and open standards</li><li>Validate the <strong>accuracy and performance</strong> of key operations</li><li>Debug <strong>edge cases</strong> discovered in large-scale, real-world workloads</li><li><strong>Evolve the API</strong> in ways that serve both Netflix and the wider industry</li><li>Create <strong>a positive feedback cycle with open standards</strong> like ACES and ASC FDL to solve for gaps when the rubber hits the road</li></ul><p>One example of this has been with the implementation of <a href="https://draftdocs.acescentral.com/background/about-aces-2/">ACES 2</a>. FilmLight’s developers quickly provided a roadmap for support. As our engineering teams collaborated on integration, we also provided feedback to the ACES technical leadership to quickly address integration challenges and test drive updates in our pipeline.</p><p>This collaborative relationship–built on open communication, joint validation, and feedback to the greater industry–is how we routinely work with FilmLight to ensure we’re not just building something that works for our shows, but also driving a healthy tooling and standards ecosystem.</p><h3>Impact</h3><p>While much of this work takes place behind the scenes, its impact is felt directly by our productions. Our goal with building MPS is for producers, post supervisors, and vendors to experience:</p><ul><li>Fewer delays caused by missing, incomplete, or incorrect media</li><li>Faster turnaround on VFX plates and other technical deliverables</li><li>More predictable, consistent handoffs between editorial, color, and VFX</li><li>Less time spent troubleshooting technical issues, and more time focused on creative review</li></ul><p>In practice, this often shows up as the absence of crisis: the time a VFX vendor doesn’t have to request a re-delivery, or the time editorial doesn’t have to wait for corrected plates, or the time the color facility doesn’t have to reinvent a tone-mapping path because the AMF and ACES pipeline are already in place.</p><h3>Looking Ahead</h3><p>As camera technology, codecs, open standards, and production workflows continue to evolve, so will MPS. The guiding principles remain:</p><ul><li>Automate what’s repeatable</li><li>Centralize what benefits from standardization</li><li>Partner where deep domain expertise already exists</li></ul><p>The integration with FilmLight API is one example of this philosophy in action. By treating image processing as a specialized discipline and collaborating with a trusted industry partner, Netflix is delivering smarter, more reliable workflows to productions worldwide.</p><p>At its core, this partnership supports a simple goal: reduce manual workflow and tool management, giving filmmakers more time to tell stories.</p><h3>Acknowledgements</h3><p>This project is the result of collaboration and iteration over many years. In addition to the authors, the following people have contributed to this work:</p><ul><li>Matthew Donato</li><li>Prabh Nallani</li><li>Andy Schuler</li><li>Jesse Korosi</li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6dab2b1e80be" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/scaling-camera-file-processing-at-netflix-6dab2b1e80be">Scaling Camera File Processing at Netflix</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[The Human Infrastructure: How Netflix Built the Operations Layer Behind Live at Scale]]></title>
            <link>https://netflixtechblog.com/the-human-infrastructure-how-netflix-built-the-operations-layer-behind-live-at-scale-33e2a311c597?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/33e2a311c597</guid>
            <category><![CDATA[netflix]]></category>
            <category><![CDATA[incident-response]]></category>
            <category><![CDATA[live-operations]]></category>
            <category><![CDATA[live-broadcast-technology]]></category>
            <category><![CDATA[live-streaming]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 17 Apr 2026 15:01:02 GMT</pubDate>
            <atom:updated>2026-04-18T14:20:23.499Z</atom:updated>
            <content:encoded><![CDATA[<p>By: <a href="https://www.linkedin.com/in/brett-axler-11577142/">Brett Axler</a>, <a href="https://www.linkedin.com/in/casper-choffat-005833b/">Casper Choffat</a>, and <a href="https://www.linkedin.com/in/alexlowry355/">Alo Lowry</a></p><p>In the three years since our first Live show, <a href="https://www.netflix.com/title/80167499"><em>Chris Rock: Selective Outrage</em></a>, we have witnessed an incredible expansion of our live content slate and the live operations that support it. From modest beginnings of streaming just one show per month, we are now capable of streaming over nine shows in a single day, reaching tens of millions of concurrent members. This post pulls back the curtain on the Live Operations teams that enable this rapid scale.</p><h3>Humble Beginnings</h3><p>In March 2023, the engineers who built Netflix’s first live streaming pipeline also operated it. There was no dedicated operations team or formal command center. All of our incident response playbooks were written for SVOD, and SLAs were not designed for the speed of live. For the first live shows on the platform, the engineers who designed what is described in <a href="https://netflixtechblog.com/behind-the-streams-live-at-netflix-part-1-d23f917c2f40">earlier parts of this series</a> monitored dashboards on laptops, coordinated over Slack, and troubleshot in real time while millions of members watched.</p><p>The physical setup matched the operational workflows: improvised. Temporary control rooms were put together in conference rooms. For larger events, Netflix rented third-party broadcast facilities, hardware control panels, multiviewers, and communication panels — the kind of infrastructure that established broadcast networks had built over decades. Every show was a team effort. Engineers and leadership at all levels were involved in every event. Each live show, regardless of size, was a massive effort to launch.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*tibjLPNWGd0PP43Y8Idjag.jpeg" /><figcaption>Netflix’s Early Live Operations</figcaption></figure><p>Last month, in March 2026, Netflix streamed the World Baseball Classic live to members in Japan. 47 matches over two weeks, with peak concurrent viewership exceeding 17.9 million for a single game, operations running 24/7 from permanent facilities in Los Gatos and Los Angeles, with international coverage extending to Tokyo. In March alone, Netflix launched approximately 70 live events. That is three events shy of the total number Netflix streamed live in all of 2024. The technical systems that make this possible have been covered in detail across this series. What hasn’t been told is the operational story: the people, procedures, and facilities Netflix built to run those systems in real time, under pressure, with no ability to pause or roll back.</p><h3>The Architecture of Live Operations</h3><p><strong>The Architecture of Live Operations: Evolving the Broadcast Operations Center</strong></p><p>When a technology company transitions into live broadcasting, it faces a unique challenge: blending traditional broadcast television practices with massive-scale live-streaming engineering. At the heart of this intersection is the <strong>Broadcast Operations Center (BOC)</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*tG3zfmPX5ciRafvtnj9phQ.jpeg" /><figcaption>The Transmission Operations Center in Los Angeles</figcaption></figure><p>The BOC serves as the critical “cockpit” for live events. It is the physical command center where a fully produced video feed is received directly from a stadium or venue and then handed off to the live streaming infrastructure. Everything from signal ingest, inspection, and conditioning to closed-captioning, graphics insertion, and ad management happens within these walls. By utilizing a hub-and-spoke model with highly redundant architectures, such as dual internet circuits and SMPTE 2022–7 seamless switching technologies, the BOC replaces direct, vulnerable paths from the venue to the live streaming pipeline, making each live event highly repeatable and far less dependent on the quirks of individual event locations.</p><p><strong>Securing the Signal: Reliability from the Venue</strong> Before the BOC can work its magic, we have to guarantee the video and audio feeds actually survive the journey from the production site to our facility. To ensure absolute reliability from the venue, Netflix enforces strict specifications for live signal contribution.</p><p>For any show-critical feed, meaning the primary feed our members will watch live, we require three completely discrete transmission paths. We utilize a strict hierarchy of approved transmission methods, prioritizing dedicated video fiber and single-feed satellite links, followed by dedicated enterprise-grade internet and robust SRT contribution systems.</p><p>We don’t just rely on redundant transport lines; we require full hardware redundancy out of the production truck itself. This includes using separate router line cards and discrete transmission hardware to prevent any single point of failure. Furthermore, every single piece of transmission hardware at the venue must be powered by two discrete power sources, protected by uninterruptible power supply (UPS) batteries, and surge-conditioned.</p><p>Finally, before we ever go live to millions of viewers, our operators execute exhaustive “FACS/FAX” (facilities checks) testing during rehearsals and before every show. This involves running specialized Audio/Video sync tests, latency tests, and quality tests to guarantee perfect audio and video synchronization, validating closed captions, and touring the backup switcher inputs.</p><p><strong>Building the Human Infrastructure:</strong> Building the human operational model to run a facility like the BOC didn’t happen overnight. For a platform scaling from its very first live comedy special to streaming over 400 global events a year, the operational strategy had to undergo a massive, multi-year evolution.</p><p><strong>Phase 1: The “All-Hands” Engineering Era.</strong> In the earliest days of live streaming, there was no dedicated operations team or formal broadcast operations center. The software engineers who wrote the code and built the live-streaming infrastructure were the same people manually operating the events on launch night. Every show was an “all-hands-on-deck” scenario. While this raw, startup-style approach worked for initial milestones, having core developers manually set up and tear down software configurations for every single broadcast was fundamentally incapable of scaling.</p><p><strong>Phase 2: The Shift to Specialized Engineering (SOEs and BOEs).</strong> To separate event execution from core software development, the operational model matured to introduce specialized engineering teams. First, the <strong>Streaming Operations Engineering (SOE)</strong> team was established. These are highly skilled streaming engineers whose sole focus is to configure the full event on the live pipeline and support it during the broadcast. By having SOEs act as the first line of escalation, the core software developers were freed up to focus on building new live-streaming pipeline features.</p><p>However, as the physical broadcast facilities grew, it became clear that supporting the streaming pipeline wasn’t enough; the physical broadcast hardware and facility workflows needed dedicated oversight too. To solve this, <strong>Broadcast Operations Engineers (BOEs)</strong> were introduced to work alongside the SOEs. The BOE acts as the primary escalation point for all physical broadcast facility and hardware issues, overseeing the operation of all shows during a given shift.</p><p><strong>Phase 3: The “Co-Pilot” Control Room Model.</strong> With specialized engineers in place to handle the deep technical infrastructure, the day-to-day operation of the actual video and audio feeds was handed over to dedicated operators. Initially, the Broadcast Control Rooms were structured much like an airplane cockpit.</p><p>This approach utilized a <strong>“first and second captain” workflow</strong>, pairing two Broadcast Control Operators (BCOs) together to run a single event, functioning exactly like a pilot and co-pilot. This collaborative model allowed for intense focus and high-quality execution, making it the ideal setup for running just one or two live events per day. However, as the ambition grew to stream up to 10 concurrent events a day for massive global tournaments, a 1:1 scale of pairing operators simply required too much space and manpower. A new model had to be adopted.</p><p><strong>Phase 4: The Transmission Operations Center (TOC) Fleet Model.</strong> To manage high-density event days and continuous tournament coverage, the workflow was completely reimagined with the launch of the <strong>Transmission Operations Center (TOC) model</strong>. Rather than treating every live broadcast as an isolated launch in its own room, the TOC treats live events like a fleet. It centralizes operations and distinctly separates the traditional broadcast functions from the streaming functions to maximize human efficiency.</p><p>The TOC model divides the labor across three highly specialized, tiered roles:</p><ul><li><strong>Transmission Control Operator (TCO):</strong> The TCO is responsible for managing all inbound signals arriving from the event venues, such as fiber optic, SRT, and satellite feeds. They ensure these incoming feeds meet strict quality, latency, and operational thresholds. Thanks to centralized dashboarding, a single TCO can manage up to <strong>five events concurrently</strong>.</li><li><strong>Streaming Control Operator (SCO):</strong> While the TCO handles what comes <em>in</em>, the SCO manages what goes <em>out</em>. They oversee all outbound feeds, including the streams heading to the live streaming pipeline and any syndication feeds sent to third parties for commercial distribution. Like the TCOs, SCOs can manage up to <strong>five events concurrently</strong>.</li><li><strong>Broadcast Control Operator (BCO):</strong> With the inbound and outbound transmission mechanics handled by the broader TOC, the BCO is able to focus entirely on the creative and qualitative execution of the event. Operating on a <strong>strict 1:1 ratio</strong> (one operator per event), the BCO seamlessly switches between backup inbound feeds if an issue arises, ensures audio and video remain in perfect synchronization, and performs rigorous quality control. They also monitor critical metadata, such as closed captions and digital ad-insertion messages (SCTE), right before the final polished feed is handed into the live streaming pipeline.</li></ul><p><strong>The Big Bet Exception.</strong> While the fleet-style TOC model enables immense concurrency for daily programming, the most critical, high-visibility events, like major holiday football games, utilize a specialized <strong>Big Bet Model</strong>. For these flagship broadcasts, an entire Broadcast Operations Center is dedicated exclusively to a single event. This hyper-focused environment strips away the multi-event ratios, providing operators with advanced instrumentation and dedicated facility engineers to ensure the absolute highest level of reliability for events where failure is simply not an option.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*n3WWOX4zg06kGyTMWE7mcg.png" /><figcaption>Operational Workflow at a Glance (Courtesy of Melissa “Mouse” Merencillo)</figcaption></figure><h3>The Live Command Center (LCC)</h3><p>The Live Command Center (LCC) is not an MCR (Master Control Room). Nor is it a traditional Network Operations Center (NOC). The LCC holds the end-to-end view of quality, health metrics, and reliability for every live stream — from signal ingest at the production venue through cloud encoding, CDN delivery, and playback on member devices — and coordinates the human response when any part of that chain breaks.</p><p>What makes this hard is the data and speed requirements. Standard monitoring tools incur propagation delays of minutes. However, during a live stream, a signal degradation that goes undetected for three minutes can affect millions of members before any mitigation begins. The LCC runs a purpose-built observability stack, the Live Control Center, that aggregates telemetry from across the entire pipeline in near real time: concurrent viewer counts, start failure rates, rebuffer ratios, CDN health, encoder status, and signal path health from the contribution feed forward.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*LN1ncCZL_HLrVfgLiBNK0g.png" /><figcaption>Live Control Center (Courtesy of Chris Carey)</figcaption></figure><p>During live events, the system ingests up to 38 million events per second. The LCC’s job is to make that volume of data meaningful and actionable for the small team of operators watching it live.</p><p>Two roles staff the LCC leading up to and during live events. LCC Operations Leads are the shift supervisors and incident commanders. They triage anomalies, make escalation decisions, and own the incident response process from detection through resolution.</p><p>Live Technical Launch Managers (TLMs) function as air traffic controllers: they maintain cross-functional context across more than 45 technical, product, and services teams from encoding, CDN, and playback to social media, customer service, and security teams. TLMs start coordinating with these teams months and sometimes years ahead of a live event to ensure escalation paths and playbooks are in place when the LCC needs to translate a CDN engineer’s concern into a product decision at 2am while a game is still in progress. Together, these roles form the operational leadership layer that keeps engineers focused on building rather than watching dashboards.</p><p>The live operations teams rank shows by three categories:</p><ul><li><strong>Low-Profile Events:</strong> These are lightweight, often lack new features, and anticipate low viewership. They are typically managed with a small team of 1–2 operators and automated alerting.</li><li><strong>High-Profile Events:</strong> These are mid-tier events that warrant more attention due to their size, unique features, or anticipated viewership.</li><li><strong>Big Bet Events:</strong> These represent the highest operational weight, such as an NFL game, with massive viewership expectations and special features. They require the full support of the LCC: a fully staffed physical operations room for the entire duration, active incident command structures, and key engineering teams on standby to support their specific product areas.</li></ul><p>In addition to a show’s event category, the TLMs deployed a Live Operational Level (LOL) model that helps engineers determine whether they need to be on standby, live online, or even in the LCC for any given show.</p><p>Based on the show’s event category, special features, expected viewership, and overall risk, non-operational teams are put into one of four categories:</p><p><strong>Red:</strong> Non-operational teams must remain online for the duration of the event. This is most often seen in large boxing matches and sporting events, such as the NFL Christmas Day games.</p><p><strong>Orange:</strong> Non-operational teams are required to check in online ~30 minutes prior to show and are asked to monitor the health of their systems through the first commercial breaks until the LCC releases them to LOL Yellow.</p><p><strong>Yellow:</strong> Non-operational teams are not required to be online, but should be reachable by page in 2 minutes. Special PagerDuty rotations and verifications are in place to ensure these teams are reachable.</p><p><strong>Grey:</strong> Business as usual. Teams will be reached out to by their normal pager rotation if their help is needed during the show.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*sgk7Y5OYAvbfxL7xRxTQbA.png" /><figcaption>Visual Representation of LOL Levels (Courtesy of Gemini Nano Banana Pro)</figcaption></figure><p>By tiering events, Netflix ensures that resource allocation is proportionate to operational needs, preventing a continuous “crisis” mentality and allowing our non-operational partners to focus on their day jobs.</p><p>As of April 2026, most engineering teams are Yellow or Grey, with Ops and Site Reliability Engineers making up most of the teams online to support shows, in addition to engineers performing feature tests.</p><h3>Building the Model</h3><p>The first lesson from 2023 was straightforward: what worked for one show a month would not work for ten shows a week. The engineers who built the pipeline were also the ones operating it, which meant the people best positioned to fix problems were also the ones most likely to be paged at 2am. There was no operational layer to absorb that load.</p><p>In 2024, Netflix streamed 72 live events and began building the team that would eventually run them. The first version of the LCC looked nothing like it does today: a cluster of desks, monitors on stands, and laptops running dashboards, set up in the middle of the office. The TLM team was stood up to own cross-functional coordination for live launches and began formalizing the runbooks, event tiering structure, and incident management protocols that would later enable Netflix to scale operations to support hundreds of shows per year.</p><p>By the time Jake Paul vs. Mike Tyson and the first NFL Christmas Games arrived, the LCC had moved into a dedicated conference room, and partnerships with device and labs teams were producing more effective monitoring tools. But the biggest operational lesson of that period came from communications.</p><p>For Tyson/Paul, Netflix had over 300 people online across engineering, product, and business functions. Some people were online because their support was needed, while many others were just excited to be part of it. Coordinating that many people over Slack and Zoom during an active event with 64 million concurrent streams was unmanageable.</p><p>That experience drove the implementation of a <strong>squad model</strong>: defined teams with clear roles, scoped communication channels, and a single escalation path into the LCC. Around the same time, the LCC began integrating with IP-based communications systems, finally bridging the gap between the command center and the Broadcast Operations Center that had been operating largely in a fractured parallel until then.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*07i_xA6sf2_m05Giu3Dqsw.png" /><figcaption>Visual Representation of Squad Operations Model (Courtesy of Gemini Nano Banana Pro)</figcaption></figure><p>2025 brought 220 live events and a permanent LCC facility, along with a dedicated operations team, the Live Command Center Operations Leads. With the growing number of shows, TLMs were getting spread thin, spending more than half their week operating shows late into the evening and over weekends, then getting called back into the office at 9 am to lead critical launch meetings. The addition of the LCC Ops Leads resolved the bandwidth issue by separating planning and operations into distinct roles within a single centralized team.</p><p>As the slate continued to grow and large series like the World Baseball Classic and FIFA Women’s World Cup were announced, the vendor-operator model was introduced, creating an elastic workforce that could scale up for large series events without carrying full-time headcount year-round to support peak capacity. <strong>The key enabler was documentation</strong>: standardized runbooks and onboarding materials detailed enough that a trained operator could reach full effectiveness within their first week. WWE RAW became a weekly operation, normalizing what had previously felt exceptional. By early 2026, multi-event days were no longer a test of capacity but had become the expected operating condition.</p><p>The next chapter is international. Netflix has begun standing up regional Live Operations Center coverage to support live events outside North America, with EMEA operations soon running out of London. The model draws on the same runbooks, tooling, and escalation structures developed in Los Gatos, with follow-the-sun shift handoffs connecting EMEA and US teams across time zones. Looking further ahead, Netflix is planning to bring the LCC and BOC under one roof — a single integrated facility that combines broadcast operations and cloud monitoring into a unified space. The physical separation between those two functions has always introduced friction at the seams. Closing it is the logical next step.</p><h3>Operational Principles for Live at Scale</h3><p>Building a live operations discipline means accepting one constraint above all others: you cannot optimize for efficiency before you have built for reliability.</p><p>Netflix designed for quality first: Standardized runbooks, tiered event structures, pre-documented failure modes, so the 50th show runs as smoothly as the fifth. Off-the-shelf monitoring tools with propagation delays don’t meet that bar. The Netflix Live Control Center and Live Control Room platforms exist because observability at live scale is a product decision that demands the same design rigor as the pipeline it monitors, turning millions of telemetry events per second into something a small team can act on in real time. Technical systems and human systems have to scale together, and the most reliable incident response plan is always the one written before anyone needs it.</p><p>The operational model is also a cultural one. Bringing contingent operators into a proprietary tech stack requires deliberate onboarding design. The vendor model only works when documentation is built to be followed confidently by someone new within their first week. <strong>Beyond process, the most durable parts of how Netflix runs live operations reflect something the </strong><a href="https://jobs.netflix.com/culture"><strong>Netflix culture memo</strong></a><strong> makes explicit: the best ideas come from anywhere.</strong> In practice, that means frontline operators catching issues that engineers miss, vendor staff surfacing workflow friction that improves the system for everyone who follows, and a team that treats candid feedback as standard practice rather than an exception. The technology, the slate, and the scale keep changing. The discipline stays current by staying curious and iterating on the tools, the runbooks, and the team.</p><h3>Conclusion: What’s Next</h3><p>With 2026 already off to a successful start in operational scaling, we’re excited to shift our focus to the upcoming launch of our new Live Broadcast Operations Center in Los Angeles and our new Live Operations Center (LOC) in West London. The LOC will initiate Netflix’s follow-the-sun coverage as live content continues to grow with over 400 live events in 2026, including the launch of 24/7 linear free-to-air broadcast channels with TF1 this summer. On the technical front, further development of automated alerting tools and monitoring by exception will continue to reduce operations’ manual workload.</p><p>In 2023, the engineers led the operations. By 2026, they had developed systems that mostly ran themselves, with a dedicated operational team ensuring they operated smoothly for millions of members. The technology behind Netflix’s Live content has been documented throughout this series, but what runs alongside the tech stack is a set of operational principles, rehearsed incident management processes, and monitoring infrastructure that had to be created from scratch and continues to develop.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*7KV_D2VSRlja_fWmHmuOBw.jpeg" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*sc4doCe3-h3mi7V9trxBbQ.jpeg" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*dZKV2OIroXQRYA6UZBZAEw.jpeg" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*w6eGAv10BZWqNBVXNCvqRA.jpeg" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*y-iRMBW0Ae2YDiDjNzwq-Q.jpeg" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*dPoOkp75Rjy5CRskFQBoQQ.jpeg" /></figure><p>A special thanks to Te-Yuan Huang, Rob Saltiel, Tara Kozuback, Chris Carey, Di Li, Patrick Li, Anne Aaron, and Melissa “Mouse” Merencillo for their support on this article.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=33e2a311c597" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/the-human-infrastructure-how-netflix-built-the-operations-layer-behind-live-at-scale-33e2a311c597">The Human Infrastructure: How Netflix Built the Operations Layer Behind Live at Scale</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Evaluating Netflix Show Synopses with LLM-as-a-Judge]]></title>
            <link>https://netflixtechblog.com/evaluating-netflix-show-synopses-with-llm-as-a-judge-6269251e6f28?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/6269251e6f28</guid>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 10 Apr 2026 16:26:01 GMT</pubDate>
            <atom:updated>2026-04-13T14:26:33.614Z</atom:updated>
            <content:encoded><![CDATA[<p>by <a href="https://www.linkedin.com/in/gabrielaalessio/">Gabriela Alessio</a>, <a href="https://www.linkedin.com/in/cameronntaylor/">Cameron Taylor</a>, and <a href="https://www.linkedin.com/in/cwolferesearch/">Cameron R. Wolfe</a></p><h3>Introduction</h3><p>When members log into Netflix, one of the hardest choices is what to watch. The challenge isn’t a lack of options — <em>there are thousands of titles</em> — but finding the most intriguing one is complex and deeply personal. To help, we surface <a href="https://netflixtechblog.com/artwork-personalization-c589f074ad76">personalized promotional assets</a>, especially the show synopsis — <em>a brief description highlighting key plot elements, with cues like genre or talent</em>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*A2j9Xni86SVnF_VyjFNwfw.png" /></figure><p>Strong synopses help members scan, understand, and choose. Poor synopses frustrate, mislead, and drive abandonment. Ensuring high-quality synopses is essential, but scaling quality validation is hard. We host hundreds of thousands of synopses, usually with multiple variants per show. We need to ensure quality at scale so every member gets a consistently great experience every time they read a synopsis. This approach helps us scale high‑quality synopsis coverage for our rapidly expanding catalog, enabling greater speed and coverage without sacrificing quality.</p><p>This report outlines our LLM-based approach for evaluating synopsis quality. Using recent advances in agents, reasoning, and LLM-as-a-Judge, we score four key synopsis quality dimensions, achieving 85%+ agreement with creative writers. Additionally, we show that higher LLM judge quality is correlated with key streaming metrics, <em>allowing us to proactively identify and fix impactful issues weeks or months before a show debuts on Netflix</em>.</p><h3>The Making of a “Good” Synopsis</h3><p>Writing high-quality synopses requires creative expertise. Our expert creative leads are best positioned to craft the creative approaches and define quality standards. However, AI can help us consistently evaluate these expert-driven quality criteria at scale. Synopsis quality at Netflix, which our system aims to predict, is viewed along two dimensions:</p><ol><li><em>Creative Quality</em>: members of our creative writing team assess synopsis quality according to our internal writing guidelines and rubrics.</li><li><em>Member Implicit Feedback</em>: we measure the relative impact of a particular show synopsis on core streaming metrics.</li></ol><p>These two definitions of quality capture distinct and important aspects of quality, one focused upon creative excellence and the other upon utility to members.</p><h4>Creative Quality</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*KzbYGovn903y_ZGysqdnYw.png" /></figure><p>For this project, we evaluate synopses against a subset of our creative writing quality rubric — <em>the same criteria to which human writers would adhere</em>. These quality rubrics change over time as quality standards evolve. Given Netflix’s distinctive voice and elevated editorial standards, the quality bar is high. Each criterion has extensive guidelines with examples across regions, genres, and synopsis types.</p><p><strong>Human evaluation.</strong> We began by partnering with a group of creative writing experts to iteratively refine our definition of creative quality. We initially labeled ~1,000 diverse synopses, where three expert writers scored each against the criteria and explained their ratings. Due to the subjectivity of the task, early instance-level agreement was low. To reach a better consensus, we conducted calibration rounds (~50 synopses per round), surfaced disagreements, and evolved our quality scoring guidelines. Key interventions that were found to improve agreement include:</p><ul><li>Using binary scores (instead of 1–4 Likert scores).</li><li>Allowing writers to reference past examples.</li><li>Maintaining a searchable taxonomy of common errors.</li></ul><p><strong>Golden evaluation data. </strong>After eight calibration rounds, writer agreement reached ~80%. To further stabilize labels, we used a model-in-the-loop consensus where:</p><ul><li>Multiple writers score each synopsis.</li><li>An LLM, guided by the rubric, aggregates to a final label.</li><li>Writers review cases with substantial disagreement.</li></ul><p>The result is a golden set of ~600 synopses with binary, criteria-level scores and explanations — <em>our North Star for aligning an LLM judge with expert opinion</em>.</p><h4>Member Implicit Feedback</h4><p>Netflix gauges implicit member feedback on a synopsis with two metrics:</p><ol><li><em>Take Fraction</em>: how often members who see a title’s synopsis choose to start watching it.</li><li><em>Abandonment Rate</em>: how often members start a title but stop watching soon after.</li></ol><p>Higher take fraction indicates more choosing, while lower abandonment suggests authentic, non-misleading presentation. Both of these metrics have been validated via A/B testing to serve as short-term behavioral proxies for long-term member retention. As part of evaluating our system, we also study the ability of LLM-derived quality scores to predict short-term engagement metrics. This step confirms that our scores capture behaviorally meaningful signals and assesses our ability to forecast member response to a given synopsis.</p><h3>Scaling Quality Scoring with LLM-as-a-Judge</h3><p>We begin our experiments by creating simple, per-criteria prompts that:</p><ol><li>Supply criterion-specific show metadata.</li><li>Summarize the relevant quality guidelines.</li><li>Use <a href="https://arxiv.org/abs/2205.11916">zero-shot chain-of-thought prompting</a> to elicit an explanation.</li><li>Request a binary decision for the synopsis.</li></ol><p>Using a single prompt to evaluate all quality criteria is found to overload the LLM and yields poor performance — <em>dedicated judges for each criteria perform better</em>. Because criteria are unique, each task has its own setup, but there are some shared components:</p><ul><li>We use the same LLM for all criteria.</li><li>The judge always outputs an explanation before its final score.</li><li>Final scores are binary.</li></ul><p>Due to our use of binary scoring, judges can be evaluated with simple accuracy metrics over the golden dataset. Next, we summarize the experiments that led to our final system.</p><p><strong>Prompt optimization.</strong> Because LLMs are sensitive to prompt phrasing, we apply <a href="https://arxiv.org/abs/2305.03495">Automatic Prompt Optimization (APO)</a> over a ~300-sample dev set. Scoring guidelines are provided as additional context to the prompt optimizer. After APO, we manually refine candidate prompts with the help of an LLM, yielding initial prompts with accuracies shown below. These prompts work well for some criteria (e.g., precision) but poorly for others (e.g., clarity), highlighting criterion-specific nuances.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*voqhaot2_6G67DSzH0rFzQ.png" /></figure><p><strong>Improved reasoning.</strong> Many failures of our initial system arise due to a lack of accurate reasoning through highly-subjective evaluation examples. To improve reasoning accuracy, we leverage two forms of inference-time scaling:</p><ul><li><em>Longer rationales</em>: increase the length of the rationale or explanation generated by the LLM prior to producing a final score.</li><li><em>Consensus scoring</em>: sample several outputs from the LLM and aggregate their scores to produce the final result.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*aEMhMA1OHAGkrvQ0Ljhesg.png" /></figure><p><strong>Tiered rationales.</strong> Using tone as an example, we tested whether longer rationales are helpful by defining three rationale length tiers (shown above) and comparing their accuracies. Accuracy rises with longer rationales but returns are diminishing. Medium rationales noticeably outperform short ones, while long rationales offer only a slight additional gain; see below.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ER-gjiA0IhaYFLr1EchOuA.png" /></figure><p>Longer rationales improve performance but degrade human-readability, which is problematic given that explanations are key pieces of evidence for creative experts. As a solution, we adopt tiered rationales: <em>the judge reasons at any length but concisely summarizes its reasoning process prior to the final score. </em>Tiered rationales preserve the benefits of extended reasoning, make outputs easier to inspect, and even benefit scoring accuracy. For example, our tone evaluator improves from 86.55% to 87.85% binary accuracy when using tiered rationales.</p><p><strong>Consensus scoring.</strong> We can also allocate more inference-time compute by sampling multiple outputs per synopsis and aggregating their scores. We aggregate via a rounded average to ensure that the final score remains binary. For tone and clarity criteria with tiered rationales, 5× consensus scoring yields a clear accuracy boost as shown below.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*U7YSyivmXC_M_kw51-tl-g.png" /></figure><p>Consensus scoring on the precision evaluator, which uses a vanilla (short) chain-of-thought, yields no benefit. As an explanation, we notice that longer rationales increase variance in scores across multiple outputs, while short rationales yield consistent scores. Consensus may be most useful for evaluators with longer rationales, where it helps to stabilize score variance. When shorter rationales are used, all scores tend to be the same, making consensus less meaningful.</p><p><strong>What about reasoning models?</strong> While our setup elicits reasoning from a standard LLM, we also explored quality scoring with true reasoning models (i.e., models that generate long reasoning trajectories prior to final output). For tone, using a reasoning model with 5× consensus yields improving accuracy with increasing reasoning effort, even outperforming tiered rationales at the highest reasoning effort; see below. However, we skip reasoning models in our final system, as they significantly increase inference costs for only a marginal performance gain.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*D9Ph4ASfjx5pzdfx5EnODQ.png" /></figure><p><strong>Agents-as-a-Judge for factuality.</strong> Synopses have four common types of factuality errors:</p><ol><li>Incorrect plot information.</li><li>Incorrect metadata (e.g., genre, location, release date).</li><li>Incorrect on- or off-screen talent.</li><li>Incorrect award information.</li></ol><p>Detecting these factuality errors requires comparing the synopsis to ground-truth context, where necessary context varies per criteria. For example, plot information requires a plot summary or script, while award information needs a list of awards. As we have learned, simplicity drives reliability: <em>too much context or too many criteria harms accuracy</em>. Motivated by this idea, we adopt factuality agents, where each agent evaluates one narrow aspect of factuality.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*oXnk59qsQPASZAgTnPi0HA.png" /></figure><p>An agent receives context tailored to one facet of factuality and produces both a rationale and a binary factuality score. The final score of the Agents-as-a-Judge system is the minimum factuality score across agents — <em>any failed aspect yields an overall fail</em>. All rationales are fed to an LLM aggregator to produce a combined rationale to accompany the final score. As shown below, leveraging factuality agents significantly benefits scoring accuracy. Further benefits are achieved by using tiered rationales and consensus scoring within each agent.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*_aZe60Sa1dwWvPKIH7YRgQ.png" /></figure><p><strong>Final system. </strong>In summary, our automatic evaluation system uses a combination of standard LLM-as-a-Judge, tiered rationales, consensus scoring, and Agents-as-a-Judge to maximize binary scoring accuracy for each criteria. A summary of the techniques used for each criteria and the associated binary scoring accuracy is provided below.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qZmJgN46QRLYpEhtlJutEQ.png" /></figure><h3>Member Validation of LLM-as-a-Judge</h3><p>Beyond expert agreement, we also study how LLM-as-a-Judge scores relate to member behavior. This analysis serves two goals:</p><ul><li>Further validating LLM-judge accuracy.</li><li>Linking creative quality to member-perceived quality.</li></ul><p>Framed as predictors of member outcomes, LLM judges help us assess how promotional assets affect viewing and determine which creative attributes matter most to members discovering content they enjoy. To perform this analysis, we take advantage of the fact that most shows have multiple, personalized synopses (i.e., a synopsis “suite”). Using this suite, we can measure the causal effect of synopsis selection on metrics like take fraction and abandonment rate.</p><p><strong>Our methodology. </strong>We correlate synopsis performance (take fraction or abandonment) with LLM quality scores. Specifically, within each show s, we relate changes in a synopsis’s LLM score to changes in its performance, normalizing by the show-level standard deviation and clustering standard errors by show; see below.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*4-wIexhYKeqcMyf-a0TqIg.png" /></figure><p>β captures the average association between within-show changes in LLM score and changes in performance. While we don’t have clean, experimental variation in LLM scores, this analysis still validates predictive value and practical utility.</p><p><strong>Member-focused results.</strong> We report correlations for individual LLM criteria and a “Weighted Score” that combines all criteria to reduce noise and maximize signal from behavioral data. As shown below, results show promising prediction of take fraction and abandonment. Precision and clarity are especially predictive, and the weighted score provides a statistically useful signal of higher take and lower abandonment. In short, LLM evaluators capture factors that matter to members, making them a valuable tool for monitoring synopsis quality and engagement.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*HnRJo-DOrH9ifgekC5xq6A.png" /></figure><h3>Closing Remarks</h3><p>The LLM-as-a-Judge system used to evaluate show synopses at Netflix is the result of extensive experimentation grounded in both creative expertise and member outcomes. Building an automatic evaluation system that works reliably in practice is hard, and the approach we have described reflects countless lessons learned through iteration to improve accuracy and scalability. We have validated the system extensively with human evaluation at both the system and component levels, and we have shown that its outputs correlate with key streaming metrics. As a result, we are confident that it captures the dimensions of synopsis quality that matter most — both creatively and from the member perspective — which has driven its widespread adoption in the Netflix synopsis authoring workflow.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6269251e6f28" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/evaluating-netflix-show-synopses-with-llm-as-a-judge-6269251e6f28">Evaluating Netflix Show Synopses with LLM-as-a-Judge</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Stop Answering the Same Question Twice: Interval-Aware Caching for Druid at Netflix Scale]]></title>
            <link>https://netflixtechblog.com/stop-answering-the-same-question-twice-interval-aware-caching-for-druid-at-netflix-scale-22fadc9b840e?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/22fadc9b840e</guid>
            <category><![CDATA[apache-druid]]></category>
            <category><![CDATA[cache]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Mon, 06 Apr 2026 22:15:14 GMT</pubDate>
            <atom:updated>2026-04-06T22:15:13.989Z</atom:updated>
            <content:encoded><![CDATA[<p><em>By </em><a href="https://www.linkedin.com/in/sykesb/"><em>Ben Sykes</em></a></p><p>In a <a href="https://netflixtechblog.com/how-netflix-uses-druid-for-real-time-insights-to-ensure-a-high-quality-experience-19e1e8568d06">previous post</a>, we described how Netflix uses Apache Druid to ingest millions of events per second and query trillions of rows, providing the real-time insights needed to ensure a high-quality experience for our members. Since that post, our scale has grown considerably.</p><p>With our database holding over 10 trillion rows and regularly ingesting up to 15 million events per second, the value of our real-time data is undeniable. But this massive scale introduced a new challenge: queries. The live show monitoring, dashboards, automated alerting, canary analysis, and A/B test monitoring that are built on top of Druid became so heavily relied upon that the repetitive query load started to become a scaling concern in itself.</p><p>This post describes an experimental caching layer we built to address this problem, and the trade-offs we chose to accept.</p><h3><strong>The Problem</strong></h3><p>Our internal dashboards are heavily used for real-time monitoring, especially during high-profile live shows or global launches. A typical dashboard has 10+ charts, each triggering one or more Druid queries; one popular dashboard with 26 charts and stats generates 64 queries per load. When dozens of engineers view the same dashboards and metrics for the same event, the query volume quickly becomes unmanageable.</p><p>Take the popular dashboard above: 64 queries per load, refreshing every 10 seconds, viewed by 30 people. That’s 192 queries per second from one dashboard, mostly for nearly identical data. We still need Druid capacity for automated alerting, canary analysis, and ad-hoc queries. And because these dashboards request a rolling last-few-hours window, each refresh changes slightly as the time range advances.</p><p>Druid’s built-in caches are effective. Both the full-result cache and the per-segment cache. But neither is designed to handle the continuous, overlapping time-window shifts inherent to rolling-window dashboards. The full-result cache misses for two reasons.</p><ul><li>If the time window shifts even slightly, the query is different, so it’s a cache miss.</li><li>Druid deliberately refuses to cache results that involve realtime segments (those still being indexed), because it values deterministic, stable cache results and query correctness over a higher cache hit rate.</li></ul><p>The per-segment cache does help avoid redundant scans on historical nodes, but we still need to collect those cached segment results from each data node and merge them in the brokers with data from the realtime nodes for every query.</p><p>During major shows, rolling-window dashboards can generate a flood of near-duplicate queries that Druid’s caches mostly miss, creating heavy redundant load. At our scale, solving this by simply adding more hardware is prohibitively expensive.</p><p>We needed a smarter approach.</p><h3>The Insight</h3><p>When a dashboard requests the last 3 hours of data, the vast majority of that data, everything except the most recent few minutes, is already settled. The data from 2 hours ago won’t change.</p><p>What if we could remember the older portions of the result and only ask Druid for the part that’s actually new?</p><p>This is the core idea behind a new caching service that understands the structure of Druid queries and serves previously-seen results from cache while fetching only the freshest portion from Druid.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*8bFAtFl8Z5pwEyoOgK54ag.png" /></figure><h3><strong>A Deliberate Trade-Off</strong></h3><p>Before diving into the implementation, it’s worth being explicit about the trade-off we’re making. Caching query results introduces some staleness, specifically, up to 5 seconds for the newest data. This is acceptable for most of our operational dashboards, which refresh every 10 to 30 seconds. In practice, many of our queries already set an end time of now-1m or now-5s to avoid the “flappy tail” that can occur with currently-arriving data.</p><p>Since our end-to-end data pipeline latency is typically under 5 seconds at P90, a 5-second cache TTL on the freshest data introduces negligible additional staleness on top of what’s already inherent in the system. We decided it was better to accept this small amount of staleness in exchange for significantly lower query load on Druid. But a 5s cache on its own is not very useful.</p><h3>Exponential TTLs</h3><p>Not all data points are equally trustworthy. In real-time analytics, there’s a well-known late-arriving data problem. Events can arrive out of order or be delayed in the ingestion pipeline. A data point from 30 seconds ago might still change as late-arriving events trickle in. A data point from 30 minutes ago is almost certainly final.</p><p>We use this observation to set cache TTLs that increase exponentially with the age of the data. Data less than 2 minutes old gets a minimum TTL of 5 seconds. After that, the TTL doubles for each additional minute of age: 10 seconds at 2 minutes old, 20 seconds at 3 minutes, 40 seconds at 4 minutes, and so on, up to a maximum TTL of 1 hour.</p><p>The effect is that fresh data cycles through the cache rapidly, so any corrections from late-arriving events in the most recent couple of minutes are picked up quickly. Older data lingers much longer, because our confidence in its accuracy grows with time.</p><p>For a 3-hour rolling window, the exponential TTL ensures the vast majority of the query is served from the cache, leaving Druid to only scan the most recent, unsettled data.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*oyLRzcOO7otVniQM6khD6g.png" /></figure><h3><strong>Bucketing</strong></h3><p>If we were to use a single-level cache key for the query and interval, similar to Druid’s existing result-level cache, we wouldn’t be able to extract only the relevant time range from cached results. A shifted window means a different key, which means a cache miss.</p><p>Instead, we use a map-of-maps. The top-level key is the query hash without the time interval; the inner keys are timestamps bucketed to the query granularity (or 1 minute, whichever is larger) and encoded as big-endian bytes so lexicographic order matches time. This enables efficient range scans; fetching all cached buckets between times A and B for a query hash. A 3-hour query at 1-minute granularity becomes 180 independent cached buckets, each with its own TTL; when the window shifts (e.g., 30 seconds later), we reuse most buckets from cache and only query Druid for the new data.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*6DL-4Lpu_CqfZBVmF2rNvg.png" /></figure><h3><strong>How It Works</strong></h3><p>Today, the cache runs as an external service integrated transparently by intercepting requests at the Druid Router and redirecting them to the cache. If the cache fully satisfies a request, it returns the result; otherwise it shrinks the time interval to the uncached portion and calls back into the Router, bypassing the redirect to query Druid normally. Non-cached requests (e.g., metadata queries or queries without time group-bys) pass straight through to Druid unchanged.</p><p>This intercepting proxy design allows us to enable or disable caching without any client changes and is a key to its adoption. We see this setup as temporary while we work out a way to better integrate this capability into Druid more natively.</p><p>When a cacheable query arrives, those that are grouping-by time (timeseries, groupBy), the cache performs the following steps.</p><p><strong>Parsing and Hashing.</strong> We parse each incoming query to extract the time interval, granularity, and structure, then compute a SHA-256 hash of the query with the time interval and parts of the context removed. That hash is the cache key: it encodes <em>what</em> is being asked (datasource, filters, aggregations, granularity) but not <em>when</em>, so the same logical query over different overlapping time windows maps to the same cache entry. There are some context properties that can alter the response structure or contents, so these are included in the cache-key.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*QaeamWjRYGEQkW6t-XePJw.png" /></figure><p><strong>Cache Lookup.</strong> Using the cache key, we fetch cached points within the requested range, but only if they’re contiguous from the start. Because bucket TTLs can expire unevenly, gaps can appear; when we hit a gap, we stop and fetch all newer data from Druid. This guarantees a complete, unbroken result set while sending at most one Druid query, rather than “filling gaps” with multiple small, fragmented queries that would increase Druid load.</p><p><strong>Fetching the Missing Tail.</strong> On a partial cache hit (e.g., 2h 50m of a 3h window), we rebuild the query with a narrowed interval for the missing 10 minutes and send only that to Druid. Since Druid then scans just the recent segments for a small time range, the query is usually faster and cheaper than the original.</p><p><strong>Combining.</strong> The cached data and fresh data are concatenated, sorted by timestamp, and returned to the client. From the client’s perspective, the response looks identical to what Druid would have returned, same JSON format, same fields.</p><p><strong>Asynchronous Caching.</strong> The fresh data from Druid is parsed into individual time-granularity buckets and written back to the cache asynchronously, so we don’t add latency to the response path.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*k3uoqCFgzbRZEuiNo_Rlzw.png" /></figure><h3>Negative Caching</h3><p>Some metrics are sparse. Certain time buckets may genuinely have no data. Without special handling, the cache would treat these empty buckets as gaps and re-query Druid for them every time.</p><p>We handle this by caching empty sentinel values for time buckets where Druid returned no data. Our gap-detection logic recognizes these empty entries as valid cached data rather than missing data, preventing needless re-queries for naturally sparse metrics.</p><p>However, we’re careful not to negative-cache trailing empty buckets. If a query returns data up to minute 45 and nothing after, we only cache empty entries for gaps <em>between</em> data points, not after the last one. This avoids incorrectly caching “no data” for time periods where events simply haven’t arrived yet, which would exacerbate the chart delays of late arriving data.</p><h3>The Storage Layer</h3><p>For the backing store, we use Netflix’s <a href="https://netflixtechblog.com/introducing-netflixs-key-value-data-abstraction-layer-1ea8a0a11b30">Key-Value Data Abstraction Layer (KVDAL)</a>, backed by Cassandra. KVDAL provides a two-level map abstraction, a natural fit for our needs. The outer key is the query hash, and the inner keys are timestamps. Crucially, KVDAL supports independent TTLs on each inner key-value pair, eliminating the need for us to manage cache eviction manually.</p><p>This two-level structure gives us efficient range queries over the inner keys, which is exactly what we need for partial cache lookups: “give me all cached buckets between time A and time B for query hash X.”</p><h3><strong>Results</strong></h3><p>The biggest win is during high-volume events (e.g., live shows): when many users view the same dashboards, the cache serves most identical queries as full hits, so the query rate reaching Druid is essentially the same with 1 viewer or 100. The scaling bottleneck moves from Druid’s query capacity to the much cheaper-to-scale cache, and with ~5.5 ms P90 cache responses, dashboards load faster for everyone.</p><p>On a typical day, 82% of real user queries get at least a partial cache hit, and 84% of result data is served from cache. As a result, the queries that reach Druid scan much narrower time ranges, touching fewer segments and processing less data, freeing Druid to focus on aggregating the newest data instead of repeatedly re-querying historical segments.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*S2tG3E2KlGIujSaEg4oetA.png" /></figure><p>An experiment validated this, showing about a 33% drop in queries to Druid and a 66% improvement in overall P90 query times. It also cut result bytes and segments queried, and in some cases, enabling the cache reduced result bytes by more than 14x. Caveat: the size of these gains depends heavily on how similar and repetitive the query workload is.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/592/1*B-j0CnKKr5caFQRI8zVZ1A.png" /></figure><h3>Looking Ahead</h3><p>This caching layer is still experimental, but results are promising and we’re exploring next steps. We’ve added partial support for templated SQL so dashboard tools can benefit without writing native Druid queries.</p><p>Longer term, we’d like interval-aware caching to be built into Druid: an external proxy adds infrastructure to manage, extra network hops, and workarounds (like SQL templating) to extract intervals. Implemented inside Druid, it could be more efficient, with direct access to the query planner and segment metadata, and benefit the broader community without custom infrastructure. We’d likely ship it as an opt-in, configurable, result-level cache in the Brokers, with metrics to tune TTLs and measure effectiveness. Please leave a comment if you have a use-case that could benefit from this feature.</p><p>More broadly, this strategy, splitting time-series results into independently cached, granularity-aligned buckets with age-based exponential TTLs, isn’t Druid-specific and could apply to any time-series database with frequent overlapping-window queries.</p><h3>Summary</h3><p>As more Netflix teams rely on real-time analytics, query volume grows too. Dashboards are essential at our scale, but their popularity can become a scaling bottleneck. By inserting an intelligent cache between dashboards and Druid, one that understands query structure, breaks results into granularity-aligned buckets, and trades a small amount of staleness for much lower Druid load, we’ve increased query capacity without scaling infrastructure proportionally, and hope to deliver these benefits to the Druid community soon as a built-in Druid feature.</p><p>Sometimes the best way to handle a flood of queries is to stop answering the same question twice.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=22fadc9b840e" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/stop-answering-the-same-question-twice-interval-aware-caching-for-druid-at-netflix-scale-22fadc9b840e">Stop Answering the Same Question Twice: Interval-Aware Caching for Druid at Netflix Scale</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Powering Multimodal Intelligence for Video Search]]></title>
            <link>https://netflixtechblog.com/powering-multimodal-intelligence-for-video-search-3e0020cf1202?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/3e0020cf1202</guid>
            <category><![CDATA[distributed-systems]]></category>
            <category><![CDATA[semantic-search]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[annotations]]></category>
            <category><![CDATA[asset-management-software]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Sat, 04 Apr 2026 00:44:32 GMT</pubDate>
            <atom:updated>2026-04-06T01:29:35.470Z</atom:updated>
            <content:encoded><![CDATA[<h3>Synchronizing the Senses: Powering Multimodal Intelligence for Video Search</h3><p>By <a href="https://www.linkedin.com/in/meenakshijindal/">Meenakshi Jindal</a> and <a href="https://www.linkedin.com/in/~munya/">Munya Marazanye</a></p><p>Today’s filmmakers capture more footage than ever to maximize their creative options, often generating hundreds, if not thousands, of hours of raw material per season or franchise. Extracting the vital moments needed to craft compelling storylines from this sheer volume of media is a notoriously slow and punishing process. When editorial teams cannot surface these key moments quickly, creative momentum stalls and severe fatigue sets in.</p><p>Meanwhile, the broader search landscape is undergoing a profound transformation. We are moving beyond simple keyword matching toward AI-driven systems capable of understanding deep context and intent. Yet, while these advances have revolutionized text and image retrieval, searching through video, the richest medium for storytelling, remains a daunting “needle in a haystack” challenge.</p><p>The solution to this bottleneck cannot rely on a single algorithm. Instead, it demands orchestrating an expansive ensemble of specialized models: tools that identify specific characters, map visual environments, and parse nuanced dialogue. The ultimate challenge lies in unifying these heterogeneous signals, textual labels, and high-dimensional vectors into a cohesive, real-time intelligence. One that cuts through the noise and responds to complex queries at the speed of thought, truly empowering the creative process.</p><h3>Why Video Search is Deceptively Complex</h3><p>Since video is a multi-layered medium, building an effective search engine required us to overcome significant technical bottlenecks. Multi-modal search is exponentially more complex than traditional indexing: it demands the unification of outputs from multiple specialized models, each analyzing a different facet of the content to generate its own distinct metadata. The ultimate challenge lies in harmonizing these heterogeneous data streams to support rich, multi-dimensional queries in real time.</p><ol><li><strong>Unifying the Timeline<br></strong>To ensure critical moments aren’t lost across scene boundaries, each model segments the video into overlapping intervals. The resulting metadata varies wildly, ranging from discrete text-based object labels to dense vector embeddings. Synchronizing these disjointed, multi-modal timelines into a unified chronological map presents a massive computational hurdle.</li><li><strong>Processing at Scale<br></strong>A standard 2,000-hour production archive can contain over 216 million frames. When processed through an ensemble of specialized models, this baseline explodes into billions of multi-layered data points. Storing, aligning, and intersecting this staggering volume of records while maintaining sub-second query latency far exceeds the capabilities of traditional database architectures.</li><li><strong>Surfacing the Best Moments<br></strong>Surface-level mathematical similarity is not enough to identify the most relevant clip. Because continuous shots naturally generate thousands of visually redundant candidates, the system must dynamically cluster and deduplicate results to surface the singular best match for a given scene. To achieve this, effective ranking relies on a sophisticated hybrid scoring engine that weighs symbolic text matches against semantic vector embeddings, ensuring both precision and interpretability.</li><li><strong>Zero-Friction Search<br></strong>For filmmakers, search is a stream-of-consciousness process, and a ten-second delay can disrupt the creative flow. Because sequential scanning of raw footage is fundamentally unscalable, our architecture is built to navigate and correlate billions of vectors and metadata records efficiently, operating at the speed of thought.</li></ol><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*wVRT7XY4C9bzNHE-lk-ViA.png" /><figcaption><em>Figure 1: Unified Multimodal Result Processing</em></figcaption></figure><h3>The Ingestion and Fusion Pipeline</h3><p>To ensure system resilience and scalability, the transition from raw model output to searchable intelligence follows a decoupled, three-stage process:</p><h4>1. Transactional Persistence</h4><p>Raw annotations are ingested via high-availability <a href="https://netflixtechblog.com/data-ingestion-pipeline-with-operation-management-3c5c638740a8">pipelines</a> and stored in our <a href="https://netflixtechblog.com/scalable-annotation-service-marken-f5ba9266d428">annotation service</a>, which leverages <a href="https://cassandra.apache.org/_/index.html">Apache Cassandra</a> for distributed storage. This stage strictly prioritizes data integrity and high-speed write throughput, guaranteeing that every piece of model output is safely captured.</p><pre>{<br>  &quot;type&quot;: &quot;SCENE_SEARCH&quot;,<br>  &quot;time_range&quot;: {<br>    &quot;start_time_ns&quot;: 4000000000,<br>    &quot;end_time_ns&quot;: 9000000000<br>  },<br>  &quot;embedding_vector&quot;: [<br>    -0.036, -0.33, -0.29 ...<br>  ],<br>  &quot;label&quot;: &quot;kitchen&quot;,<br>  &quot;confidence_score&quot;: 0.72<br>}</pre><p><em>Figure 2: Sample Scene Search Model Annotation Output</em></p><h4>2. Offline Data Fusion</h4><p>Once the annotation service securely persists the raw data, the system publishes an event via <a href="https://kafka.apache.org/">Apache Kafka</a> to trigger an asynchronous processing job. Serving as the architecture’s central logic layer, this offline pipeline handles the heavy computational lifting out-of-band. It performs precise temporal intersections, fusing overlapping annotations from disparate models into cohesive, unified records that empower complex, multi-dimensional queries.</p><p>Cleanly decoupling these intensive processing tasks from the ingestion pipeline guarantees that complex data intersections never bottleneck real-time intake. As a result, the system maintains maximum uptime and peak responsiveness, even when processing the massive scale of the Netflix media catalog.</p><p>To achieve this intersection at scale, the offline pipeline normalizes disparate model outputs by mapping them into fixed-size temporal buckets. This discretization process unfolds in three steps:</p><ul><li><strong>Bucket Mapping:</strong> Continuous detections are segmented into discrete intervals. For example, if a model detects a character <em>“Joey”</em> from seconds 2 through 8, the pipeline maps this continuous span of frames into seven distinct one-second buckets.</li><li><strong>Annotation Intersection:</strong> When multiple models generate annotations for the same temporal bucket, such as character recognition <em>“Joey”</em> and scene detection <em>“kitchen” </em>overlapping in second 4, the system fuses them into a single, comprehensive record.</li><li><strong>Optimized Persistence:</strong> These newly enriched records are written back to Cassandra as distinct entities. This creates a highly optimized, second-by-second index of multi-modal intersections, perfectly associating every fused annotation with its source asset.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ft36HZpYbqUBfhP8p0iShw.png" /><figcaption><em>Figure 3: Temporal Data Fusion with Fixed-Size Time Buckets</em></figcaption></figure><p>The following record shows the overlap of the character<em> “Joey”</em> and scene <em>“kitchen” </em>annotations during a 4 to 5 second window in a video asset:</p><pre>{<br>  &quot;associated_ids&quot;: {<br>    &quot;MOVIE_ID&quot;: &quot;81686010&quot;,<br>    &quot;ASSET_ID&quot;: &quot;01325120–7482–11ef-b66f-0eb58bc8a0ad&quot;<br>  },<br>  &quot;time_bucket_start_ns&quot;: 4000000000,<br>  &quot;time_bucket_end_ns&quot;: 5000000000,<br>  &quot;source_annotations&quot;: [<br>    {<br>      &quot;annotation_id&quot;: &quot;7f5959b4–5ec7–11f0-b475–122953903c43&quot;,<br>      &quot;annotation_type&quot;: &quot;CHARACTER_SEARCH&quot;,<br>      &quot;label&quot;: &quot;Joey&quot;,<br>      &quot;time_range&quot;: {<br>        &quot;start_time_ns&quot;: 2000000000,<br>        &quot;end_time_ns&quot;: 8000000000<br>      }<br>    },<br>    {<br>      &quot;annotation_id&quot;: &quot;c9d59338–842c-11f0–91de-12433798cf4d&quot;,<br>      &quot;annotation_type&quot;: &quot;SCENE_SEARCH&quot;,<br>      &quot;time_range&quot;: {<br>        &quot;start_time_ns&quot;: 4000000000,<br>        &quot;end_time_ns&quot;: 9000000000<br>      },<br>      &quot;label&quot;: &quot;kitchen&quot;,<br>      &quot;embedding_vector&quot;: [<br>        0.9001, 0.00123 ....<br>      ]<br>    }<br>  ]<br>}</pre><p><em>Figure 4: Sample Intersection Record for Character + Scene Search</em></p><h4>3. Indexing for Real Time Search</h4><p>Once the enriched temporal buckets are securely persisted in Cassandra, a subsequent event triggers their ingestion into Elasticsearch.</p><p>To guarantee absolute data consistency, the pipeline executes upsert operations using a composite key <em>(asset ID + time bucket)</em> as the unique document identifier. If a temporal bucket already exists for a specific second of video, perhaps populated by an earlier model run, the system intelligently updates the existing record rather than generating a duplicate. This mechanism establishes a single, unified source of truth for every second of footage.</p><p>Architecturally, the pipeline structures each temporal bucket as a nested document. The root level captures the overarching asset context, while associated child documents house the specific, multi-modal annotation data. This hierarchical data model is precisely what empowers users to execute highly efficient, cross-annotation queries at scale.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Mt1fbK8AvwgNk-LY55wFqg.png" /><figcaption><em>Figure 5: Simplified Elasticsearch Document Structure</em></figcaption></figure><h3>Multimodal Discovery and Result Ranking</h3><p>The search service provides a high-performance interface for real-time discovery across the global Netflix catalog. Upon receiving a user request, the system immediately initiates a query preprocessing phase, generating a structured execution plan through three core steps:</p><ul><li><strong>Query Type Detection:</strong> Dynamically categorizes the incoming request to route it down the most efficient retrieval path.</li><li><strong>Filter Extraction:</strong> Isolates specific semantic constraints such as character names, physical objects, or environmental contexts to rapidly narrow the candidate pool.</li><li><strong>Vector Transformation:</strong> Converts raw text into high-dimensional, model-specific embeddings to enable deep, context-aware semantic matching.</li></ul><p>Once generated, the system compiles this structured plan into a highly optimized Elasticsearch query, executing it directly against the pre-fused temporal buckets to deliver instantaneous, frame-accurate results.</p><h4>Fine-Tuning Semantic Search</h4><p>To support the diverse workflows of different production teams, the system provides fine-grained control over search behavior through configurable parameters:</p><ul><li><strong>Exact vs. Approximate Search:</strong> Users can toggle between exact k-Nearest Neighbors (k-NN) for uncompromising precision, and Approximate Nearest Neighbor (ANN) algorithms (such as <a href="https://en.wikipedia.org/wiki/Hierarchical_navigable_small_world">HNSW</a>) to maintain blazing speed when querying massive datasets.</li><li><strong>Dynamic Similarity Metrics:</strong> The system supports multiple distance calculations, including <a href="https://en.wikipedia.org/wiki/Cosine_similarity">cosine similarity</a> and <a href="https://en.wikipedia.org/wiki/Euclidean_distance">Euclidean distance</a>. Because different models shape their high-dimensional vector spaces distinctly based on their underlying training architectures, the flexibility to swap metrics ensures that mathematical closeness perfectly translates to true semantic relevance.</li><li><strong>Confidence Thresholding:</strong> By establishing strict minimum score boundaries for results, users can actively prune the long tail of low-probability matches. This aggressively filters out visual noise, guaranteeing that creative teams are not distracted and only review results that meet a rigorous standard of mathematical similarity.</li></ul><h4>Textual Analysis &amp; Linguistic Precision</h4><p>To handle the deep nuances of dialogue-heavy searches, such as isolating a character’s exact catchphrase amidst thousands of hours of speech, we implement a sophisticated text analysis strategy within Elasticsearch. This ensures that conversational context is captured and indexed accurately.</p><ul><li><strong>Phrase &amp; Proximity Matching:</strong> To respect the narrative weight of specific lines (e.g., <em>“Friends don’t lie”</em> in Stranger Things), we leverage <a href="https://www.elastic.co/docs/reference/query-languages/query-dsl/query-dsl-match-query-phrase">match-phrase</a> queries with a configurable <em>slop</em> parameter. This guarantees the system retrieves the correct scene even if the user’s memory slightly deviates from the exact transcription.</li><li><strong>N-Gram Analysis for Partial Discovery:</strong> Because video search is inherently exploratory, we utilize edge <a href="https://en.wikipedia.org/wiki/N-gram">N-gram</a> tokenizers to support search-as-you-type functionality. By actively indexing dialogue and metadata substrings, the system surfaces frame-accurate results the moment an editor begins typing, drastically reducing cognitive load.</li><li><strong>Tokenization and Linguistic Stemming:</strong> To seamlessly support the global scale of the Netflix catalog, our analysis chain applies sophisticated <a href="https://en.wikipedia.org/wiki/Stemming">stemming</a> across multiple languages. This ensures a query for <em>“running”</em> automatically intersects with scenes tagged with <em>“run”</em> or <em>“ran”</em> collapsing grammatical variations into a single, unified search intent.</li><li><strong>Levenshtein Fuzzy Matching:</strong> To account for transcription anomalies or phonetic misspellings, we incorporate fuzzy search capabilities based on <a href="https://en.wikipedia.org/wiki/Levenshtein_distance">Levenshtein</a> distance algorithms. This intelligent soft-matching approach ensures that high-value shots are never lost to minor data-entry errors or imperfect queries.</li></ul><h4>Aggregations and Flexible Grouping</h4><p>The architecture operates at immense scale, seamlessly executing queries within a single title or across thousands of assets simultaneously. To combat result fatigue, the system leverages custom aggregations to intelligently cluster and group outputs based on specific parameters, such as isolating the top 5 most relevant clips of an actor per episode. This guarantees a diverse, highly representative return set, preventing any single asset from dominating the search results.</p><h4>Search Response Curation</h4><p>While temporal buckets are the internal mechanism for search efficiency, the system post-processes Elasticsearch results to reconstruct original time boundaries. The reconstruction process ensures results reflect narrative scene context rather than arbitrary intervals. Depending on the query intent, the system generates results based on two logic types:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*JamOz6AmYyLfsCOBtUY3Bg.png" /><figcaption><em>Figure 6: Depiction of Temporal Union vs Intersection</em></figcaption></figure><ul><li><strong>Union:</strong> Returns the full span of all matching annotations <em>(3–8 sec),</em> which prioritizes breadth, capturing any instance where a specified feature occurs.</li><li><strong>Intersection:</strong> Returns only the exact overlapping duration of matching signals <em>(4–6 sec).</em> The intersection logic focuses on co-occurrence, isolating moments when multiple criteria align.</li></ul><pre>{<br>  &quot;entity_id&quot;: {<br>    &quot;entity_type&quot;: &quot;ASSET&quot;,<br>    &quot;id&quot;: &quot;1bba97a1–3562–4426–9cd2-dfbacddcb97b&quot;<br>  },<br>  &quot;range_intervals&quot;: [<br>    {<br>      &quot;intersection_time_range&quot;: {<br>        &quot;start_time_ns&quot;: 4000000000,<br>        &quot;end_time_ns&quot;: 8000000000<br>      },<br>      &quot;union_time_range&quot;: {<br>        &quot;start_time_ns&quot;: 2000000000,<br>        &quot;end_time_ns&quot;: 9000000000<br>      },<br>      &quot;source_annotations&quot;: [<br>        {<br>          &quot;annotation_id&quot;: &quot;fc1525d0–93a7–11ef-9344–1239fc3a8917&quot;,<br>          &quot;annotation_type&quot;: &quot;SCENE_SEARCH&quot;,<br>          &quot;metadata&quot;: {<br>            &quot;label&quot;: &quot;kitchen&quot;<br>          }<br>        },<br>        {<br>          &quot;annotation_id&quot;: &quot;5974fb01–93b0–11ef-9344–1239fc3a8917&quot;,<br>          &quot;annotation_type&quot;: &quot;CHARACTER_SEARCH&quot;,<br>          &quot;metadata&quot;: {<br>            &quot;character_name&quot;: [<br>              &quot;Joey&quot;<br>            ]<br>          }<br>        }<br>      ]<br>    }<br>  ]<br>}</pre><p><em>Figure 7: Sample Response for “Joey” + “Kitchen” Query</em></p><h3><strong>Future Extensions</strong></h3><p>While our current architecture establishes a highly resilient and scalable foundation, it represents only the first phase of our multi-modal search vision. To continuously close the gap between human intuition and machine retrieval, our roadmap focuses on three core evolutions:</p><ul><li><strong>Natural Language Discovery:</strong> Transitioning from structured JSON payloads to fluid, conversational interfaces (e.g., <em>“Find the best tracking shots of Tom Holland running on a roof”</em>). This will abstract away underlying query complexity, allowing creatives to interact with the archive organically.</li><li><strong>Adaptive Ranking:</strong> Implementing machine learning feedback loops to dynamically refine scoring algorithms. By continuously analyzing how editorial teams interact with and select clips, the system will self-tune its mathematical definition of semantic relevance over time.</li><li><strong>Domain-Specific Personalization:</strong> Dynamically calibrating search weights and retrieval behaviors to match the exact context of the user. The platform will tailor its results depending on whether a team is cutting high-action marketing trailers, editing narrative scenes, or conducting deep archival research.</li></ul><p>Ultimately, these advancements will elevate the platform from a highly optimized search engine into an intelligent creative partner, fully equipped to navigate the ever-growing complexity and scale of global video media.</p><h3><strong>Acknowledgements</strong></h3><p>We would like to extend our gratitude to the following teams and individuals whose expertise and collaboration were instrumental in the development of this system:</p><ul><li><strong>Data Science Engineering:</strong> <a href="https://www.linkedin.com/in/nagendrak/">Nagendra Kamath</a>, <a href="https://www.linkedin.com/in/chao-pan-02791824/">Chao Pan</a>, <a href="mailto:prachees@netflix.com">Prachee Sharma</a>, <a href="https://www.linkedin.com/in/ying-liao-nyu/">Ying Liao</a> and <a href="mailto:csoo@netflix.com">Carolyn Soo</a> for the critical media model insights that informed our architectural design.</li><li><strong>Product Management:</strong> <a href="https://www.linkedin.com/in/nimesh-narayan/">Nimesh Narayan</a>, <a href="https://www.linkedin.com/in/ian-krabacher-06816710/">Ian Krabacher</a>, <a href="https://www.linkedin.com/in/ananya-ani-poddar/">Ananya Poddar</a>, <a href="https://www.linkedin.com/in/meghanbailey02/">Meghan Bailey</a> and <a href="https://www.linkedin.com/in/anitakuc/">Anita Kuc</a> for defining the user requirements and product vision.</li><li><strong>Media Production Suite Team:</strong> <a href="https://www.linkedin.com/in/szymon-borodziuk/">Szymon Borodziuk</a>, <a href="https://www.linkedin.com/in/mike-czarnota/">Mike Czarnota</a>, <a href="https://www.linkedin.com/in/dsarkowicz/?locale=en">Dominika Sarkowicz</a>, <a href="mailto:bkoval@netflix.com">Bohdan Koval</a> and <a href="https://www.linkedin.com/in/sabov/">Sasha Sabov</a> for their work in engineering the end-user search experience.</li><li><strong>Asset Management Platform Team:</strong> For their collaborative efforts in operationalizing this design and bringing the system into production.</li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=3e0020cf1202" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/powering-multimodal-intelligence-for-video-search-3e0020cf1202">Powering Multimodal Intelligence for Video Search</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Smarter Live Streaming at Scale: Rolling Out VBR for All Netflix Live Events]]></title>
            <link>https://netflixtechblog.com/smarter-live-streaming-at-scale-rolling-out-vbr-for-all-netflix-live-events-c8f833b238cc?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/c8f833b238cc</guid>
            <category><![CDATA[encoding]]></category>
            <category><![CDATA[content-delivery-platform]]></category>
            <category><![CDATA[adaptive-bitrate]]></category>
            <category><![CDATA[live-streaming]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Thu, 02 Apr 2026 21:46:27 GMT</pubDate>
            <atom:updated>2026-04-02T21:46:26.036Z</atom:updated>
            <content:encoded><![CDATA[<p>By Renata Teixeira, Zhi Li, Reenal Mahajan, and Wei Wei</p><p>On January 26, 2026, we flipped an important switch for Live at Netflix: <strong>all Live events are now encoded using VBR (Variable Bitrate) instead of CBR (Constant Bitrate)</strong>. It sounds like a small configuration change, but it required us to revisit some of the foundational assumptions behind how we deliver Live video at global scale.</p><p>VBR lets us tailor the bitrate to the actual complexity of the scene, instead of sending every second of video at roughly the same bitrate. When a scene is simple, VBR “shaves off” bits that wouldn’t improve what you see on screen; when a scene is complex, it spends more bits to preserve quality. (The more general idea is often referred to as capped variable bitrate, or capped VBR.) That makes our encodes more efficient and our network more scalable. But it also makes traffic much less predictable: large bitrate swings can overload servers and CDNs, and our old assumptions about “what bitrate equals what quality” no longer hold. As a result, we have to rethink both how we manage delivery and capacity, and which bitrates we offer for each version of the stream. In our live pipeline, we currently use AWS Elemental MediaLive, where this “capped” VBR is implemented using the QVBR (Quality‑Defined Variable Bitrate) setting.</p><h3>| Why Move Live from CBR to VBR?</h3><p>Our initial Live encoding pipeline used constant bitrate (CBR). For each encoded stream, we configured a resolution and a nominal bitrate — for example, a 1080p stream targeting 5 Mbps — and the actual bitrate stayed close to that target over time. This predictability made both capacity planning and day‑to‑day operations easier. If a server could safely deliver around 100 Gbps of Live traffic, and each stream averaged close to its nominal rate, we could admit on the order of twenty thousand concurrent sessions per server and be confident we were operating within limits. During an event, the total traffic sent by a server would change mainly when members joined or left; as long as concurrency was stable, traffic stayed relatively flat. The network saw a smooth, easy‑to‑reason‑about load profile, and large changes in throughput almost always reflected a real change in usage, not just a different scene on screen.</p><p>The problem is that content isn’t constant. A talking‑head segment in a studio or a simple animation is much easier to compress than a sequence of rapid camera moves and fast‑moving athletes in front of a highly detailed crowd. With CBR, both the easy and the hard segments get the same bitrate. In simple scenes, we spend more bits than we need; in complex scenes, we sometimes don’t spend enough.</p><p>VBR flips the objective. Rather than aiming for a fixed bitrate, the encoder aims for a target quality and is allowed to raise or lower the bitrate according to scene complexity. When the picture is easy to encode, VBR can drop the bitrate substantially below the old CBR level while keeping quality constant. When the action heats up, it can temporarily use more bits to avoid visible artifacts.</p><p>The figure below shows the per‑segment bitrate over time for the same episode of WWE RAW, encoded once with CBR and once with VBR at a nominal 8 Mbps. With CBR (blue), the bitrate wobbles a bit from segment to segment but stays close to the target; if you average it over a minute, it’s practically a straight line. We end up spending roughly the same number of bits on simple scenes, like the waiting room at the start of the stream (shaded region), as on complex scenes, like the confetti‑filled shot later in the show. With VBR (orange), the encoder can drop the bitrate for the waiting room to a small fraction of the nominal rate, while allowing the confetti‑filled shot to use a much higher bitrate to preserve quality.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*_MoqLhu01wArcD1n58SZyg.png" /><figcaption>Per-segment bitrate over a WWE RAW episode for CBR and VBR encodes at a nominal of 8 Mbps.</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/512/1*huWI8zAfmmwWq5IsxKIL3Q.png" /><figcaption><strong><em>“Waiting room” scene:</em></strong><em> visually simple and easy to compress, so VBR can safely use a low bitrate.</em></figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/512/1*jBCcIjhCgSi9kPXLrsd1FA.png" /><figcaption><strong><em>“Confetti-filled” shot:</em></strong><em> visually complex and noisy, so VBR spends many more bits to maintain quality.</em></figcaption></figure><p>At Netflix scale, the shift to VBR has several important effects. The most important is efficiency on the network: we reduce the average number of bytes we need to deliver a full event, which in turn reduces the traffic needed to fill all of the servers in Open Connect, our content delivery network (CDN), and the traffic needed to serve segments to members. The second is quality of experience. Because VBR sends fewer bits for similar video quality, we see fewer rebuffers and lower start‑up delay. Across multiple A/B tests on different Live events, we observed about 5% fewer rebuffers per hour, while transferring roughly 15% fewer bytes on average and around a 10% reduction in traffic at the peak minute.</p><h3>| When Efficiency Fights Stability</h3><p>The central challenge with VBR for Live is not that bitrate varies at all (it does under CBR as well), but that VBR can have much deeper, longer dips in bitrate, tightly coupled to what’s happening in the content.</p><p>Under CBR, a 5 Mbps stream is effectively that: on a per‑minute basis, traffic is remarkably flat. A server that is comfortably handling, say, ten thousand such sessions now is likely to still be comfortable in a minute, barring a wave of new joins. It’s safe for our steering logic to look at current traffic, see plenty of headroom, and route additional sessions to that server.</p><p>Under VBR, the same stream behaves very differently. During a slow, easy sequence, the encoder might only generate 2 Mbps for the 5 Mbps stream — or even less — to maintain its target quality, and it can stay at that lower level for an extended period. The server then appears to have plenty of unused capacity: per‑session bitrate is low and aggregate traffic is well below its limits. Our steering systems naturally interpret this as a signal that the server is under‑utilized and can accept more sessions.</p><p>The problem surfaces when the content changes. A fight starts, confetti begins to fall, or the camera cuts to a highly detailed, fast‑moving shot. To preserve quality, VBR may increase the bitrate to 6, 7, or 8 Mbps on the very next segments. If the server has admitted many additional sessions during the preceding low‑bitrate period, the aggregate traffic can suddenly exceed what the network link or NIC can sustain. Latency rises, packets are dropped, and devices start to experience stalls or quality downshifts. In extreme cases, this pattern of “bitrate dips followed by spikes” can destabilize parts of the system.</p><h3>| Making Servers Aware of Bitrate Variability</h3><p>These long VBR bitrate dips are great for efficiency, but — as we just saw — they can trick our delivery systems into thinking servers are under‑utilized and safe to load up. Under CBR, that behavior was predictable enough that current traffic was a good proxy for how “full” a server was; under VBR, it isn’t.</p><p>Our fix was to change how we decide whether a server can take more sessions. Instead of basing that decision only on current traffic, we reserve capacity based on each stream’s nominal bitrate, not just what it happens to be using at that moment. Even if a VBR stream is currently in a very cheap, low‑bitrate phase, we still treat it as something that can quickly return to its nominal rate.</p><p>This keeps our traffic‑steering behavior consistent between CBR and VBR and avoids the key failure mode where a server accepts too many sessions during a long low‑bitrate period and then becomes overloaded when bitrate rises again.</p><h3>| Tuning VBR Nominal Bitrates to Match CBR Quality</h3><p>The WWE example above already hints at why this isn’t automatic. We looked at a single 8 Mbps stream, encoded once with CBR and once with VBR. Both encodes have the same nominal bitrate, but the figure shows how differently they behave. The CBR encode stays clustered around 8 Mbps with frequent short spikes, while the VBR encode often drops far below that level and only spikes up when the content gets complex. That’s more efficient, but it also means that “same nominal bitrate” does not imply “same average number of bits” anymore — so simply reusing our CBR settings risks giving VBR less bitrate on average and losing some quality.</p><p>In practice, of course, we don’t just encode a single stream; we produce a set of streams at different resolutions and nominal bitrates — often called a bitrate ladder — so devices can adapt to their current network conditions by switching between them. When we first applied VBR using the existing CBR ladder, offline analysis with <a href="https://github.com/Netflix/vmaf">VMAF</a> (a <a href="https://netflixtechblog.com/toward-a-practical-perceptual-video-quality-metric-653f208b9652">perceptual video quality metric</a>) confirmed the concern from the WWE example: time‑averaged quality dropped slightly on a few streams, especially at the lowest bitrates. Early A/B tests showed the same pattern: overall VMAF about one point lower than CBR, with most of the gap at the bottom of the ladder.</p><p>To fix this, we compared CBR and VBR encodes rung by rung and looked at per‑stream VMAF. Wherever VBR fell more than about one VMAF point below CBR, we increased its nominal bitrate just enough to close the gap. Higher‑bitrate streams, where VBR quality was already very close to CBR, were left largely unchanged, including the 8 Mbps stream from the figure.</p><p>The result is a VBR ladder with slightly higher nominal bitrates on a few low‑end streams, but lower overall traffic, because VBR still drops the bitrate on simple scenes. This lets us match the quality of our CBR ladder while keeping the efficiency and stability gains that motivated the switch to VBR in the first place.</p><h3>| What’s Next for Live VBR</h3><p>With VBR in production for all Live events, our focus now is on using it more intelligently.</p><p>First, we are testing how to use the actual sizes of upcoming segments in our adaptive bitrate algorithms on devices, instead of relying only on nominal bitrates. This should help devices pick streams that better match how VBR will behave in the next few seconds, not just how a stream is labeled on paper.</p><p>Second, we are experimenting with making our capacity reservation less conservative. Today we reserve based on nominal bitrates to keep servers safe; by carefully applying a “discount” informed by real VBR behavior, we hope to free up additional headroom without sacrificing stability.</p><p>This work was the result of a broad, cross‑team effort. We’d like to thank Mariana Afonso, Dave Andrews, Mark Brady, Jake Freeland, Te-Yuan Huang, Ivan Ivanov, Yeshwenth Jayaraman, Patrick Kunka, Zheng Lu, Anirudh Mendiratta, Chris Pham, David Pfitzner, Jon Rivas, Garett Singer, Brenda So, Stan Surmay, Bowen Tan, Devashish Thakur, and Allan Zhou for their many contributions to making Live VBR a reality at Netflix.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c8f833b238cc" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/smarter-live-streaming-at-scale-rolling-out-vbr-for-all-netflix-live-events-c8f833b238cc">Smarter Live Streaming at Scale: Rolling Out VBR for All Netflix Live Events</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Scaling Global Storytelling: Modernizing Localization Analytics at Netflix]]></title>
            <link>https://netflixtechblog.com/scaling-global-storytelling-modernizing-localization-analytics-at-netflix-816f47290641?source=rss----2615bd06b42e---4</link>
            <guid isPermaLink="false">https://medium.com/p/816f47290641</guid>
            <category><![CDATA[localization]]></category>
            <category><![CDATA[big-data]]></category>
            <category><![CDATA[analytics]]></category>
            <category><![CDATA[data]]></category>
            <category><![CDATA[tech-debt]]></category>
            <dc:creator><![CDATA[Netflix Technology Blog]]></dc:creator>
            <pubDate>Fri, 06 Mar 2026 15:01:27 GMT</pubDate>
            <atom:updated>2026-03-06T15:01:28.509Z</atom:updated>
            <content:encoded><![CDATA[<p><a href="https://www.linkedin.com/in/valentingeffrier/">Valentin Geffrier</a>, <a href="https://www.linkedin.com/in/tanguycornuau/">Tanguy Cornuau</a></p><p><em>Each year, we bring the Analytics Engineering community together for an Analytics Summit — a multi-day internal conference to share analytical deliverables across Netflix, discuss analytic practice, and build relationships within the community. This post is one of several topics presented at the Summit highlighting the breadth and impact of Analytics work across different areas of the business.</em></p><p>At Netflix, our goal is to entertain the world, which means we must speak the world’s languages. Given the company’s growth to serving 300 million+ members in more than 190+ countries and 50+ languages, the Localization team has had to scale rapidly in creating more dubs and subtitle assets than ever before. However, this growth created technical debt within our systems: a fragmented landscape of analytics workflows, duplicated pipelines, and siloed dashboards that we are now actively modernizing.</p><h4>The Challenge: “Who Made This Dub?”</h4><p>Historically, business logic for localization metrics was replicated across isolated domains. A question as simple as “<em>Who made this dub/subtitle?”</em> is actually complex — it requires mapping multiple data sources through intricate and constantly changing logic, which varies depending on the specific language asset type and creation workflow.</p><p>When this logic is copied into isolated pipelines for different use cases it creates two major risks: inconsistency in reporting and a massive maintenance burden whenever upstream logic changes. We realized we needed to move away from these vertical silos.</p><h4>Our Modernization Strategy</h4><p>To address this, we defined a vision centered on consolidation, standardization, and trust, executed through three strategic pillars:</p><p>1. The Audit and Consolidation Playbook</p><p>We initiated a comprehensive audit of over 40 dashboards and tools to assess usage and code quality. Our focus has shifted from patching frontend visualizations to consolidating backend pipelines. For example, we are currently merging three legacy dashboards related to dubbing partner KPIs (around operational performance, capacity, and finances), focusing first on a unified data and backend layer that can support a variety of future frontend iterations.</p><p>2. Reducing “Not-So-Tech” Debt</p><p>Technical debt isn’t just about code; it is also about the user experience. We define “Not-So-Tech Debt” as the friction stakeholders feel when tools are hard to interpret or can benefit from better storytelling. To fix this, we revamped our Language Asset Consumption tool — instead of reporting dub and subtitle metrics independently, we combine audio and text languages into one consumption language that helps differentiate Original Language versus Localized Consumption and measure member preferences between subtitles, dubs, or a combination of both for a given language. This unlocks more intuitive insights based on actual recurring stakeholder use cases.</p><p>3. Investing in Core Building Blocks</p><p>We are shifting to a <em>write once, read many</em> architecture. By centralizing business logic into unified tables — such as a “Language Asset Producer” table — we solve the “<em>Who made this dub?”</em> problem once. This centralized source now feeds into multiple downstream domains, including our Dub Quality and Translation Quality metrics, ensuring that any logic update propagates instantly across the ecosystem.</p><h4>The Future: Event-Level Analytics</h4><p>Looking ahead, we are moving beyond asset-level metrics to event-level analytics. We are building a generic data model to capture granular timed-text events, such as individual subtitle lines. This data helps us understand how subtitle characteristics (e.g. reading speed) affect member engagement and, in turn, refine the style guidelines we provide to our subtitle linguists to improve the member experience with localized content.</p><p>Ultimately, this modernization effort is about scaling our ability to measure and enhance the joy and entertainment we deliver to our diverse global audience, ensuring that every member, regardless of their language, has the best possible Netflix experience.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=816f47290641" width="1" height="1" alt=""><hr><p><a href="https://netflixtechblog.com/scaling-global-storytelling-modernizing-localization-analytics-at-netflix-816f47290641">Scaling Global Storytelling: Modernizing Localization Analytics at Netflix</a> was originally published in <a href="https://netflixtechblog.com">Netflix TechBlog</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>