Entertainment Analytics Conference

Past Events

The conference focuses on data science, analytics, and machine learning as they relate to media and entertainment. Topics span causal inference, social data analysis, synthetic control models, and the challenges of using data to drive decision-making in entertainment companies.

Thirty data and analytics leaders from major UK and European media organisations gathered for two days at The Lowry in Salford. This was the first Entertainment Analytics conference to comprehensively address the impact of language models on media businesses. Sessions ran 30 minutes each, split equally between presentation and Q&A under Chatham House Rules.

Talks

  • “Estimating Cross-Media Reach Using Copulas”A public broadcaster showed how copulas (a multivariate technique from finance) de-duplicate audience reach across TV, radio and online for 42 local areas, feeding an automated weekly pipeline that local newsrooms use in near real-time.
  • “Optimising YouTube Distribution Through Genre-Specific Channels”A public service broadcaster found 72% of YouTube views came from algorithmic recommendations but only one or two shows per channel ever broke through. Restructuring into niche genre channels using demographic clustering drove 77% year-on-year growth.
  • “Outcome Measurement for Television Advertising”A commercial broadcaster modelled 640+ brands to show TV ads produce a large short-term spike in website visits plus a tiny long-term effect lasting six months that roughly equals the short-term impact in total value.
  • “Structuring Unstructured Recipe Data with Classical ML”A major newspaper’s data team extracted structured data from 30,000 free-text recipes using fine-tuned named entity recognition, achieving 96–99% accuracy while deliberately avoiding generative AI so they could diagnose every error precisely.
  • “Multi-Agent Systems for Audience Insights”An assessment of where multi-agent AI systems genuinely work in the audience insights pipeline. The standout finding: persona simulation generated approximately $12 million in net profit across 58 projects in four months, where companies simulate their existing segmentation answering hundreds of additional questions.
  • “Z-Score Audience Analysis in Sports Media”A European media group with 230 million monthly devices used Z-scores to analyse reader behaviour across content categories. The simplicity of the metric was its greatest strength: editorial teams immediately understood and acted on it. A parallel “zero Google” analysis showed 80%+ of sports traffic arrives directly.
  • “The AI Transformation Chain”Interviews with 22 insight leaders revealed 19 are failing at AI transformation. Success requires a six-link chain: executive sponsorship, app approvals, foundational skills, standardised tasks, process orchestration, and sustained reinforcement. One bright spot: a consumer goods firm delivering 36% more insights with 20% fewer people.
  • “A Completion-Based Model for Fairer Streaming Royalties”A former chief economist proposed adding a completion threshold to music royalties. Analysis of 15,000 songs showed completion rates are consistent across song lengths (85–95%), so the model does not punish longer songs. It also eliminates click-farm fraud engineered to last just 31 seconds.
  • “Customer Segmentation Across Linear and Streaming”A pay-TV operator simplified 16-segment clustering into four primary consumption cohorts to enable cross-platform comparison for the first time. The key finding: ~25% of the base relies heavily on linear partner channels, making channel-dropping decisions riskier than assumed.
  • “AI-Generated Long-Form Sports Content at Scale”A sports publisher produces 15,000 AI-generated match previews per year across 49 football leagues. The architectural insight: turn structured data into text strings first using traditional ML, then use the language model purely for fluency. AI content showed ~20% lower engagement on major leagues but near-parity on obscure leagues where no human alternative existed.
  • “AI-Powered Product Discovery in Luxury Retail”A luxury retailer demonstrated text-to-image search, image-to-product matching, and an LLM outfit stylist, all built by a single data scientist. The image search required bespoke training to bridge the domain gap between clean e-commerce shots and distorted real-world photography.
  • “Visual Exploration of Content Catalogues Using AI Embeddings”An interactive tool maps 20,000 streaming titles into navigable 2D space using language model embeddings. Filtering by date reveals the emergence of genre clusters over time. The presenter argued chat interfaces alone are insufficient for exploration when you do not yet know the question.
  • “Estimating Impact of LLMs on Traffic”A major international news publication shared data on millions of weekly visits lost to AI platforms, with scrape-to-referral ratios reaching 60,000:1 for some AI services.
  • “AI in Academic Publishing”A global academic publisher explored how language models are reshaping scholarly research workflows and discovery.
  • “Data Democratisation in News Media”A European media company described building a self-service analytics culture across newsrooms.
  • “Using LLMs to Analyse Survey Responses at Scale”A UK broadcaster applied language models to process a massive volume of open-text survey responses, dramatically reducing the time from collection to insight.
  • “Candid Truths about AI in Insights”A practitioner guide to the specific risks of using language models for data analytics, where compounding errors through iterative analysis can be far more costly than a poorly written email.

What we learned

The AI Disruption Reality: Publishers face millions of weekly visits lost to AI platforms, with scrape-to-referral ratios reaching 60,000:1 for some AI services. Traditional discovery mechanisms are fundamentally breaking as users get answers without clicking through.

19 of 22 Are Failing: Off-the-record interviews with 22 insight and analytics leaders found 19 are failing at AI transformation despite three years since ChatGPT’s launch. The bright spots were dramatic: one firm delivers 36% more insights with 20% fewer people; another in-sourced work previously costing $1.5 million per year.

Algorithm Dependence: A broadcaster found 72% of YouTube views came from algorithmic recommendations, yet only one or two shows per channel ever broke through. Reorganising content into niche genre channels based on audience clustering achieved 77% year-on-year growth.

Persona Simulation at Scale: Multi-agent AI persona simulation generated approximately $12 million in net profit across 58 projects in four months. Companies with existing segmentation pay to simulate those personas answering hundreds of additional questions via language models.

Completion-Based Royalties: Analysis of 15,000 songs showed completion rates are consistent across song lengths (85–95% regardless of duration), supporting a model that redistributes royalty revenue from skipped songs to completed ones. The approach also eliminates click-farm fraud engineered to last just 31 seconds.

AI Sports Content: Parity on the Long Tail: A sports publisher producing 15,000 AI-generated match previews per year found ~20% lower engagement on major leagues but near-parity on obscure leagues where no human alternative existed. Audiences showed little pushback when AI authorship was disclosed, except when predictions contradicted conventional wisdom.

Language models and AI have moved from experimental to existential for media organisations. Technical capabilities exist but human adoption, political barriers, and fundamental business model questions remain largely unresolved.

The UK conference returned after a five-year gap. Two days of 50-minute talks followed by 50-minute discussions, with attendees from major broadcasters, publishers, music companies, data consultancies, and regulators. A suggestion was made to create a penalty jar for every mention of “Gen AI” during the two days.

Talks

  • “Building a Data-Driven Culture at a 180-Year-Old Publisher”A media analytics leader joined a major global publication and found it roughly five years behind a comparable title in data maturity, despite nearly identical tech stacks. The gap was cultural, not technical: the fix was getting data people into leadership meetings, not buying new tools.
  • “Principles for Responsible AI Deployment”A major financial newspaper warned that organisations experimenting with generative AI were creating a “Pandora’s POC box”: small experiments proliferating with no central visibility, each carrying infrastructure costs, and in one case duplicating work the internal team had already completed.
  • “AI Applications in News Publishing”A data science leader returning from parental leave found a third-party consultancy had been contracted to rebuild a recommendation model that duplicated her team’s own work, using the organisation’s own content taxonomy. The contract was signed before the data science team knew about it.
  • “Cross-Platform Audience Analytics and Attribution”All tests optimised for a single metric: total customer lifetime value of incoming subscribers. This removed endless debates about conversion rate vs. product mix vs. average order value, making testing fast and decisive.
  • “Data Governance and Regulatory Perspectives”Discussion of how media data regulation is evolving, with perspectives from regulators and practitioners on the challenges of balancing innovation against privacy obligations.
  • “Advertising Effectiveness and Outcome-Based Measurement”Calculating the return on investment for a data science initiative frequently took longer than building and deploying the model itself. One speaker described spending extensive effort just to justify £50,000 of infrastructure spend.
  • “Music Industry Data: Streaming Economics”An examination of how streaming royalties are calculated, distributed, and disputed, with fresh data on the economic impact of playlist curation on independent artists.
  • “Regional Broadcaster Perspectives on Data Maturity”Even a deliberately simplified approach to AI required about an hour a day of experimentation to avoid falling behind. Attendees from major broadcasters described mandatory eight-hour AI training courses as a prerequisite just to access basic tools.
  • “Measuring UX Impact Without A/B Testing”A TV platform provider that doesn’t own the customer relationship developed causal inference alternatives to randomised testing. The key learning: getting product owners to write down hypotheses and decision trees upfront mattered far more than the measurement methodology itself.
  • “The Data and AI Playbook from Private Equity”Research across 35+ PE-backed data leaders found compensation has risen from £250,000 to £400,000 (now including carried interest) even as data teams elsewhere face cuts. Generative AI had produced zero change in enterprise valuations so far because it has only delivered productivity gains, not new revenue.
  • “Glocalization: Local Artists Dominating Global Platforms”On-demand streaming is driving local artists to the top of local charts. In Germany, zero of the top 100 radio songs were German-language, but a third of the top 100 streaming songs were. In Denmark, nine of the top ten streaming tracks were Danish. The UK has not produced a true global pop star since 2017.
  • “Simplifying the NLP Stack with Generative AI”A major publisher found that prompt-engineered named entity extraction outperformed established NLP tools (spaCy, NLTK) in blind editorial evaluations, and reduced training data requirements from 100 articles per topic to 10 using a human-in-the-loop feedback system.
  • “Neuroscience-Based Creative Optimisation”Predictive models trained on 120,000+ neuroscience test subjects (eye tracking, EEG, pupil dilation) evaluate marketing creative effectiveness at millisecond-level subconscious responses, avoiding the self-reporting bias of traditional survey methods.

What we learned

Ten Million Pounds From Testing: One publication estimated that optimisation tests run against customer lifetime value had generated roughly £10 million per year in incremental value. The breakthrough was a single decision: all tests optimise for total CLTV of incoming subscribers, removing endless internal debates.

The CEO Stopped the Meeting: During a quarterly budget review, three changes were launched simultaneously. When the analytics team could not isolate the impact of each, the CEO halted the meeting and refused to continue until data could explain what had actually happened. Data culture is a behaviour, not a slogan.

Eight Hours Before You Can Touch AI: One broadcaster required an eight-hour mandatory AI training course before staff could access even basic tools. The practical result: most employees never completed it, creating a de facto opt-out for AI adoption across the organisation.

Five Years Behind on Identical Technology: Two publications with nearly identical tech stacks (Snowflake, Salesforce, Tableau, GA) were five years apart in data maturity. The gap was entirely cultural. The finance team had become the de facto data team, controlling the narrative. The fix was relationship building, not technology investment.

The ROI Trap: Calculating ROI on data science projects often becomes a project in itself. One speaker needed to demonstrate £500,000 of projected value to justify £50,000 of spend. The comparison: nobody asks a UX designer to cost up twenty prototypes before deciding which one to build.

Pandora’s POC Box: Small generative AI experiments were proliferating across business units with no central visibility. In one case, a third-party consultancy was contracted to build a model that duplicated the internal team’s own work. The contract was signed before the data science team even knew about it.

Local Artists Are Winning on Global Platforms: On-demand streaming is reversing decades of Anglo-American chart dominance. In Germany, zero of the top 100 radio songs were German-language, but a third of the top 100 streaming songs were. In Brazil, 86 positions above the first English-language track were all Brazilian Portuguese acts. The UK has not produced a true global pop star since 2017.

Gen AI: Zero Impact on Valuations: Research across 35+ PE-backed companies found generative AI had produced zero change in enterprise valuations because it has only delivered internal productivity gains, not new revenue streams. Data leader compensation has risen from £250,000 to £400,000 even as data teams elsewhere face budget cuts.

Twelve talks covering content valuation, streaming forecasting, genomic targeting, release window optimisation, playlist economics, and the emergence of language models in entertainment analytics. A live demonstration showed a non-programmer running a full clustering analysis of music industry data through natural language conversation.

Talks

  • “The Editor vs. the Algorithm”A field experiment at a major news outlet found the algorithm outperformed human editors, but a combination of both would have delivered 13% more clicks than either alone. Pure algorithmic personalisation also reduced consumption diversity over five months, creating a feedback loop that narrows the content a platform can justify producing.
  • “Estimating Incremental Acquisition of Content Launches”An econometric model valued individual titles within a streaming bundle by measuring their effect on subscriber churn. Tentpole releases increased total platform consumption by roughly 6%, and subscribers whose consumption dropped below their personal average were disproportionately likely to cancel.
  • “Forecasting Long-Term Streaming of Emerging Artists”A record label’s data science team trained their model on 7,000 songs with eight years of data. One case study: a song released in 2019 doing 16,000 streams per week jumped to five million per week by May 2022. TikTok virality has decoupled a song’s discovery moment from its release date.
  • “Combining AI and Storytelling Expertise to Describe Content”A human-coded dataset of clearly defined traits was used to train models that predict complex movie traits with high accuracy. Identifying trait combinations proved more valuable than examining genres alone.
  • “Measuring the Value of a Title in a Subscription Bundle”A structural demand model allocated tens of billions of dollars of subscription revenue across bundled services. The Shapley value method was abandoned because removing a single benefit entirely produced such extreme counterfactuals that results were not credible.
  • “Large Language Models Will Revolutionise Entertainment Analytics”A live demo showed a non-coder using natural language to run a full clustering analysis across dozens of countries, complete with named segments, marketing recommendations, and visualisations. Of the room, only about six attendees had corporate-endorsed language model access, while at least five admitted to using unapproved tools.
  • “The Effect of Linear Television Airing on Digital Channels”Airings of movies on linear TV produced a small but positive lift in streaming sign-ups. Yet co-licensing films to similar streaming services caused cannibalization of viewer engagement.
  • “Enhancing Advertisements Through Genomic Targeting”Content genome data was applied to film and television advertising targeting, matching ad creative attributes to audience taste profiles at scale.
  • “Optimising Release Windows Using Structural Equation Modelling”A structural model examined how consumers choose between theatrical, home video, and streaming formats, extended to include piracy, allowing simulation of alternative release scenarios.
  • “How Playlists Affect Off-Platform Behaviour”Playlisting emerging artists led to a noteworthy increase in both on-platform activity and off-platform live concert bookings. The persistent effects suggest playlisting is a potent tool for amplifying emerging careers.
  • “An Insights-Driven Narrative to Support Talent Initiatives”Data-driven approaches to talent representation decisions, quantifying the commercial and cultural impact of diversity initiatives in the entertainment industry.
  • “Entertainment Economics”An examination of economic models in entertainment, covering pricing, bundling, and consumer behaviour across digital and physical formats.

What we learned

Algorithms Beat Humans, But Not By Much: A field experiment found algorithms outperform human editors at news curation, but the effect was modest. A combination of both would have delivered 13% more clicks than either alone. When a bug caused the algorithm’s data to go stale for one week, the human editor outperformed it. Humans also won during breaking news events.

Nobody Predicted Barbie’s $162 Million Opening: Every forecast model failed to predict Barbie’s opening weekend. The concern: if teams increasingly rely on language models trained on historical data, they will systematically produce lowest-common-denominator ideas. The models work within the box and cannot get out of it.

Songs Can Sleep for Years, Then Explode: A song released in 2019 doing 16,000 streams per week jumped to 350,000 by late March 2022 and five million by May. The forecasting team had to redefine “week one” as the week streaming growth first hit 10% of its eventual maximum, not the release date. TikTok virality has decoupled discovery from release.

Shadow AI Before the Rules Arrived: A show of hands revealed roughly six attendees from major entertainment companies were using non-endorsed AI tools at work, tools they knew would get them in trouble. Only a similar number had any corporate-approved way to use a language model.

The $60,000 Certification Problem: A graduate student learned everything he needed about machine learning from a free online resource. As one panellist put it: “The information is free, and the certification costs $60,000 a year. That can’t be sustainable.” Language models may accelerate the unbundling of education from credentialing.

The conference returned after a two-year COVID hiatus with around 70 attendees, up from 31 in 2016. Topics included Thompson Sampling for ticket pricing, Markov Chain subscriber lifetime value, structural modelling for release windows, Spotify playlist power, the SVOD-to-AVOD transition, and the first live demonstration of language models for entertainment analytics.

Talks

  • “Pricing Movie Tickets with Bayesian Bandits”A cinema chain abandoned traditional regression models and adopted Thompson Sampling instead, treating each theatre like a multi-armed bandit running 10,000 simulated iterations per pricing decision. Movie demand shifts so drastically week to week that traditional price elasticity requires stable demand cinema never has.
  • “Measuring the Value of Content on Screens and Streams”A content analytics company scored over 2,000 characteristics per film using human analysts. Self-organising maps projected this data onto a 2D plane, creating a topographic map of cinema. One superhero film independently clustered apart from its franchise peers because it was actually a spy thriller in disguise.
  • “Subscription Lifetime Value Using Markov Chains”A streaming platform showed that the standard industry method for valuing subscriber acquisition overstates the true value by approximately five times. The error: it ignores the baseline probability that people would have subscribed anyway.
  • “Optimising Release Windows with Counterfactual Simulation”Structural modelling examined how consumers choose between formats, extended to include piracy, allowing simulation of alternate release scenarios of interest to the business.
  • “Understanding Audiences Through Content Genome Data”Overlaying audience taste profiles on a content genome map revealed that marketing a film’s core genre might alienate the target demographic, while a secondary content thread could be a stronger hook.
  • “The Power of Streaming Playlists”Research on the 4,000 most-followed new music playlists found that playlist inclusion causally determines roughly 50% of a song’s total streams. Major-label share on platform-curated playlists fell by ten percentage points, largely explaining their declining streaming revenue share.
  • “Marketing to Fandoms”Data-driven approaches to understanding and activating fan communities across entertainment properties.
  • “Communicating Data Science to Executives”The gap between analytical sophistication and business impact was framed as a four-link chain: right data, right method, right recommendation, right adoption. Almost every conference presentation addressed only the second link.
  • “Unlocking the Hidden Value of Large AI Models”A live demonstration where a non-coder used natural language to run a clustering analysis of music industry data. Only about six of 70 attendees had corporate-endorsed language model access. Five admitted to using non-endorsed tools that could get them in trouble.
  • “Pricing Against Piracy”Economic modelling of how pricing strategies interact with piracy rates across different release windows and geographic markets.
  • “Improving Causal Inference Using Data-Mined Variables”Novel approaches to strengthening causal claims in observational entertainment data by identifying and controlling for algorithmically discovered confounders.
  • “Social Data for Unlocking Net Audiences”Using social graph analysis to identify and reach net-new audiences beyond existing marketing databases.
  • “Transitioning from Ad-Free to Ad-Supported Streaming”Average revenue per user was higher for ad-supported content, but survival rates were lower. Adding linear channels proved to be an important growth driver.

What we learned

The Five-Times Overvaluation Trap: The standard industry method for valuing subscriber acquisition overstates the true value by approximately five times. It ignores the baseline probability that people would have subscribed anyway. Marketing teams systematically claim credit for revenue that would have materialised regardless.

Playlists Control Half of Streaming Success: Playlist inclusion causally determines roughly 50% of a song’s total streams. The platform’s own curated playlists account for about 30%, with major label playlists adding 20%. Major-label share on platform-curated playlists fell by ten percentage points, largely explaining their declining streaming revenue share.

Slot Machines for Movie Tickets: A cinema chain used Bayesian Thompson Sampling to optimise ticket pricing, treating each location like a multi-armed bandit running 10,000 simulations per decision. Movie demand resets so drastically each week that traditional price elasticity simply does not apply.

2,000 Genes per Film: A content genome with over 2,000 characteristics per film, scored by human analysts, was projected onto a 2D map using self-organising maps. One superhero film clustered apart from its franchise peers because it was actually a spy thriller in disguise. The same technique revealed that marketing a film’s core genre can alienate the target audience.

The Better-the-AI, the-Worse-the-Human Effect: Students given the option to use AI tools in a university course produced lower quality work than those who did not. The pattern: “they got lazy, they thought it sounds good, but it’s good fluff.” A separate study on recruiters using language models to screen CVs found the same dynamic. The better the AI, the worse the human performed.

The Analytics Impact Chain: The gap between analytical sophistication and business impact was framed as a four-link chain: right data, right method, right recommendation, right adoption. Almost every conference presentation addressed only the second link. As one organiser noted, “nobody ever suggests an ROI case study” because companies will not share what happened after the insight was delivered.

The fourth US conference continued to build the bridge between entertainment industry practitioners and academic researchers. Two days of presentations and discussion covered content valuation, audience segmentation, causal inference, and platform economics.

Talks

  • Content valuation methodologies for subscription services
  • Machine learning approaches to audience segmentation
  • Causal inference methods in entertainment industry experiments
  • Platform economics and marketplace dynamics
  • Social media analytics for content marketing
  • Streaming behaviour and subscriber retention

The third US conference featured a full programme of presentations and a panel discussion. Eight sponsors supported the event, reflecting the conference’s growing reputation in the entertainment data science community.

Talks

  • Content genome applications for audience understanding
  • Pricing experimentation in digital media
  • Demand prediction and forecasting models
  • Release window optimisation strategies
  • Advertising effectiveness measurement
  • Social data for audience insights
  • Panel discussion on the future of analytics in entertainment

The second US conference grew to 49 attendees, with sponsorship from universities, streaming services, and premium cable networks. The two-day event maintained the intimate, discussion-heavy format established in the first year.

Talks

  • Content demand measurement and prediction at global scale
  • Audience analytics for streaming platforms
  • Digital pricing experiments in publishing
  • Marketing effectiveness measurement using search data
  • Social network analysis for audience understanding
  • Release strategy optimisation for film distribution

The first UK edition of the conference, inspired by the success of the 2016 LA event. Held in January 2017 in Salford, it brought the same format of intimate presentations and discussion to a UK and European audience, with a full day of presentations plus a social dinner, followed by an optional half day of deeper discussion.

Talks

  • How content genome data drove increased reach and engagement for a major franchise on YouTube
  • Optimisation of content pricing using randomised experiments
  • Mining Google Trends for automated global marketing impact analysis
  • Audience analytics and segmentation for UK broadcasters
  • Social graph analysis for content strategy

The inaugural Entertainment Analytics Conference. Around 30 attendees from TV, SVOD, film studios, music, data and technology companies, and universities gathered for an intimate day of sharing and learning about entertainment analytics and data science.

Talks

  • “A Moneyball Moment for Hollywood?”The keynote framed the conference’s founding question: entertainment has historically been a gut-feel industry, but these firms now compete against technology companies that make quantitative decisions. In the fight between data and gut feel, it’s not even a fair fight.
  • “Measuring and Predicting Global Demand for Content”A method for quantifying audience demand for entertainment content across multiple markets using digital signals, enabling cross-country comparison of content appetite.
  • “A Global Cross-Media Buzz Reporting System”Using Google Trends data to build an automated system for tracking and comparing marketing impact across campaigns, territories and media types.
  • “Testing Print and Digital Pricing Using Randomised Experiments”Randomised controlled experiments applied to media pricing decisions, demonstrating how experimental design can replace intuition in subscription pricing.
  • “Global Streaming Distribution and Frictionless Digital Trade”An examination of how digital distribution is reshaping international content trade, reducing friction but creating new competitive dynamics.
  • “Improving the Economics of Measuring Advertising Effectiveness”Ghost ads (showing a control group what they would have seen without the ad campaign) offered a cost-effective alternative to traditional brand lift studies.
  • “Using Content Genome to Predict TV Content Sales”A system decomposing TV shows into 30+ attributes to predict international demand across 500+ seasons and 20 countries. Synopses predicted accurately; scripts confused the analysis with stage directions and dialogue filler.
  • “The Insight Contained in the Social Graph”Graph-based audience clustering using follower networks (not what people say) revealed distinct audience segments invisible to traditional analytics. Among 300,000 followers of major comic-book brands, superfans skewed 60% toward one publisher while film fans skewed 56% toward the rival.
  • “The Impact of Free Streaming on Paid Download Sales”A public broadcaster found free streaming reduced paid digital sales by 7–8%, but the advertising revenue earned from free streaming almost exactly cancelled out the lost sales revenue, creating a genuine strategic stalemate.
  • “Early International Digital Movie Releases and Domestic Box Office”Releasing films on Chinese streaming platforms triggered high-quality pirated copies within days, causing a roughly 13% decline in US theatrical revenue starting from week four of the box office run.
  • “How Messaging Apps Are Changing Online Video”For Generation Z, the primary way they encounter a song is a 15-second clip. Messaging platforms like WeChat and QQ had 1.6 billion monthly users but offered essentially zero analytics. Ignoring platforms because metrics don’t exist was described as dangerous.

What we learned

The Moneyball Moment: The keynote framed the conference’s founding question: entertainment has historically been a gut-feel industry where people rise by backing hits. But these firms now compete against technology companies that make quantitative decisions. In the fight between data and gut feel, it’s not even a fair fight.

Piracy Leaks Within Days: Releasing films on Chinese streaming platforms triggered high-quality pirated copies within days. Of 23 tracked titles, 22 had a leak appear within a week. The damage: a roughly 13% decline in US theatrical revenue, beginning at week four of the box office run.

Free Streaming Costs Exactly What It Earns: A public broadcaster found free streaming reduced paid digital sales by 7–8%. But the advertising revenue earned from streaming almost exactly cancelled out the lost sales, creating a genuine strategic stalemate between public service mission and commercial subsidiary.

Scripts Mislead, Synopses Predict: Researchers building a content genome found that machine learning on scripts produced inconsistent results (too much non-plot material: stage directions, camera notes, filler), while synopses predicted international demand accurately across 500+ seasons and 20 countries.

The Social Graph Revelation: Clustering 300,000 followers of major entertainment brands by follower networks (not what they say) revealed distinct segments invisible to traditional analytics. A “moms” cluster appeared on Twitter but not on Pinterest, where “dads” emerged instead, each self-identifying on the platform where they are the minority.

15 Seconds Is the Whole Song: For Generation Z, the primary way they encounter and internalise a song is a 15-second clip. Messaging platforms had 1.6 billion monthly users but offered essentially zero analytics. Ignoring platforms because metrics don’t exist was described as dangerous.