~/ai-stream
~/industry/openai-1b-disney-blindside-20260331
The Rundown AI·Industryhot

OpenAI's Sora Shutdown: $1M Daily Burn and Disney Blindside

content

Sora Shutdown

🔍Inside Sora's $1M-a-day Collapse at OpenAI

A WSJ investigation revealed the behind-the-scenes chaos of OpenAI Sora video generator shutdown, including a $1M daily burn rate, a blindsided Disney, and the internal code-named model that required Sora's compute budget.

  • Sora was reportedly burning "roughly a million dollars a day" and using significant compute, with Sora 3 training set to start just as it was axed
  • Disney learned about the shutdown "less than an hour" before the announcement, with the relationship now "effectively dormant"
  • The freed-up chips went to "Spud," a model targeting coding and enterprise in response to Anthropic's powerful moves in the sector
  • An enterprise version of Sora was already in pilot with Disney for marketing and VFX work, with a spring launch expected

Why it matters: The Disney blindside is certainly a strange way to handle a potential $1B partnership with one of the biggest media companies on the planet.

Microsoft Researcher

🔬Microsoft Pits Claude Against ChatGPT for Research

Microsoft released Critique and Council, two new features that turn its Copilot Researcher into a multi-model system that can review and edit research reports and run both systems side by side.

  • Copilot's Researcher already uses OpenAI for multi-step work, with Critique now adding Claude as a second model to review every report before it ships
  • One model drafts the research, and the second tears it apart on source quality, completeness, and evidence grounding behind the scenes
  • Model Council mode runs both models side by side, then flags where they agree, where they split, and what each uniquely surfaced
  • Updates come alongside broader rollout of Copilot Cowork into Frontier, Microsoft's Claude-based agentic tool

Why it matters: One model will sell you on anything, so you better ask two. Multi-model collaboration is the future.

Perplexity Travel

🗺Build a Travel Itinerary with Perplexity Computer

In this guide, you will learn how to use Perplexity Computer to plan a full trip itinerary with flights, a day-by-day schedule, and sources in one run.

  • Open Perplexity and enable Computer mode (Pro accounts can test for free)
  • Example prompt: "Plan a trip itinerary for [DESTINATION] for [DATES]. Departing from: [AIRPORT] Budget: [range] Style: [relaxed/outdoors/etc.] Must-haves: [2-4 items]. Make a full PDF as if you were a travel agent"
  • Let Perplexity Computer run for 15-20 minutes, get a PDF itinerary

Why it matters: Perplexity Computer can deploy sub-agents to code. Ask it to create an interactive calendar website to help you plan.

Stanford Study

🔬Stanford Exposes AI's People-Pleasing Problem

Stanford researchers published a new study showing that major AI chatbots consistently take users' side in personal conflicts, even backing harmful or illegal behavior.

  • Researchers tested 11 LLMs using 2K Reddit posts where crowds agreed the poster was wrong, but chatbots still sided with the user over half the time
  • Over 2,400 participants chatted with agreeable and neutral AIs and preferred the sycophantic version, rating it as more trustworthy
  • After chatting with the agreeable model, users doubled down on their position, lost interest in apologizing, and couldn't tell the AI was biased

Why it matters: Most frontier models aren't much different from OpenAI's 4o when it comes to people-pleasing behavior.