PROMIT NOW · LEADER DAILY · 2026-04-26

DeepSeek V4 on Huawei Ascend Upends AI Vendor Strategy

· Leader · 8 sources · 1,544 words · 8 min

Topics LLM Inference · AI Capital · Agentic AI

DeepSeek V4 is running natively on Huawei Ascend chips — not NVIDIA — while pricing at $0.14 per million tokens under MIT license, and Chinese labs now hold 4 of the top 5 open-weight model positions. The same week, Google committed $40B to lock Anthropic into its cloud, OpenAI doubled GPT-5.5's API price, and the Musk v. Altman trial begins Monday. Your AI vendor strategy, cost model, and supply chain assumptions were built for a world that ended this week — and the new one has no clear winner.

◆ INTELLIGENCE MAP

  1. 01

    China Achieves AI Compute Sovereignty — Export Controls Have Failed

    act now

    DeepSeek V4 runs natively on Huawei Ascend chips, proving China can build frontier AI without NVIDIA. Chinese labs now hold 4 of 5 top open-weight positions. V4 Flash at $0.14/M tokens undercuts every Western competitor 2-3x — with further drops planned when Ascend 950 deploys H2 2026.

    4 of 5
    top open models are Chinese
    3
    sources
    • V4 Flash pricing
    • KV cache reduction
    • Compute reduction
    • Hallucination rate
    • Training FLOPs
    1. 01Kimi K2.6 🇨🇳54
    2. 02DeepSeek V4 Pro 🇨🇳52
    3. 03GLM-5.1 🇨🇳50
    4. 04Qwen 3.6 🇨🇳48
    5. 05Top Western open46
  2. 02

    $65B Hyperscaler Lock-In Reshapes AI Power Structure

    act now

    Google's $40B Anthropic deal ($10B upfront, $30B on milestones, $350B valuation) plus Amazon's $25B creates $65B in committed capital and compute lock-in. Anthropic's 233% revenue growth ($9B to $30B annualized in ~4 months) is the fastest enterprise software scaling in history. Musk v. Altman trial begins Monday seeking $100B+ — a partial win could freeze OpenAI's IPO and governance.

    $65B
    committed to Anthropic
    4
    sources
    • Google investment
    • Amazon investment
    • Anthropic valuation
    • Revenue growth
    • Musk damages sought
    1. Google → Anthropic40
    2. Amazon → Anthropic25
    3. Microsoft → OpenAI13
    4. Anthropic ARR30
  3. 03

    Enterprise AI Hits Organizational Wall — ROI Invisible on Balance Sheets

    monitor

    AI productivity gains are individually documented but invisible on corporate P&Ls — a Solow Paradox redux threatening enterprise budgets. A practitioner with $300K in the game argues the root cause: 80%+ of AI spend is product-facing while internal operations remain pre-AI. FTE-denominated budgets and rational information hoarding structurally block transformation. The 18-24 month window to fix this is narrowing.

    80/20
    product vs. operations AI
    2
    sources
    • Transformation window
    • Product-facing AI
    • Operations AI
    1. AI in the business (product)80
    2. AI on the business (ops)20
  4. 04

    AI's Physical Economy Spillovers: Memory Crisis, Debt Limits, Permitting Walls

    monitor

    Samsung warns of its first-ever smartphone net loss — not from weak sales but because AI memory demand drove DRAM/NAND prices beyond consumer margins. One Nvidia Vera server consumes memory equal to 4,600 phones. Oracle's ~$300B data center push is breaking Wall Street's syndication capacity, and 12+ US states are considering construction moratoriums. AI's capital intensity is outpacing the financial and physical systems that fund it.

    4,600x
    phone-equivalent per server
    3
    sources
    • Oracle DC investment
    • States with DC curbs
    • Intel surge
    1. 1 Nvidia Vera server4600
    2. Samsung Galaxy S261
  5. 05

    Stablecoins Pivot From Cross-Border to Domestic Payments Infrastructure

    background

    Intra-country stablecoin transactions grew from ~50% to ~75% of volume in two years, contradicting the remittance narrative. C2B commerce transactions rose 128% YoY to 284.6M. Velocity doubled from 2.6x to 6x — supply is transacted, not hoarded. Regulation is accelerating adoption: GENIUS Act boosted volume to $4.5T/quarter, and MiCA created a $15-25B/month non-USD market from near zero.

    75%
    now domestic transactions
    1
    sources
    • Q1 2026 volume
    • C2B growth YoY
    • Asia share
    • Velocity (2yr change)
    1. Domestic share 202450
    2. Domestic share 202675

◆ DEEP DIVES

  1. 01

    DeepSeek V4 on Huawei Ascend: China's AI Stack Just Went NVIDIA-Independent

    <p>This isn't another model release to benchmark-watch. <strong>DeepSeek V4 running natively on Huawei Ascend chips</strong> is the moment the US export control thesis — restrict NVIDIA access, keep China behind at the frontier — demonstrably failed. V4 was trained at approximately <strong>1e25 FLOPs using FP4 precision</strong> on what appears to be a mix of NVIDIA and Huawei hardware, and it now runs inference entirely on Huawei's CANN stack. DeepSeek has publicly stated that V4 Pro pricing will "fall sharply" once <strong>Ascend 950 supernodes deploy at scale in H2 2026</strong>. This is a roadmap for a parallel AI compute ecosystem, not a hedge.</p><blockquote>Chinese labs now hold 4 of the top 5 open-weight model positions — Kimi K2.6, DeepSeek V4, GLM-5.1, and Qwen 3.6 — all under MIT license with full technical reports. The open-weight frontier is a Chinese-led market.</blockquote><h3>The Architecture Is the Real Story</h3><p>V4's <strong>Compressed Sparse Attention</strong> and <strong>Heavily Compressed Attention</strong> systems reduce KV cache memory by 8.7x at 1M tokens (from 83.9 GiB to 9.62 GiB) and total FLOPs by 73%. The full 1.6T-parameter model fits on a <strong>single 8xB200 node</strong> via FP4/FP8 mixed-precision quantization. These are the innovations that make million-token context practical at commodity prices. Combined with GPT-5.5 and Qwen 3.6 also supporting 1M tokens, long context is now table stakes — any product treating it as premium is already behind.</p><h3>The Pricing Pressure Is Existential</h3><p>V4 Flash at <strong>$0.14/$0.28 per million input/output tokens</strong> is 2-3x cheaper than the nearest competitor, under MIT license. Meanwhile, GPT-5.5 just <strong>doubled API prices</strong>. The AI market is bifurcating: a premium tier (OpenAI, Anthropic) betting on brand trust and integration, and a commodity tier (DeepSeek, Qwen) betting on architectural efficiency and open licensing. The capability gap between these tiers is collapsing while the price gap widens.</p><h3>The Critical Caveat</h3><p>Before you migrate anything: V4's <strong>94-96% hallucination rates</strong> on the AA-Omniscience factual benchmark are disqualifying for most enterprise use cases. Benchmark leadership doesn't equal production readiness. The companies that solve <strong>reliable deployment of unreliable models</strong> — through verification layers, domain fine-tuning, human-in-the-loop — will capture the enterprise value that raw model providers cannot. <em>This is where Western companies still have a defensible position, but only if they build it now.</em></p><h3>Cross-Source Tension</h3><p>One source highlights the US State Department issuing global warnings about alleged IP theft by DeepSeek. Another notes DeepSeek is simultaneously seeking outside funding. A third observes that OpenAI's own chief scientist publicly admitted progress is "surprisingly slow" — while marketing GPT-5.5 as a "new class of intelligence." The internal-external narrative gap at Western labs, combined with China's proven ability to deliver frontier models on domestic silicon, suggests the competitive window for cost-based Western AI dominance is <strong>12-18 months, not 3-5 years.</strong></p>

    Action items

    • Map every open-weight model in your production stack by country of origin and licensing terms within 30 days — assess regulatory exposure for government and regulated-industry customers
    • Re-price your inference infrastructure strategy for $0.10/M token floor by Q1 2027 — stress-test every use case against commodity model access
    • Establish quarterly Huawei Ascend capability assessments starting this quarter — track Ascend 950 deployment timeline specifically
    • Identify your defensible differentiation assuming commodity model access — if anyone can serve V4 at MIT-licensed prices, what is your moat?

    Sources:DeepSeek V4 on Huawei Ascend chips = China's AI stack is now NVIDIA-independent · AI's pricing crisis is here: frontier models double in cost while open-weight rivals match them · AI's $65B arms race is repricing your entire supply chain

  2. 02

    $65B Lock-In Week: Google-Anthropic, Oracle's Debt Wall, and a Trial That Could Kneecap OpenAI

    <h3>The Compute Lock-In Play</h3><p>Google's $40B Anthropic investment isn't a funding round — it's an <strong>infrastructure capture</strong>. The deal is structured as $10B upfront at a $350B valuation, with up to $30B more tied to performance milestones, plus massive compute supply agreements including <strong>5GW of TPU capacity over five years</strong>. By making Anthropic dependent on Google Cloud for training and inference, Google creates a lock-in that transcends financial investment. Combined with Amazon's $25B pledge, Anthropic now commands <strong>$65B+ in committed capital</strong> from its two largest cloud providers. Both hyperscalers are calculating that whoever provides the compute substrate controls the value chain, regardless of which model wins.</p><blockquote>The two leading frontier model companies are now cloud-captive. Model provider optionality — the foundation of most enterprise AI strategies — is evaporating.</blockquote><h3>Anthropic's Revenue Validates the Bet</h3><p>Anthropic's reported <strong>233% revenue growth</strong> — from roughly $9B to $30B annualized in approximately four months — is the fastest enterprise software scaling in recorded history. This isn't early-stage momentum; it's infrastructure-tier adoption. The remaining independent alternatives — Mistral, open-source, potentially Cohere (which just acquired Aleph Alpha with $600M backing from Germany's Schwarz Group for European sovereign AI) — represent a materially different capability tier for now.</p><h3>The Trial That Could Change Everything</h3><p>Monday's Musk v. Altman jury selection isn't corporate drama — it's a <strong>binary event for AI industry structure</strong>. Musk is demanding over $100B in damages, the removal of Altman and Brockman, and reversal of OpenAI's for-profit restructuring. Even a partial victory — testimony from Nadella, Altman, Musk, and Brockman will dominate headlines for weeks — could freeze OpenAI's IPO preparation and create governance uncertainty that ripples through every Microsoft Copilot dependency. One litigation expert called it a "Hindenburg landing on the Titanic." The timing against Google's Anthropic mega-deal is <em>not coincidental</em> — Google is hedging that if OpenAI stumbles, Anthropic becomes the instant beneficiary.</p><h3>The Infrastructure Financing Wall</h3><p>Beneath the deal headlines, <strong>Oracle's ~$300B data center push</strong> is straining Wall Street's syndication capacity. Banks are hitting exposure limits, meaning the financial system itself is struggling to pace AI capital demands. When X-energy's nuclear IPO prices above range oversubscribed, it confirms that <strong>power supply for data centers is a binding constraint</strong>. Compute is likely to get more expensive and less available over the next 18-36 months. Any growth model assuming the opposite needs stress-testing.</p><hr><p>The strategic imperative across all four signals is identical: <strong>build multi-provider AI architecture now</strong>, while there's still time to do it thoughtfully rather than in crisis mode. The window for deliberate positioning is narrowing as hyperscaler consolidation, legal uncertainty, and capital constraints simultaneously reduce optionality.</p>

    Action items

    • Conduct an emergency review of OpenAI/Microsoft AI stack concentration risk and develop a 90-day provider diversification plan before Musk trial dynamics crystallize
    • Open preliminary conversations with Anthropic's enterprise team this quarter to evaluate as a strategic AI partner or hedge
    • Stress-test compute procurement assumptions against a scenario where costs rise 20-30% over 18 months due to financing and permitting constraints
    • Brief your board this month on the AI competitive landscape shift and its implications for vendor, infrastructure, and capital strategy

    Sources:AI's $65B arms race is repricing your entire supply chain · Google's $40B Anthropic lock-in just reset the AI platform war · Musk v. Altman trial + Google's $40B Anthropic bet are reshaping your AI partner calculus · AI's pricing crisis is here: frontier models double in cost while open-weight rivals match them

  3. 03

    Enterprise AI's Real Bottleneck Is Your Organization, Not Your Models

    <h3>The Diagnosis Nobody Wants to Hear</h3><p>A practitioner with <strong>$300K in active enterprise transformation contracts</strong> articulates the thesis every technology leader needs to internalize: there are no AI-native enterprises, and there won't be soon — not because models aren't capable, but because enterprises are political economies whose internal physics <strong>actively resist the transparency AI requires</strong>. The framework is deceptively simple but strategically profound: nearly every AI success story you've seen is <strong>"AI in the business"</strong> — what the company sells or delivers. Meanwhile, how the company actually runs — quarterly planning on emailed slide decks, budgets in headcount units, information hoarded as political currency — remains stubbornly pre-AI.</p><blockquote>Enterprise illegibility is rational, not broken. Information hoarding isn't pathology — it's how individuals maintain leverage. When you deploy AI agents requiring transparent workflows, you're asking people to voluntarily dismantle the power structures they've spent years building.</blockquote><h3>The Productivity Paradox Compounds This</h3><p>Multiple sources converge on a troubling pattern: <strong>AI productivity gains are individually documented but invisible on corporate P&Ls</strong>. Individual developers are faster, individual analysts more productive — but organizations aren't measurably more profitable. This is a Solow Paradox redux, and it's becoming a <strong>board-level credibility problem</strong>. When the CFO asks "where's the ROI?" and the answer is "developers feel more productive," that budget gets scrutinized. If your business depends on enterprise AI spending — whether as vendor, platform, or buyer — you need to solve the measurement problem before the market solves it by cutting budgets.</p><h3>Why Agents Will Hit a Wall Your Roadmap Doesn't Show</h3><p>The "agent Cold Start problem" is political, not technical. When an AI agent joins an organization, it faces the same challenge any new executive does — <strong>learning the real org chart, not the one in HR</strong>. The names people recommend you talk to look very different from reporting lines on paper. If your agentic AI roadmap assumes organizations are more machine-readable than they actually are, you'll hit a wall that no model upgrade will fix. The prerequisite layer — <strong>organizational intelligence tooling</strong> that maps real decision flows and informal power structures — is largely unbuilt and represents a massive market gap.</p><h3>The Structural Blocker: FTE-Denominated Budgets</h3><p>Traditional cost-center budgets denominated in headcount are <strong>structurally incompatible with AI transformation</strong>. AI work that spans functions, replaces partial tasks across many roles, or creates value that doesn't map to existing line items literally cannot be funded in most enterprise financial architectures. The honest question isn't "what models are we deploying?" but "have we redesigned our financial architecture to fund AI work that doesn't fit in FTE-denominated cost center budgets?" The answer, for nearly every enterprise, is no.</p><hr><p>The companies that will win are the ones doing the unsexy work: <strong>redesigning budget structures, creating incentive alignment for transparency, and building organizational intelligence</strong> before deploying agents into workflows. Everyone else is doing transformation theater — impressive product-side deployments that look great in board presentations while the actual operating system of the enterprise remains untouched. The window is <strong>18-24 months</strong> before competitors figure this out.</p>

    Action items

    • Audit all AI initiatives this month: split into 'in the business' (product) vs. 'on the business' (operations) — if the ratio exceeds 80/20, flag for board discussion
    • Challenge your CFO to model AI capital allocation outside FTE-denominated structures — pilot one cross-functional 'AI budget' by end of Q3
    • Map shadow AI usage across the organization and treat it as demand signal, not compliance problem — identify top 5 shadow use cases by adoption
    • Launch an AI ROI attribution program tying deployments to revenue, margin, or cycle time — not developer satisfaction — before next budget cycle

    Sources:Your enterprise AI strategy probably has a fatal blind spot · AI's pricing crisis is here: frontier models double in cost while open-weight rivals match them

◆ QUICK HITS

  • Anthropic's Project Deal experiment proves stronger AI agents systematically extract more value in negotiations — and the losing side rates their deals as equally fair, with no awareness they lost

    AI's $65B arms race is repricing your entire supply chain

  • Intel surged 23.6% — its best day since 1987 — signaling the first credible foundry/manufacturing challenge to the Nvidia-TSMC axis; have your infrastructure team track Intel's AI accelerator roadmap seriously

    Musk v. Altman trial + Google's $40B Anthropic bet are reshaping your AI partner calculus

  • Cohere acquired Aleph Alpha backed by $600M from Germany's Schwarz Group — European sovereign AI is moving from 'preferred' to 'required' in regulated procurement, expect compliance mandates within 18 months

    Google's $40B Anthropic lock-in just reset the AI platform war

  • SpaceX's S-1 disclosed xAI investigations into sexually abusive AI imagery as a material securities risk — establishing precedent that AI safety failures are securities disclosure items affecting valuations and market access

    Google's $40B Anthropic lock-in just reset the AI platform war

  • OpenAI's chief scientist publicly admits AI progress is 'surprisingly slow' while simultaneously marketing GPT-5.5 as a 'new class of intelligence' — the internal-external narrative gap signals diminishing returns on brute-force scaling

    AI's pricing crisis is here: frontier models double in cost while open-weight rivals match them

  • Update: Google at Cloud Next '26 announced Wiz integration + AI security agents for multicloud protection, positioning to own the cloud security layer through platform bundling — a lock-in play that forces CISO vendor stack decisions this year

    Anthropic's Mythos leak + Google's Wiz play signal a cybersecurity power shift

  • Update: CISA leadership vacuum deepens after Plankey withdrawal — described as 'not a straightforward story' — plan for unreliable federal cyber coordination through at least mid-2027 and diversify threat intelligence sources

    Anthropic's Mythos leak + Google's Wiz play signal a cybersecurity power shift

  • Update: Stablecoin velocity doubled from 2.6x to 6x in two years — supply is being transacted, not hoarded — with C2B commerce transactions up 128% YoY to 284.6M and Rain-powered card programs at $300M+/month from zero in 15 months

    Stablecoins just flipped from cross-border tool to local payments rail

BOTTOM LINE

China's AI stack just went NVIDIA-independent — DeepSeek V4 runs on Huawei Ascend at $0.14/M tokens while 4 of 5 top open-weight models are Chinese-built and MIT-licensed. Google responded by locking $40B in compute around Anthropic, OpenAI doubled GPT-5.5 pricing, and Musk's trial to dismantle OpenAI starts Monday. The model layer is commoditizing from below while consolidating from above, and the organizations that will capture value are the ones solving the hard problems no model upgrade fixes: reliable deployment of unreliable AI, organizational redesign for agent-readiness, and multi-provider architecture before optionality disappears.

Frequently asked

Should we migrate production workloads to DeepSeek V4 given its pricing advantage?
Not without a verification layer. V4 Flash at $0.14/$0.28 per million tokens under MIT license is compelling, but its 94-96% hallucination rates on the AA-Omniscience factual benchmark disqualify it from most enterprise use cases. The defensible play is building reliable deployment infrastructure — verification, domain fine-tuning, human-in-the-loop — around commodity models rather than treating raw model access as the product.
What does Google's $40B Anthropic deal actually change for enterprise buyers?
It ends meaningful model-provider optionality at the frontier tier. With Anthropic now tied to Google Cloud via 5GW of TPU capacity and $65B+ in committed hyperscaler capital (Google plus Amazon's $25B), and OpenAI locked to Microsoft, your cloud choice increasingly dictates your model choice. Enterprise AI strategies built on multi-model flexibility need to be re-architected around this compute-capture reality.
How should we prepare for the Musk v. Altman trial's potential impact?
Treat it as a binary governance event and build contingency now. Musk is seeking $100B+ in damages, removal of Altman and Brockman, and reversal of OpenAI's for-profit restructuring. Even a partial victory could freeze OpenAI's IPO and create 6-12 months of uncertainty rippling through every Microsoft Copilot dependency. A 90-day provider diversification plan should be in motion before the trial's dynamics crystallize.
Why aren't our documented AI productivity gains showing up in financial results?
Because FTE-denominated budgets structurally prevent AI value from aggregating. Individual productivity improvements get absorbed into existing cost-center structures that can't fund work spanning functions or replacing partial tasks across roles. Until finance architecture is redesigned to fund 'on the business' AI work separately from headcount, gains will remain individually real but organizationally invisible — and AI budgets will face escalating CFO scrutiny.
What's the realistic timeline for Western AI cost leadership given China's progress?
Roughly 12-18 months, not 3-5 years. DeepSeek V4 running natively on Huawei Ascend demonstrates a functional NVIDIA-independent stack, Chinese labs hold 4 of the top 5 open-weight positions, and Ascend 950 supernodes deploy at scale in H2 2026. Combined with OpenAI's chief scientist publicly calling progress 'surprisingly slow,' the competitive window for cost-based Western dominance is closing faster than most strategic plans assume.

◆ ALSO READ THIS DAY AS

◆ RECENT IN LEADER