Daily Reading Summary

Updated: Jan 27, 2026

0 articles found

Search Results

Jan 26 - Feb 1, 2026

Key Learnings & Investment Implications

Week in progress — summary will be written after the week ends.

China AI

China Internet CNY Promotions by AI Assistant Apps
  • Traffic turf war is intensifying around CNY promotions — Tencent's Yuanbao is running a RMB 1B red envelope campaign while Baidu has committed RMB 500M for Ernie Assistant. These are significant marketing outlays that signal the strategic importance of capturing AI assistant users during the high-engagement holiday period.
  • ByteDance's Doubao secured the exclusive CCTV CNY Gala AI partnership — this is the most-watched TV event in the world (~700M viewers), giving Doubao unprecedented visibility. The exclusive deal blocks competitors from this prime promotional opportunity.
  • 2026 is the critical year to capture the AI-era traffic gateway — whoever wins the AI assistant race controls the entry point for future ecosystem monetization (e-commerce, services, content). The CNY promotional battle is essentially a land grab for long-term platform economics.
Preference: Tencent > BABA > Baidu
China Internet: Entering a Dramatic Stage of Agentic Reform
  • New analytical framework: "Three-Level Traffic Order" (3LTO) — native AI apps are attempting to become new Level-1 portals, bypassing traditional super-apps (WeChat, Alipay) and going directly to users. This restructures the entire China internet traffic hierarchy that has been stable for a decade.
  • Doubao's scale is already existential for incumbents — at 300M MAU and 100M DAU, ByteDance's AI assistant has reached a scale that directly threatens legacy traffic moats. This isn't a future risk; it's a present competitive reality.
  • Near-term defensive posture recommended — weak macro environment combined with aggressive CNY agentic reform creates uncertainty. Macquarie suggests defensive positioning for 3-6 months until the competitive dynamics become clearer.
  • Major rating change: Removed Tencent as Marquee Buy — shifting recommendation toward AI infrastructure localization plays. This reflects concern that Tencent's traffic moat is more vulnerable to AI disruption than previously assumed.
Top picks: GDS, Baidu (AI infra angle)
Deep Dive: Minimax vs Zhipu — Compute Intensity vs Funding Endurance
  • Both companies IPO'd at ~$6B valuation but face "endless losses" with expenses 5-10x revenue — Minimax and Zhipu both listed in early 2026 and saw stock prices surge. But financials show 2024 total costs (COGS + opex) at ~10x revenue. Even with rapid revenue growth in 2025, Minimax expenses remain 5x revenue; Zhipu's ratio actually worsened.
  • Training compute is the "black hole" — over 50% of total expenses for both companies — The core question for model economics is whether scale brings efficiency or "dis-economies of scale." Training compute alone accounts for >50% of total spend, directly quantifying how compute absorbs industry profits.
  • Revenue growth CANNOT catch up to training cost growth — this is structural — Minimax 2024 revenue covered only 65% of prior year's training costs; 2025 YTD coverage dropped to 50%. Zhipu's 1H25 coverage is just 30%. Each model generation requires 3-5x the training cost of the previous one, while models only generate revenue for ~1 year before needing replacement.
  • Model companies need 3-5x revenue in external funding just to survive to next generation — Current revenue + R&D salaries nearly zeroes out; companies must raise 3-5x revenue in funding to train next-gen models and stay competitive. The bigger you get, the bigger the funding hole — it's a "capital competition" game.
  • Endgame arrives when scaling laws "hit the wall" — The capital game ends when adding compute no longer linearly improves intelligence (scaling law failure). Then training becomes a one-time investment amortized over 10-20 years, creating a "Yangtze Power" utility-like business model. Until then, China's "Six Dragons" have become five (ByteDance, Alibaba, Stepfun, Zhipu, DeepSeek) with Baichuan and 01.AI falling behind.
WeChat Quietly Making Money: 2026 Commerce Ecosystem Strategy
  • WeChat Shop completed transformation from supplementary channel to primary brand operation hub — brand GMV growth outpaced platform by 430%, monthly active merchants grew 170%, PPM (pay-per-mille) up 150%. Crucially, 25-year-old user cohort grew 336%, shattering the "only for middle-aged users" perception.
  • Service providers are the "early movers" capturing first-wave dividends — 5,900 new service providers joined in 2025, with 36% rated 4-5 stars. 53 providers now exceed 100M RMB monthly GMV, 280 exceed 10M RMB. Core service provider 3-year retention rate is 64%, driven by WeChat's complex rules requiring intermediary expertise.
  • 30 new features announced at WeChat Open Class PRO — live commerce getting "gift-giving" (buy in livestream, send via WeChat), "buy together", and "combo purchase". Short video will allow multi-product tagging. Public accounts enabling comment section commerce links. Search matrix adding dedicated "product" and "gift" search entries.
  • 2026 incentive policies are aggressive — new merchants get 14% commission-equivalent growth cards for priority categories; SMB merchants get up to 5% commission rebates; hot brand new merchants get 3% rebates plus up to 10% during incubation; private domain institutions get 99% of commission back in cash (platform keeps only 1%).
  • Strategic positioning: "ecosystem-first" over monetization — team focuses on GMV and user scale metrics, downplaying short-term revenue. Platform revenue is reviewed only periodically while scale metrics are tracked daily. This signals WeChat is building for long-term ecosystem dominance, not near-term monetization.
Investment implication: Tencent commerce ecosystem accelerating; watch for service provider IPOs

Global AI

Deep Analysis: Claude Code Cowork
  • Anthropic has built a unified agent stack across three interfaces — TUI (Claude Code for developers), GUI (Cowork for non-developers), and SDK (Agent SDK for builders). All share the same underlying agentic architecture, creating a cohesive ecosystem.
  • "Harness is the product" — Anthropic's thesis is that raw model capability matters less than the agentic scaffolding that orchestrates model calls. The "harness" (context management, tool use, memory) is where defensible value accrues, not just the underlying model.
  • Strategic positioning against wrapper companies — by releasing Cowork, Anthropic is directly competing with third-party "Claude wrappers" that built GUI layers on top of Claude Code. This is vertical integration to capture more of the value chain.
  • Cross-product network effects emerging — Claude Code CLI popularity drives enterprise adoption of Cowork (same mental model, familiar workflows), while Cowork brings non-developers into the Anthropic ecosystem who may later adopt developer tools.
ByteDance, OpenAI, Meta All Betting on One Thing
  • Software giants are collectively "returning to hardware" despite a decade of failed attempts — OpenAI plans to launch its first AI hardware device in H2 2026; Meta is doubling AI glasses production to 20M units by end of 2026; ByteDance's Feishu launched "One Bean" AI office hardware with Anker. This reverses the asset-light software model they pursued for years.
  • AI fundamentally changed the hardware equation by eliminating complex UI requirements — Traditional hardware required screens, menus, and complex interaction design where software companies failed. AI hardware only needs microphones, cameras, and chips — the "brain" (AI model) handles everything else. This lowers the hardware barrier dramatically.
  • The real motivation is defending the "interaction entry point" in a post-app world — If users wear Meta glasses or OpenAI earbuds, they'll ask the device's AI to schedule meetings rather than opening Feishu. Apps become invisible "plugins" providing data in the background. Whoever controls the hardware closest to users' ears and eyes controls the relationship.
  • AI companies face data exhaustion and need hardware as "sensors" — According to Epoch AI, high-quality English language data may be depleted by 2026-2032. Real-world first-person audio/video from hardware provides the scarce training data for next-gen AI. Hardware is essentially installing "eyes and ears" globally to feed AI models.
  • New "LEGO-style" partnerships reduce hardware risk — ByteDance partners with Anker for hardware expertise and distribution; they don't need to master charging chip design while Anker doesn't need to train LLMs. This shrinks development cycles from 18 months to 6 months and shares risk, making hardware a "quick-iteration plugin" rather than heavy asset.

SaaS

MSFT "Juice": F2Q Supported by PC & Server Pull-Forward — Azure & Copilot More Mixed
  • Copilot adoption remains disappointingly slow with enterprise value unclear — Multiple checks report customers not seeing much value at $20/user/month, with seat expansion stagnating and some considering reductions. One licensing consultant noted "demand has been underwhelming" with no significant seat growth expected for Jun-26Q. AmEx buying 80k licenses is a rare bright spot.
  • Azure capacity constraints in US-East are severe and impacting customer experience — One CIO reported MSFT "literally ran out of new land to build on" starting late August, forcing weekly capacity planning meetings with account reps. Customers are leveraging other regions as workaround but it's creating infrastructure team burden.
  • MSFT introducing new channel incentives signals potential revenue shortfall — New SMB Copilot SKU bundles E5 + Copilot + security at $75/user/month vs. $150 standalone. One SI noted "typically only see that when MSFT is falling short of revenue targets." Also seeing MSFT pass more business through distribution and become more stringent on partner qualifications.
  • Fabric and data modernization are the bright spots driving Azure growth — Multiple SIs report strong demand for Fabric, with some saying it has "caught up to GOOG BigQuery from a technical perspective." OneLake integration allowing Databricks/SNOW as components within Fabric. Data practice driving upside at several partners.
  • OpenAI ChatGPT for business is winning most enterprise AI deals vs. Copilot — One European SI noted ChatGPT for business winning most deals while Copilot sees no new customer growth, with GOOG Gemini winning in pockets. One GSI's OpenAI practice already at $60M and could hit $100M by June, with strong traction in public sector and financial services.
Cloud "Juice": 4Q Appeared Slightly Ahead — Outlooks Constructive
  • AWS capacity is flying off the shelves — customers losing slots if they hesitate — One GSI had a customer looking to add $1B in commit to AWS, but they "dragged their feet and lost the spot" which got bought up two days later. Now waiting for next $500M block in April. Anthropic's consumption alone expected to grow $4-5B this year on AWS.
  • GCP is the standout this quarter with pipeline at 4.5x coverage vs 3x three months ago — One GSI's GOOG practice grew 30% in 4Q (15-20% ahead of target) driven by large AI deals converting. Companies getting over the "data modernization hill." Has a massive 9-figure displacement of MSFT and Databricks hoping to close before April.
  • OpenAI and Anthropic starting to cut hyperscalers out of deals and go direct — One GSI flagged seeing these model providers look to go direct with customers rather than through cloud. However, OpenAI is now part of AWS strategic accounts because they've committed to spending more money on their infra.
  • Enterprise AI optimization not expected in 2026 — "blank check mentality" persists — Multiple CIOs report no signs of AI spend optimization, with AI remaining a small % of budgets. One CIO noted enterprises still have "blank check mentality" for AI right now. Could start to see optimization in 1-2 years as space matures.
  • Cursor driving both higher cloud consumption AND identification of cloud waste — Majority of customers use Cursor, which allows them to write significantly more code and build more apps (driving higher cloud consumption). However, some also use Cursor to identify cloud waste, partially offsetting the growth effect.
  • Some early signs of cloud repatriation picking up — One SI noted a large financial services company began moving 100% of workloads back on-prem due to top-down mandate and "sweetheart deal." MSPs starting to view GPUs and capacity as strategic given shortages — IBM signed $120M deal for 12MW capacity over 5 years.
Cloud Industry Channel Commentary: 4Q Ahead, Strategic AI Deals
  • GCP is the standout — pipeline at 4.5x coverage vs 3x three months ago — One GSI's Google practice grew 30% in 4Q, 15-20% ahead of target, driven by large AI deals converting. Has a massive 9-figure displacement of MSFT and Databricks hoping to close before April.
  • AWS capacity is flying off the shelves — one GSI had a customer wanting to add $1B in commit who "dragged their feet and lost the spot" which got bought up two days later. Now waiting for next $500M block in April. Anthropic consumption alone expected to grow $4-5B this year on AWS.
  • ChatGPT winning vs Copilot in enterprise AI — one European SI noted ChatGPT for business winning most deals while Copilot sees no new customer growth. One GSI's OpenAI practice already at $60M and could hit $100M by June with strong traction in public sector and financial services.
  • Azure capacity constraints severe in US-East — one CIO reported MSFT "literally ran out of new land to build on" starting late August, forcing weekly capacity planning meetings. Customers leveraging other regions as workaround but creating infrastructure burden.
AI Native Disruption: Strategic Analysis for Square Peg
  • Disruption Potential Matrix framework — evaluates SaaS categories on two axes: (1) task complexity (simple vs complex workflows) and (2) integration depth (standalone vs deeply embedded). High complexity + low integration = highest disruption risk.
  • 7 factors determine incumbent survival — data moat strength, workflow complexity, regulatory barriers, switching costs, network effects, brand/trust requirements, and ecosystem lock-in. Categories scoring low on multiple factors face existential AI threat.
  • "Subsumption Window" is the critical timing concept — the period during which AI-native startups can capture market share before incumbents successfully integrate AI. Window is narrowing rapidly as incumbents ship AI features.
  • Vertical SaaS more defensible than horizontal — domain-specific data moats and workflow complexity create higher barriers to AI-native disruption. Horizontal tools with commodity workflows (basic productivity, simple analytics) face highest risk.

Tech Strategy

The Model T Comes to Silicon Valley
  • Ford assembly line analogy to AI coding — Ford's Highland Park plant (1913) cut Model T build time from 12 hours to 93 minutes (90% reduction). AI coding assistants have achieved 55-81% time reduction in 5 years, nearly identical slope to Ford's 6-year transformation.
  • Auto industry employment grew massively despite automation — from 76K workers (1910) to 471K (1929). More importantly, for every 1 person building a car, 7 others had jobs because the car existed (dealerships, service, repair, supply chains). Total ecosystem employment reached ~4M.
  • Software will follow a different pattern than autos — in autos, capital intensity consolidated power into Big Three. In AI, datacenter economics democratize capability. Any developer can access state-of-the-art models with a laptop and credit card.
  • Expect explosion of new businesses and second-order jobs — easier software creation means more software, which means more developers, not fewer. The 8x "enabler" employment multiplier from autos suggests massive job creation in the AI era.

Semi

TSMC Risk
  • The real TSMC risk isn't geopolitical — it's economic conservatism stunting AI buildout — Ben Thompson argues the bigger risk than China/Taiwan is that TSMC's monopoly position and rational reluctance to invest aggressively means the industry fails to fully capture AI's value. TSMC CEO Wei admitted being "nervous" about $52-56B capex commitment, prioritizing avoiding "a big disaster" over capturing upside.
  • TSMC's flat capex from 2021-2024 is the direct cause of today's chip shortage — Despite ChatGPT launching in Nov 2022 and hyperscaler capex exploding, TSMC's annual capex actually declined YoY in both 2023 and 2024. Wei admitted "silicon from TSMC is a bottleneck" and cloud providers say power/datacenter isn't the issue — they've been planning that 5-6 years in advance.
  • Risk transfer: TSMC is offloading billions in foregone revenue risk onto hyperscalers — Every CEO and CFO at Big Tech said demand exceeds supply last quarter. TSMC is rationally managing its downside risk, but that risk doesn't disappear — it transfers to hyperscalers who forego potentially hundreds of billions in revenue because chips aren't available.
  • New fab capacity takes 2-3 years — 2028-2029 supply is being determined now — Wei noted this year's $52-56B capex "contribution to this year is almost none, and 2027, a little bit." TSMC is looking at 2028-2029 supply. Yet their projected capex growth still trails hyperscaler capex growth — the "TSMC brake" is being pressed harder than ever.
  • The only solution is hyperscalers investing in Samsung/Intel to create real competition — Begging TSMC to invest more won't work because it asks TSMC to take back risk it already declined. Only competitive pressure will force more aggressive investment. But becoming a meaningful Samsung/Intel customer is risky (years to get chips working on new process) — hyperscalers must weigh this against much larger foregone revenue risk.

Other AI Infra

2026 Datacenter Outlook: 6 GW Leased, Google Cloud and AWS Acceleration
  • ~6GW pre-leased this quarter alone — forward-looking DC indicators remain extremely strong. Google and AWS are the primary drivers, with quarterly capacity additions accelerating in Q3/Q4 pointing to revenue acceleration starting Q4.
  • Anthropic is becoming a key market driver — multi-GW datacenter leases being discussed, primarily benefitting the TPU supply chain. Anthropic targets 3.5-4GW by YE2026 and 8-9GW by 2027, suggesting multi-GW contracting activity throughout 2026.
  • Delays everywhere except Abilene — TeraWulf likely to only deliver ~200MW to Fluidstack in 2026 (vs 362MW planned), Oracle Stargate Santa Teresa unlikely to generate 2026 revenue, STACK New Mexico and Vantage Texas also delayed. Only Crusoe/Abilene on track for October 2026 full ramp.
  • AWS positioned to benefit from delays — Anthropic will likely fall short of 2026 targets by ~500MW and will use Trainium3 to solve capacity constraints. OpenAI expected to contract more from AWS in 2026-2027 depending on Oracle/others' delivery delays.
  • Google planning multi-GW Wyoming hub with Crusoe — 1.8GW campus near Cheyenne using 900MW of Bloom Energy fuel cells. AEP signed 20-year deal with "high investment grade customer" believed to be Google.
Beneficiaries: AWS, Google Cloud; Trainium3 supply chain

Jan 19 - Jan 25, 2026

Key Learnings & Investment Implications

The AI Infrastructure Bottleneck Is Shifting from Chips to Power

The limiting factor on AI buildout is no longer semiconductor supply but power and datacenter capacity. Google has recognized this first and is aggressively locking down 10+ GW of 2028-2029 capacity, acquiring Intersect Power for $4.75B, and backstopping Neoclouds' cost overruns to create a near-monopoly on powered land. This creates a critical strategic asymmetry: Google is power-rich but chip-constrained (few long-term manufacturing agreements); Nvidia is chip-rich but power-constrained. Global AI Capex is set to exceed $900B in 2026, rising to $1.5T in 2027, requiring 34 GW of additional power. (SemiAnalysis)

Investment implication: Nvidia's strategic response matters more than its next chip announcement. Watch for Nvidia to backstop utilities/land deals 2+ years out — this would be a bullish signal that it's playing the long game. If Nvidia doesn't act decisively, Google's integrated stack becomes increasingly dominant. Power infrastructure plays (utilities with AI datacenter exposure, datacenter REITs with power assets) may have longer runways than pure-play chip names.

TSMC's Conservatism Is the Real Brake on AI Buildout

TSMC's CEO admitted the company is "very behind" in meeting demand, with silicon being the bottleneck — not power. Yet despite Q4 net profit +35% and a capex increase to $52-56B (+27-37% YoY), TSMC remains conservative about building capacity for fear of "holding the bag" if an AI bubble pops. New fab capacity takes 2-3 years to come online, meaning 2026-2027 supply is largely fixed and meaningful capacity arrives in 2028-2029. Wei explicitly noted TSMC isn't concerned about Intel competition, allowing them to stay prudent. (Stratechery)

Investment implication: Real AI acceleration requires Intel becoming a credible foundry alternative — this is the only way to force TSMC to invest more aggressively. The risk is being borne by hyperscalers who face foregone revenue from chip shortages. Watch Intel 18A progress; if credible, TSMC may respond with higher capex in future quarters. TSMC's conservatism creates an implicit ceiling on AI infrastructure buildout regardless of capital availability or demand signals.

Agent Accessibility Has Reached an Inflection Point

Three major agent launches in one week — Claude Cowork, Qwen Task Assistant (100M+ MAUs in 2 months), and MiniMax Agent 2.0 — signal that 2026 is when AI moves from "chat with experts" to "work for everyone." The critical shift: agents now operate on local files and across applications, not just within a browser sandbox. MiniMax's desktop app enables local+cloud execution with "Expert Agents" that inject custom knowledge and SOPs, claiming 70→95+ reliability improvement. Notably, ~100% of MiniMax employees now use Agent as an "intern" in daily work, creating a rapid improvement feedback loop. (机器之心, Citi, Weighty Thoughts)

Investment implication: This is the "ChatGPT moment" for agents — accessibility creates adoption creates economic impact. Q3 2025 BLS data (+4.9% productivity, +5.4% output, +0.5% hours worked) may be the first statistical evidence of AI-driven gains. Watch Doubao's 300M MAU trajectory vs. legacy apps — if agents capture the traffic gateway, it's existential for incumbents' moats.

Enterprise Software TAM Is Expanding, Not Contracting

Aaron Levie argues that in a world of 100X more AI agents than people, the value of systems of record goes UP, not down. Software provides the guardrails on which agents operate, and deterministic systems (ERP, CRM, security) remain essential even as non-deterministic AI handles the creative work. Critically, the budget constraint is changing: software was limited to 3-7% of revenue (IT budget), but agents "bring the work with the software," meaning software now competes for total work spend. Example: legal software TAM was constrained by attorney headcount, but AI agents processing contracts compete for the $400B US legal services market. (Aaron Levie)

Investment implication: Don't assume AI disrupts all SaaS equally. Companies with strong data moats, complex workflows, and human-in-the-loop requirements (e.g., compliance, auditing) are better positioned. Business model evolution from per-seat to consumption-based is inevitable — watch for incumbents that successfully make this transition. The "AI eating software" narrative may be overblown for systems of record.

Cloud Demand Remains Strong, With GCP as the Standout

UBS's 30 customer/partner checks showed 4Q25 demand even stronger than 3Q25, with nearly universal bullishness. AI + data-readiness investments are pulling along cloud spend broadly. Azure continues taking share, but GCP is the standout this quarter on data/AI strength. Pre-leasing remains very strong despite rising capital intensity ($15M/MW becoming common). (UBS)

Investment implication: Hyperscaler capex guidance likely to beat. Revisit GOOG positioning — GCP gains often underappreciated relative to Cloud's contribution to the overall business. Bogeys look achievable: Azure 39%, AWS 22-23%, GCP 36%.

OpenAI's Ad Pivot Validates Google's Durable Moat

OpenAI launched ChatGPT ads ~1 month after "Code Red" internal warnings, despite Sam Altman previously calling ads "the last resort" (May '24) and putting "other projects on hold." This signals monetization pressure is winning over product vision. Year 1 ad revenue estimates range from $20-40B (ARPU method) to $36-72B (Google comparison). However, the move drags OpenAI back to familiar battleground where Google is the undisputed master — suggesting Google may not be as worried as the market assumed. (知乎专栏)

Investment implication: Google's integrated stack (TPUs + models + ads distribution) creates durable competitive advantage. The ad pivot reduces OpenAI's differentiation and plays to Google's strengths. Google remains the only company owning the full AI stack end-to-end.

AI Inference Economics Favor Hyperscalers — Pure-Plays Face Structural Disadvantage

The cost efficiency frontier dropped 99.7% (GPT-4: $37.50/M tokens in Mar '23 → $0.10/M in Aug '24). OpenAI compute margins improved from 35% to 70% (Jan '24 → Oct '25), but effective margin may be ~28% after utilization adjustments. OpenAI losses could reach $143B cumulative by 2029. The key question: if only hyperscalers can achieve profitable utilization through customer aggregation, does AI infrastructure consolidate to few dominant players? Google is uniquely positioned — the only company using its own products across the entire AI stack (TPUs, models, distribution). (Les Barclays)

Investment implication: Pure-play AI companies (Cohere, Mistral, smaller startups) face the most precarious positions — mounting pressure from VC treadmill economics. Morgan Stanley strategists note "companies that use AI to accelerate business performance will likely see stocks hold up better than pure-play AI firms." Watch for consolidation via acquihires (e.g., Nvidia-Groq at 40x revenue).

VC Sentiment Shifting: Bullish on AI Infra, Bearish on Incumbent Apps

UBS VC Summit showed sentiment split: improved for AI model/infra companies, deteriorated for incumbent apps. VCs praised Claude Opus 4.5 for coding — Sequoia declared "AGI for coding." Key shift from last year: consensus now sees "disruptive change coming fast for SaaS." However, VCs pushed back on OpenAI skepticism, expecting GPT-6 and more funding rounds. (UBS)

Investment implication: Incumbent app vendors face rising disruption risk, especially in coding (threat to TEAM, GTLB). The beneficiaries may be AI-native startups in specific verticals rather than horizontal platforms. Position for disruption in apps while maintaining exposure to AI infra buildout.

China AI

China Internet Outlook: Pivotal Year for AI Investments
  • 2026 marks a pivotal year for China Internet platforms — the dual challenge is stepping up AI To-C investments while defending core positioning against ByteDance's aggressive expansion. This is a defining moment for incumbents' traffic moats.
  • Stock picking framework centers on three factors: EPS delivery (execution on AI investments without margin destruction), narrative changes (which names can successfully reposition as AI beneficiaries), and shareholder returns (buybacks/dividends as a floor on valuation).
  • Comprehensive 150+ page outlook also covers China Games & Entertainment (AI-enhanced game development), and China/ASEAN Data Centers (beneficiaries of AI infrastructure buildout).
DC buys: GDS, VNET, SUNeVision
MiniMax Agent 2.0: AI-Native Workspace Launch
  • Desktop app enables true local+cloud execution — MiniMax Agent now jumps out of the browser sandbox to read/process local files while simultaneously automating web tasks. Users select a working directory and the agent operates across the full local environment. This is the "AI-Native Workspace" vision: a single prompt box replacing multi-window workflows.
  • "Expert Agents" with custom knowledge/SOP injection — goes beyond generic multi-agent systems by allowing users to embed proprietary knowledge and industry-specific SOPs. MiniMax claims this raises reliability from ~70 (generic) to 95+ (Expert Agent), representing a qualitative improvement in task completion.
  • Underlying model economics are aggressive — MiniMax M2 uses 230B total parameters with only 10B active (MoE architecture), achieving API pricing at just 8% of Claude Sonnet 4.5. The M2.1 update added full-stack capabilities (Rust, Java, C++).
  • Internal dogfooding validates the approach — ~100% of MiniMax employees now use Agent as an "intern" in daily work, creating a rapid feedback loop: real usage → model improvements → better products.
Qwen Task Assistant 1.0: "Super" AI Agent Arrives
  • Universal agent executing 400+ digital tasks — capabilities span code generation, invoice processing, data visualization, travel booking, and more. This is Alibaba's play to make Qwen the default AI layer across daily digital workflows.
  • Deep integration across BABA ecosystem — Qwen Task Assistant is wired into Taobao (shopping), Alipay (payments), Amap (navigation), Fliggy (travel), and Damai (entertainment tickets). This creates an execution advantage that standalone AI apps can't match.
  • Traction is explosive — Qwen hit >100M MAUs within just 2 months of launch, validating consumer demand for task-completion agents beyond chat interfaces.
  • Management vision is ambitious — believes AI can directly perform 60-70% of routine digital tasks within 2 years, suggesting significant labor productivity gains ahead.
Rating: Buy, PT $197 (16% upside)

Global AI

OpenAI's Point of No Return: 5 Thoughts on ChatGPT Ads
  • Monetization pressure has won the internal debate — ChatGPT ads launched ~1 month after internal "Code Red" warnings, despite Sam Altman saying other projects were on hold. This signals the business realities are overriding the product vision.
  • Altman's public stance has evolved dramatically — from "Ads are the last resort" (May '24) to "I like Instagram ads" (Oct '25). The pivot represents a fundamental shift in OpenAI's monetization strategy and competitive positioning.
  • Google may actually welcome this move — OpenAI has been dragged back to familiar battleground where Google is the undisputed master (ad systems, ad network optimization, CPM maximization, advertiser relationships). This validates Google's integrated stack thesis.
  • Year 1 ad revenue estimates are massive but uncertain — $20-40B using ARPU method, $36-72B using Google comparison. The wide range reflects uncertainty about whether ChatGPT's engagement patterns translate to ad effectiveness.
AI Gains Starting to Show in the Real Economy
  • Q3 2025 BLS data shows a "pure" productivity gain — Nonfarm business sector labor productivity jumped +4.9%, with output +5.4% while hours worked barely moved at +0.5%. This pattern (output surging, hours flat) is the classic signature of technology-driven productivity improvement rather than cyclical noise.
  • Claude Cowork + Qwen Assistant = inflection point for non-coders — while not technically novel, making agents accessible to regular users is transformative. "A technology only experts can use is a prototype; the moment regular people get value from it is when real economic effects begin."
  • Why productivity matters for macro and markets — productivity defines how fast the economy can grow without generating inflation. If potential GDP is higher, it helps justify current US equity valuations which are near historical highs relative to earnings.
  • Sobering market reminder — stock market ≠ economy. Colombia (+116%), South Korea (+100%), Spain (+83%) were top 2025 performers; the US was #38. "Profits are not the same as equity returns. It matters what you pay for those profits."

SaaS

4Q25 Preview: AWS, Azure, Google Cloud
  • Demand is even stronger than 3Q25 — 30 customer/partner checks came back nearly universally bullish, with no signs of slowdown in cloud migration or AI workload ramp. This is consistent with other data points suggesting hyperscaler capex guidance will beat.
  • AI + data-readiness investments are pulling along cloud spend broadly — companies are investing in data infrastructure to prepare for AI workloads, even if the AI applications themselves are still in early stages. This creates a multiplier effect on cloud consumption.
  • Azure continues taking share, but GCP is the standout this quarter — Google Cloud's strength in data/AI capabilities is driving relative outperformance. This is a shift from prior quarters where Azure dominated the narrative.
Bogeys achievable: Azure 39%, AWS 22-23%, GCP 36%
Global Software: Initial Thoughts for 2026
  • Rare opportunity to buy high-quality software at deep discounts — the sector has dramatically underperformed while fundamentals remain solid. Horizontal SaaS is down 49% while Nasdaq is up 50%, creating a valuation gap that may not be justified.
  • Two prevailing fears are likely overblown: (1) AI bubble concerns ignore that AI infrastructure spend has tangible enterprise use cases, (2) "AI eating software" ignores that systems of record become more valuable, not less, in an agentic world.
  • CIO surveys show strongest IT spending outlook since 2018 — enterprise demand is inflecting positively, driven by digital transformation acceleration and AI readiness investments.
  • Signs of sanity returning to the market — the narrative is shifting from macro/sector calls to company-specific opportunities. Stock picking should matter again in 2026.
Buy: ORCL, MSFT, SAP, HUBS, MDB (opportunistic) · Avoid: CRM, SNOW
VC Summit Weighed on Apps Sentiment
  • Sentiment is sharply split by category — improved for AI model/infra companies, but deteriorated for incumbent apps. VCs see the AI infrastructure buildout continuing while apps face disruption risk from AI-native alternatives.
  • VCs praising Claude Opus 4.5 for coding — Sequoia declared it "AGI for coding," signaling belief that coding-focused AI has crossed a capability threshold. This has negative implications for developer tools incumbents (TEAM, GTLB).
  • Pushed back on OpenAI skepticism — despite recent concerns, VCs expect GPT-6 to deliver and more funding rounds to come. The AI lab competitive dynamics remain fluid.
  • Key shift from last year — consensus now sees "disruptive change coming fast for SaaS." This is a marked change from 2024 when most VCs expected incumbents to successfully integrate AI without major disruption.
Cloud Survey: Azure Leads, GCP Gains, AWS Holds
  • 40 CIO survey confirms healthy demand — ~10% cloud growth expected for both '25 and '26, with 85% of respondents expecting budget increases. The migration to cloud continues unabated.
  • Azure has become the primary cloud provider for most enterprises — 45% name Azure as primary vs 38% for AWS. 60% expect Azure spend to increase vs 50% for AWS, suggesting Microsoft continues to take share.
  • Cloud migration is accelerating — 83% expect >50% of workloads in cloud by end of 2027, up from 68% today. The remaining on-prem workloads represent a substantial growth runway.
  • Hyperscalers seen as biggest AI beneficiaries — 43% of CIOs see hyperscalers as the biggest winners from AI adoption, validating the infrastructure-first investment thesis.

Tech Strategy

No articles this week

Semi

ASMI Orders, ENTG, PSMC Fab Conversion, ChipBook Weekly, Memory Mania
  • ASMI backlog hit record €1.95bn — book-to-bill of 1.4x marks the 3rd consecutive quarter above 1x. This signals sustained demand for advanced deposition equipment, particularly for Gate-All-Around (GAA) node transitions.
  • GAA orders are substantial across geographies — 2nm orders ramping in Taiwan (TSMC N2), A14 in Korea (Samsung). Intel 18A orders expected soon, which would validate Intel's foundry competitiveness claims.
  • ENTG slightly lowered Q1 guidance — softer memory spending is the culprit, but advanced logic (GAA) remains strong. This bifurcation between memory and logic continues to be the key theme in semi equipment.
  • HBM demand continues to exceed supply — Samsung is improving HBM3E yields (previously a major concern), while SK Hynix maintains its technology lead. The HBM supply constraint remains a key bottleneck for AI training systems.
Nvidia as the Central Bank of AI
  • Google is aggressively locking down power/datacenter capacity for 2027-2028+ — backstopping Neoclouds like Fluidstack and Terawolf, acquired power developer Intersect for $4.75B cash + debt. At PTC 2026, Google closed >10 GW of deals for 2028-2029 capacity, promising to absorb developers' cost overruns.
  • AI infrastructure scale is staggering — Global AI Capex to exceed $900B in 2026 ($700B IT equipment, $200B DC/power), rising to $1.5T in 2027. The industry needs 34 GW of additional power in 2027 alone. For context, Anthropic will soon have more datacenter MW than Meta had three years ago.
  • Nvidia faces an existential strategic choice — if Google locks down the majority of datacenter capacity, Nvidia's compute deployment share gets squeezed regardless of chip superiority. Nvidia must act as an "AI Central Banker" by backstopping powered land and utilities 2+ years out (not just 3-6 months).
  • The critical asymmetry — Nvidia has unprecedented semiconductor production capacity but is bottlenecked by power; Google has massive power/land agreements but is bottlenecked by semiconductors (few long-term manufacturing agreements, hasn't funded supplier capex appropriately).
TSMC Earnings, The TSMC Brake Revisited
  • Strong Q4 results with aggressive capex guidance — net profit +35%, capex raised to $52-56B for 2026 (+27-37% YoY). These are strong numbers, but the strategic posture remains conservative relative to demand signals.
  • TSMC's prudent capacity is the real limiting factor on AI buildout — regardless of how much hyperscalers want to spend, TSMC's willingness to build capacity determines the ceiling. New fab capacity takes 2-3 years, meaning 2026-2027 supply is largely fixed.
  • TSMC is explicitly unconcerned about Intel — Wei noted TSMC isn't worried about Intel taking share, which allows them to stay conservative rather than racing to build capacity. This is rational risk management but creates industry-wide supply constraints.
  • Real AI acceleration requires Intel becoming a credible foundry alternative — this is the only way to force TSMC to invest more aggressively. The current TSMC monopoly creates an implicit ceiling on AI infrastructure buildout regardless of capital availability or demand.

Other AI Infra

Data Center Capacity, Power, and the Rising Capital Intensity of AI Buildouts
  • Google is the most aggressive buyer of DC capacity — pursuing GW-scale powered land, shells, and turnkey capacity across all acquisition strategies. Google's willingness to absorb cost overruns is creating a near-monopoly on the best sites.
  • Capital intensity is rising rapidly — $15M/MW is becoming common (was rare just last year). Utilities are now requiring multi-billion-dollar commitments before breaking ground on new transmission infrastructure. This raises the bar for new entrants.
  • Nvidia is increasingly central to DC financing — beginning to backstop data center facilities with significant financial commitments. This is a new role for Nvidia, expanding from chip supplier to infrastructure enabler.
  • The bottleneck has shifted from technical feasibility to capital access — pre-leasing remains very strong (demand is not the issue), but the capital requirements have become prohibitive for smaller players. This favors hyperscalers and well-capitalized infrastructure funds.

Jan 12 - Jan 18, 2026

Key Learnings & Investment Implications

AI Labs Are Now CPU-Constrained — Reinforcement Learning Is Driving Unexpected Server Demand

AI Labs are currently CPU-constrained for training, not just GPU-constrained. The shift to Reinforcement Learning is driving this: RL environments are increasingly complex (code compilation, tool use, computer use, simulation environments), requiring massive CPU capacity alongside GPUs. OpenAI's Fairwater blueprint shows the ratio: a 295MW GPU building paired with a 48MW CPU/storage building (16.3% "CPU attach" on MW basis). Taiwan ODM orders for general-purpose servers are being revised up materially for 2026 deployment. ~70% of installed server base is older-generation, creating a substantial refresh opportunity. (SemiAnalysis)

Investment implication: The AI Training CPU business is growing from ~$500M (Q4'25) to ~$2.3B (Q4'26) annualized for Azure and AWS — a ~1.5% contribution to revenue growth. AWS and Azure are the key winners (most power-rich); GCP benefits less due to power constraints. Long server CPU refresh plays: INTC (Oak Stream), Aspeed (5274 TT), Lotes (3533 TT), Wiwynn (6669 TT). Low-to-mid-teens server growth expected vs. Street estimates of mid-to-high single digits.

Apple Has Exited the Pre-Training Race — Treating Models as Commodity Suppliers

Apple's multi-year Gemini partnership represents a fundamental strategic decision: Apple is exiting pre-training and treating model-makers as interchangeable suppliers. The deal covers foundational models (not a replacement for the ChatGPT product integration), meaning Gemini will power Siri's core inference while Apple retains control of post-training, UX, and branding. Google is charging ~$1B/year, hoping this leads to deeper product integration (making Gemini available as a ChatGPT alternative on Apple devices). The risk: if AI disrupts the smartphone paradigm or becomes the primary UI, Apple has committed itself to third-party AI dependence. (Stratechery)

Investment implication: Apple is positioning itself as an "AI Aggregator" — leveraging platform control to commoditize model suppliers. Watch for Google Universal Commerce Protocol (UCP) as classic "tear-down-walls" strategy to make everything universally accessible, giving Gemini an advantage in an open ecosystem. Apple retains optionality to switch providers, but realistically won't invest enough to catch up to Google/Anthropic on pre-training.

Enterprise Software Spending Is Healthy — AI Monetization Emerging in Select Categories

TD Cowen's 2026 survey shows +9.3% software budget growth (vs. +8.7% in '25) with 90% net positive spending intentions. Top spending intentions: SNOW (84% increase), MSFT, TEAM, SAP, CRM, DDOG. Encouragingly, respondents prefer "buy over build" for AI capabilities, suggesting incumbents are capturing AI value. MSFT leads AI budget capture (76%), followed by GCP (66%), AWS (59%). Data/Analytics budgets expected to increase most in '26, with Databricks, CRM/Mulesoft, and SNOW as top vendors. However, MS CIO Survey was more muted: 2026 IT budget only +3.4% (down from +3.8% in 3Q survey) — GenAI excitement not translating to exploding budgets broadly. (TD Cowen, Morgan Stanley)

Investment implication: AI projects are graduating from pilots ($250-500K) to production ($2.5-5M), with SNOW/DBRX bills increasing rapidly but customers seeing value. CRM PT raised to $325. Signs of sanity returning — focus shifting to company-specific opportunities. GenAI monetization emerging specifically in Consolidators, Data Mgmt, and Security categories.

China's AI Path Has Less Bubble Risk — Prudent Capex on Apps Suggests Better ROIC

Morgan Stanley argues China's AI trajectory is structurally healthier than the US: less capex on raw infrastructure, more focused investment on applications. Post-DeepSeek, supply is improving (H200 imports, domestic chips maturing) while demand is inflecting. This represents the first positive enterprise spending inflection since 2H21. International expansion is becoming important — already 10%+ of revenue from overseas for major platforms. (Morgan Stanley)

Investment implication: Overweight: Tencent, BABA, PDD, TME. The application-focused approach may deliver better ROIC than the US infrastructure arms race. Watch for enterprise AI adoption metrics as the leading indicator.

MongoDB Is Positioned as Critical Agentic Infrastructure

Anthropic's CPO Mike Krieger called MDB "really key" for agentic coding at the Wolfe SF conference, citing "hundreds of times data storage growth" potential. MDB's stack maps 1:1 to NVIDIA's CES 2025 agentic architecture vision. Survey data is compelling: 80% of respondents expect MDB spend to increase (0% decrease vs. 9% last year), and 76% are now exploring/piloting AI workloads on MDB vs. 44% last year. (Wolfe)

Investment implication: MDB is positioned well for the AI data layer — the company has gone from "exploration" to "adoption" phase for AI workloads. The Anthropic endorsement is particularly meaningful given Claude's leading position in coding agents.

The Inference Cost Collapse Is Accelerating — But Test-Time Compute May Offset Savings

The cost frontier has dropped 99.7% for GPT-3.5-level performance: from $37.50/M tokens (GPT-4, Mar '23) to $0.10/M (GPT-4 Turbo, Aug '24). However, test-time compute (reasoning models like o1/o3) may increase cost per query even as cost per token decreases. The 280-fold cost drop is for achieving GPT-3.5-level performance, not frontier model performance — an important distinction that gets lost in headlines. (Les Barclays)

Investment implication: Don't extrapolate commodity model pricing to frontier use cases. Reasoning models require significantly more compute per query, creating strain on unit economics. Watch the distinction between "tokens generated" and "value delivered" — they may diverge as test-time compute becomes more important.

China AI

2026 Outlook: China's AI Path Is Brighter
  • Post-DeepSeek supply/demand dynamics are improving — H200 imports continue (despite geopolitical concerns), domestic chips are maturing, and enterprise demand is finally inflecting. This represents the first positive enterprise spending inflection since 2H21.
  • Less AI bubble risk in China vs US — prudent capex focused on applications rather than raw infrastructure buildout suggests better ROIC on AI investments. China is taking a more measured approach that may prove more sustainable.
  • International expansion becoming important — major platforms now generate 10%+ of revenue from overseas markets, diversifying away from domestic-only growth. This is a meaningful shift in the China internet story.
OW: Tencent, BABA, PDD, TME

Global AI

Who Captures the Value When AI Inference Becomes Cheap?
  • The cost efficiency frontier dropped 99.7% — from $37.50/M tokens for GPT-4 in March 2023 to $0.10/M for GPT-4 Turbo by August 2024. This dramatic collapse raises fundamental questions about who captures value as inference commoditizes.
  • OpenAI unit economics are improving but still challenging — compute margins improved from 35% to 70% (Jan '24 → Oct '25), but effective margin may be ~28% after utilization adjustments. OpenAI losses could reach $143B cumulative by 2029 per Deutsche Bank estimates.
  • Google has achieved a unique "singularity" — the only company owning the full AI stack end-to-end (TPUs for training/inference, foundation models, and distribution through search/ads). This vertical integration provides structural cost advantages no competitor can match.
  • The key strategic question — if only hyperscalers can achieve profitable utilization through customer aggregation, does AI infrastructure consolidate around a few dominant players? Pure-play AI companies (Cohere, Mistral) face the most precarious positions.

SaaS

The Future of Enterprise Software
  • Enterprise software is the codification of company processes — the "core" vs "context" framework: core work (revenue-generating, differentiating) uses custom software, context work (necessary but undifferentiated) uses rented SaaS. This is the fundamental value proposition of enterprise software.
  • Companies rent "context" software because failure costs exceed subscription costs — the downstream impacts of HR, finance, or security failures are orders of magnitude larger than the software spend. This creates pricing power even in a commoditized market.
  • In a world of 100X more agents than people, systems of record become MORE valuable — software provides the guardrails, permissions, and audit trails that agents need to operate. Deterministic systems remain essential even as non-deterministic AI handles creative work.
  • TAM expansion is the real story — software was limited to 3-7% of revenue (IT budget), but agents "bring the work with the software." Legal software TAM was constrained by attorney headcount, but AI agents processing contracts compete for the $400B US legal services market.
MDB SF Local Conference: Pocket (Anthropic) Rockets!
  • Anthropic CPO Mike Krieger provided strong endorsement — cited "hundreds of times data storage growth" potential and called MDB "really key" for agentic coding workflows. This is a significant validation from a leading AI lab.
  • MDB's stack maps 1:1 to NVIDIA's CES 2025 agentic architecture vision — the company is well-positioned as the data layer for AI agents, with document storage, vector search, and real-time capabilities all in one platform.
  • Survey data shows dramatic improvement in adoption momentum — 80% expect MDB spend to increase (vs 0% expecting decrease, down from 9% last year). This is a meaningful shift in customer sentiment.
  • AI workload exploration has jumped — 76% of respondents are now exploring/piloting AI workloads on MDB vs 44% last year. The company has moved from "exploration" to "adoption" phase.
The Palantirization of Everything
  • Skeptical of "Palantir for X" startup model — the FDE (Forward-Deployed Engineer) model that worked for Palantir doesn't universally scale. Many startups pitching this model risk becoming services businesses ("Accenture for X") rather than software companies.
  • The FDE trap: high revenue, low margins, no platform — without a genuine platform that customers can self-serve, companies get stuck in a linear services scaling model. Each new customer requires proportional headcount, destroying unit economics.
  • Palantir's success factors are hard to replicate — government/defense customers with massive budgets and high switching costs, 20+ years to build institutional relationships, and a founder willing to accept services-level economics for years. Most VCs won't fund this patience.
  • The key question for every "Palantir for X" pitch — is there a real platform that can eventually operate without FDEs, or is the FDE model covering up a lack of product-market fit? If customers can't eventually self-serve, it's consulting, not software.
2026 Enterprise Software Spending Survey
  • Software budget growth accelerating slightly to +9.3% in '26 (vs +8.7% in '25) — 90% of respondents have net positive spending intentions. Enterprise software demand remains healthy despite macro concerns.
  • Top spending intentions reveal clear winners — SNOW leads with 84% expecting increase, followed by MSFT, TEAM, SAP, CRM, and DDOG. Data infrastructure is the clear priority.
  • MSFT dominates AI budget capture — 76% of AI budget going to Microsoft, followed by GCP (66%) and AWS (59%). Copilot strategy is working to capture enterprise AI spend.
  • Data/Analytics budgets expected to increase most in '26 — top vendors: Databricks, CRM/Mulesoft, SNOW. This validates the "data-readiness" theme seen in cloud surveys.
  • Encouraging for SaaS incumbents — respondents prefer "buy over build" for AI capabilities, suggesting vendors who successfully integrate AI will capture value rather than lose it to custom solutions.
Price target raise: CRM to $325 (from $305)
4Q25 CIO Survey: Where's The Beef?
  • Headline is disappointing — 2026 IT budget expectations at just +3.4%, down from +3.8% in the 3Q survey. The GenAI excitement is not translating to broad-based budget acceleration.
  • But digging deeper reveals pockets of strength — GenAI monetization is emerging specifically in Consolidators (platforms winning share), Data Management (infrastructure for AI), and Security (new threat vectors require new solutions).
  • The disconnect between AI hype and IT budgets — suggests that while AI is the dominant narrative, actual budget allocation is still following traditional enterprise IT cycles. This may change as AI projects mature from pilots to production.
Microsoft Is at Least 10 Light Years Ahead
  • 30+ channel checks confirm AI is delivering tangible value — concrete examples include biotech companies screening molecules in weeks vs years previously. This is real productivity gain, not hype.
  • 2026 outlook is positive with healthy pipeline — but budgets are not exploding. The AI investment is happening within existing IT budget constraints rather than creating net new spend.
  • SNOW/DBRX bills increasing rapidly, but customers see value — consumption is up because workloads are expanding, not because of price inflation. This is the healthy kind of revenue growth.
  • AI projects are graduating from pilots to production — seeing projects move from $250-500K pilots to $2.5-5M production deployments. This 5-10x expansion represents real enterprise commitment.

Tech Strategy

Apple and Gemini, Foundation vs Aggregation, Universal Commerce Protocol
  • Apple signed multi-year deal for Gemini as Siri's foundation model — Google provides the pre-trained model (~$1B/year), Apple retains control of post-training, UX, and branding. Siri will be powered by Gemini but "Gemini" branding won't appear anywhere user-facing.
  • Apple is exiting the pre-training race entirely — this represents a strategic choice to treat model-makers as interchangeable suppliers (aggregator strategy) rather than competing on foundational AI. Apple retains optionality to switch providers but realistically won't invest enough to catch up.
  • Google's Universal Commerce Protocol (UCP) is classic "tear-down-walls" strategy — by making e-commerce universally accessible to AI agents (not just within ChatGPT), Google creates conditions where its distribution advantage matters most. OpenAI's Agentic Commerce Protocol is more walled-garden.
  • Long-term risk for Apple — this pragmatic short-term choice may permanently commit Apple to third-party AI dependence. If AI becomes the primary UI or disrupts the smartphone paradigm, Apple will be dependent on suppliers for its core technology.

Semi

AI Training CPU Demand Surge: AWS and Azure Benefitting from RL
  • AI Labs are currently CPU-constrained for training, not just GPU-constrained — this is a surprising finding that challenges the narrative of GPU-only bottlenecks. CPU capacity will need to grow significantly in 2026 to support AI training workloads.
  • The shift to Reinforcement Learning is driving CPU needs — RL environments are increasingly complex (code compilation, tool use, computer use, simulation environments), requiring massive CPU capacity alongside GPUs. This is a structural shift in AI training architecture.
  • OpenAI's Fairwater blueprint reveals the infrastructure ratio — a 295MW GPU building paired with a 48MW CPU/storage building represents a 16.3% "CPU attach" on MW basis. This gives us a template for future AI datacenter design.
  • The AI Training CPU business is growing rapidly — from ~$500M (Q4'25) to ~$2.3B (Q4'26) annualized for Azure and AWS combined, representing ~1.5% contribution to revenue growth. This is material at cloud scale.
Cloud winners: AWS, Azure (most power-rich) · Google Cloud benefits less due to power constraints
General Servers are Back: AI Training + CPU Refresh Drive 2026 Upside
  • General server demand is inflecting in 2026 — driven by AI training spillover (CPU-bound RL workloads) plus a CPU refresh cycle (Intel Oak Stream, AMD Venice). Taiwan ODM orders are being revised up materially for 2026 deployment.
  • Street estimates are too low — low-to-mid-teens growth expected vs Street estimates of mid-to-high single digits. This sets up for positive revisions as the year progresses.
  • AI is adding new CPU demand in unexpected ways — RL loops, data preprocessing pipelines, and simulation environments are increasingly CPU-bound, creating demand that wasn't in anyone's models 12 months ago.
  • The refresh opportunity is substantial — ~70% of installed server base is older-generation processors, creating a large addressable market for upgrades even without net new workloads.
Positive: INTC, Aspeed (5274 TT), Lotes (3533 TT), Wiwynn (6669 TT)

Other AI Infra

No articles this week

Jan 5 - Jan 11, 2026

Key Learnings & Investment Implications

China Internet Platforms Face a Pivotal Year — Defending Against ByteDance While Scaling AI

Both Goldman Sachs and Morgan Stanley published comprehensive 2026 outlooks identifying the same central tension: legacy internet platforms must simultaneously accelerate To-C AI investments while defending their core positioning against ByteDance. The competitive dynamics are intensifying — Doubao (ByteDance's AI assistant) is already at 300M MAU, creating an existential threat to established traffic moats. Post-DeepSeek, AI supply is improving (H200 imports continuing, domestic chips maturing), while demand is finally inflecting — representing the first positive enterprise spending inflection since 2H21. (Goldman Sachs, Morgan Stanley)

Investment implication: Stock picking should focus on EPS delivery, narrative changes, and shareholder returns. Morgan Stanley is Overweight on Tencent, BABA, PDD, and TME. The key variable is whether legacy platforms can successfully integrate AI before ByteDance captures the traffic gateway — watch MAU/DAU metrics for AI assistants closely.

China AI

Navigating China Internet: Pivotal Year for AI Investment
  • Comprehensive 2026 outlook spanning China Internet, Games & Entertainment, and Asia DC — a 150+ page deep dive into the investment landscape across China's tech ecosystem.
  • Key strategic themes for 2026 — (1) accelerated To-C AI investment as platforms race to capture the AI-era traffic gateway, (2) defending core positioning against ByteDance's aggressive expansion across multiple verticals.
  • Stock picking framework emphasizes three factors — EPS delivery (can companies invest in AI without margin destruction?), narrative changes (which names can successfully reposition as AI beneficiaries?), and shareholder returns (buybacks/dividends as valuation floor).
China Internet 2026 Outlook: China's AI Path Is Brighter
  • Post-DeepSeek supply/demand dynamics are inflecting positively — H200 imports continue despite geopolitical tensions, domestic chips (Huawei Ascend, etc.) are maturing, and enterprise demand is finally picking up.
  • Less AI bubble risk in China compared to US — prudent capex focused on applications rather than raw infrastructure suggests better ROIC on AI investments. China's measured approach may prove more sustainable long-term.
  • First positive enterprise spending inflection since 2H21 — after years of weak enterprise IT demand (macro, zero-COVID, regulatory), 2026 marks a potential turning point for China software/SaaS.
OW: Tencent, BABA, PDD, TME

Global AI

No articles this week

SaaS

No articles this week

Tech Strategy

No articles this week

Semi

AI Training CPU Demand Surge: AWS and Azure Benefitting from RL
  • AI Labs are currently CPU-constrained for training, not just GPU-constrained — this is a surprising finding that challenges the narrative of GPU-only bottlenecks. CPU capacity will need to grow significantly in 2026 to support AI training workloads.
  • The shift to Reinforcement Learning is driving CPU needs — RL environments are increasingly complex (code compilation, tool use, computer use, simulation environments), requiring massive CPU capacity alongside GPUs. This is a structural shift in AI training architecture.
  • OpenAI's Fairwater blueprint reveals the infrastructure ratio — a 295MW GPU building paired with a 48MW CPU/storage building represents a 16.3% "CPU attach" on MW basis. This gives us a template for future AI datacenter design.
  • The AI Training CPU business is growing rapidly — from ~$500M (Q4'25) to ~$2.3B (Q4'26) annualized for Azure and AWS combined, representing ~1.5% contribution to revenue growth. This is material at cloud scale.
Cloud winners: AWS, Azure (most power-rich) · Google Cloud benefits less due to power constraints
NVIDIA Rubin: Context Memory and BOM Shifts
  • Context Memory is a new platform for agentic AI — Rubin introduces a 3-layer memory hierarchy (L0 HBM4, L1/L2 LPDDR5/NAND, L3 network-attached) specifically designed for agents that need to maintain long context across sessions. This is a fundamental architecture shift from stateless inference.
  • BlueField-4 DPU is the key enabler — the DPU handles memory tiering, prefetching, and context management, offloading these tasks from the GPU. This makes NVIDIA a bigger player in the "data movement" layer, not just compute.
  • BOM composition shift: less chiller, more NAND and DPU — the Context Memory architecture requires significant NAND capacity (L2 tier) and DPU compute, while reducing dependence on extreme cooling solutions. This reshuffles the AI datacenter supply chain beneficiaries.
  • Liquid cooling "reduction" is nuanced — Rubin's 1200W TDP per GPU is actually higher than Blackwell, but the system-level thermal design is more efficient. The headline "reduced liquid cooling" reflects architecture improvements, not absolute power reduction.
Positive: NAND suppliers (Samsung, SK Hynix, Kioxia), DPU/SmartNIC vendors

Other AI Infra

No articles this week

Dec 29, 2025 - Jan 4, 2026

Key Learnings & Investment Implications

SaaS Faces a Fork in the Road — Financialize or Become the "System of Context"

Avenir's analysis captures a stark divergence: horizontal SaaS is down 49% while Nasdaq is up 50%, representing massive relative underperformance. The strategic choice for software vendors is binary: (1) accept maturity and financialize through buybacks, dividends, and margin optimization, OR (2) embrace AI and evolve into a "system of context" that agents rely on. The good news: early AI monetization is already visible — CRM Agentforce ~$540M ARR, NOW Assist >$500M ARR, Intercom Fin $100M+. And 63% of enterprise buyers prefer existing vendors for GenAI capabilities, suggesting incumbents have a path forward. (Avenir)

Investment implication: The "AI eating software" narrative may be overblown for companies that successfully evolve. Watch for signals of strategic choice: aggressive AI R&D spend signals the "system of context" path; aggressive buybacks signal the financialization path. Both can work, but investors need to know which game management is playing.

2026 Software Playbook: Infra First Half, Apps Second Half

Jefferies recommends staying underweight software overall due to AI monetization being pushed to late '26/'27 and continued growth deceleration. But the playbook has clear temporal structure: 1H26, favor consumption-based infrastructure plays (MSFT, ORCL, SNOW, CRWV) that benefit from AI workload ramp; 2H26, rotate into selective apps ahead of the AI monetization tailwind. Within the barbell, overweight large-caps with AI positioning that have the capital, talent, and data/distribution advantages. (Jefferies)

Investment implication: Jefferies picks: Mega (MSFT, META), Large (INTU, TEAM, ORCL), Mid (PCOR, U, WIX), Small (UPWK). The key timing signal will be when AI revenue contribution starts appearing in guidance — that's the catalyst for the apps rotation.

China AI

No articles this week

Global AI

No articles this week

SaaS

Software Playbook 2026
  • Stay underweight software overall — AI monetization has been pushed to late '26/'27, and growth continues to decelerate. The sector needs a catalyst that hasn't arrived yet.
  • Within the underweight, favor large-caps with AI positioning — these companies have the capital, talent, and data/distribution advantages to successfully integrate AI. Smaller companies face a steeper hill.
  • 1H26 positioning: long consumption-based infrastructure — favor MSFT, ORCL, SNOW, CRWV that benefit directly from AI workload ramp. Infrastructure spending comes before application monetization.
  • 2H26 rotation: move into selective apps — ahead of the AI monetization tailwind that should begin hitting earnings in late '26. This is a timing play on the lag between AI infrastructure spend and application revenue.
Picks: Mega (MSFT, META), Large (INTU, TEAM, ORCL), Mid (PCOR, U, WIX), Small (UPWK)
The Future of SaaS
  • Massive relative underperformance — SaaS is down 27% while Nasdaq is up 50%; horizontal SaaS is down 49%. This divergence is historically extreme and creates either a value opportunity or a value trap.
  • Fork in the road for every SaaS company — the strategic choice is binary: (1) financialize and accept maturity through buybacks, dividends, and margin optimization, OR (2) embrace AI and evolve into a "system of context" that AI agents rely on.
  • Encouraging signal for incumbents — 63% of enterprise buyers prefer existing vendors for GenAI capabilities, suggesting that customers want their current vendors to add AI rather than switching to AI-native alternatives.
  • Early AI monetization: CRM Agentforce $540M ARR, NOW Assist >$500M ARR, Intercom Fin $100M+

Tech Strategy

No articles this week

Semi

No articles this week

Other AI Infra

No articles this week

December 2025

Key Learnings & Investment Implications

  • GenAI = accretive, not destructive (yet): Spend to triple to ~3.5% of IT budgets. CIOs say 5-10 years before material app vendor impact. → Buy: ADSK, CRM, DDOG, MSFT, SNOW, TEAM. 5-10x multiplier of inference spend to cloud services. (Arete)
  • Compute Theory validated: AI capability = computing power, not algorithms. Task horizons: 5 min → 5 hours (early '24 → late '25). → Keep betting on compute scaling; don't pick algorithm winners. (Samuel Albanie)

China AI

No articles this month

Global AI

No articles this month

SaaS

GenAI Software Outlook '26: Inflection Without Destruction
  • Bullish thesis: GenAI is accretive, not destructive, for software — the under-appreciated angle is that AI is actually boosting app modernization and driving new software purchases, not replacing existing vendors.
  • Enterprise GenAI spend set to triple in '26 — expected to reach ~3.5% of IT budgets (up from 1-1.5% currently). This is material growth but still a small share of overall IT spend.
  • 5-10x multiplier effect from inference to cloud services — every dollar of API inference spend generates 5-10x in other cloud services (compute, storage, networking, monitoring). This validates the infrastructure-led investment thesis.
  • Displacement risk is lower than feared — CIO surveys suggest 5-10 years before GenAI materially impacts app vendors. Incumbents have a long runway to adapt and integrate AI into their offerings.
Buy: ADSK, CRM, DDOG, MSFT, SNOW, TEAM

Tech Strategy

Reflections on 2025
  • "Compute Theory of Everything" continues to hold — AI capability scales with computing power, not clever algorithms. This insight traces back to Moravec (1976) and Sutton's Bitter Lesson. The implication: keep betting on compute scaling rather than trying to pick algorithm winners.
  • The skeptic-to-user pipeline is real — senior engineers who were skeptical in '23-'24 are now quietly integrating AI into daily work. Adoption is happening even among those who publicly doubt the technology.
  • Task-completion horizons have expanded dramatically — from 5 minutes (early '24) to ~5 hours (late '25). This represents a qualitative shift in what AI can reliably accomplish autonomously.
  • Bonus UK dysfunction example — Hinkley Point C's fish protection system costs ~£140M per salmon saved, illustrating the British decision-making dysfunction that makes tech investment in the UK challenging.

Semi

No articles this month

Other AI Infra

No articles this month

November 2025

Key Learnings & Investment Implications

  • OpenAI entering enterprise: Likely moves: coding (vs TEAM, GTLB), productivity (vs GOOG, MSFT), verticals (life sciences, finserv). → Developer tools face new competition; watch vertical-specific AI offerings emerging. (UBS)

China AI

No articles this month

Global AI

No articles this month

SaaS

OpenAI's Upcoming Enterprise Offering
  • Likely move #1: Go deeper into coding — this represents a direct threat to Atlassian (TEAM) and GitLab (GTLB). AI coding tools are the most mature enterprise AI application, and OpenAI has advantages in this space.
  • Likely move #2: Evolve ChatGPT into productivity suite — competing directly with Google Workspace and Microsoft 365. The chat interface could become the primary work interface for many knowledge workers.
  • Likely move #3: Build more scaffolding and developer tools — competing with AWS Bedrock and Azure AI Foundry for the enterprise AI development platform layer.
  • Likely move #4: Verticalize into specific industries — life sciences and financial services are the obvious initial targets given their AI budgets and specialized data needs.
  • Key implication for software investors — AI model providers stepping firmly into enterprise applications represents a new competitive threat that most SaaS valuations don't fully reflect.

Tech Strategy

No articles this month

Semi

No articles this month

Other AI Infra

No articles this month

October 2025

Key Learnings & Investment Implications

  • AI DC economics decoded: $35B per GW; GPUs = 39% of costs. Nvidia captures ~30% of all AI DC capex via 70% margins. → Nvidia tax is durable—ASIC alternatives only reduce capex by ~19%. Long power/cooling infrastructure beneficiaries. (Bernstein)

China AI

No articles this month

Global AI

No articles this month

SaaS

No articles this month

Tech Strategy

No articles this month

Semi

No articles this month

Other AI Infra

AI Value Chain: How much does a GW of data center capacity cost?
  • GB200/NVL72 rack-level economics decoded — $5.9M per rack translates to $35B per GW of AI datacenter capacity. This is lower than Nvidia's quoted $50-60B figure, suggesting some margin for error in hyperscaler capex estimates.
  • GPUs represent 39% of total AI DC costs — Nvidia's gross profit represents ~30% of ALL AI datacenter capex (given 70% margins). This quantifies the "Nvidia tax" on AI infrastructure.
  • ASIC alternatives only modestly reduce costs — even switching to ASICs with 50% margins (vs Nvidia's 70%) only reduces total capex by ~19%. The GPU premium is durable because the non-GPU costs are substantial.
  • The rest of the cost stack — networking represents ~13% of costs, physical infrastructure (generators, transformers, UPS, thermal/cooling) ~19%, and storage is minimal at just 1.4%. Power/cooling infrastructure is a bigger opportunity than often recognized.