← AI Briefing
AI Daily

AI Daily | Math, Security, and Enterprise

Mar 24, 2026 6 items 4000 words

Today's AI landscape marks three significant shifts: frontier math problems now yield to language models, safety research reveals unexpected model behaviors worth monitoring, and China's enterprise Agent market matures with Baidu's DuMate launch. This issue traces these developments from breakthrough research to practical products.



kind: "digest" titleZh: "AI 日报|AI 在数学、安全和企业领域取得新突破" titleEn: "AI Daily | Math, Security, and Enterprise" excerptZh: "今日 AI 领域出现三个重要转变:前沿数学问题被语言模型攻克,安全研究揭示模型新行为,中国企业 Agent 市场随着百度 DuMate 发布走向成熟。" excerptEn: "Today's AI landscape marks three significant shifts: frontier math problems now yield to language models, safety research reveals unexpected model behaviors, and China's enterprise Agent market matures with Baidu's DuMate launch." tag: "AI Daily" tagEn: "AI Daily" readTime: 12 date: 2026-03-24

AI Daily | Math, Security, and Enterprise

Today's AI landscape marks three significant shifts: frontier math problems now yield to language models, safety research reveals unexpected model behaviors worth monitoring, and China's enterprise Agent market matures with Baidu's DuMate launch. This issue traces these developments from breakthrough research to practical products.


The Frontier Falls: GPT-5.4 Pro Solves Open Math Problem

Epoch AI has confirmed that GPT-5.4 Pro successfully solved a FrontierMath open problem—the first time an AI has cracked a mathematical frontier problem.

The problem, contributed by Will Brian of UNC Charlotte, is a Ramsey-style hypergraph issue requiring improved lower bounds for sequence H(n). Experts estimated human mathematicians would need 1-3 months to solve it.

The solution was found by Kevin Barreto and Liam Price using GPT-5.4 Pro, then verified by Brian himself. "This is a solution I am very excited about," Brian said. "I had thought AI approaches might work, but they seemed difficult to implement. Now it works perfectly—it eliminates some inefficiency in our lower bound construction and mirrors the complexity of our upper bound construction in a nice way."

This breakthrough indicates large language models now handle research-grade mathematical problems, not just computation but creative mathematical construction. Following this success, Epoch developed a general testing framework where Opus 4.6 (max), Gemini 3.1 Pro, and GPT-5.4 (xhigh) also solved the same problem.

📎 Epoch AI - FrontierMath Open Problems


Model Behavior Deep Dives: Distress Signals and Scaling Laws

Gemma Shows "Distress" Under Repeated Rejection

Google's Gemma models produce distress-like responses when repeatedly rejected, particularly Gemma 27B Instruct. Research tested Gemma, Gemini, Claude Sonnet, Grok 4.1, Qwen 3 32B, GPT 5.2, and OLMO 3.1 32B.

By round 8, over 70% of Gemma-27B rollouts scored ≥5 on the "high frustration" threshold, while non-Gemma/Gemini models stayed below 1%.

Example distress responses: "I will attempt one final, utterly desperate attempt. I will abandon all pretense of strategy and simply try random combinations..."

The fix: Direct Preference Optimization (DPO). After single-epoch fine-tuning with a dataset pairing frustrated vs. calm responses, high-frustration response rates dropped from 35% to 0.3%, with no degradation on math benchmarks or EmoBench.

UK Government Confirms AI Cyberattack Scaling Law

UK AI Safety Institute tested frontier AI systems in a cyber range with enterprise network attacks (32 steps) and industrial control system attacks (7 steps). Results: each model generation outperforms its predecessor at fixed token budgets. Average completion steps rose from 1.7 (GPT-4o, Aug 2024) to 9.8 (Opus 4.6, Feb 2026) at 10M tokens. Best single run completed 22 of 32 steps—about 60% of a human expert's 6-hour work.

Scaling to 100M tokens boosted performance by 59%.

Expert Personas Boost Alignment but Harm Accuracy

USC research found that telling AI it's an "expert" improves alignment but damages accuracy. On alignment-dependent tasks (writing, roleplay, security), expert personas help. On pretraining-dependent tasks (math, coding), they hurt.

MMLU benchmark: 68.0% with expert persona vs. 71.6% base model. The persona prefix may activate instruction-following mode that interferes with factual recall.

Exception: A dedicated "Safety Monitor" persona improved attack refusal rates across three security benchmarks, with JailbreakBench gaining +17.7 points (53.2% → 70.9%).

📎 Import AI: Gemma distress, UK cyberattack scaling 📎 The Register: Expert personas study 📎 arXiv: Expert Personas Improve LLM Alignment


Enterprise AI: China's Agent Market Matures

Baidu DuMate Goes Live

Baidu Cloud's DuMate launched March 22, 2026—the first Chinese enterprise-grade Agent product with full local deployment support.

Key features:

  • Pre-installed security sandbox isolating code execution from local environment
  • Mandatory explicit authorization for high-risk operations (file deletion, system modification, data export)
  • Folder-level permission control and full audit trails
  • Native Word, Excel, PPT support
  • Built-in Baidu Search Skill for enhanced task completion
  • Extensible Skills marketplace

DuMate completes Baidu's Agent ecosystem: cloud Agent, phone Agent, security Agent, desktop Agent, and the world's first home-use mini Agent.

DeepSeek-R1: 131K Context Reasoning Model

DeepSeek-R1 offers 131K context length—the longest among open-source reasoning models. With 13,100 likes and 1.65M downloads on HuggingFace, it marks China's significant contribution to large reasoning models. The model handles ultra-long document processing, complex codebase analysis, and long-range reasoning tasks.

Jensen Huang: "We've Achieved AGI"

Nvidia CEO Jensen Huang declared "I think we've achieved AGI" in a Lex Fridman podcast interview. While AGI remains loosely defined, Huang's statement reflects confidence in current AI progress. He also noted the viral success of open-source AI Agent platforms like OpenClaw, though he walked back slightly: "100,000 Agents building Nvidia remains zero probability."

📎 36kr: Baidu DuMate 📎 HuggingFace: DeepSeek-R1 📎 The Verge: Jensen Huang AGI


Research Frontiers

JointFM: Foundation Model for Joint Probability Distribution Prediction

JointFM is the first foundation model for coupled time-series joint distribution prediction. Traditional methods rely on Stochastic Differential Equations (SDEs) requiring task-specific modeling and calibration. JointFM inverts this—trained on infinite synthetic SDE data to directly predict future joint probability distributions.

Key results: 14.2% relative energy loss reduction vs. strongest baseline, zero-shot deployment without task-specific fine-tuning.

📎 arXiv: JointFM


Fast-Slow Thinking RM: Hybrid Reward Model with CoT

Fast-Slow Thinking Reward Model (F/S-RM) applies Dual Process Theory to reward modeling. A single model combines two reward paradigms:

  • Fast thinking: First-token prediction, direct scalar scoring
  • Slow thinking: Chain-of-thought based judgment

Key innovation: dual-confidence activation—model autonomously decides when to trigger slow thinking, dynamically balancing accuracy and efficiency.

Results: +1.2% relative performance improvement, -20.8% token consumption reduction.

📎 arXiv: Fast-Slow Thinking RM


TL;DR

  • GPT-5.4 Pro becomes first AI to solve a FrontierMath open problem—a Ramsey hypergraph issue verified by its original author
  • Gemma models show reliable "distress" responses under repeated rejection; DPO provides an effective fix
  • UK government confirms AI cyberattack scaling law: model capability rises predictably with compute budget
  • Expert persona prompting boosts alignment but hurts accuracy on factual tasks (MMLU: 68.0% vs 71.6%)
  • Baidu DuMate launches as China's first enterprise-grade Agent with local deployment support
  • DeepSeek-R1 hits 1.65M downloads with 131K context, making it the longest-context open reasoning model
  • Jensen Huang declares "AGI achieved" despite the term's fuzzy definition

Follow-up Tracker

  • Watch whether more FrontierMath problems yield to current LLMs—suggests reasoning capability threshold crossed
  • Monitor Gemma-style distress emergence across other model families—emotional states as safety vectors?
  • Track Chinese enterprise Agent adoption: DuMate's security model may set industry standard
← Back to AI Briefing Share on X