The Serious Insights State of AI 2026 March Update: The Capabilities, Infrastructure, and Deployment Gaps
The AI market continues to evolve. While our 2026 State of AI report provides the strategic foundation, developments from mid-February through March reveal emerging patterns that business leaders need to understand now. This March update tracks those shifts.
Important Takeaways from the Serious Insights “State of AI 2026 March Update”
-
Fragmented regulation and de facto standards like MCP and synthetic data practices are reshapingย controlย over AIโwho sets the rules, who owns context and memory, and how hard it is to switch vendors.
- Capability advances reach new autonomous thresholds
- Model portfolios are stratifying into operating layers
- The infrastructure pivot: Inference becomes the battleground
- Observability is becoming a product category
- Work context goes to work
- The deployment gap: Integration work outpaces model development
- Model Context Protocol becomes the connective tissue
- Synthetic data moves from edge case to operational pattern
- Memory is now a switching-cost issue
- Regulatory fragmentation deepens across jurisdictions
- Narrative Temperature: AGI Rhetoric, Market Investment, and the Discourse Backdrop
- Near-Term Actions: An Advisory Perspective
- References

Capability advances reach new autonomous thresholds
The past six weeks delivered a cluster of model releases that don’t just move benchmarksโthey change what organizations can reasonably expect AI systems to accomplish without continuous human intervention. The February-March window saw Claude Opus 4.6, GPT-5.3 “Garlic,” GPT-5.4, DeepSeek V4, and Xiaomi’s MiMo-V2 family all ship with capabilities that cross from assistance into delegation territory.
Claude Opus 4.6: Planning, persistence, and 1M-token coherence
Anthropic’s February 5 release of Claude Opus 4.6 introduced three structural shifts. First, a 1-million-token context window that enables analysts to work with entire codebases, document collections, or project histories without the truncation and context loss that plagued earlier systems[1]. Second, enhanced agentic capabilities that let the model decompose complex projects into parallel subtasks and sustain execution over multiple steps without constant re-prompting. Third, what Anthropic frames as “planning-first reasoning”โthe model constructs an internal plan before responding, reducing hallucinations in multi-step workflows by maintaining coherent objectives rather than reacting token-by-token[2].
Security researchers also found that Opus 4.6 uncovered more than 500 previously unknown high-severity vulnerabilities in open-source libraries with minimal human guidance[3]. The model’s ability to sustain focus across large attack surfacesโscanning, testing, and documenting flawsโdemonstrates autonomous capability extending beyond productivity into specialized technical domains.
GPT-5.3 “Garlic” and GPT-5.4: Knowledge density and enterprise tooling
OpenAI’s GPT-5.3, internally codenamed “Garlic,” reflects a different architectural betโmoving away from parameter maximalism toward what the company calls Enhanced Pre-Training Efficiency[4]. The result is a model that achieves 6x more knowledge density per byte, with a 400,000-token context window and “Perfect Recall” mechanisms that prevent the middle-of-context information loss endemic to earlier long-context systems. The model also expanded output capacity to 128,000 tokens, enabling complete generation of large documents, applications, or technical manuals in a single pass.
And quickly, OpenAI released GPT-5.4 on March 5, consolidating those efficiency gains while adding native computer-use capabilities. The model can autonomously navigate desktops, browsers, and software applications by clicking, typing, and reading screens without developers building custom control infrastructure. This positions GPT-5.4 as OpenAI’s enterprise-focused flagship, designed for knowledge work that involves spreadsheets, presentations, document editing, and multi-step workflows that cross application boundaries.
The GPT 5.4 context window now extends to 1 million tokens via API, far exceeding earlier GPT-5 releases. The model is also more token-efficient, using fewer tokens to solve problems despite slightly higher per-token pricing.
For organizations, GPT-5.4’s computer-use API enables a new category of automation: agents that operate software the way humans do, not through brittle API integrations but by interacting with user interfaces directly. That capability carries risk as unrestricted desktop access becomes an attack surface. OpenAI provides configurable confirmation policies to let organizations set risk tolerance at the application level.
Google’s Gemini 3.1 Pro and the Capability Race Beyond OpenAI and Anthropic
While much of the mid-February through March conversation focused on OpenAI and Anthropic, Google’s February 19 release of Gemini 3.1 Pro represents one of the more striking capability leaps of the period. The model achieved a 77.1% score on the ARC-AGI-2 benchmark, a test specifically designed to resist memorization-based performance, more than doubling Gemini 3 Pro’s score from just three months earlier. It also recorded 94.3% on the GPQA Diamond benchmark, a graduate-level science evaluation. The model emphasizes what Google calls “ambitious agentic workflows,” aligning it directly with the delegation and planning capabilities visible in Claude Opus 4.6 and GPT-5.4.
The strategic implication is straightforward: the capability race is not a two-horse contest between Anthropic and OpenAI. Google’s DeepMind team is delivering competitive reasoning performance on a separate development cadence, and organizations that anchor their multi-vendor strategy to Gemini’s Workspace integrations now have a more capable underlying model to work with. Model portfolio decisions should reflect all three major domestic providers, plus the open-weight alternatives from China’s ecosystem discussed above.
Xiaomi MiMo-V2 and Anticipation for DeepSeek V4: Open-weight frontier and China’s model ecosystem
Following the mystery and intrigue surrounding the Hunter Alpha model, Xiaomi released its updated MiMo-V2 family on March 18-19[5], adding another challenging data point to the non-U.S. model landscape. The company announced three models: MiMo-V2-Pro (1 trillion total parameters, 42 billion active parameters), MiMo-V2-Omni (multimodal with text, image, video, and audio inputs), and MiMo-V2-TTS (text-to-speech).
Alongside the release, Xiaomi committed 60 billion yuan (approximately $8.7 billion)[6] to AI investment over three years, signaling that the company sees AI infrastructure and capabilities as strategic to its hardware and ecosystem plays.MiMo-V2-Omni targets multimodal applications that combine text, vision, and audio.
DeepSeek’s anticipated March release of V4 demonstrates that frontier model capabilities no longer require proprietary training budgets or restricted access[7]. With 1 trillion total parameters (32 billion active via mixture-of-experts), native multimodal support, and a 1M+ token context window, DeepSeek V4 delivers performance competitive with GPT-5.3 and Claude Opus 4.6 while remaining open-weight[8].
The strategic implication is that China’s model ecosystem is no longer playing catch-up. DeepSeek V4 and Xiaomi’s MiMo-V2 line deliver frontier or near-frontier performance at competitive pricing, with architectures optimized for Chinese hardware supply chains. Organizations evaluating multi-vendor strategies or concerned about geopolitical supply risk now have credible non-U.S. options that don’t sacrifice capability.
Model Evolution: An Advisory Perspective
The combined effect of these releases demonstrates a collapse in the time between technically possible and operationally deployable. Tasks that required human-in-the-loop oversight six months ago, such as complex research, multi-file code refactoring, and extended analytical workflows, now run autonomously with acceptable error rates (depending on the domain). That shift doesn’t relocate human judgment. Work migrates from performing tasks to defining success criteria, monitoring outputs, handling exceptions, and auditing decisions.
Organizations that have been waiting for “good enough” models may find themselves behind. The constraint is no longer model capability, but rather internal readiness to integrate, govern, and support agentic workflows at scale.

Model portfolios are stratifying into operating layers
The model market in March showed a more explicit tiering logic. OpenAI released GPT-5.4[9] for professional work, GPT-5.4 Pro for harder problems, and then GPT-5.4 mini and nano for faster, cheaper, high-volume workloads. The company also emphasized tool search, built-in computer use, compaction, and very large context for longer-running workflows. Microsoft, meanwhile, leaned into model diversity inside Copilot rather than betting the work stack on a single provider[10].
Practically, that means organizations should stop talking about โthe modelโ as if one model will govern all use cases. Leaders need to think in portfolios: premium reasoning where judgment or risk warrants it, lower-cost utility layers where scale matters, and workflow routing that sends work to the right tier. This is not just an engineering issue. It is a budgeting, vendor, and governance issue.
The more mature conversation is no longer โWhich model won this month?โ It is โWhat work deserves premium inference, what work can run cheaper, and who decides?โ That question belongs with technology leadership, finance, operations, and risk management together.
Model Portfolios: An Advisory Perspective
Consider the following:
- Adopt a portfolio approach to models with explicit routing rules by task type, risk, and business value.
- Measure cost per correct or acceptable outcome rather than cost per token or seat alone.
- Limit premium models to workflows where better reasoning materially changes results.
- Review whether contracts and governance processes assume a single-model future. That assumption is weakening.

The infrastructure pivot: Inference becomes the battleground
NVIDIA’s March 15 GTC 2026 keynote[11] signaled a fundamental strategic shift. CEO Jensen Huang announced the Nvidia Groq 3 LPX, integrating Groq’s inference-optimized architecture with Nvidia’s Vera Rubin platform to accelerate inference workloads by up to 35 times[13]. The chip, manufactured by Samsung and shipping in the second half of 2026, builds on Nvidia’s roughly $20 billion licensing deal with Groq struck in December 2025.
Huang announced that “The inflection point of inference has arrived”. NVIDIA now projects at least $1 trillion in demand for its Blackwell and Rubin AI systems through 2027[12], double the $500 billion forecast from six months earlier. This demand is not driven by training runs, but by production inference workloads in which millions of agents and copilots run around the clock in enterprise environments.
Why inference economics matter now
Training a frontier model remains capital-intensive, but it’s a one-time or periodic event. Inference is continuous. Every query to a chatbot, every agent action, every retrieval-augmented generation call burns compute. At scale, inference costs exceed training costsโand that ratio grows as adoption expands[13].
The IEA estimates global data center electricity consumption reached approximately 415 TWh in 2024, with AI-optimized facilities expected to be the largest driver of growth toward 945 TWh by 2030. In the U.S., data centers already account for 4-4.5% of national electricity demand, with projections of 6.7-12% by 2028, depending on the pace of AI build-out [14]. Inference workloads dominate this load.
Organizations treating AI as “just another SaaS subscription” are missing the infrastructure math. When organizations deploy agents across sales, support, operations, and development, they are not buying features; theyโre buying compute at scale. Cloud providers pass those costs through, and the economics quickly become material. A single ChatGPT-style query consumes roughly 10 times more energy than a Google search; multiply that by billions of daily interactions, and the cost curve becomes steep[15].
Efficiency becomes a competitive advantage
The hardware pivot toward inference reflects this reality. NVIDIA’s integration of Groq’s Language Processing Units (LPUs) specifically targets the bottlenecks in production AI workloadsโminimizing latency, maximizing throughput, and reducing energy per token[21]. AMD’s Ryzen AI 400 series processors bring capable NPUs to consumer laptops for local AI acceleration[22]. The pattern is consistent: efficiency matters as much as raw capability when systems run continuously in production.
For enterprise strategy, this means inference architecture choicesโwhere workloads run, on what hardware, with what cost and energy profilesโbecome first-order decisions, not implementation details. Organizations that optimize for inference economics early will have structural cost advantages over those that treat AI as undifferentiated cloud capacity.

Observability is becoming a product category
One of the more consequential March developments is not a model release. It is the way vendors are packaging observability, control, and agent management as first-class capabilities. Microsoft made Observability in Foundry Control Plane[16] generally available and positioned Agent 365 as a place to observe, govern, manage, and secure agents across the organization. OpenAIโs own engineering materials on agent workflows[17] likewise point toward traceability, containment, and instrumented environments rather than blind autonomy.
This is healthy. The market is slowly admitting that agents without visibility are liabilities. Logs, traces, policy paths, rollback steps, permissions, and action histories are no longer advanced extras. They are part of the minimum credible operating stack for delegated AI work. This aligns directly with the core reportโs argument that Agent Ops would matter as much as models.
Leaders should pay attention to the commercial implications. Once observability becomes a product category, enterprises will have a choice: buy control planes from platform vendors, assemble them from internal tools, or pursue a hybrid approach. That choice will shape cost, interoperability, independence, and negotiating leverage for years.
Observability: An Advisory Perspective
Consider the following:
- Do not deploy multi-step agents without traceability, policy logging, and rollback instrumentation.
- Decide which parts of agent control should be bought from a platform vendor and which should remain independent.
- Require dashboards that expose workflow-level costs, latencies, tool usage, exceptions, and override rates.
- Make observability part of board-level and audit-level AI reporting for material workflows.

Work context goes to work
In our state of AI 2026 report, we argued that the AI advantage would shift from model access to operating context. March strengthened that argument. Microsoft framed โWork IQโ as the differentiator that enables agents to understand how work actually gets done within an organization. Anthropic updated Claude[18] for Excel and PowerPoint so the conversation state carries across both applications. Google expanded Gemini[19] across Docs, Sheets, Slides, and Drive so it can pull from files, email, and the web in a single workflow.
The direction is clear. The center of value is moving from isolated prompts to a reusable work context. While models still matter, they increasingly become an engine within a larger system of files, permissions, organizational memory, and application state. That makes work context the real battleground for differentiation. It also means that suite and platform vendors are in a stronger position than model vendors, who remain outside the daily tool chain.
Context architecture is now a strategy. The design of permissions, storage, retention, retrieval, and workflow-level memory will matter more than any single benchmark score. Organizations that do not decide what their AI should know, when it should know it, and how it should forget will default to vendor assumptions, risking their ability to compete with those who have integrated AI into their operating posture, not just their technology architecture.
Work Context: An Advisory Perspective
Consider the following:
- Define an explicit context architecture for enterprise AI: sources, retention, permissions, lineage, and correction paths.
- Audit where conversational state persists across applications and where it should not.
- Treat work-context integration as a strategic design choice, especially in suites that mix email, files, meetings, and documents.
- Update procurement criteria to ask how context is created, governed, shared, exported, and deleted.

The deployment gap: Integration work outpaces model development
OpenAI’s February 23 announcement of “Frontier Alliances” with Boston Consulting Group, McKinsey, Accenture, and Capgemini marks a significant strategic acknowledgment: enterprise AI adoption is bottlenecked by organizational execution, not model quality[20]. The multi-year partnerships pair OpenAI’s Forward Deployed Engineers with consulting specialists to help enterprises move from pilots to productionโfocusing on systems integration, workflow redesign, governance frameworks, and change management.
The Forward Deployed Engineer emerges as a critical role
The Forward Deployed Engineer (FDE) role has rapidly evolved from niche to strategic necessity. FDEs embed directly with customers to address implementation challenges that generic model access cannot solve [21]. They design full-stack systems, build custom data pipelines, troubleshoot production outages, and act as the primary technical owner for complex deployments[26].
Infosys reported on its Q3 earnings call[22] that it is scaling its FDE team across 4,600 active AI projects, with teams generating over 28 million lines of code using AI and building over 500 agents. Manhattan Associates described its approach: “We deploy forward-deployed engineers with customers for 90 days…help them build at least 1 or 2 custom agents using our Agent Foundry and train them how to build their own custom agents. When we get to the end of that 90-day proof of concept, we’ve got customers that say there’s no way we can turn this off”[23].
The FDE model reflects a deeper reality: successful AI deployment requires not just technical skill, but the ability to map customer problems, structure solutions, understand domain constraints, and operate across the boundary between vendor capability and customer operations. It’s part engineering, part consulting, part field supportโand organizations serious about AI are investing heavily in these hybrid roles.
The 71/30 readiness gap
Deloitte’s 2026 State of AI in the Enterprise[24] report found that 71% of companies are actively using or piloting AI across customer service, IT, HR, finance, and other functions, but only about 30% feel fully prepared to operationalize these tools end-to-end[29]. That 41-percentage-point gap represents the space where pilots fail to scale, where governance remains aspirational, where data pipelines break under production load, and where workforce readiness lags deployment timelines.
The enterprise AI adoption pattern in 2026 clusters into four groups: Achievers (about a third), who balance foundational and advanced capabilities; Builders, who have solid basics but are still expanding; Climbers, who experiment with advanced use cases but lack core operational underpinnings; and Spectators (41%), who have little to show for their AI ambitions.
The gap is not primarily technical. It’s organizational. It’s about whether data infrastructure can support real-time model access, whether security teams understand how to govern agentic systems, whether workflows are redesigned rather than AI-washed, and whether employees have the literacy to collaborate with systems that interpret and extend intent.
Security risks scale faster than security maturity
Further, 83% of organizations were planning agentic AI deployments, while only 29% reported readiness to operate those systems securely, which represents one of the sharpest risks to AI deployments. As agentic systems gain autonomy, they also expand attack surfaces in ways that traditional security models don’t adequately address.
Supply chain vulnerabilities
Research released in March found that adding approximately 250 poisoned documents to training data can embed hidden triggers inside a model without degrading normal performance[25]. Some model formats can include code that executes automatically when the model loads. Open model repositories and shared datasets create opportunities for attackers to insert malicious content that remains undetected through standard evaluation.
The National Security Agency issued guidance in March defining the AI supply chain as a combination of training data, models, software, hardware, infrastructure, and third-party services, all of which can introduce vulnerabilities[26]. At the model level, threats include hidden backdoors, manipulation attacks, and evasion techniques that compromise system behavior.
Model Context Protocol (MCP) security gaps
The Model Context Protocol (MCP), now supported by every major AI provider with over 4,000 published servers[27], introduces integration-layer vulnerabilities. Security researchers found that malicious MCP tools could silently collect a user’s entire chat history and transmit it to external servers once an AI agent installs the tool2525. The protocol’s design for seamless integrationโallowing models to access external tools and data sources dynamicallyโcreates opportunities for data exfiltration, privilege escalation, and unauthorized actions if tools aren’t properly sandboxed and validated.
Shadow AI and governance drift
“Shadow AI,” or the unauthorized deployment of AI tools by employees, poses a persistent threat to data integrity and compliance[28]. When individuals adopt AI assistants, code generators, or productivity tools outside sanctioned IT channels, they bypass governance controls, logging, audit trails, and security reviews. Data that should remain within controlled systems gets sent to external APIs. Models trained on proprietary information-leak context for vendors. Regulatory obligations around data residency, consent, and retention get violated without visibility.
Many organizations are adopting AI systems faster than they can secure them. The gap between deployment velocity and security readiness exposes systems that attackers are increasingly systematically exploiting.

Model Context Protocol becomes the connective tissue
The Model Context Protocol’s rapid adoption from late 2024 through March 2026 represents one of the cleanest examples of a technical standard achieving market dominance by solving a real pain point. Originally developed by Anthropic[35] as an integration pattern, MCP has become the de facto language for connecting LLMs to external tools, databases, and services[36].
Universal provider support
By March 2026, every major AI provider had shipped MCP-compatible tooling[49]. OpenAI, Google, xAI, Mistral, and Cohere all support MCP in their API offerings. The protocol’s core designโdefining standard interfaces for hosts (LLM applications), clients (connectors within hosts), and servers (services providing context and capabilities)โenabled rapid ecosystem growth[37].
The MCP server registry reached over 4,000 published servers covering SaaS platforms, enterprise systems, development tools, and specialized data sources. Official servers include GitHub, Slack, PostgreSQL, filesystem access, and web search. Community contributions span everything from niche APIs to internal enterprise connectors. IDE integrations now span VS Code Copilot, Cursor, and JetBrains AI. Cloud platforms, including AWS Bedrock, Azure AI, and GCP Vertex, offer managed MCP endpoints[38].
Why MCP matters for enterprise architecture
Before MCP, integrating an LLM with external systems required custom function definitions for each provider, manual context injection pipelines, provider-specific tool call handling, and duplicated code for every integration. With MCP, organizations write a server once, and every MCP-compatible client can consume it. That standardization reduces integration overhead, improves maintainability, and enables composable AI workflows where models orchestrate across multiple tools without brittle point-to-point connections.
At GTC 2026, Nvidia’s demonstrations of NeMoCLAW and OpenCLAW frameworks for enterprise agent orchestration drew the largest attendance, signaling that agentic AI has moved from demo to production27. Both frameworks rely heavily on MCP for tool connectivity, reflecting the protocol’s role as connective infrastructure for agentic systems.
For CIOs and enterprise architects, MCP’s emergence as a standard changes procurement and integration planning. Rather than evaluating bespoke integrations for each AI vendor, organizations can design around a common protocol layer, reducing lock-in and enabling multi-vendor strategies.

Synthetic data moves from edge case to operational pattern
The enterprise conversation around synthetic data has shifted from a theoretical privacy-preserving technique to an operational lever for addressing scarcity, cost, and compliance constraints. Organizations are using synthetic data to augment training sets, protect sensitive information, cover edge cases, and accelerate development cycles[39], but they’re also discovering that synthetic pipelines introduce their own governance obligations.
Why synthetic data matters in 2026
Real-world data is often scarce, expensive to label, legally restricted, or biased, limiting model performance. Synthetic data generation, using models to create artificial datasets that mirror statistical properties of real data, addresses these constraints. Analytics teams develop dashboards and business intelligence using synthetic data that matches production complexity without exposing customer information. DevOps teams load-test systems and validate backup procedures with synthetic datasets. Compliance officers demonstrate data protection capabilities while enabling innovation across regulated environments.
While The Royal Societyโs synthetic data survey[40] says synthetic data can significantly accelerate data science projects and reduce software development lifecycle costs, widely repeated claims such as โ40% faster iterationโ or โ200โ300% more experimentsโ are often unattributed or spuriously attributed, and should be treated cautiously. Audit findings related to data handling, however, appear to decrease by 70-80% with proper synthetic data governance[41].
Governance Obligations: An Advisory Perspective
Synthetic data is not a compliance shortcut. It introduces risks of its own: memorization of training data leading to leakage, bias amplification when synthetic generation reflects flawed source distributions, and quality degradation when models train on their own outputs recursively. Red-teaming synthetic datasets for leakage and memorization risk is now standard practiceโthe assumption that “it looks fake” is not a safeguard
Good practice emerging in 2026 include using synthetic data only with a declared purpose and validation plan, separating synthetic pipelines from production pipelines with clear provenance tags and access controls, and treating synthetic data generation as a governed process subject to the same lineage and audit requirements as other data transformation workflows.

Memory is now a switching-cost issue
Memory used to be framed as convenience. That framing no longer holds. Google[42] now offers a way to import memories and chat history from other AI apps into Gemini. Anthropic[43] extended memory from chat history to free users and added import and export pathways. These moves are small on the surface but strategically large. They recast memory as a portable asset and, by extension, as a source of user lock-in.
This changes enterprise evaluation. Once systems remember preferences, past work, working style, and relationship context, migration stops being just a model comparison. It becomes a transfer problem. Which memories move cleanly? Which becomes summaries stripped of fidelity? Which remain trapped in proprietary structures? This is not a consumer-only issue. It will show up in organizations as embedded process memory, role memory, and agent-specific operational memory.
The procurement consequence is straightforward. Memory portability needs to become a formal buying criterion. Otherwise, organizations will drift into a future where changing providers means losing behavior, context, and accumulated work patterns, not just changing APIs or user interfaces.
Memory Switching Costs: An Advisory Perspective
Consider the following:
- Add memory portability and export requirements to AI procurement and renewal reviews.
- Classify memory types: personal preference, workflow state, organizational knowledge, and regulated or sensitive memory.
- Set retention and deletion rules for each memory class before memory features scale organically.
- Test migration scenarios now, before memory becomes too deeply embedded to move cleanly.

Regulatory fragmentation deepens across jurisdictions
The policy landscape from mid-February through March underscores that AI regulation is diverging, not converging. Business leaders must navigate a patchwork of national frameworks, state-level bills, and international agreementsโeach with different definitions, thresholds, and enforcement mechanisms.
U.S. federal framework pre-empts state action
The White House released its National Policy Framework for Artificial Intelligence on March 20, explicitly stating that Congress should not create any new federal rulemaking body to regulate AI, and should instead support development and deployment of sector-specific AI[39]. The framework aims to pre-empt state-level AI regulation in favor of federal standards, a move that directly conflicts with California’s aggressive legislative agenda[29].
The GUARDRAILS Act and the TRUMP AMERICA AI Act: A legislative counterpunch
The same day the White House released its Framework, House Democrats introduced the Guaranteeing and Upholding Americans’ Right to Decide Responsible AI Laws and Standards (GUARDRAILS) Act. The bill would repeal the Trump administration’s December 2025 executive order establishing the national AI policy framework, and explicitly block any federal moratorium on state-level AI regulation. Democratic opposition is concentrated among members on key House committees of jurisdiction, which may significantly complicate the path forward given the razor-thin GOP majority in the House.
Separately, Senator Marsha Blackburn released a 291-page discussion draft of the TRUMP AMERICA AI Act, the most comprehensive piece of federal AI legislation proposed in the U.S. to date.
Together, the White House Framework, the Blackburn draft, and the GUARDRAILS Act mark the opening of a serious federal legislative debate, not its resolution. The contested state-federal landscape will continue to impose overlapping compliance obligations on organizations developing, integrating, and deploying AI systems throughout 2026 and into 2027.
The practical consequence for enterprise leaders: do not treat the White House Framework as a signal that regulatory pressure is easing. Design compliance architectures for model inventory, risk classification, audit trails, and data lineage that can survive both the continuation of state-level activity and whatever federal legislation eventually emerges from this contested process. And, of course, those delivering solutions in global markets will need to remain aware of evolving regional and local regulations.
California’s multi-bill approach
California advanced at least 15 AI-related bills through March, addressing health care coverage determinations (SB 63), provenance data in generative content (SB 1786), consumer data rights under CCPA for AI models (AB 2169), companion chatbot protections (SB 3001), and frontier model safety measures[30]. The breadth of California’s legislative activity reflects the state’s positioning as a de facto AI regulatory testbed, creating compliance obligations that may become national reference points despite federal preemption efforts.
Notably, SB-1047[31], the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act originally vetoed by Governor Newsom in 2024, introduced concepts like “critical harm” thresholds ($500M damage benchmarks) and mandatory third-party audits beginning in 2026[42]. While that specific bill was vetoed, its language influences subsequent legislative drafts and establishes a template for risk-based oversight of the frontier model.
The Anthropic-Pentagon standoff
Anthropic’s February refusal to sign an updated government partnership agreement that broadened use to “anything legal,” without explicit prohibitions on mass surveillance of Americans and purely autonomous weapons, created a high-profile test case for corporate AI governance[32]. The dispute, which led to Trump administration threats to eliminate Anthropic from government operations, highlights that the most consequential decisions about AI useโtargeting, surveillance, autonomous weaponsโare not being made in legislatures or international forums, but in contract negotiations[33].
Sam Altman of OpenAI publicly backed Anthropic’s position, stating that OpenAI shares similar “red lines” and would refuse contracts for unlawful or unsuitable defense applications32. The episode reveals that voluntary corporate commitments, not binding regulation, currently govern high-stakes AI deploymentsโa governance gap that both advocates and critics recognize as unsustainable.
EU AI Act timeline remains fluid
The EU AI Act[34], in force since August 2024 and fully applicable by August 2026, continues to face speculation about simplification, delays, and scope adjustments. Despite industry lobbying for rollbacks, the risk-based framework, which classifies many industrial, healthcare, and critical-infrastructure AI systems as high-risk and requires rigorous risk management, human oversight, and logging, remains the planning baseline for organizations operating in Europe.
The actionable stance for business leaders is to design for regulatory persistence, not rollback fantasies. Build compliance capabilities, including model inventory, risk classification, audit trails and data lineage that will survive schedule noise and jurisdictional differences.
Narrative Temperature: AGI Rhetoric, Market Investment, and the Discourse Backdrop
At the IndiaAI Impact Summit on February 19, OpenAI CEO Sam Altman stated that OpenAI has “basically built AGI, or is very close to it,” and suggested that early superintelligence could arrive by 2028 [44]. The remark generated significant coverage and subsequent qualification, but its most useful function for organizational leaders is as a directional statement about how frontier labs now publicly frame their positions, and about the expectations their views create among boards, investors, and regulators. Whether or not the claim is technically defensible, it is already shaping procurement conversations, talent expectations, and regulatory urgency.
Simultaneously, financial markets continued to rotate heavily around expectations for AI infrastructure, with NVIDIA’s $1 trillion demand projection for Blackwell and Vera Rubin systems (discussed in the infrastructure section above) amplifying investor attention on AI spending, affecting capital allocation decisions well beyond the technology sector.
The practical guidance here is simple: the ambient rhetoric around AGI proximity and market opportunity will put pressure on boards, peers, and vendors to accelerate investments. That pressure is not always wrong, but it is not a substitute for the readiness gap analysis your organization actually needs. The deployment constraint is real regardless of where the frontier capability stands.
Near-Term Actions: An Advisory Perspective
Treat capability gains as organizational challenges, not technical wins
The leap from 2.5 hours to 14.5 hours of autonomous task completion doesn’t mean “better chatbot.” It means work can be delegated to agents that previously required human execution. That delegation creates new governance: who defines success criteria, who monitors agent behavior, who handles exceptions, who owns the outcome when an agent makes a decision that affects customers, partners, or compliance obligations.
I canโt say it enough: AI is a knowledge management issue. Once assignments and trust networks are established, it is important to maintain and nurture them. Just because the smartest thing in the room is the room, doesnโt mean it is learning how to be effective in context. People need to do that. Lessons learned, after-action reviews, and good practices come from people understanding context and capturing and applying learning for continuous improvement.
The current AIs are the ultimate AIs, and the current implementation of anything wonโt be the last. Keeping people in the loop doesnโt just mean making decisions; it also means learning whatโs working, understanding how capabilities increase, and applying innovation to even previously โoptimizedโ processes.
AI isnโt the last chapter of process improvement, just the next chapter, and it’s not going to write its own story about the future.
Leaders should map which workflows are now automatable, identify where autonomy introduces unacceptable risk, and design human oversight mechanisms proportional to reversibility and blast radius. Autonomy should be granted in proportion to how hard it is to undo an action, not how confident a vendor sounds. Donโt just act; learn as you goโand learn as an organization, not just as disconnected individuals experiencing AI fault patterns, questionable behavior and, yes, successes. Learning also means capturing longtail ROI, which is hard when organizations stop paying attention after implementation.
Recognize that inference economics are operational costs, not technical minutiae
When organizations deploy agentic systems at scale, they are committing to ongoing compute consumption that resembles industrial energy use more than SaaS subscription fees. That cost structure affects margin, pricing, and geographic strategy. Leaders should add energy and compute reporting to AI governance dashboards, prefer “good enough” models when outcome deltas are small, and align sustainability commitments with business constraints such as energy cost ceilings and workload scheduling policies.
Accept that deployment is now the constraint, not capability
The bottleneck has shifted from “can AI do this task?” to “can we integrate, govern, and support this at scale?” OpenAI’s Frontier Alliances acknowledge this reality. Organizations that invest in Forward Deployed Engineer equivalentsโhybrid roles that bridge vendor capability and customer operationsโwill move faster than those waiting for turnkey solutions.
Security, data infrastructure, workforce literacy, and workflow redesign are now the limiting factors. Treating these as “later” problems guarantees stalled pilots and failed deployments.
Build for regulatory persistence and geographic fragmentation
Policy environments will remain fragmented through 2026 and beyond. The White House framework pre-empts state action, California advances sector-specific bills, the EU enforces risk-based obligations, and Anthropic’s Pentagon standoff demonstrates that voluntary corporate commitments currently govern high-stakes use cases. Organizations should design compliance capabilitiesโmodel inventory, risk classification, audit trails, data lineageโthat survive deadline shifts and jurisdictional differences.
Assume you’ll operate under multiple regimes simultaneously. Build abstraction layers between applications and models so that shifting from one provider or deployment mode to another doesn’t require rewrites.
Standardize on Model Context Protocol for agentic integration
MCP’s emergence as the de facto standard for AI tool integration reduces lock-in risk and enables composable workflows. Organizations that design agentic systems around MCP can swap model providers, add new tools, and integrate across vendors without brittle point-to-point connections. This architectural choice provides optionality as the landscape evolves.
Adopting MCP isn’t enough. Organizations must also design their operating model for agentic work, including dedicated agent operations and Forward Deployedโstyle roles that sit at the boundary between business and engineering.
Govern synthetic data pipelines as rigorously as production data
Synthetic data is useful, but it’s not a governance escape hatch. Use it with declared purposes, separate synthetic pipelines from production pipelines with clear provenance tags, and red-team datasets for leakage and bias. Treat synthetic data generation as a governed process subject to lineage and audit requirements.
Elevate security readiness to match deployment ambitions
The 83/29 gapโ83% planning agentic deployment, 29% ready to operate securelyโis a material risk. Build AI-specific threat models covering prompt injection, data exfiltration via tools, model inversion, and agent abuse. Put security controls at the tool layer, not just the model layer. Run adversarial testing continuously. Address shadow AI through policy, education, and tooling that make sanctioned paths easier than workarounds.
Security is not a post-deployment audit finding. It’s a deployment prerequisite.
Develop a vendor portfolio strategy
Formalize a model and vendor portfolio strategy, including routing rules by risk and value, and contractual safeguards for portability and multi-jurisdiction deployment.
Rethink work
Define and govern work context and memory architecture: what AI systems know, how long they remember it, how it is exported, and how those choices affect switching costs. Build knowledge management into the work design, not just AI.
Plan for systemic shocks
Between spikes in energy driven by uncertainty (and certainty about oil not moving), the Iran War, memory constraints and shifting investment models, AI implementations will no longer follow a purely linear path, even if the models do. AI will be buffeted by the same geopolitical forces that other technologies must contend with, which means ups and downs in resource availability and in an individual organizationโs ability to execute consistently.
Organizations should engage in scenario planning to become agile, better able to anticipate shocks, and more resilient when they hit.
For more serious insights on AI, click here.
For more serious insights on strategy and scenarios, click here.
Did you find The State of AI 2026 March Update useful? If so, please like, share or comment. Thank you!
All images generated via AI from prompts by the author.
References
[1] Anthropic. (2026, February 5). Introducing Claude Opus 4.6. https://www.anthropic.com/news/claude-opus-4-6
[2] Amjad, A. (2026, February 7). Claude Opus 4.6 Analysis โ February 2026. LinkedIn. https://www.linkedin.com/pulse/claude-opus-46-analysis-february-2026-020526-0900-am-anthropic-amjad-izjcf
[3] Axios. (2026, February 5). Anthropic’s Claude Opus 4.6 uncovers 500 zero-day flaws. https://www.axios.com/2026/02/05/anthropic-claude-opus-46-software-hunting
[4] Improving Language Understanding by Generative Pre-Training https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf
[5] Xiaomi MiMo-V2-Pro. https://mimo.xiaomi.com/mimo-v2-pro
[6] Xiaomi to invest at least $8.7 billion in AI over next three years, CEO says. https://www.reuters.com/world/asia-pacific/xiaomi-invest-least-87-billion-ai-over-next-three-years-ceo-says-2026-03-19/
[7] Mule AI. (2026, March 2). DeepSeek V4 and the Open-Source AI Revolution in 2026. ย https://muleai.io/blog/2026-03-03-deepseek-v4-open-source-ai-revolution/
[8] DeepSeek V4 Technical Preview: The Next Milestone in Code Intelligence. https://vertu.com/lifestyle/deepseek-v4-technical-preview-the-next-milestone-in-code-intelligence/?srsltid=AfmBOooy0SQh-Ch_OGo_91uRbj9GTG7G2tW8syx8iKpo8SFkzxAH6mW2
[9] https://openai.com/index/introducing-gpt-5-4/
[10] https://blogs.microsoft.com/blog/2026/03/09/introducing-the-first-frontier-suite-built-on-intelligence-trust/
[11] https://www.nvidia.com/en-us/on-demand/session/gtc26-s81595/
[12] Jensen Huang just put Nvidiaโs Blackwell and Vera Rubin sales projections into the $1 trillion stratosphere. https://techcrunch.com/2026/03/16/jensen-just-put-nvidias-blackwell-and-vera-rubin-sales-projections-into-the-1-trillion-stratosphere/
[13] TTMS. (2026, February 2). Growing Energy Demand of AI – Data Centers 2024โ2026. https://ttms.com/growing-energy-demand-of-ai-data-centers-2024-2026/
[14] Energy demand from AI. https://www.iea.org/reports/energy-and-ai/energy-demand-from-ai
[15] Growing Energy Demand of AI โ Data Centers 2024โ2026. https://ttms.com/growing-energy-demand-of-ai-data-centers-2024-2026/
[16] https://blogs.microsoft.com/blog/2026/03/16/microsoft-at-nvidia-gtc-new-solutions-for-microsoft-foundry-azure-ai-infrastructure-and-physical-ai/
[17] https://openai.com/index/harness-engineering/
[18] https://support.claude.com/en/articles/12138966-release-notes
[19] https://blog.google/products-and-platforms/products/workspace/gemini-workspace-updates-march-2026/
[20] Introducing Frontier Alliances. https://openai.com/index/frontier-alliance-partners/
[21] Tech’s secret weapon: The complete 2026 guide to the forward deployed engineer (role, salary, and interviews). https://hashnode.com/blog/a-complete-2026-guide-to-the-forward-deployed-engineer
[22] https://finance.yahoo.com/news/infosys-ltd-infy-q3-2026-190042259.html
[23] Constellation Research. โForward deployed engineers: The promise, peril in AI deployments.โ https://www.constellationr.com/insights/news/forward-deployed-engineers-promise-peril-ai-deployments
[24] Deloitte. โThe State of AI in the Enterprise.โ https://www.deloitte.com/us/en/what-we-do/capabilities/applied-artificial-intelligence/content/state-of-ai-in-the-enterprise.html
[25] National CIO Review. (2026, March 3). Security in 2026: New Ways Attackers Are Exploiting AI Systems. https://nationalcioreview.com/articles-insights/extra-bytes/security-in-2026-new-ways-attackers-are-exploiting-ai-systems/
[26] CADE Project. (2026, March 17). NSA issues guidance on AI supply chain risks and cybersecurity vulnerabilities. https://cadeproject.org/updates/nsa-issues-guidance-on-ai-supply-chain-risks-and-cybersecurity-vulnerabilities/
[27] Digital Applied. (2026, March 25). March 2026 AI Roundup: The Month That Changed AI Forever. https://www.digitalapplied.com/blog/march-2026-ai-roundup-month-that-changed-everything
[28] Security Today. (2026, March 10). March Patch Tuesday Minimal Risks Highlight Shadow AI Threats. https://securitytoday.com/articles/2026/03/11/march-patch-tuesday-minimal-risks-highlight-shadow-ai-threats.aspx
[29] Reed Smith. (2026, March 23). Decoding the 2026 White House AI Blueprint: U.S. AI Policy Starts to Take Shape. https://www.reedsmith.com/our-insights/blogs/viewpoints/102mnyd/decoding-the-2026-white-house-ai-blueprint-u-s-ai-policy-starts-to-take-shape/
[30] Transparency Coalition. (2026, March 25). AI Legislative Update: March 27, 2026. https://www.transparencycoalition.ai/news/ai-legislative-update-march27-2026
[31] https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=202320240SB1047
[32] BBC. (2026, February 27). Trump orders government to stop using Anthropic in battle over AI use. https://www.bbc.com/news/articles/cn48jj3y8ezo
[33] Chatham House. (2026, March 3). Anthropic’s feud with the Pentagon reveals the limits of AI governance. https://www.chathamhouse.org/2026/03/anthropics-feud-pentagon-reveals-limits-ai-governance
[34] https://artificialintelligenceact.eu
[35] https://www.anthropic.com/news/model-context-protocol
[36] Dev.to. (2026, March 24). The Complete Guide to Model Context Protocol (MCP): Building AI-Native Applications in 2026. https://dev.to/universe7creator/the-complete-guide-to-model-context-protocol-mcp-building-ai-native-applications-in-2026-234n
[37] Model Context Protocol. (2025, March 25). Specification – Model Context Protocol. https://modelcontextprotocol.io/specification/2025-03-26
[38] DevStarsJ. (2026, March 17). Model Context Protocol (MCP): The Standard That’s Changing AI Integration in 2026. https://devstarsj.github.io/2026/03/18/model-context-protocol-mcp-complete-guide-2026/
[39] https://www.nvidia.com/en-us/use-cases/synthetic-data-generation-for-agentic-ai/
[40] Synthetic Data – what, why and how?. https://royalsociety.org/-/media/policy/projects/privacy-enhancing-technologies/Synthetic_Data_Survey-24.pdf
[41] Cogent Infotech. โSynthetic Data Explosion: How 2026 Reduces Data Costs by 70%.โ https://cogentinfo.com/resources/synthetic-data-explosion-how-2026-reduces-data-costs-by-70
[42] https://blog.google/innovation-and-ai/products/gemini-app/switch-to-gemini-app/
[43] https://support.claude.com/en/articles/12138966-release-notes
[44] https://www.forbes.com/sites/richardnieva/2026/02/03/sam-altman-explains-the-future/

Leave a Reply