Category: AI News

  • Anthropic’s Claude Can Now Control Your Mac: The Agent Era Just Got Real

    Anthropic’s Claude Can Now Control Your Mac: The Agent Era Just Got Real

    Anthropic has just made the abstract concept of AI agents viscerally concrete. On Monday, the company launched the ability for Claude to directly control a user’s Mac — clicking buttons, opening applications, typing into fields, and navigating software on the user’s behalf while they step away from their desk.

    Available immediately as a research preview for paying subscribers (Claude Pro at /month or Max at -/month), the feature transforms Claude from a conversational assistant into something closer to a remote digital operator. And it’s available inside both Claude Cowork — the company’s agentic productivity tool — and Claude Code, its developer-focused command-line agent.

    How Computer Use Actually Works

    The system operates through a layered priority hierarchy that reveals how Anthropic is thinking about reliability versus reach. When a user assigns Claude a task, it first checks whether a direct connector exists — integrations with Gmail, Google Drive, Slack, or Google Calendar. These connectors are the fastest and most reliable path. If no connector is available, Claude falls back to navigating Chrome via Anthropic’s browser extension. Only as a last resort does Claude interact directly with the user’s screen — clicking, typing, scrolling, and opening applications the way a human operator would.

    This hierarchy matters. As Anthropic’s documentation explains, pulling messages through your Slack connection takes seconds, but navigating Slack through your screen takes much longer and is more error-prone. Screen-level interaction is the most flexible mode — it can theoretically work with any application — but it’s also the slowest and most fragile.

    Dispatch: Your iPhone as a Remote Control

    The strategic play may not be computer use itself, but how Anthropic is pairing it with Dispatch — a feature that launched last week and now extends to Claude Code. Dispatch creates a persistent conversation between Claude on your phone and Claude on your desktop. A user pairs their mobile device with their Mac by scanning a QR code, and from that point forward, they can text Claude instructions from anywhere. Claude executes those instructions on the desktop — which must remain awake — and sends back the results.

    The use cases Anthropic envisions range from mundane to ambitious: having Claude check your email every morning, pull weekly metrics into a report, organize a cluttered Downloads folder, or compile a competitive analysis from local files into a formatted document. Scheduled tasks allow users to set a cadence once — every Friday, every morning — and let Claude handle the rest without further prompting.

    The Production Reality

    Anthropic is calling this a research preview for a reason. Early hands-on testing suggests the feature works well for information retrieval and summarization but struggles with more complex, multi-step workflows that require interacting with multiple applications. Screen-level interaction is inherently fragile — anything that changes the UI can derail a task. And the fact that Claude takes screenshots of your desktop to navigate raises obvious privacy considerations, even with Anthropic’s documented guardrails.

    But the trajectory is clear. One early user on social media put it well: combine this with scheduling and you’ve basically got a background worker that can interact with any app on a cron job. That’s not an AI assistant anymore, that’s infrastructure.

    The competition is heating up accordingly. Reuters reported that OpenAI is actively courting private equity firms in what it described as an enterprise turf war with Anthropic — a battle in which the ability to ship working agents is becoming the decisive weapon. With Claude now physically capable of operating your desktop, Anthropic has fired a significant shot.

  • Beyond LLMs: The Three Architectural Approaches Teaching AI to Understand Physics

    Beyond LLMs: The Three Architectural Approaches Teaching AI to Understand Physics

    Large language models excel at writing poetry and debugging code, but ask them to predict what happens when you drop a ball and you’ll quickly discover their limitations. Despite mastering chess, generating art, and passing bar exams, today’s most powerful AI systems fundamentally don’t understand physics.

    This gap is becoming increasingly apparent as companies try to deploy AI in robotics, autonomous vehicles, and manufacturing. The solution? World models鈥攊nternal simulators that let AI systems safely test hypotheses before taking physical action. And investors are paying attention: AMI Labs raised a billion-dollar seed round, while World Labs secured funding from backers including Nvidia and AMD.

    The Problem with Next-Token Prediction

    LLMs work by predicting the next token in a sequence. This approach has been remarkably successful for text, but it has a critical flaw when applied to physical tasks. These models cannot reliably predict the physical consequences of real-world actions, according to AI researchers.

    Turing Award recipient Richard Sutton warned that LLMs just mimic what people say instead of modeling the world, which limits their capacity to learn from experience. DeepMind CEO Demis Hassabis calls this jagged intelligence鈥擜I that can solve complex math olympiads but fails at basic physics.

    The industry is responding with three distinct architectural approaches, each with different tradeoffs.

    1. JEPA: Learning Abstract Representations

    The Joint Embedding Predictive Architecture, endorsed by AMI Labs and pioneered by Yann LeCun, takes a fundamentally different approach. Instead of trying to predict what the next video frame will look like at the pixel level, JEPA models learn a smaller set of abstract, or latent, features.

    Think about how humans actually observe the world. When you watch a car driving down a street, you track its trajectory and speed鈥攜ou don’t calculate the exact reflection of light on every leaf in the background. JEPA models reproduce this cognitive shortcut.

    The benefits are substantial: JEPA models are highly compute and memory efficient, require fewer training examples, and run with significantly lower latency. These characteristics make it suitable for applications where real-time inference is non-negotiable鈥攔obotics, self-driving cars, high-stakes enterprise workflows.

    Healthcare company Nabla is already using this architecture to simulate operational complexity in fast-paced medical settings, reducing cognitive load for healthcare workers.

    2. Gaussian Splats: Building Spatial Worlds

    The second approach, adopted by World Labs led by AI pioneer Fei-Fei Li, uses generative models to build complete 3D spatial environments. The process takes an initial prompt鈥攅ither an image or textual description鈥攁nd uses a generative model to create a 3D Gaussian splat.

    A Gaussian splat represents 3D scenes using millions of tiny mathematical particles that define geometry and lighting. Unlike flat video generation, these 3D representations can be imported directly into standard physics and 3D engines like Unreal Engine, where users and AI agents can freely navigate and interact from any angle.

    World Labs founder Fei-Fei Li describes LLMs as wordsmiths in the dark鈥攑ossessing flowery language but lacking spatial intelligence and physical experience. The company’s Marble model aims to give AI that missing spatial awareness.

    Industrial design giant Autodesk has backed World Labs heavily, planning to integrate these models into their design applications. The approach has massive potential for spatial computing, interactive entertainment, and building training environments for robotics.

    3. End-to-End Generation: Physics Native

    The third approach uses an end-to-end generative model that continuously generates the scene, physical dynamics, and reactions on the fly. Rather than exporting to an external physics engine, the model itself acts as the engine.

    DeepMind’s Genie 3 and Nvidia’s Cosmos fall into this category. These models ingest an initial prompt alongside continuous user actions and generate subsequent environment frames in real-time, calculating physics, lighting, and object reactions natively.

    The compute cost is substantial鈥攃ontinuously rendering physics and pixels simultaneously requires significant resources. But the investment enables synthetic data factories that can generate infinite interactive experiences and massive volumes of synthetic training data.

    Nvidia Cosmos uses this architecture to scale synthetic data and physical AI reasoning. Waymo built its world model on Genie 3 for training self-driving cars, synthesizing rare, dangerous edge-case conditions without the cost or risk of physical testing.

    The Hybrid Future

    LLMs will continue serving as the reasoning and communication interface, but world models are positioning themselves as foundational infrastructure for physical and spatial data pipelines. We’re already seeing hybrid architectures emerge.

    Cybersecurity startup DeepTempo recently developed LogLM, integrating LLMs with JEPA elements to detect anomalies and cyber threats from security logs. The boundary between AI that thinks and AI that understands the physical world is beginning to dissolve.

    As world models mature, expect AI systems that can not only tell you how to change a tire, but actually understand what happens when you apply torque to a rusted bolt. The physical world is finally coming into focus for artificial intelligence.

  • Luma AI’s Uni-1 Claims to Outscore Google and OpenAI — At 30% Lower Cost

    Luma AI’s Uni-1 Claims to Outscore Google and OpenAI — At 30% Lower Cost

    A new challenger has entered the multimodal AI arena — and it’s making bold claims about performance and cost. Luma AI, known primarily for its AI-powered 3D capture technology, has launched Uni-1, a model that the company says outscores both Google and OpenAI on key benchmarks while costing up to 30 percent less to run.

    The announcement represents Luma AI’s most ambitious move yet from 3D reconstruction into the broader world of general-purpose multimodal intelligence. Uni-1 reportedly tops Google’s Nano Banana 2 and OpenAI’s GPT Image 1.5 on reasoning-based benchmarks, and nearly matches Google’s Gemini 3 Pro on object detection tasks.

    What’s Different About Uni-1?

    Unlike models that specialize in a single modality, Uni-1 is architected as a true multimodal system — capable of reasoning across text, images, video, and potentially 3D data. This positions it as a competitor not just to image generation models but to the full spectrum of frontier multimodal systems.

    The cost claim is particularly significant. Luma AI says Uni-1 achieves its performance benchmarks at a 30 percent lower operational cost compared to comparable offerings from Google and OpenAI. For enterprises watching their inference budgets, this could be a game-changer — especially if the performance claims hold up in real-world deployments.

    Benchmark Performance Breakdown

    According to Luma AI’s published results:

    • Uni-1 outperforms Google’s Nano Banana 2 on reasoning-based benchmarks
    • Uni-1 outperforms OpenAI’s GPT Image 1.5 on the same reasoning-based evaluations
    • Uni-1 nearly matches Google’s Gemini 3 Pro on object detection tasks

    These results, if independently verified, would place Uni-1 among the top-tier multimodal models — a remarkable achievement for a company that hasn’t traditionally competed in this space.

    Luma AI’s Broader Vision

    Luma AI initially gained recognition for its neural radiance field (NeRF) technology, which could reconstruct 3D scenes from 2D images captured on any smartphone. The company’s Dream Machine product brought AI-powered video generation to a mass audience. Uni-1 represents a significant expansion of ambitions.

    The move into general-purpose multimodal AI puts Luma AI in direct competition with some of the largest and best-funded AI labs in the world. The company’s ability to deliver competitive performance at lower cost suggests either a breakthrough in model efficiency, a novel architecture, or a different approach to training data — all of which would be noteworthy.

    Enterprise Implications

    The cost-performance combination is what makes Uni-1 potentially disruptive. Enterprise AI adoption has been slowed in part by the high cost of running state-of-the-art models at scale. If a new entrant can reliably deliver frontier-level performance at a 30 percent discount, it could accelerate adoption in cost-sensitive industries and use cases.

    Of course, benchmark performance doesn’t always translate to real-world superiority. The AI industry has seen numerous models that excel on standard benchmarks but underperform in production environments. Independent evaluations and enterprise pilots will be the true test of Uni-1’s capabilities.

    Availability and Access

    Luma AI has begun rolling out access to Uni-1 through its existing platform. Developers and enterprises interested in evaluating the model can sign up through the Luma AI website. The company has indicated plans for API access and enterprise custom deployment options.

    The multimodal AI market is heating up rapidly, and Luma AI’s entry with Uni-1 adds another dimension to an already competitive landscape. Whether Uni-1 can live up to its ambitious claims remains to be seen — but the company has made a clear statement of intent.

  • OpenAI’s Nuclear Ambitions: Sam Altman’s Fusion Energy Deal Raises AI-Power Questions

    In a move that underscores the massive energy appetite of artificial intelligence systems, OpenAI is reported to be in advanced talks to purchase electricity from Helion Energy, the nuclear fusion startup where CEO Sam Altman previously served as board chair.

    The revelation comes as no surprise to industry observers who have watched AI development increasingly constrained by energy availability. Training large language models requires enormous computational resources, and the subsequent inference operations鈥攔unning those models for millions of users鈥攃onsume power continuously.

    The AI-Energy Connection

    Sam Altman’s dual roles at OpenAI and Helion have long raised questions about potential conflicts of interest and strategic alignments. His recent departure from Helion’s board, announced via social media, appears designed to address those concerns as the two companies explore a commercial relationship.

    Helion Energy has been working toward what many consider the Holy Grail of clean energy: practical nuclear fusion. The technology promises virtually unlimited, clean power generation, though significant scientific challenges remain before commercial viability.

    The Timing Matters

    OpenAI’s interest in fusion energy reflects a broader recognition within the AI industry that power availability could become the defining constraint on AI advancement. Data centers are already straining electrical grids in many regions, and the trend shows no signs of slowing.

    Major tech companies are exploring various solutions:

    • Nuclear Power: Microsoft’s agreement with Constellation Energy to restart Three Mile Island’s nuclear plant
    • Solar and Wind: Large-scale renewable installations for data center complexes
    • Fusion Research: Investments in next-generation technologies like Helion
    • Grid Infrastructure: Upgrades to transmission and distribution systems

    What This Means for AI Development

    The convergence of AI and energy industries represents a fundamental shift in how we think about computational infrastructure. AI systems are no longer purely digital endeavors鈥攖hey have become physical installations requiring substantial real-world resources.

    For OpenAI, securing long-term power agreements could provide strategic advantages in the increasingly competitive AI landscape. Companies that can guarantee power supply may be better positioned to scale their operations and train even larger models.

    The Bigger Picture

    While fusion power remains years away from commercial deployment, OpenAI’s interest signals the company’s long-term thinking about infrastructure needs. The fact that one of the world’s leading AI companies is looking toward nuclear fusion鈥攖raditionally considered decades from practical application鈥攗nderscores the scale of resources AI is expected to require.

    As AI capabilities continue to advance, the question of sustainable power supply will only become more pressing. OpenAI’s move toward fusion energy may prove prescient鈥攐r perhaps premature. Either way, it marks an important moment in the evolution of the AI industry.

    Industry analysts will be watching closely as more details emerge about the OpenAI-Helion discussions and what they might mean for the future of AI development and energy consumption.

  • Mark Zuckerberg Is Training an AI to Do His Job — and It Might Be Better at It Than You Think

    The idea that AI will eventually replace human workers is no longer a fringe prediction — it is a live strategic project at some of the world’s largest companies. According to a Wall Street Journal report, that project has now reached the corner office. Mark Zuckerberg, CEO of Meta Platforms, is actively building an AI agent to assist him in performing the duties of a chief executive.

    What the AI CEO Agent Does

    The agent, still in development according to sources familiar with the project, is not being designed to replace Zuckerberg entirely — at least not yet. Instead, it is currently serving as a kind of ultra-efficient executive assistant that can:

    • Retrieve information that would normally require Zuckerberg to go through multiple layers of subordinates
    • Synthesize data from across Meta’s numerous business units without scheduling meetings or waiting for reports
    • Draft responses to strategic questions by pulling together real-time information from internal systems
    • Act as a rapid-response information retrieval layer between Zuckerberg and the company’s sprawling organizational hierarchy

    In short, the agent is doing what CEOs are supposed to do — making decisions based on comprehensive information — except potentially faster and without the organizational friction that typically slows executive decision-making.

    The “Who Needs CEOs?” Question Gets Real

    Surveys consistently show that the American public holds CEOs in relatively low esteem — a 2025 poll found that 74% of Americans disapprove of Mark Zuckerberg’s performance. If AI agents can perform the core informational and decision-making functions of a CEO without the ego, compensation controversies, and reputational baggage, the economic case for AI CEOs becomes harder to dismiss.

    AI CEOs do not need to sleep. They do not need million in annual compensation. They do not generate PR disasters through personal behavior. They do not play golf.

    Of course, they also cannot do everything a CEO does. Building consensus among human board members, managing the emotional dynamics of a workforce, navigating political landscapes both inside and outside the company — these are areas where human judgment still matters enormously. Whether the AI CEO agent is a genuine strategic asset or a sophisticated administrative tool remains to be seen.

    The Meta AI Strategy

    For Meta, building an AI CEO agent is also a demonstration of capability. If Meta’s AI can handle the information complexity of running one of the world’s largest technology companies, that is a powerful proof of concept for enterprise AI products. The company has been aggressively integrating AI across its product portfolio — from Instagram recommendation systems to Meta AI assistants — and an internal CEO agent would be the ultimate stress test.

    Zuckerberg’s agent project also reflects a broader reality about how AI is being deployed in practice: not as dramatic replacements, but as layered augmentations that handle the routine and information-intensive parts of high-skill work. The pattern is familiar from other domains — radiologists are not being replaced wholesale by AI, but AI is increasingly doing the initial scan analysis while humans handle the nuanced cases. The same dynamic may apply to CEOs.

    What This Means for the Future of Work

    The Zuckerberg AI agent is significant not because it represents a completed transformation, but because it signals the direction of travel. The highest-paid, most powerful knowledge workers are now in the AI replacement conversation, not just junior employees whose tasks are more easily automated.

    If an AI can function as a CEO — or even as a highly capable executive assistant to one — the implications for executive compensation, corporate governance, and the distribution of economic power are profound. The technology is moving faster than the policy conversation, and incidents like the Zuckerberg AI agent project are forcing a reckoning with questions that used to belong in science fiction.

    Mark Zuckerberg Meta AI agent

  • Jensen Huang Says We Have Already Achieved AGI. The Problem? Nobody Agrees What That Means.

    Nvidia CEO Jensen Huang has declared that artificial general intelligence — AGI — has already been achieved. There is just one small problem: no one in the AI field can agree on what AGI actually means, making Huang is claim either historic, vacuous, or both.

    The statement, reported by The Verge, came during a public appearance where Huang was asked about the state of AGI development. Huang’s response was characteristically confident: the industry has achieved AGI. When pressed on what exactly he meant, Huang seemed to suggest that the definition is flexible enough to accommodate current AI capabilities — a framing that critics say sidesteps the harder question entirely.

    What Is AGI, Exactly?

    The term artificial general intelligence has been used so broadly, so inconsistently, and so strategically that it has become nearly meaningless as a technical benchmark. Depending on who you ask, AGI means:

    • Any AI that can perform any intellectual task a human can
    • An AI that can reason across domains without task-specific training
    • A system that achieves self-improvement capability
    • A system that passes a broad cognitive benchmark (like the Turing Test, or more modern equivalents)
    • Something vague but clearly impressive that AI companies can claim credit for

    That last definition is the one that seems to matter most in practice. When Jensen Huang says AGI has been achieved, the most charitable interpretation is that Nvidia’s AI products have reached a level of capability that, by some definition, qualifies as general intelligence. The less charitable reading is that Huang is redefining AGI downward to mean whatever current AI does, and then claiming victory.

    Why the Definition Problem Matters

    The definitional ambiguity around AGI is not just an academic concern. It has real consequences:

    • Investment decisions are made on the basis of AGI milestones — if everyone defines those milestones differently, capital allocation becomes irrational
    • Safety research depends on having clear benchmarks — you cannot evaluate whether an AI is safe if nobody agrees on what it should do
    • Regulatory frameworks require definitional clarity — policymakers drafting AGI rules need to know what they are regulating
    • Public trust in AI companies suffers when executives make grand claims that subsequent events contradict

    The Industry’s Incentives

    Part of the reason AGI keeps being declared — and undeclared — is that the term has enormous marketing value. For Nvidia, claiming AGI has been achieved is implicitly a claim that Nvidia’s chips and infrastructure are powering that achievement. For OpenAI, Google, and others, being first to AGI would represent the most significant technological milestone in human history.

    These incentives create pressure to claim AGI as soon as possible, and to define it loosely enough to claim it plausibly. Critics of the AI industry argue that this definitional inflation devalues the concept and makes serious evaluation impossible.

    What Huang Actually Said

    According to The Verge’s coverage, Huang’s actual claim was hedged enough to be almost unfalsifiable. He essentially argued that the boundary between narrow AI and AGI is blurry, and that modern AI systems have crossed so many specific capability thresholds that the aggregate effect is indistinguishable from AGI by any reasonable definition.

    This framing is not entirely without merit. Modern large language models can write code, analyze legal documents, diagnose medical conditions, generate creative content, and engage in multi-step reasoning — all capabilities that would have been considered AGI milestones a decade ago. Whether doing all of these things without further training constitutes general intelligence is the crux of the debate.

    Until the AI field develops consensus around what AGI actually means — and establishes rigorous, independently verifiable benchmarks — CEO declarations of its achievement will remain more about public relations than scientific progress.

    Nvidia CEO Jensen Huang AGI claim

  • OpenAI’s Fusion Energy Gambit: Sam Altman, Helion, and the AI Power Problem

    In a move that sounds like science fiction but might just be the next chapter of the AI arms race, OpenAI is reportedly in advanced talks to purchase electricity directly from Helion Energy — the nuclear fusion startup where Sam Altman serves as board chair. The twist: Altman himself founded Helion, raising immediate questions about conflict of interest, corporate structure, and the unprecedented energy demands of modern AI.

    The Story So Far

    On March 23, 2026, it emerged that Sam Altman has stepped down from Helion Energy’s board while simultaneously being in discussions about a major commercial electricity deal between OpenAI and the fusion company. Axios first reported the talks, describing them as advanced — though significant scientific hurdles remain before fusion can deliver a single watt to the grid.

    The timing is notable. Altman announced his departure from Helion’s board publicly on X (formerly Twitter), a move that appears designed to distance himself from the deal-making process while the talks continue. Whether this recusal is legally sufficient remains a question for regulators to answer.

    Why Fusion? Why Now?

    The AI industry is facing an energy crisis — literally. Data centers powering large language models consume enormous amounts of electricity. Microsoft, OpenAI’s primary investor and cloud provider, has been exploring every conceivable energy source to fuel its expanding AI infrastructure. Nuclear power — both fission and fusion — has moved from fringe discussion to mainstream strategic planning at every major tech company.

    Helion Energy has been developing compact nuclear fusion technology for over a decade, positioning itself as potentially the first company to achieve net energy from fusion (producing more energy than it consumes). The company has a long-term power purchase agreement with Microsoft, and now appears to be courting OpenAI as an additional anchor customer.

    Fusion has long been called \”the Holy Grail of clean energy\” — capable of generating enormous power from abundant hydrogen isotopes, with virtually no carbon emissions and minimal radioactive waste. But despite decades of research and billions in investment, commercial fusion power remains perpetually \”about 20 years away.\”

    The Conflict of Interest Problem

    What makes this situation particularly thorny is the overlapping roles. Altman was not just an early investor in Helion — he is arguably its most high-profile champion. His dual roles as OpenAI CEO and Helion board chair created an obvious structural conflict when OpenAI started shopping for massive amounts of clean power.

    Corporate ethics experts have raised concerns about the propriety of negotiating a power deal between two entities where the same person holds influential positions on both sides. Altman’s step back from the Helion board is being watched closely by legal analysts — his financial and reputational interests remain deeply intertwined with Helion’s success regardless of his formal board status.

    What This Means for AI’s Future

    Setting aside the governance questions, the OpenAI-Helion story illuminates something important about the AI industry’s self-perception: the largest AI labs increasingly see themselves as infrastructure companies, not just software companies. They are making 20-year bets on energy technology because they believe AI compute demand will continue growing at a rate that makes conventional power sources inadequate.

    This has profound implications:

    • AI companies are becoming energy companies in all but name
    • Grid stability will become a public policy concern as AI data centers compete with residential and industrial users
    • Fusion’s potential viability as a commercial power source is being taken seriously for the first time by serious capital

    The Road Ahead

    Fusion still faces enormous scientific challenges. The laws of physics are not known for responding to market pressure. But the fact that OpenAI — one of the most capital-efficient companies in the world — is willing to lock in long-term power purchase agreements with fusion startups tells us something significant about how the AI industry is thinking about its future.

    The Helion-OpenAI talks, regardless of their outcome, mark a milestone: the moment AI companies stopped treating energy as a utility cost and started treating it as a strategic war.

    OpenAI and nuclear fusion energy