Logo

Why AI Gains Are Happening at the Edge

By Simone Larsson, Head of Enterprise AI, EMEA, Lenovo.

  • Monday, 30th March 2026 Posted 1 hour ago in by Phil Alsop

Over the past few years, the AI conversation has been dominated by hyperscale data centres and GPU supply constraints. Training ever-larger models has required unprecedented compute density, and organisations have understandably focused on securing capacity in centralised environments.

But that focus risks missing an equally consequential shift. The next wave of enterprise AI value will not come from scaling central infrastructure alone, but also by distributing intelligence closer to where data is generated, decisions are made and outcomes are measured.

This year and beyond, edge AI will move from selective deployments to a core component of enterprise architecture. This is not about replacing cloud. It is about building infrastructure that reflects how AI is actually used in production environments.

Creating AI Value Where Data Lives

Most enterprise data is not created in hyperscale facilities. It originates in factories, hospitals, retail stores, logistics hubs and office buildings. Sensors, cameras, connected devices and operational systems continuously generate high volumes of information.

Historically, organisations moved that data to central clouds or data centres for processing. For AI training, that model still makes sense. But for inference and real-time action, repeatedly transporting large data sets introduces latency, cost and complexity.

As AI use cases mature, enterprises are prioritising responsiveness. Predictive maintenance systems cannot wait for round-trip cloud processing. Computer vision systems in manufacturing need instant feedback to prevent defects. Retail analytics platforms must respond in real time to customer behaviour.

Processing data at or near the source addresses these requirements directly. It reduces delay, limits unnecessary data movement and allows organisations to act immediately on insight. In many cases, only summarised or exception-based data needs to be sent upstream.

At Lenovo, we see this shift reflected in how customers are designing hybrid AI environments. AI training may occur in centralised clusters, but inference increasingly happens at regional data centres or edge locations. The architecture is becoming distributed by design.

Right-Sized Infrastructure for Delivering Practical AI

There is a persistent assumption that AI success depends on access to unlimited cloud compute. That may be true for frontier model development. It is not true for most enterprise applications.

The majority of production AI workloads rely on inference against established models. These workloads require reliability, security and predictable performance more than massive scale. Overprovisioning central infrastructure for every use case increases cost without improving outcomes.

Right-sized infrastructure means aligning compute resources with specific business requirements. In many scenarios, that means compact, energy-efficient systems deployed at the edge or in regional facilities. These deployments can be tailored to workload intensity, environmental constraints and regulatory needs.

This approach also improves cost control. Data movement is expensive. Transmitting raw video streams or industrial telemetry to a distant cloud for constant analysis consumes bandwidth and energy. By filtering and processing locally, organisations reduce network dependency and operational expense.

Edge AI is not about shrinking ambition. It is about applying compute where it delivers measurable value.

Performance, Power and Proximity

Latency is often discussed in technical terms, but its impact is operational. In industrial environments, milliseconds can determine whether a defect is corrected or a production line is halted. In healthcare, rapid processing of diagnostic data can accelerate clinical decisions. In logistics, real-time optimisation improves throughput and asset utilisation.

Proximity between data and compute reduces that delay. It also improves resilience. Many edge environments operate with variable or limited connectivity. Local processing ensures systems continue functioning even when external networks are disrupted.

Energy efficiency is another factor shaping AI infrastructure strategy. Centralised facilities will remain essential, but distributing certain workloads can improve overall power utilisation. Regional and edge deployments allow enterprises to balance density with efficiency, avoiding unnecessary concentration of compute in a single location.

As AI adoption expands, infrastructure planning must consider performance per watt and performance per dollar, not just peak compute capability.

A Strategic Shift 

The enterprise AI journey is entering a new phase. Early experimentation focused on proving technical feasibility. The next phase is about operational integration.

That transition requires CIOs and infrastructure leaders to reassess how they design AI environments. With nearly half (46%) of AI proof-of-concepts already progressing into production, the shift from experimentation to scalable deployment is well underway. Instead of defaulting to centralisation, organisations must evaluate where intelligence drives the greatest business impact. In many industries, that answer points to distributed architectures that combine central training with edge inference.

This shift also influences governance and security. Keeping sensitive data within local or regional boundaries can simplify compliance with data sovereignty requirements. It allows organisations to apply consistent policy controls across hybrid environments without excessive data replication.

At Lenovo, we are working with customers to build hybrid AI frameworks that integrate cloud, data centre and edge seamlessly. That includes validated infrastructure designs, scalable HPC systems for model development and ruggedised edge platforms capable of supporting inference in demanding environments. The objective is not to push workloads to one location or another, but to give enterprises flexibility grounded in business outcomes.

The Edge as a Multiplier

The next gains in AI will not come solely from larger models or additional GPUs in central facilities. They will come from architectural decisions that align compute with context.

Enterprises that distribute AI intelligently can accelerate insight, control cost and improve reliability. They can deploy models where action happens, rather than routing every decision through a distant core.

The most effective AI strategies will reflect this balance. Training at scale will remain important. But inference at the edge will increasingly determine whether AI delivers practical value across operations.

The conversation is moving beyond how much compute an organisation can access. The more important question is where that compute should reside. For many enterprises, the answer is closer than they think.

By Sujatha S Iyer, Head of AI Security at Zoho Corp.
By Rick Vanover, Vice President, Product Strategy at Veeam.
By Peter Manta, AI Strategy and Practice Director, Informatica by Salesforce.
By Dmitry Panenkov, CEO and founder of emma
By Apurva Kadakia, Global Head, Cloud and Partnerships, Hexaware.
By David Trossell, CEO and CTO at Bridgeworks.