Loading...

When AI hits 100kW+ per rack, data center design must be rewritten

India’s data center industry is entering a new phase, one defined not by capacity alone, but by the ability to support growing AI Factories. As AI workloads move into production, they are revealing a gap between what traditional data centers were designed for and what modern compute actually demands. And for enterprise IT leaders, this shift is already visible.

AI deployments are moving into production across analytics, automation, customer platforms, and real-time decision-making systems.  What is emerging is not an incremental shift, but a structural one that challenges the very physics data centers were built on.

 Amit Agrawal, President, Techno Digital

The breaking point of traditional design

Conventional data centers were designed for stability under moderate, predictable loads. Rack densities typically ranged between 6 to 10 kW, cooling systems were designed for distributed heat loads, and power systems were engineered for gradual variations.

Today, GPU clusters are pushing rack densities to 100 kW and beyond. More importantly, these workloads introduce sharp load variability, continuous high thermal output, and far tighter tolerance requirements across both power and cooling systems.

At these densities, inefficiencies that were once negligible begin to compound rapidly. Voltage drops across distribution paths, minor electrical losses, or airflow imbalances can directly impact performance. Thermal instability is no longer an operational inconvenience, it is becoming a constraint on compute output.

Why Chennai is emerging as an AI infrastructure hub

India’s data center growth is expanding beyond traditional concentration in a single metro. Chennai has emerged as one of the most strategically positioned destinations for AI-scale data centers.

The reasons are structural.

  • A strong industrial-grade power backbone capable of supporting sustained high-density loads.
  • Proximity to submarine cable landing stations enables low-latency connectivity across Asia-Pacific markets.
  • Policy momentum in Tamil Nadu is accelerating approvals and infrastructure readiness.

For enterprises deploying AI workloads particularly those requiring a balance between latency, power availability, and scalability, Chennai represents a compelling combination.

But location alone is not enough. What matters is how infrastructure is engineered within that location.

Designing for AI from first principles

One of the key lessons in building for AI-scale environments is that legacy models cannot simply be stretched to accommodate new demands. They must be chosen and rethought as per the current requirements.

Power architecture: Stability under dynamic load

In AI environments, power is no longer just about availability. It is about behaviour under stress.

GPU-driven workloads do not follow predictable patterns. They create sudden spikes and sustain high demand, which expose weaknesses in traditional power systems. At high densities, even minor inconsistencies in voltage or distribution can affect performance.

This changes how power architecture is approached. The focus shifts toward tighter control, reduced losses, and ensuring consistent electrical conditions across the system. Stability becomes critical not at peak load alone, but across every fluctuation cycle.

Distribution strategy: Stability closer to the rack

One of the key challenges in high-density environments is maintaining voltage stability at the point of consumption.

In high-density environments, voltage instability often originates closer to the rack rather than at the source. The longer the distribution path, the higher the chances of fluctuation and inefficiency.

Modern infrastructure design is moving toward tighter distribution, where control is maintained closer to the point of consumption. This reduces variability, improves efficiency, and ensures that compute systems operate within stable electrical limits, particularly during peak demand.

Resilience reimagined: Continuity without disruption

Traditional resilience models focus on surviving failures. AI workloads require something more continuity without interruption.

AI workloads require a different approach. Even short transitions during power events can interrupt processing and affect outcomes. Recovery is no longer sufficient.

The focus is shifting toward continuity. Infrastructure needs to ensure that disruptions do not translate into workload instability. This requires coordinated systems where backup power, energy storage, and transition mechanisms operate seamlessly, without visible impact on compute.

Cooling architecture: Designed for thermal intensity

At densities approaching 100 kW per rack, cooling becomes a primary constraint. It is one of the first systems to reach its limit as AI workloads scale. At densities approaching and exceeding 100 kW per rack, heat is no longer a secondary consideration. It becomes a defining constraint that directly influences how efficiently compute can operate over time. 

  • Cooling systems must handle continuous and concentrated thermal loads, not intermittent peaks 
  • Approaches such as adiabatic cooling help reduce water dependency while maintaining performance under high ambient conditions 
  • Airflow design needs to ensure uniform distribution to avoid hotspots across densely packed racks 
  • Thermal systems must be built to scale with increasing rack densities without compromising stability 
  • Consistent thermal conditions are critical, as even small variations can impact workload performance 

This ensures that thermal conditions remain stable, allowing AI workloads to operate at their intended performance levels.

Redefining the role of data centers

The industry must now move beyond viewing data centers as physical shells that house compute.

In AI-driven environments, performance depends on how well power, cooling, and distribution work together under continuous load. This shifts the focus from building capacity to sustaining efficiency, stability, and control at scale. As AI adoption accelerates, infrastructure is no longer a background layer.

Because at AI scale, infrastructure does not simply support compute. It is the system that defines performance.

To know more about Techno Digital, visit: https://technodigital.in/ 

– Author is Amit Agrawal, President, Techno Digital

About The Author