19 March 2026
Siemens, Rittal partner on data centre power
 
19 March 2026
Report: AI boom driving US data centres off grid
 
18 March 2026
Schneider, NVIDIA to advance AI data centre design
 
18 March 2026
SUBCO expands Australia network route diversity
 
18 March 2026
NetApp launches new EF-Series storage systems
 

Latest News


Panduit expands fibre portfolio with fusion splice connectors
Panduit, a manufacturer of electrical and network infrastructure solutions, has introduced OmniSplice, a new range of fusion-spliced fibre optic connectors designed for data centres, edge environments, and enterprise networks. The addition expands the company’s fibre optic portfolio with connectors aimed at supporting high-performance connectivity and faster installation in modern network infrastructure. OmniSplice connectors are designed for use with standard fusion splicing equipment, allowing integration into existing installation and maintenance workflows without requiring additional tools or modifications. Panduit says the connectors are intended to support consistent performance while reducing installation time. Integrated design for simplified deployment A key feature of the OmniSplice range is the integration of the splice point within the connector housing. This removes the need for additional components such as pigtails, helping to reduce space requirements and simplify installation. The connectors include pre-assembled fibre stubs and a holder design intended to support the fusion splicing process, aiming to improve consistency and reduce the likelihood of installation errors. According to Panduit, the design is suited to environments where rapid deployment or maintenance is required, including moves, adds, and changes, as well as repair work under time constraints. The launch reflects continued growth in fibre optic infrastructure across data centres, enterprise LANs, and edge applications, where there is increasing demand for solutions that can be integrated efficiently into existing systems. For more from Panduit, click here.

Nscale, Microsoft partner on large-scale campus in West Virginia
Nscale, a UK developer of AI data centres and cloud infrastructure, has signed a letter of intent with Microsoft to deliver 1.35GW of AI compute capacity at the Monarch AI campus in West Virginia, in collaboration with NVIDIA and Caterpillar. The development will deploy NVIDIA’s next-generation Vera Rubin NVL72 GPU systems, based on the NVIDIA DSX AI Factory reference design, with the undertaking expected to begin in phases from late 2027. In addition to this news, Nscale has also announced the acquisition of American Intelligence & Power Corporation (AIPCorp), which includes the Monarch Compute Campus in Mason County. The site spans up to 2,250 acres (9.1 km²) and is designed as a state-certified AI microgrid, with the potential to scale beyond 8GW of power capacity. Hyperscale AI infrastructure and power integration Under the agreement, Nscale will construct and operate the data centre infrastructure, with Microsoft supporting long-term compute services and lease arrangements. The campus is intended to support large-scale AI training and inference workloads, with high-speed connectivity to major US data centre hubs, including Ashburn and Chicago. As part of the project, Caterpillar will supply G3500 series natural gas generator sets, with plans to deliver up to 2GW of on-site power generation by the first half of 2028. The microgrid design enables the facility to operate independently of the local grid, while also allowing for potential future grid integration. The development reflects increasing demand for AI-driven data centre capacity, with industry forecasts indicating significant growth in global power requirements over the coming years. The Monarch campus is expected to build on Nscale’s existing capacity and support expansion of large-scale AI infrastructure in the US. For more from Nscale, click here.

Siemens expands data centre ecosystem for AI infrastructure
German multinational technology company Siemens has expanded its data centre partner ecosystem to support the growth of next-generation artificial intelligence infrastructure, focusing on the integration of compute, power, and operational systems. The expansion includes a strategic investment in Emerald AI, a collaboration with PhysicsX, and the integration of energy storage technologies from Fluence. As AI adoption accelerates, data centre operators are facing increasing constraints around power availability and grid connection timelines. Siemens says the expanded ecosystem is intended to improve flexibility across infrastructure, helping operators scale capacity while maintaining reliability in power-constrained environments. Coordinating compute and energy systems Emerald AI’s technology enables AI workloads to shift in time and location to align with grid conditions, allowing data centre demand to respond dynamically to available power. This approach is designed to reduce peak demand pressures and support faster grid connections. Fluence’s battery energy storage systems (BESS) are intended to help operators manage large-scale AI workloads by shaping energy demand and supporting more predictable load profiles. The systems can also provide on-site power during grid constraints or outages, supporting operational continuity. In addition, Siemens is working with PhysicsX to apply physics-based AI modelling to data centre power distribution systems. Using simulation data, the approach enables engineers to model thermal behaviour in real time, reducing design times and supporting optimisation for dynamic AI workloads. Siemens said the combined ecosystem brings together workload orchestration, energy infrastructure, and AI-driven modelling to address the growing complexity of data centre design and operation as AI demand increases. For more from Siemens, click here.

Barriers to colocation could hold back DC market
Speaking from the Space Comm Expo Europe event in London, Strategies in Satellite Ground Segment (SSGS), the organiser of the world’s only conference dedicated exclusively to the satellite ground segment, has warned that the global success of data centres is at risk if barriers to their colocation with satellite ground stations are not removed. SSGS argues that the advantages to placing a satellite ground station next to a data centre are already clear. It suggests that, for example, keeping the antenna systems physically close to compute and storage functions eliminates long terrestrial backhaul links and helps to reduce latency. Both the data centre and ground station can benefit from cost-savings and can become scalable digital hubs as capacity can be quickly added to meet increased demand. However, securing the necessary planning permissions, obtaining sufficient levels of funding, obtaining the necessary spectrum licenses in already crowded frequency bands, and event issues relating to data sovereignty can all hold up the process of colocation. Colocating data centres with satellite ground systems Kevin French, Director of the SSGS conference and exhibition, says, “Satellite ground systems are increasingly being colocated with data centres, and it’s not a coincidence; it’s a structural shift. "As demand for high-capacity, low-latency services accelerate, bringing ground infrastructure closer to compute isn’t just efficient; it’s transformative. “Colocation enables virtualised architectures, cloud native workflows, and far more resilient end-to-end operations. It ensures that satellite data isn’t just collected, but processed, stored, and delivered with the speed and reliability that modern networks now require.” While there are successful colocation projects in operation - such as the recent approval given to Amazon's Project Kuiper to operate a satellite earth station gateway at the National Space Centre (NSC) in Cork, Ireland - there are fears that the speed of approvals in some jurisdictions could lag behind demand and create a two-tier system. The issue of colocation, data centres, cloud, and edge computing is set to be a major topic of discussion at the SSGS conference being held at the Park Plaza Hotel in London on 30 September 2026. Registration is now open. Find out more by clicking here.

'One in four DC operators fails to track energy usage'
A late‑2025 451 Research study, commissioned by Janitza, a German manufacturer of energy measurement and power quality monitoring equipment, reveals that nearly one in four data centre operators does not monitor the power consumption of their primary sites, even as AI workloads drive unprecedented pressure on electrical and cooling infrastructure. Without precise, real‑time energy data, Janitza argues, operators cannot safely scale AI‑ready capacity or protect their investments. Energy consumption without control 451 Research, the technology market intelligence unit of S&P Global, surveyed 208 data centre professionals to assess how efficiently business‑critical facilities operate today, using power usage effectiveness (PUE) as a key metric. Just over half of respondents reported a PUE between 1.5 and 2.0, while 23% admitted they are not tracking this fundamental performance indicator at all. The study highlights a structural business risk: power has become the limiting factor in building, scaling, and monetising AI‑capable infrastructure. Highly dynamic AI workloads drive power fluctuations of up to 40–70% within milliseconds, creating new challenges for power quality and increasing the risk of outages and equipment damage. The report notes, “In an environment where milliseconds matter, flexibility and data expertise are the critical differentiators.” The findings suggest that reliable, high‑resolution energy data now underpins predictive maintenance, capacity planning, and revenue optimisation in modern data centres. Janitza says operators who capture and analyse detailed power and power‑quality data can detect emerging faults earlier, extend the lifetime of critical components, and avoid unplanned downtime. As rack power densities rise towards 40–120 kW and AI models continue to grow, the study finds that comprehensive monitoring across the entire power chain, from grid connection to individual racks, is becoming a decisive competitive factor. For more from Janitza, cick here.

Why DC-powered lighting matters for modern data centres
In this exclusive article for DCNN, Ton van de Wiel, Global Segment Manager, Intelligent Buildings at Signify, outlines why DC-powered LED lighting is emerging as a key consideration in making data centre infrastructure more efficient and resilient: Building resilience from the ground up The digital services that underpin modern economies – from media streaming to cloud computing – depend on a rapidly expanding global network of data centres. These facilities are not only critical to digital connectivity; they represent significant sources of employment, infrastructure investment, and tax revenue through construction and long-term operation. Today, data centre operators face a convergence of challenges. Capacity requirements are accelerating due to AI-driven workloads, energy prices are rising, and expectations around sustainability and carbon reduction are becoming more stringent. In response, the industry is re-examining its electrical infrastructure. Direct current (DC) power architectures, once limited to niche applications, are gaining traction as a foundation for higher efficiency and greater operational resilience. Within this shift, lighting – often treated as a peripheral system – can play a strategic role. DC-powered LED lighting combines high energy efficiency with relatively low implementation risk, making it an effective starting point for broader DC adoption. Beyond energy savings, lighting can also function as an intelligent layer within next-generation data centre infrastructure. How power architectures are changing Operating a data centre requires tight coordination between IT equipment, networking, cooling, security, and electrical distribution. Historically, alternating current (AC) has been the default for power distribution. However, as facilities' scale and power densities increase, electrical efficiency has become a primary design concern. Early facilities relied on 48V DC for backup systems – safe but capacity-constrained. This gave way to 230/277V AC distribution, followed by 380V DC for internal systems. Today, the extreme power demands of AI servers are driving another transition towards 650V DC and even 800V DC architectures. According to the Open Direct Current Alliance (ODCA), 650V DC represents the optimal level for building-wide distribution, balancing efficiency with safety, while organisations such as NVIDIA and the Open Compute Project are investigating 800V DC. Despite promising high-power IT loads, these higher voltages do not yet deliver the same system-wide efficiency benefits as a facility-level 650V DC approach. Outside the data centre sector, industrial sites are already deploying 650V DC systems to improve energy efficiency and resilience. One key advantage is the ability to capture regenerative energy from motor drives and robotics – energy that would otherwise be dissipated as heat. Because lighting is a continuous base load, it can readily absorb this recovered energy, reducing grid dependency and operating costs. Integrating lighting, motors, renewables, and storage on a shared DC grid reduces conversion losses, cuts copper usage through fewer conductors, and lowers transmission losses compared with 400V AC systems. When paired with solar PV and batteries, DC grids also improve self-consumption, backup capability, and flexible energy management. What’s driving the move? The momentum behind DC power in data centres is rooted in both engineering logic and economics: • Lower conversion losses — Conventional AC systems require multiple conversion steps, resulting in energy losses of up to 18%. • Alignment with IT equipment — Servers and GPUs operate natively on DC power. • Simpler renewable integration — Solar panels and battery systems produce DC, enabling more efficient connections. • Reduced system complexity — Fewer transformers and rectifiers mean simpler installation and improved reliability. • Preparedness for AI growth — Rising AI workloads are accelerating the shift towards DC-based power systems. DC power is therefore not just an alternative distribution method, but a pathway to smarter, more resilient infrastructure. Lighting as the first step Among all building systems, lighting is often the most practical candidate for early DC adoption. Connected LED lighting allows operators to pilot DC distribution with limited risk before extending it to mission-critical IT loads. The benefits are tangible: • Capital expenditure savings — DC lighting cables reduce copper use by 40%. Three-conductor DC cables (L+, L-, PE) can transmit the same power as five-conductor 400V three-phase AC cables. • Operational cost reductions — With only two current-carrying conductors, DC lighting avoids approximately 33% of cable losses compared with three-phase AC at the same current. • Improved resilience — DC lighting can operate directly from on-site solar generation or battery storage, strengthening microgrid performance during outages. DC-compatible luminaires and components are already commercially available. For example, Signify offers a 100W Xitanium LED driver designed for 620–750V DC operation, integrated into the Pacific LED Gen5 and Maxos Fusion luminaire families. These solutions achieve up to 165lm/W efficacy and can be paired with systems such as Signify Interact and Philips Dynalite. Driver-level efficiency can exceed 95%, with future potential to reach 200lm/W through ultra-high-efficiency LED modules. Sustainability and ESG outcomes DC-powered lighting supports measurable sustainability objectives: • Lower carbon emissions through reduced conversion losses and material usage • Support for certifications such as LEED Zero and BREEAM • Energy optimisation with connected lighting systems, cutting lighting energy use by up to 75% For hyperscalers like Amazon Web Services and Microsoft Azure, as well as colocation providers, these outcomes translate directly into stronger ESG reporting and progress towards carbon neutrality. DC lighting can also be implemented incrementally. Some facilities deploy rack-level DC lighting while retaining an AC backbone. Others adopt facility-wide DC grids that integrate lighting, renewables, storage, and IT infrastructure. In larger deployments, centralised emergency lighting connected to the DC backbone ensures continuous illumination during outages, reinforcing safety in mission-critical spaces. A strategic role for lighting As operators prepare for the next phase of digital expansion, DC-powered lighting offers a practical, high-impact entry point into efficient, renewable-ready DC infrastructure. Modern connected lighting systems extend far beyond illumination. With embedded sensors measuring occupancy, daylight, temperature, humidity, and air quality, luminaires form a dense, facility-wide sensing network without the need for additional hardware. Using open protocols such as DALI, BACnet, and MQTT, DC lighting networks integrate with building management systems and DCIM platforms, enabling predictive maintenance, enhanced operational intelligence, and optimised cooling and space utilisation. By simplifying cabling, reducing losses, and enabling intelligent energy management, DC lighting transforms illumination from a passive load into an active contributor to resilient, sustainable data centre operations.

DSE delivers data centre energy resilience
Deep Sea Electronics (DSE), a UK-based manufacturer with over 50 years of engineering expertise, delivers advanced paralleling and ATS controllers ideally suited to the demanding requirements of modern data centre environments. In facilities where uptime is critical and resilience is non-negotiable, DSE solutions provide precise control, seamless synchronisation, and dependable automatic transfer between mains and standby power sources. Designed and manufactured in the UK, DSE’s advanced paralleling controllers enable reliable load sharing, complex multi set configurations, and fast, stable response to load changes. Complementing this, its ATS controllers ensure smooth and accurate mains failure detection and transfer, minimising risk and protecting critical infrastructure. With robust monitoring, clear diagnostics, and flexible communications integration, DSE systems support full visibility and control across standby power architectures. Backed by global technical support and long-term product availability, DSE provides data centre operators with trusted technology that strengthens energy resilience and safeguards operational continuity.

Keysight expands validation for 1.6T AI DC interconnects
Keysight Technologies, a manufacturer of electronic test and measurement equipment and software, has introduced the Functional Interconnect Test Solutions (FITS) portfolio, alongside the first product in the range, FITS-8CH, designed to validate digital-layer error performance for high-speed optical and copper interconnects used in network infrastructure. The platform provides bit error ratio (BER) and forward error correction (FEC) validation for interconnect technologies supporting modern ethernet architectures, including 400GE, 800GE, and emerging 1.6T deployments. As interconnect speeds increase and designs become more complex, manufacturers of chips, interconnects, and networking equipment face greater pressure to ensure reliability before mass production and during manufacturing. While traditional physical-layer test tools validate electrical lanes against industry specifications, system-level testing provides additional insight into how fully integrated interconnect assemblies perform under operational conditions. Digital-layer testing for high-speed interconnects The FITS-8CH platform provides multi-lane error performance validation at the digital layer, supporting PAM4 signalling speeds from 53Gb/s to 212.5Gb/s. The system enables simultaneous bi-directional testing across eight transmit and eight receive channels, allowing complete optical or copper interconnect assemblies to be validated during development, manufacturing, and system-level qualification. The platform also integrates with Keysight’s physical-layer testing systems, enabling validation across a broader range of network configurations and topologies. According to the company, the platform includes automated lane-by-lane tuning to optimise PAM4 signal output and improve measurement consistency. This capability can help identify potential manufacturing or configuration issues earlier in the process, including mechanical misalignment, thermal failures, or incorrect digital signal processor settings. Keysight says the FITS portfolio is intended to support testing requirements across the full product lifecycle, from research and development through to production and deployment in large-scale network environments. For more from Keysight, click here.

Pure DC, AVK deploy 'Europe’s first' data centre microgrid
Pure Data Centres Group (Pure DC), a designer, developer, and operator of hyperscale data centres, together with AVK, a provider of power systems and electrical infrastructure for data centres, have announced the launch of what they describe as Europe’s first, large-scale, 110MW on-site microgrid, developed to support early‑phase site operational resilience. Located within Pure DC’s Dublin campus, the on‑site energy system provides the opportunity for dispatchable capacity to support data centre operations during initial development phases, prior to full integration with the national electricity system as grid connection capacity becomes available. Over time, the campus is intended to operate as part of a hybrid energy configuration, combining grid‑supplied electricity with on‑site infrastructure designed to enhance flexibility, resilience, and system stability. What AVK describes as a "first-of-its-kind deployment in Europe" showcases the ability to use its microgrid technology for on-site power generation, and the transitional and complementary role it can play in supporting the delivery of strategically important digital infrastructure. This is particularly relevant for regions where grid reinforcement and renewable generation are being delivered on a phased basis under national planning frameworks. A replicable model The microgrid also represents a blueprint for energy generation and showcases how large-scale microgrids can be replicated across Europe - with Germany, the Netherlands, and the UK having been identified as key target markets for the technology. The Mayor of Fingal County Council, Councillor Tom O'Leary, comments, “Fingal wants to remain a champion for breakthrough technologies, but we also understand that progress must be delivered in a way that is climate friendly, resilient, and aligned with Ireland’s energy transition. That’s why this project is so important. "A microgrid that can generate and manage its own power supports future integration into the national grid, integrates renewable energy, enables storage, and trials new low‑carbon fuels like biomethane. This is innovation with purpose.” Gary Wojtaszek, Pure DC’s Executive Chairman and interim CEO, notes, “The biggest barrier to deploying AI infrastructure in Europe today isn’t technology; it’s power. This microgrid proves that even the most constrained markets can unlock new digital capacity, giving Ireland the opportunity to lead Europe’s next chapter of AI infrastructure. "The future of AI infrastructure will be built where energy and compute come together, and that’s exactly what we’re building at Pure.” Speaking about the project, Ben Pritchard, CEO of AVK-SEG, adds, “We are delighted to have worked with Pure DC to deliver this groundbreaking project. While several microgrids are already in operation in the US, until today there were none of these deployments in Europe. This project demonstrates how carefully designed onsite energy infrastructure can complement national energy planning frameworks. “This recognises that power is now the new differentiator for data centres, and that energy has shifted from being a utility to a strategic asset - shaping the location, design, economics, and competitiveness for operators. "The first of many in Europe, this microgrid has the capability to revolutionise the data centre power race as we know it, providing a complementary solution that will ease gridlock and pave the way for greater take-up of AI and cloud.” Powering the digital economy Pure DC’s microgrid is comprised of three, interconnected energy centres, with each building generating up to 30 MW of power. Energy Centre 1 (EC1) and EC2 will be fully operational by the end of 2026 and will be followed by EC3 at a later stage. The design includes combined heat and power (CHP) capability, with infrastructure in place to enable heat recovery and potential future connection to district heating networks, subject to third‑party demand and regulatory approvals. Waste heat recovery systems are also used to improve operational efficiency within the energy centres. Future water management measures include rainwater harvesting and on‑site treatment, reducing reliance on mains water for engine‑related processes. The system is engineered to accommodate incremental changes in fuel composition - including hydrogen blending - supporting future decarbonisation of the gas network in line with national policy developments. Pure DC’s battery energy storage system (BESS) is integrated to manage load fluctuations and enhance operational efficiency, improving response times and enabling more optimal engine operation. The BESS is designed to support future renewable energy integration as part of a broader transition pathway. For more from Pure DC, click here, and for more from AVK, click here.

Legrand's UPS wins Data Centre World award
French multinational infrastructure products manufacturer Legrand’s Keor FLEX modular uninterruptible power supply (UPS) has won the Best Reuse or Recycling of Products, Energy, or Data Centre Infrastructure category at the Data Centre World Awards 2026. The award was presented during Data Centre World London, held on 4–5 March at ExCel London, and recognises projects and technologies that support resource reuse, waste reduction, and improved sustainability across the data centre sector. Keor FLEX was recognised for its modular architecture and design approach aimed at extending the operational lifespan of critical power infrastructure. Modular design focused on lifecycle extension Unlike traditional UPS systems that require replacement of the entire unit at the end of its lifecycle, the Keor FLEX system allows individual power or bypass modules to be replaced or refurbished independently. The system uses a hot-swappable modular design, allowing capacity to be expanded or maintained without taking the entire system offline. According to Legrand, the system achieves 98.6% efficiency in online double conversion mode and more than 99% efficiency in ECO mode. It also has an 85% recyclability rate under IEC/TR 62635, with more than 69% recyclable metal content and packaging that includes 50% recycled material. The UPS integrates silicon carbide technology and a low-impedance internal busbar architecture, designed to reduce thermal stress on components and extend the lifespan of power modules. Keor FLEX also supports a universal battery interface that allows existing VRLA, lithium-ion, or nickel-zinc battery systems to be retained during upgrades. Marc Marazzi, Vice President at Legrand Data Center Solutions Europe, says, “Data centres are under pressure to deliver more compute power while reducing environmental impact. “Keor FLEX proves that sustainability and performance are not mutually exclusive. By designing circularity into the core architecture, we’ve created a UPS platform that extends asset life, reduces waste, lowers energy consumption, and supports evolving AI workloads - all while improving total cost of ownership. This reflects Legrand’s broader sustainability commitments, including being awarded an ‘A’ rating by CDP for the second consecutive year.” The system is designed to scale from 100kW to 1.2MW per frame, with up to 4.8MW available in parallel configurations for larger data centre environments. For more from Legrand, click here.



Translate »