Breakthrough in computer chip energy efficiency could cut data center electricity use

MAY 24, 2023

by Steve Lundeberg, 

(a) 3D schematic of the on-chip WDM filters consisting of four cascaded tunable Si-MRRs with different radii. Inset: zoom-in view of microring waveguide with ITiO/HfO2/Si MOS capacitor. (b) Simulated mode profiles in the active region with waveguide widths of 300 and 400 nm, respectively. (c) Simulated Q-factor (blue) and tunability (orange) versus waveguide widths by assuming that the gate material has a carrier concentration of 3 × 1019 cm−3 and mobility of 25 cm2V−1 s−1 in the simulation. (d) Simulated Q-factor as a function of mobility with different waveguide widths, assuming that the bent Si waveguide has a radius of 8 μm and a height of 250 nm with a 50 nm slab in the simulation. Credit: Scientific Reports (2023). DOI: 10.1038/s41598-023-32313-0

Researchers at Oregon State University and Baylor University have made a breakthrough toward reducing the energy consumption of the photonic chips used in data centers and supercomputers. The paper is published in the journal Scientific Reports.

The findings are important because a  can consume up to 50 times more energy per square foot of floor space than a typical office building, according to the U.S. Department of Energy.

A data center houses an organization’s information technology operations and equipment; it stores, processes and disseminates data and applications. Data centers account for roughly 2% of all electricity use in the United States, the DOE says.

According to the U.S. International Trade Commission, the number of  has risen rapidly as data demand has soared. In the United States, home to many firms that produce and consume vast amounts of data including Facebook, Amazon, Microsoft and Google, there are more than 2,600 data centers.

The advance by John Conley of the OSU College of Engineering, former Oregon State colleague Alan Wang, now of Baylor, and OSU graduate students Wei-Che Hsu, Ben Kupp and Nabila Nujhat involves a new, ultra-energy-efficient method to compensate for  that degrade photonic chips. Such chips “will form the high-speed communication backbone of future data centers and supercomputers,” Conley said.

The circuitry in photonic chips uses photons—particles of light—rather than the electrons that course through conventional computer chips. Moving at the speed of light, photons enable the extremely rapid, energy-efficient transmission of data.

The issue with  is that up until now, significant energy has been required to keep their temperature stable and performance high. The team led by Wang, however, has shown that it’s possible to reduce the energy needed for  by a factor of more than 1 million.

“Alan is an expert in photonic materials and devices and my area of expertise is  and ,” Conley said. “We were able to make working prototypes that show temperature can be controlled via gate voltage, which means using virtually no electric current.”

Presently, Wang said, the photonics industry exclusively relies on components known as “thermal heaters” to fine tune the working wavelengths of high-speed, electro-optic devices and optimize their performance. These thermal heaters consume several milliwatts of electricity per device.

“That might not sound like much considering that a typical LED lightbulb uses six to 10 watts,” Wang said. “However, multiply those several milliwatts by millions of devices and they add up quickly, so that approach faces challenges as systems scale up and become bigger and more powerful.”

“Our method is much more acceptable for the planet,” Conley added. “It will one day allow data centers to keep getting faster and more powerful while using less energy so that we can access ever more powerful applications driven by machine learning, such as ChatGPT, without feeling guilty.”

SOURCE:

Leave a reply