Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

Electrostatic Motors Reach the Macro Scale



It’s a pretty sure bet that you couldn’t get through a typical day without the direct support of dozens of electric motors. They’re in all of your appliances not powered by a hand crank, in the climate-control systems that keep you comfortable, and in the pumps, fans, and window controls of your car. And although there are many different kinds of electric motors, every single one of them, from the 200-kilowatt traction motor in your electric vehicle to the stepper motor in your quartz wristwatch, exploits the exact same physical phenomenon: electromagnetism.

For decades, however, engineers have been tantalized by the virtues of motors based on an entirely different principle: electrostatics. In some applications, these motors could offer an overall boost in efficiency ranging from 30 percent to close to 100 percent, according to experiment-based analysis. And, perhaps even better, they would use only cheap, plentiful materials, rather than the rare-earth elements, special steel alloys, and copious quantities of copper found in conventional motors.

“Electrification has its sustainability challenges,” notes Daniel Ludois, a professor of electrical engineering at the University of Wisconsin in Madison. But “an electrostatic motor doesn’t need windings, doesn’t need magnets, and it doesn’t need any of the critical materials that a conventional machine needs.”

Such advantages prompted Ludois to cofound a company, C-Motive Technologies, to build macro-scale electrostatic motors. “We make our machines out of aluminum and plastic or fiberglass,” he says. Their current prototype is capable of delivering torque as high as 18 newton meters and power at 360 watts (0.5 horsepower)—characteristics they claim are “the highest torque and power measurements for any rotating electrostatic machine.”

The results are reported in a paper, “Synchronous Electrostatic Machines for Direct Drive Industrial Applications,” to be presented at the 2024 IEEE Energy Conversion Congress and Exposition, which will be held from 20 to 24 October in Phoenix, Ariz. In the paper, Ludois and four colleagues describe an electrostatic machine they built, which they describe as the first such machine capable of “driving a load performing industrial work, in this case, a constant-pressure pump system.”

Making Electrostatic Motors Bigger

The machine, which is hundreds of times more powerful than any previous electrostatic motor, is “competitive with or superior to air-cooled magnetic machinery at the fractional [horsepower] scale,” the authors add. The global market for fractional horsepower motors is more than US $8.7 billion, according to consultancy Business Research Insights.

3/4 view cutaway inside the motor C-Motive’s 360-watt motor has a half dozen each of rotors and stators, shown in yellow in this cutaway illustration.C-Motive Technologies

Achieving macro scale wasn’t easy. Electrostatic motors have been available for years, but today, these are tiny units with power output measured in milliwatts. “Electrostatic motors are amazing once you get below about the millimeter scale, and they get better and better as they get smaller and smaller,” says Philip Krein, a professor of electrical engineering at the University of Illinois Urbana-Champaign. “There’s a crossover at which they are better than magnetic motors.” (Krein does not have any financial connection to C-Motive.)

For larger motors, however, the opposite is true. “At macro scale, electromagnetism wins, is the textbook answer,” notes Ludois. “Well, we’ve decided to challenge that wisdom.”

For this quest he and his team found inspiration in a lesser-known accomplishment of one of the United States’ founding fathers. “The fact is that Benjamin Franklin built and demonstrated a macroscopic electrostatic motor in 1747,” says Krein. “He actually used the motor as a rotisserie to grill a turkey on a riverbank in Philadelphia” (a fact unearthed by the late historian I. Bernard Cohen for his 1990 book Benjamin Franklin’s Science ).

Krein explains that the fundamental challenge in attempting to scale electrostatic motors to the macro world is energy density. “The energy density you can get in air at a reasonable scale with an electric-field system is much, much lower—many orders of magnitude lower—than the density you can get with an electromagnetic system.” Here the phrase “in air” refers to the volume within the motor, called the “air gap,” where the machine’s fields (magnetic for the conventional motor, electric for the electrostatic one) are deployed. It straddles the machine’s key components: the rotor and the stator.

Let’s unpack that. A conventional electric motor works because a rotating magnetic field, set up in a fixed structure called a stator, engages with the magnetic field of another structure called a rotor, causing that rotor to spin. The force involved is called the Lorentz force. But what makes an electrostatic machine go ‘round is an entirely different force, called the Coulomb force. This is the attractive or repulsive physical force between opposite or like electrical charges.

Overcoming the Air Gap Problem

C-Motive’s motor uses nonconductive rotor and stator disks on which have been deposited many thin, closely spaced conductors radiating outward from the disk’s center, like spokes in a bicycle wheel. Precisely timed electrostatic charges applied to these “spokes” create two waves of voltage, one in the stator and another in the rotor. The phase difference between the rotor and stator waves is timed and controlled to maximize the torque in the rotor caused by this sequence of attraction and repulsion among the spokes. To further wring as much torque as possible, the machine has half a dozen each of rotors and stators, alternating and stacked like compact discs on a spindle.

The C-Motive motor sitting on a desk in front of a black background The 360-watt motor is hundreds of times more powerful than previous electrostatic motors, which have power output generally measured in milliwatts.C-Motive Technologies

The machine would be feeble if the dielectric between the charges was air. As a dielectric, air has low permittivity, meaning that an electric field in air can not store much energy. Air also has a relatively low breakdown field strength, meaning that air can support only a fairly weak electric field before it breaks down and conducts current in a blazing arc. So one of the team’s greatest challenges was producing a dielectric fluid that has a much higher permittivity and breakdown field strength than air, and that was also environmentally friendly and nontoxic. To minimize friction, this fluid also had to have very low viscosity, because the rotors would be spinning in it. A dielectric with high permittivity concentrates the electric field between oppositely charged electrodes, enabling greater energy to be stored in the space between them. After screening hundreds of candidates over several years, the C-Motive team succeeded in producing an organic liquid dielectric with low viscosity and a relative permittivity in the low 20s. For comparison, the relative permittivity of air is 1.

Another challenge was supplying the 2,000 volts their machine needs to operate. High voltages are necessary to create the intense electric fields between the rotors and stators. To precisely control these fields, C-Motive was able to take advantage of the availability of inexpensive and stupendously capable power electronics, according to Ludois. For their most recent motor, they developed a drive system based on readily available 4.5-kilovolt insulated-gate bipolar transistors, but the rate of advancement in power semiconductors means they have many attractive choices here, and will have even more in the near future.

Ludois reports that C-Motive is now testing a 750-watt (1 hp) motor in applications with potential customers. Their next machines will be in the range of 750 to 3,750 watts (1 to 5 hp), he adds. These will be powerful enough for an expanded range of applications in industrial automation, manufacturing, and heating, ventilating, and air conditioning.

It’s been a gratifying ride for Ludois. “For me, a point of creative pride is that my team and I are working on something radically different that, I hope, over the long term, will open up other avenues for other folks to contribute.”

Could Advanced Nuclear Reactors Fuel Terrorist Bombs?



Various scenarios to getting to net zero carbon emissions from power generation by 2050 hinge on the success of some hugely ambitious initiatives in renewable energy, grid enhancements, and other areas. Perhaps none of these are more audacious than an envisioned renaissance of nuclear power, driven by advanced-technology reactors that are smaller than traditional nuclear power reactors.

What many of these reactors have in common is that they would use a kind of fuel called high-assay low-enriched uranium (HALEU). Its composition varies, but for power generation, a typical mix contains slightly less than 20 percent by mass of the highly fissionable isotope uranium-235 (U-235). That’s in contrast to traditional reactor fuels, which range from 3 percent to 5 percent U-235 by mass, and natural uranium, which is just 0.7 percent U-235.

Now, though, a paper in Science magazine has identified a significant wrinkle in this nuclear option: HALEU fuel can theoretically be used to make a fission bomb—a fact that the paper’s authors use to argue for the tightening of regulations governing access to, and transportation of, the material. Among the five authors of the paper, which is titled “The Weapons Potential of High-Assay Low-Enriched Uranium,” is IEEE Life Fellow Richard L. Garwin. Garwin was the key figure behind the design of the thermonuclear bomb, which was tested in 1952.

The Science paper is not the first to argue for a reevaluation of the nuclear proliferation risks of HALEU fuel. A report published last year by the National Academies, “Merits and Viability of Different Nuclear Fuel Cycles and Technology Options and the Waste Aspects of Advanced Nuclear Reactors,” devoted most of a chapter to the risks of HALEU fuel. It reached similar technical conclusions to those of the Science article, but did not go as far in its recommendations regarding the need to tighten regulations.

Why is HALEU fuel concerning?

Conventional wisdom had it that U-235 concentrations below 20 percent were not usable for a bomb. But “we found this testimony in 1984 from the chief of the theoretical division of Los Alamos, who basically confirmed that, yes, indeed, it is usable down to 10 percent,” says R. Scott Kemp of MIT, another of the paper’s authors. “So you don’t even need centrifuges, and that’s what really is important here.”

Centrifuges arranged very painstakingly into cascades are the standard means of enriching uranium to bomb-grade material, and they require scarce and costly resources, expertise, and materials to operate. In fact, the difficulty of building and operating such cascades on an industrial scale has for decades served as an effective barrier to would-be builders of nuclear weapons. So any route to a nuclear weapon that bypassed enrichment would offer an undoubtedly easier alternative. The question now is, how much easier?

“It’s not a very good bomb, but it could explode and wreak all kinds of havoc.”

Adding urgency to that question is an anticipated gold rush in HALEU, after years of quiet U.S. government support. The U.S. Department of Energy is spending billions to expand production of the fuel, including US $150 million awarded in 2022 to a subsidiary of Centrus Energy Corp., the only private company in the United States enriching uranium to HALEU concentrations. (Outside of the United States, only Russia and China are producing HALEU in substantial quantities.) Government support also extends to the companies building the reactors that will use HALEU. Two of the largest reactor startups, Terrapower (backed in part by Bill Gates) and X-Energy, have designed reactors that run on forms of HALEU fuel, and have received billions in funding under a DOE program called Advanced Reactor Demonstration Projects.

The difficulty of building a bomb based on HALEU is a murky subject, because many of the specific techniques and practices of nuclear weapons design are classified. But basic information about the standard type of fission weapon, known as an implosion device, has long been known publicly. (The first two implosion devices were detonated in 1945, one in the Trinity test and the other over Nagasaki, Japan.) An implosion device is based on a hollow sphere of nuclear material. In a modern weapon this material is typically plutonium-239, but it can also be a mixture of uranium isotopes that includes a percentage of U-235 ranging from 100 percent all the way down to, apparently, around 10 percent. The sphere is surrounded by shaped chemical explosives that are exploded simultaneously, creating a shockwave that physically compresses the sphere, reducing the distance between its atoms and increasing the likelihood that neutrons emitted from their nuclei will encounter other nuclei and split them, releasing more neutrons. As the sphere shrinks it goes from a subcritical state, in which that chain reaction of neutrons splitting nuclei and creating other neutrons cannot sustain itself, to a critical state, in which it can. As the sphere continues to compress it achieves supercriticality, after which an injected flood of neutrons triggers the superfast, runaway chain reaction that is a fission explosion. All this happens in less than a millisecond.

The authors of the Science paper had to walk a fine line between not revealing too many details about weapons design while still clearly indicating the scope of the challenge of building a bomb based on HALEU. They acknowledge that the amount of HALEU material needed for a 15-kiloton bomb—roughly as powerful as the one that destroyed Hiroshima during the second World War—would be relatively large: in the hundreds of kilograms, but not more than 1,000 kg. For comparison, about 8 kg of Pu-239 is sufficient to build a fission bomb of modest sophistication. Any HALEU bomb would be commensurately larger, but still small enough to be deliverable “using an airplane, a delivery van, or a boat sailed into a city harbor,” the authors wrote.

They also acknowledged a key technical challenge for any would-be weapons makers seeking to use HALEU to make a bomb: preinitiation. The large amount of U-238 in the material would produce many neutrons, which would likely result in a nuclear chain reaction occurring too soon. That would sap energy from the subsequent triggered runaway chain reaction, limiting the explosive yield and producing what’s known in the nuclear bomb business as a “fizzle.“ However, “although preinitiation may have a bigger impact on some designs than others, even those that are sensitive to it could still produce devastating explosive power,” the authors conclude.

In other words, “it’s not a very good bomb, but it could explode and wreak all kinds of havoc,” says John Lee, professor emeritus of nuclear engineering at the University of Michigan. Lee was a contributor to the 2023 National Academies report that also considered risks of HALEU fuel and made policy recommendations similar to those of the Science paper.

Critics of that paper argue that the challenges of building a HALEU bomb, while not insurmountable, would stymie a nonstate group. And a national weapons program, which would likely have the resources to surmount them, would not be interested in such a bomb, because of its limitations and relative unreliability.

“That’s why the IAEA [International Atomic Energy Agency], in their wisdom, said, ‘This is not a direct-use material,’” says Steven Nesbit, a nuclear-engineering consultant and past president of the American Nuclear Society, a professional organization. “It’s just not a realistic pathway to a nuclear weapon.”

The Science authors conclude their paper by recommending that the U.S. Congress direct the DOE’s National Nuclear Security Administration (NNSA) to conduct a “fresh review” of the risks posed by HALEU fuel. In response to an email inquiry from IEEE Spectrum, an NNSA spokesman, Craig Branson, replied: “To meet net-zero emissions goals, the United States has prioritized the design, development, and deployment of advanced nuclear technologies, including advanced and small modular reactors. Many will rely on HALEU to achieve smaller designs, longer operating cycles, and increased efficiencies over current technologies. They will be essential to our efforts to decarbonize while meeting growing energy demand. As these technologies move forward, the Department of Energy and NNSA have programs to work with willing industrial partners to assess the risk and enhance the safety, security, and safeguards of their designs.”

The Science authors also called on the U.S. Nuclear Regulatory Commission (NRC) and the IAEA to change the way they categorize HALEU fuel. Under the NRC’s current categorization, even large quantities of HALEU are now considered category II, which means that security measures focus on the early detection of theft. The authors want weapons-relevant quantities of HALEU reclassified as category I, the same as for quantities of weapons-grade plutonium or highly enriched uranium sufficient to make a bomb. Category I would require much tighter security, focusing on the prevention of theft.

Nesbit scoffs at the proposal, citing the difficulties of heisting perhaps a metric tonne of nuclear material. “Blindly applying all of the baggage that goes with protecting nuclear weapons to something like this is just way overboard,” he says.

But Lee, who performed experiments with HALEU fuel in the 1980s, agrees with his colleagues. “Dick Garwin and Frank von Hipple [and the other authors of the Science paper] have raised some proper questions,” he declares. “They’re saying the NRC should take more precautions. I’m all for that.”

How Large Language Models Are Changing My Job



Generative artificial intelligence, and large language models in particular, are starting to change how countless technical and creative professionals do their jobs. Programmers, for example, are getting code segments by prompting large language models. And graphic arts software packages such as Adobe Illustrator already have tools built in that let designers conjure illustrations, images, or patterns by describing them.

But such conveniences barely hint at the massive, sweeping changes to employment predicted by some analysts. And already, in ways large and small, striking and subtle, the tech world’s notables are grappling with changes, both real and envisioned, wrought by the onset of generative AI. To get a better idea of how some of them view the future of generative AI, IEEE Spectrum asked three luminaries—an academic leader, a regulator, and a semiconductor industry executive—about how generative AI has begun affecting their work. The three, Andrea Goldsmith, Juraj Čorba, and Samuel Naffziger, agreed to speak with Spectrum at the 2024 IEEE VIC Summit & Honors Ceremony Gala, held in May in Boston.

Click to read more thoughts from:

  1. Andrea Goldsmith, dean of engineering at Princeton University.
  2. Juraj Čorba, senior expert on digital regulation and governance, Slovak Ministry of Investments, Regional Development
  3. Samuel Naffziger, senior vice president and a corporate fellow at Advanced Micro Devices

Andrea Goldsmith

Andrea Goldsmith is dean of engineering at Princeton University.

There must be tremendous pressure now to throw a lot of resources into large language models. How do you deal with that pressure? How do you navigate this transition to this new phase of AI?

A woman with brown shoulder length hair smiles for a portrait in a teal jacket in an outside scene Andrea J. Goldsmith

Andrea Goldsmith: Universities generally are going to be very challenged, especially universities that don’t have the resources of a place like Princeton or MIT or Stanford or the other Ivy League schools. In order to do research on large language models, you need brilliant people, which all universities have. But you also need compute power and you need data. And the compute power is expensive, and the data generally sits in these large companies, not within universities.

So I think universities need to be more creative. We at Princeton have invested a lot of money in the computational resources for our researchers to be able to do—well, not large language models, because you can’t afford it. To do a large language model… look at OpenAI or Google or Meta. They’re spending hundreds of millions of dollars on compute power, if not more. Universities can’t do that.

But we can be more nimble and creative. What can we do with language models, maybe not large language models but with smaller language models, to advance the state of the art in different domains? Maybe it’s vertical domains of using, for example, large language models for better prognosis of disease, or for prediction of cellular channel changes, or in materials science to decide what’s the best path to pursue a particular new material that you want to innovate on. So universities need to figure out how to take the resources that we have to innovate using AI technology.

We also need to think about new models. And the government can also play a role here. The [U.S.] government has this new initiative, NAIRR, or National Artificial Intelligence Research Resource, where they’re going to put up compute power and data and experts for educators to use—researchers and educators.

That could be a game-changer because it’s not just each university investing their own resources or faculty having to write grants, which are never going to pay for the compute power they need. It’s the government pulling together resources and making them available to academic researchers. So it’s an exciting time, where we need to think differently about research—meaning universities need to think differently. Companies need to think differently about how to bring in academic researchers, how to open up their compute resources and their data for us to innovate on.

As a dean, you are in a unique position to see which technical areas are really hot, attracting a lot of funding and attention. But how much ability do you have to steer a department and its researchers into specific areas? Of course, I’m thinking about large language models and generative AI. Is deciding on a new area of emphasis or a new initiative a collaborative process?

Goldsmith: Absolutely. I think any academic leader who thinks that their role is to steer their faculty in a particular direction does not have the right perspective on leadership. I describe academic leadership as really about the success of the faculty and students that you’re leading. And when I did my strategic planning for Princeton Engineering in the fall of 2020, everything was shut down. It was the middle of COVID, but I’m an optimist. So I said, “Okay, this isn’t how I expected to start as dean of engineering at Princeton.” But the opportunity to lead engineering in a great liberal arts university that has aspirations to increase the impact of engineering hasn’t changed. So I met with every single faculty member in the School of Engineering, all 150 of them, one-on-one over Zoom.

And the question I asked was, “What do you aspire to? What should we collectively aspire to?” And I took those 150 responses, and I asked all the leaders and the departments and the centers and the institutes, because there already were some initiatives in robotics and bioengineering and in smart cities. And I said, “I want all of you to come up with your own strategic plans. What do you aspire to in these areas? And then let’s get together and create a strategic plan for the School of Engineering.” So that’s what we did. And everything that we’ve accomplished in the last four years that I’ve been dean came out of those discussions, and what it was the faculty and the faculty leaders in the school aspired to.

So we launched a bioengineering institute last summer. We just launched Princeton Robotics. We’ve launched some things that weren’t in the strategic plan that bubbled up. We launched a center on blockchain technology and its societal implications. We have a quantum initiative. We have an AI initiative using this powerful tool of AI for engineering innovation, not just around large language models, but it’s a tool—how do we use it to advance innovation and engineering? All of these things came from the faculty because, to be a successful academic leader, you have to realize that everything comes from the faculty and the students. You have to harness their enthusiasm, their aspirations, their vision to create a collective vision.

Juraj Čorba

Juraj Čorba is senior expert on digital regulation and governance, Slovak Ministry of Investments, Regional Development, and Information, and Chair of the Working Party on Governance of AI at the Organization for Economic Cooperation and Development.

What are the most important organizations and governing bodies when it comes to policy and governance on artificial intelligence in Europe?

Portrait of a clean-shaven man with brown hair wearing a blue button down shirt. Juraj Čorba

Juraj Čorba: Well, there are many. And it also creates a bit of a confusion around the globe—who are the actors in Europe? So it’s always good to clarify. First of all we have the European Union, which is a supranational organization composed of many member states, including my own Slovakia. And it was the European Union that proposed adoption of a horizontal legislation for AI in 2021. It was the initiative of the European Commission, the E.U. institution, which has a legislative initiative in the E.U. And the E.U. AI Act is now finally being adopted. It was already adopted by the European Parliament.

So this started, you said 2021. That’s before ChatGPT and the whole large language model phenomenon really took hold.

Čorba: That was the case. Well, the expert community already knew that something was being cooked in the labs. But, yes, the whole agenda of large models, including large language models, came up only later on, after 2021. So the European Union tried to reflect that. Basically, the initial proposal to regulate AI was based on a blueprint of so-called product safety, which somehow presupposes a certain intended purpose. In other words, the checks and assessments of products are based more or less on the logic of the mass production of the 20th century, on an industrial scale, right? Like when you have products that you can somehow define easily and all of them have a clearly intended purpose. Whereas with these large models, a new paradigm was arguably opened, where they have a general purpose.

So the whole proposal was then rewritten in negotiations between the Council of Ministers, which is one of the legislative bodies, and the European Parliament. And so what we have today is a combination of this old product-safety approach and some novel aspects of regulation specifically designed for what we call general-purpose artificial intelligence systems or models. So that’s the E.U.

By product safety, you mean, if AI-based software is controlling a machine, you need to have physical safety.

Čorba: Exactly. That’s one of the aspects. So that touches upon the tangible products such as vehicles, toys, medical devices, robotic arms, et cetera. So yes. But from the very beginning, the proposal contained a regulation of what the European Commission called stand-alone systems—in other words, software systems that do not necessarily command physical objects. So it was already there from the very beginning, but all of it was based on the assumption that all software has its easily identifiable intended purpose—which is not the case for general-purpose AI.

Also, large language models and generative AI in general brings in this whole other dimension, of propaganda, false information, deepfakes, and so on, which is different from traditional notions of safety in real-time software.

Čorba: Well, this is exactly the aspect that is handled by another European organization, different from the E.U., and that is the Council of Europe. It’s an international organization established after the Second World War for the protection of human rights, for protection of the rule of law, and protection of democracy. So that’s where the Europeans, but also many other states and countries, started to negotiate a first international treaty on AI. For example, the United States have participated in the negotiations, and also Canada, Japan, Australia, and many other countries. And then these particular aspects, which are related to the protection of integrity of elections, rule-of-law principles, protection of fundamental rights or human rights under international law—all these aspects have been dealt with in the context of these negotiations on the first international treaty, which is to be now adopted by the Committee of Ministers of the Council of Europe on the 16th and 17th of May. So, pretty soon. And then the first international treaty on AI will be submitted for ratifications.

So prompted largely by the activity in large language models, AI regulation and governance now is a hot topic in the United States, in Europe, and in Asia. But of the three regions, I get the sense that Europe is proceeding most aggressively on this topic of regulating and governing artificial intelligence. Do you agree that Europe is taking a more proactive stance in general than the United States and Asia?

Čorba: I’m not so sure. If you look at the Chinese approach and the way they regulate what we call generative AI, it would appear to me that they also take it very seriously. They take a different approach from the regulatory point of view. But it seems to me that, for instance, China is taking a very focused and careful approach. For the United States, I wouldn’t say that the United States is not taking a careful approach because last year you saw many of the executive orders, or even this year, some of the executive orders issued by President Biden. Of course, this was not a legislative measure, this was a presidential order. But it seems to me that the United States is also trying to address the issue very actively. The United States has also initiated the first resolution of the General Assembly at the U.N. on AI, which was passed just recently. So I wouldn’t say that the E.U. is more aggressive in comparison with Asia or North America, but maybe I would say that the E.U. is the most comprehensive. It looks horizontally across different agendas and it uses binding legislation as a tool, which is not always the case around the world. Many countries simply feel that it’s too early to legislate in a binding way, so they opt for soft measures or guidance, collaboration with private companies, et cetera. Those are the differences that I see.

Do you think you perceive a difference in focus among the three regions? Are there certain aspects that are being more aggressively pursued in the United States than in Europe or vice versa?

Čorba: Certainly the E.U. is very focused on the protection of human rights, the full catalog of human rights, but also, of course, on safety and human health. These are the core goals or values to be protected under the E.U. legislation. As for the United States and for China, I would say that the primary focus in those countries—but this is only my personal impression—is on national and economic security.

Samuel Naffziger

Samuel Naffziger is senior vice president and a corporate fellow at Advanced Micro Devices, where he is responsible for technology strategy and product architectures. Naffziger was instrumental in AMD’s embrace and development of chiplets, which are semiconductor dies that are packaged together into high-performance modules.

To what extent is large language model training starting to influence what you and your colleagues do at AMD?

Portrait of a brown haired man in a dark blue shirt. Samuel Naffziger

Samuel Naffziger: Well, there are a couple levels of that. LLMs are impacting the way a lot of us live and work. And we certainly are deploying that very broadly internally for productivity enhancements, for using LLMs to provide starting points for code—simple verbal requests, such as “Give me a Python script to parse this dataset.” And you get a really nice starting point for that code. Saves a ton of time. Writing verification test benches, helping with the physical design layout optimizations. So there’s a lot of productivity aspects.

The other aspect to LLMs is, of course, we are actively involved in designing GPUs [graphics processing units] for LLM training and for LLM inference. And so that’s driving a tremendous amount of workload analysis on the requirements, hardware requirements, and hardware-software codesign, to explore.

So that brings us to your current flagship, the Instinct MI300X, which is actually billed as an AI accelerator. How did the particular demands influence that design? I don’t know when that design started, but the ChatGPT era started about two years ago or so. To what extent did you read the writing on the wall?

Naffziger: So we were just into the MI300—in 2019, we were starting the development. A long time ago. And at that time, our revenue stream from the Zen [an AMD architecture used in a family of processors] renaissance had really just started coming in. So the company was starting to get healthier, but we didn’t have a lot of extra revenue to spend on R&D at the time. So we had to be very prudent with our resources. And we had strategic engagements with the [U.S.] Department of Energy for supercomputer deployments. That was the genesis for our MI line—we were developing it for the supercomputing market. Now, there was a recognition that munching through FP64 COBOL code, or Fortran, isn’t the future, right? [laughs] This machine-learning [ML] thing is really getting some legs.

So we put some of the lower-precision math formats in, like Brain Floating Point 16 at the time, that were going to be important for inference. And the DOE knew that machine learning was going to be an important dimension of supercomputers, not just legacy code. So that’s the way, but we were focused on HPC [high-performance computing]. We had the foresight to understand that ML had real potential. Although certainly no one predicted, I think, the explosion we’ve seen today.

So that’s how it came about. And, just another piece of it: We leveraged our modular chiplet expertise to architect the 300 to support a number of variants from the same silicon components. So the variant targeted to the supercomputer market had CPUs integrated in as chiplets, directly on the silicon module. And then it had six of the GPU chiplets we call XCDs around them. So we had three CPU chiplets and six GPU chiplets. And that provided an amazingly efficient, highly integrated, CPU-plus-GPU design we call MI300A. It’s very compelling for the El Capitan supercomputer that’s being brought up as we speak.

But we also recognize that for the maximum computation for these AI workloads, the CPUs weren’t that beneficial. We wanted more GPUs. For these workloads, it’s all about the math and matrix multiplies. So we were able to just swap out those three CPU chiplets for a couple more XCD GPUs. And so we got eight XCDs in the module, and that’s what we call the MI300X. So we kind of got lucky having the right product at the right time, but there was also a lot of skill involved in that we saw the writing on the wall for where these workloads were going and we provisioned the design to support it.

Earlier you mentioned 3D chiplets. What do you feel is the next natural step in that evolution?

Naffziger: AI has created this bottomless thirst for more compute [power]. And so we are always going to be wanting to cram as many transistors as possible into a module. And the reason that’s beneficial is, these systems deliver AI performance at scale with thousands, tens of thousands, or more, compute devices. They all have to be tightly connected together, with very high bandwidths, and all of that bandwidth requires power, requires very expensive infrastructure. So if a certain level of performance is required—a certain number of petaflops, or exaflops—the strongest lever on the cost and the power consumption is the number of GPUs required to achieve a zettaflop, for instance. And if the GPU is a lot more capable, then all of that system infrastructure collapses down—if you only need half as many GPUs, everything else goes down by half. So there’s a strong economic motivation to achieve very high levels of integration and performance at the device level. And the only way to do that is with chiplets and with 3D stacking. So we’ve already embarked down that path. A lot of tough engineering problems to solve to get there, but that’s going to continue.

And so what’s going to happen? Well, obviously we can add layers, right? We can pack more in. The thermal challenges that come along with that are going to be fun engineering problems that our industry is good at solving.

❌
❌