Normal view

There are new articles available, click to refresh the page.
Yesterday — 16 September 2024New on MIT Technology Review

Google is funding an AI-powered satellite constellation that will spot wildfires faster

16 September 2024 at 15:00

Early next year, Google and its partners plan to launch the first in a series of satellites that together would provide close-up, frequently refreshed images of wildfires around the world, offering data that could help firefighters battle blazes more rapidly, effectively, and safely.

The online search giant’s nonprofit and research arms have collaborated with the Moore Foundation, the Environmental Defense Fund, the satellite company Muon Space, and others to deploy 52 satellites equipped with custom-developed sensors over the coming years. 

The FireSat satellites will be able to spot fires as small as 5 by 5 meters (16 by 16 feet) on any speck of the globe. Once the full constellation is in place, the system should be capable of updating those images about every 20 minutes, the group says.

Those capabilities together would mark a significant upgrade over what’s available from the satellites that currently provide data to fire agencies. Generally, they can provide either high-resolution images that aren’t updated rapidly enough to track fires closely or frequently refreshed images that are relatively low-resolution.

The Earth Fire Alliance collaboration will also leverage Google’s AI wildfire tools, which have been trained to detect early indications of wildfires and track their progression, to draw additional insights from the data.

The images and analysis will be provided free to fire agencies around the world, helping to improve understanding of where fires are, where they’re moving, and how hot they’re burning. The information could help agencies stamp out small fires before they turn into raging infernos, place limited firefighting resources where they’ll do the most good, and evacuate people along the safest paths.

“In the satellite image of the Earth, a lot of things can be mistaken for a fire: a glint, a hot roof, smoke from another fire,” says Chris Van Arsdale, climate and energy research lead at Google Research and chairman of the Earth Fire Alliance. “Detecting fires becomes a game of looking for needles in a world of haystacks. Solving this will enable first responders to act quickly and precisely when a fire is detected.”

Some details of FireSat were unveiled earlier this year. But the organizations involved will announce additional information about their plans today, including the news that Google.org, the company’s charitable arm, has provided $13 million to the program and that the inaugural launch is scheduled to occur next year. 

Reducing the fog of war

The news comes as large fires rage across millions of acres in the western US, putting people and property at risk. The blazes include the Line Fire in Southern California, the Shoe Fly Fire in central Oregon, and the Davis Fire south of Reno, Nevada.

Wildfires have become more frequent, extreme, and dangerous in recent decades. That, in part, is a consequence of climate change: Rising temperatures suck the moisture from trees, shrubs, and grasses. But fires increasingly contribute to global warming as well. A recent study found that the fires that scorched millions of acres across Canada last year pumped out 3 billion tons of carbon dioxide, four times the annual pollution produced by the airline industry.

GOOGLE

Humans have also increased fire risk by suppressing natural fires for decades, which has allowed fuel to build up in forests and grasslands, and by constructing communities on the edge of wilderness boundaries without appropriate rules, materials, and safeguards

Observers say that FireSat could play an important role in combating fires, both by enabling fire agencies to extinguish small ones before they grow into large ones and by informing effective strategies for battling them once they’re crossed that point.

“What these satellites will do is reduce the fog of war,” says Michael Wara, director of the climate and energy policy program at Stanford University’s Woods Institute for the Environment, who is focused on fire policy issues. “Like when a situation is really dynamic and very dangerous for firefighters and they’re trying to make decisions very quickly about whether to move in to defend structures or try to evacuate people.” 

(Wara serves on the advisory board of the Moore Foundation’s Wildfire Resilience Initiative.)

Some areas, like California, already have greater visibility into the current state of fires or early signs of outbreaks, thanks to technology like Department of Defense satellites, remote camera networks, and planes, helicopters, and drones. But FireSat will be especially helpful for “countries that have less-well-resourced wildland fighting capability,” Wara adds.

Better images, more data, and AI will not be able to fully counter the increased fire dangers. Wara and other fire experts argue that regions need to use prescribed burns and other efforts to more aggressively reduce the buildup of fuel, rethink where and how we build communities in fire-prone areas, and do more to fund and support the work of firefighters on the ground. 

Sounding an earlier alarm for fires will only help reduce dangers when regions have, or develop, the added firefighting resources needed to combat the most dangerous ones quickly and effectively. Communities will also need to put in place better policies to determine what types of fires should be left to burn, and under what conditions.

‘A game changer’

Kate Dargan Marquis, a senior wildfire advisor to the Moore Foundation who previously served as state fire marshal for California, says she can “personally attest” to the difference that such tools will make to firefighters in the field.

“It is a game changer, especially as wildfires are becoming more extreme, more frequent, and more dangerous for everyone,” she says. “Information like this will make a lifesaving difference for firefighters and communities around the globe.”

Kate Dargan Marquis, senior advisor, Moore Foundation.
GOOGLE

Google Research developed the sensors for the satellite and tested them as well as the company’s AI fire detection models by conducting flights over controlled burns in California. Google intends to work with Earth Fire Alliance “to ensure AI can help make this data as useful as possible, and also that wildfire information is shared as widely as possible,” the company said.

Google’s Van Arsdale says that providing visual images of every incident around the world from start to finish will be enormously valuable to scientists studying wildfires and climate change. 

“We can combine this data with Google’s existing models of the Earth to help advance our understanding of fire behavior and fire dynamics across all of Earth’s ecosystems,” he says. “All this together really has the potential to help mitigate the environmental and social impact of fire while also improving people’s health and safety.”

Specifically, it could improve assessments of fire risk, as well as our understanding of the most effective means of preventing or slowing the spread of fires. For instance, it could help communities determine where it would be most cost-effective to remove trees and underbrush. 

Figuring out the best ways to conduct such interventions is another key goal of the program, given their high cost and the limited funds available for managing wildlands, says Genny Biggs, the program director for the Moore Foundation’s Wildfire Resilience Initiative.

The launch

The idea for FireSat grew out of a series of meetings that began with a 2019 workshop hosted by the Moore Foundation, which provided the first philanthropic funding for the program. 

The first satellite, scheduled to be launched aboard a SpaceX rocket early next year, will be fully functional aside from some data transmission features. The goals of the “protoflight” mission include testing the onboard systems and the data they send back. The Earth Fire Alliance will work with a handful of early-adopter agencies to prepare for the next phases. 

The group intends to launch three fully operational satellites in 2026, with additional deployments in the years that follow. Muon Space will build and operate the satellites. 

Agencies around the world should be able to receive hourly wildfire updates once about half of the constellation is operational, says Brian Collins, executive director of the Earth Fire Alliance. It hopes to launch all 52 satellites by around the end of this decade.

Each satellite is designed to last about five years, so the organization will eventually need to deploy 10 more each year to maintain the constellation.

The Earth Fire Alliance has secured about two-thirds of the funding it needs for the first phase of the program, which includes the first four launches. The organization will need to raise additional money from government agencies, international organizations, philanthropies, and other groups  to deploy, maintain, and operate the full constellation. It estimates the total cost will exceed $400 million, which Collins notes “is 1/1000th of the economic losses due to extreme wildfires annually in the US alone.”

Asked if commercial uses of the data could also support the program, including potentially military ones, Collins said in an email: “Adjacent applications range from land use management and agriculture to risk management and industrial impact and mitigation.” 

“At the same time, we know that as large agencies and government agencies adopt FireSat data to support a broad public safety mandate, they may develop all-hazard, emergenc[y] management, and security related uses of data,” he added. “As long as opportunities are in balance with our charter to advance a global approach to wildfire and climate resilience, we welcome new ideas and applications of our data.”

‘Living with fire’

A wide variety of startups have emerged in recent years promising to use technology to reduce the frequency and severity of wildfires—for example, by installing cameras and sensors in forests and grasslands, developing robots to carry out controlled burns, deploying autonomous helicopters that can drop suppressant, and harnessing AI to predict wildfire behavior and inform forest and fire management strategies

So far, even with all these new tools, it’s still been difficult for communities to keep pace with the rising dangers.

Dargan Marquis—who founded her own wildfire software company, Intterra—says she is confident the incidence of disastrous fires can be meaningfully reduced with programs like FireSat, along with other improved technologies and policies. But she says it’s likely to take decades to catch up with the growing risks, as the world continues warming up.

“We’re going to struggle in places like California, these Mediterranean climates around the world, while our technology and our capabilities and our inventions, etc., catch up with that level of the problem,” she says. 

“We can turn that corner,” she adds. “If we work together on a comprehensive strategy with the right data and a convincing plan over the next 50 years, I do think that by the end of the century, we absolutely can be living with fire.”

The Download: an AI safety hotline, and tech for farmers

16 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Why we need an AI safety hotline

—Kevin Frazier is an assistant professor at St. Thomas University College of Law and senior research fellow in the Constitutional Studies Program at the University of Texas at Austin.

In the past couple of years, regulators have been caught off guard again and again as tech companies compete to launch ever more advanced AI models. As it stands, it seems there’s little anyone can do to delay or prevent the release of a model that poses excessive risks.

Existing measures to mitigate AI risks aren’t enough to protect us, so we need new approaches. One could be a kind of AI safety hotline tasked with expert volunteers. Read more about how the hotline could work.

African farmers are using private satellite data to improve crop yields

In many developing countries, farming is impaired by lack of data. For centuries, farmers relied on native intelligence rooted in experience and hope.

Now, farmers in Africa are turning to technology to avoid cycles of heavy crop losses that could spell financial disaster. They’re partnering with EOS Data Analytics, a California-based provider of satellite imagery and data for precision farming, which allows them to track where or when specific spots needed attention on various farms—and even to anticipate weather warnings. Read the full story.

—Orji Sunday

This piece is from the latest print issue of MIT Technology Review, which is celebrating 125 years of the magazine! If you don’t already, subscribe now to get 25% off future copies once they land.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 TikTok is heading to court today 
An appeals court will hear whether the app should be banned in the US. (The Verge)
+ The app is fighting the potential ban, which would kick in early next year. (NYT $)
+ The depressing truth behind US attempts to ban TikTok. (MIT Technology Review)

2 China has made a chipmaking equipment breakthrough
A new machine should lessen its reliance on suppliers sanctioned by the US. (Bloomberg $)
+ What’s next in chips. (MIT Technology Review)

3 OpenAI’s newest AI models could be used to create bioweapons
The company itself has given its latest releases its highest safety warning to date. (FT $)
+ In theory, it could aid experts with reproducing a biological threat. (Vox)
+ To avoid AI doom, learn from nuclear safety. (MIT Technology Review)

4 Big Tech’s carbon footprint is likely way bigger than they say
Like, 662% bigger. (The Guardian)
+ Google, Amazon and the problem with Big Tech’s climate claims. (MIT Technology Review)

5 Donald Trump is launching a new crypto business
He’s set to launch it during a livestream today. (NYT $)

6 SpaceX’s private space mission has touched down safety on Earth
The first commercial mission was a big success—and opened the doors for future non-government projects. (BBC)
+ The crew reached a higher altitude than any human has traveled in 50 years. (CNN)

7 We need to stop building in the ocean
It’s severely affecting the way marine life navigates. (The Atlantic $)

8 We’re still learning about the benefits of breast milk 
Its antimicrobial properties could help to treat cancer and other conditions. (Economist $)
+ Startups are racing to reproduce breast milk in the lab. (MIT Technology Review)

9 What the future of food holds
From robot chefs to healthier potatoes. (WSJ $)
+ Robot-packed meals are coming to the frozen-food aisle. (MIT Technology Review)

10 Meet Silicon Valley’s seriously pampered pets 🐕
Pet tech doesn’t come cheap. (The Information $)

Quote of the day

“Welcome back to planet Earth.”

—The host of the live SpaceX broadcast tracking the return of the company’s Polaris Dawn crew greets their homecoming after five days in orbit, the Washington Post reports.

The big story

The future of open source is still very much in flux

August 2023

When Xerox donated a new laser printer to MIT in 1980, the company couldn’t have known that the machine would ignite a revolution.

While the early decades of software development generally ran on a culture of open access, this new printer ran on inaccessible proprietary software, much to the horror of Richard M. Stallman, then a 27-year-old programmer at the university.

A few years later, Stallman released GNU, an operating system designed to be a free alternative to one of the dominant operating systems at the time: Unix. The free-software movement was born, with a simple premise: for the good of the world, all code should be open.

Forty years later, tech companies are making billions on proprietary software, and much of the technology around us is inscrutable. But while Stallman’s movement may look like a failed experiment, the free and open-source software movement is not only alive and well; it has become a keystone of the tech industry. Read the full story.

—Rebecca Ackermann

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ The Hotel California solo series continues, this time with a very intense recorder edition (thanks Niall!)
+ Avocados should be extinct!? Say it ain’t so!
+ This quick and easy peach cobbler recipe looks like an absolute treat. 
+ Nothing but love for Moo Deng the tiny viral baby hippo. 🦛

Why we need an AI safety hotline

16 September 2024 at 11:00

In the past couple of years, regulators have been caught off guard again and again as tech companies compete to launch ever more advanced AI models. It’s only a matter of time before labs release another round of models that pose new regulatory challenges. We’re likely just weeks away, for example, from OpenAI’s release of ChatGPT-5, which promises to push AI capabilities further than ever before. As it stands, it seems there’s little anyone can do to delay or prevent the release of a model that poses excessive risks.

Testing AI models before they’re released is a common approach to mitigating certain risks, and it may help regulators weigh up the costs and benefits—and potentially block models from being released if they’re deemed too dangerous. But the accuracy and comprehensiveness of these tests leaves a lot to be desired. AI models may “sandbag” the evaluation—hiding some of their capabilities to avoid raising any safety concerns. The evaluations may also fail to reliably uncover the full set of risks posed by any one model. Evaluations likewise suffer from limited scope—current tests are unlikely to uncover all the risks that warrant further investigation. There’s also the question of who conducts the evaluations and how their biases may influence testing efforts. For those reasons, evaluations need to be used alongside other governance tools. 

One such tool could be internal reporting mechanisms within the labs. Ideally, employees should feel empowered to regularly and fully share their AI safety concerns with their colleagues, and they should feel those colleagues can then be counted on to act on the concerns. However, there’s growing evidence that, far from being promoted, open criticism is becoming rarer in AI labs. Just three months ago, 13 former and current workers from OpenAI and other labs penned an open letter expressing fear of retaliation if they attempt to disclose questionable corporate behaviors that fall short of breaking the law. 

How to sound the alarm

In theory, external whistleblower protections could play a valuable role in the detection of AI risks. These could protect employees fired for disclosing corporate actions, and they could help make up for inadequate internal reporting mechanisms. Nearly every state has a public policy exception to at-will employment termination—in other words, terminated employees can seek recourse against their employers if they were retaliated against for calling out unsafe or illegal corporate practices. However, in practice this exception offers employees few assurances. Judges tend to favor employers in whistleblower cases. The likelihood of AI labs’ surviving such suits seems particularly high given that society has yet to reach any sort of consensus as to what qualifies as unsafe AI development and deployment. 

These and other shortcomings explain why the aforementioned 13 AI workers, including ex-OpenAI employee William Saunders, called for a novel “right to warn.” Companies would have to offer employees an anonymous process for disclosing risk-related concerns to the lab’s board, a regulatory authority, and an independent third body made up of subject-matter experts. The ins and outs of this process have yet to be figured out, but it would presumably be a formal, bureaucratic mechanism. The board, regulator, and third party would all need to make a record of the disclosure. It’s likely that each body would then initiate some sort of investigation. Subsequent meetings and hearings also seem like a necessary part of the process. Yet if Saunders is to be taken at his word, what AI workers really want is something different. 

When Saunders went on the Big Technology Podcast to outline his ideal process for sharing safety concerns, his focus was not on formal avenues for reporting established risks. Instead, he indicated a desire for some intermediate, informal step. He wants a chance to receive neutral, expert feedback on whether a safety concern is substantial enough to go through a “high stakes” process such as a right-to-warn system. Current government regulators, as Saunders says, could not serve that role. 

For one thing, they likely lack the expertise to help an AI worker think through safety concerns. What’s more, few workers will pick up the phone if they know it’s a government official on the other end—that sort of call may be “very intimidating,” as Saunders himself said on the podcast. Instead, he envisages being able to call an expert to discuss his concerns. In an ideal scenario, he’d be told that the risk in question does not seem that severe or likely to materialize, freeing him up to return to whatever he was doing with more peace of mind. 

Lowering the stakes

What Saunders is asking for in this podcast isn’t a right to warn, then, as that suggests the employee is already convinced there’s unsafe or illegal activity afoot. What he’s really calling for is a gut check—an opportunity to verify whether a suspicion of unsafe or illegal behavior seems warranted. The stakes would be much lower, so the regulatory response could be lighter. The third party responsible for weighing up these gut checks could be a much more informal one. For example, AI PhD students, retired AI industry workers, and other individuals with AI expertise could volunteer for an AI safety hotline. They could be tasked with quickly and expertly discussing safety matters with employees via a confidential and anonymous phone conversation. Hotline volunteers would have familiarity with leading safety practices, as well as extensive knowledge of what options, such as right-to-warn mechanisms, may be available to the employee. 

As Saunders indicated, few employees will likely want to go from 0 to 100 with their safety concerns—straight from colleagues to the board or even a government body. They are much more likely to raise their issues if an intermediary, informal step is available.

Studying examples elsewhere

The details of how precisely an AI safety hotline would work deserve more debate among AI community members, regulators, and civil society. For the hotline to realize its full potential, for instance, it may need some way to escalate the most urgent, verified reports to the appropriate authorities. How to ensure the confidentiality of hotline conversations is another matter that needs thorough investigation. How to recruit and retain volunteers is another key question. Given leading experts’ broad concern about AI risk, some may be willing to participate simply out of a desire to lend a hand. Should too few folks step forward, other incentives may be necessary. The essential first step, though, is acknowledging this missing piece in the puzzle of AI safety regulation. The next step is looking for models to emulate in building out the first AI hotline. 

One place to start is with ombudspersons. Other industries have recognized the value of identifying these neutral, independent individuals as resources for evaluating the seriousness of employee concerns. Ombudspersons exist in academia, nonprofits, and the private sector. The distinguishing attribute of these individuals and their staffers is neutrality—they have no incentive to favor one side or the other, and thus they’re more likely to be trusted by all. A glance at the use of ombudspersons in the federal government shows that when they are available, issues may be raised and resolved sooner than they would be otherwise.

This concept is relatively new. The US Department of Commerce established the first federal ombudsman in 1971. The office was tasked with helping citizens resolve disputes with the agency and investigate agency actions. Other agencies, including the Social Security Administration and the Internal Revenue Service, soon followed suit. A retrospective review of these early efforts concluded that effective ombudspersons can meaningfully improve citizen-government relations. On the whole, ombudspersons were associated with an uptick in voluntary compliance with regulations and cooperation with the government. 

An AI ombudsperson or safety hotline would surely have different tasks and staff from an ombudsperson in a federal agency. Nevertheless, the general concept is worthy of study by those advocating safeguards in the AI industry. 

A right to warn may play a role in getting AI safety concerns aired, but we need to set up more intermediate, informal steps as well. An AI safety hotline is low-hanging regulatory fruit. A pilot made up of volunteers could be organized in relatively short order and provide an immediate outlet for those, like Saunders, who merely want a sounding board.

Kevin Frazier is an assistant professor at St. Thomas University College of Law and senior research fellow in the Constitutional Studies Program at the University of Texas at Austin.

Before yesterdayNew on MIT Technology Review

The Download: conspiracy-debunking chatbots, and fact-checking AI

13 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Chatbots can persuade people to stop believing in conspiracy theories

The internet has made it easier than ever before to encounter and spread conspiracy theories. And while some are harmless, others can be deeply damaging, sowing discord and even leading to unnecessary deaths.

Now, researchers believe they’ve uncovered a new tool for combating false conspiracy theories: AI chatbots. Researchers from MIT Sloan and Cornell University found that chatting about a conspiracy theory with a large language model (LLM) reduced people’s belief in it by about 20%—even among participants who claimed that their beliefs were important to their identity

The findings could represent an important step forward in how we engage with and educate people who espouse baseless theories. Read the full story.

—Rhiannon Williams

Google’s new tool lets large language models fact-check their responses

The news: Google is releasing a tool called DataGemma that it hopes will help to reduce problems caused by AI ‘hallucinating’, or making incorrect claims. It uses two methods to help large language models fact-check their responses against reliable data and cite their sources more transparently to users. 

What next: If it works as hoped, it could be a real boon for Google’s plan to embed AI deeper into its search engine. But it comes with a host of caveats. Read the full story.

—James O’Donnell

Neuroscientists and architects are using this enormous laboratory to make buildings better

Have you ever found yourself lost in a building that felt impossible to navigate? Thoughtful building design should center on the people who will be using those buildings. But that’s no mean feat.

A design that works for some people might not work for others. People have different minds and bodies, and varying wants and needs. So how can we factor them all in?

To answer that question, neuroscientists and architects are joining forces at an enormous laboratory in East London—one that allows researchers to build simulated worlds. Read the full story.

—Jessica Hamzelou

This story is from The Checkup, our weekly biotech and health newsletter. Sign up to receive it in your inbox every Thursday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 OpenAI has released an AI model with ‘reasoning’ capabilities
It claims it’s a step toward its broader goal of human-like artificial intelligence. (The Verge)
+ It could prove particularly useful for coders and math tutors. (NYT $)
+ Why does AI being good at math matter? (MIT Technology Review)

2 Microsoft wants to lead the way in climate innovation
While simultaneously selling AI to fossil fuel companies. (The Atlantic $)
+ Google, Amazon and the problem with Big Tech’s climate claims. (MIT Technology Review)

3 The FDA has approved Apple’s AirPods as hearing aids
Just two years after the body first approved over-the-counter aids. (WP $)
+ It could fundamentally shift how people access hearing-enhancing devices. (The Verge)

4 Parents aren’t using Meta’s child safety controls 
So claims Nick Clegg, the company’s global affairs chief. (The Guardian)
+ Many tech execs restrict their own childrens’ exposure to technology. (The Atlantic $)

5 How AI is turbo boosting legal action
Especially when it comes to mass litigation. (FT $)

6 Low-income Americans were targeted by false ads for free cash
Some victims had their health insurance plans changed without their consent. (WSJ $)

7 Inside the stratospheric rise of the ‘medical’ beverage
Promising us everything from glowier skin to increased energy. (Vox)

8 Japan’s police force is heading online
Cybercrime is booming, as criminal activity in the real world drops. (Bloomberg $)

9 AI can replicate your late loved ones’ handwriting ✍
For some, it’s a touching reminder of someone they loved. (Ars Technica)
+ Technology that lets us “speak” to our dead relatives has arrived. Are we ready? (MIT Technology Review)

10 Crypto creators are resorting to dangerous stunts for attention
Don’t try this at home. (Wired $)

Quote of the day

“You can’t have a conversation with James the AI bot. He’s not going to show up at events.”

—A former reporter for Garden Island, a local newspaper in Hawaii, dismisses the company’s decision to invest in new AI-generated presenters for its website, Wired reports.

The big story

AI hype is built on high test scores. Those tests are flawed.

August 2023

In the past few years, multiple researchers claim to have shown that large language models can pass cognitive tests designed for humans, from working through problems step by step, to guessing what other people are thinking.

These kinds of results are feeding a hype machine predicting that these machines will soon come for white-collar jobs. But there’s a problem: There’s little agreement on what those results really mean. Read the full story.
 
—William Douglas Heaven

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ It’s almost time for Chinese mooncake madness to celebrate the Moon Festival! 🥮
+ Pearl the Wonder Horse isn’t just a therapy animal—she’s also an accomplished keyboardist.
+ We love you Peter Dinklage!
+ Money for Nothing sounds even better on a lute.

Neuroscientists and architects are using this enormous laboratory to make buildings better

13 September 2024 at 11:00

Have you ever found yourself lost in a building that felt impossible to navigate? Thoughtful building design should center on the people who will be using those buildings. But that’s no mean feat.

It’s not just about navigation, either. Just think of an office that left you feeling sleepy or unproductive, or perhaps a health center that had a less-than-reviving atmosphere. A design that works for some people might not work for others. People have different minds and bodies, and varying wants and needs. So how can we factor them all in?

To answer that question, neuroscientists and architects are joining forces at an enormous laboratory in East London—one that allows researchers to build simulated worlds. In this lab, scientists can control light, temperature, and sound. They can create the illusion of a foggy night, or the tinkle of morning birdsong.

And they can study how volunteers respond to these environments, whether they be simulations of grocery stores, hospitals, pedestrian crossings, or schools. That’s how I found myself wandering around a fake art gallery, wearing a modified baseball cap with a sensor that tracked my movements.

I first visited the Person-Environment-Activity Research Lab, referred to as PEARL, back in July. I’d been chatting to Hugo Spiers, a neuroscientist based at University College London, about the use of video games to study how people navigate. Spiers had told me he was working on another project: exploring how people navigate a lifelike environment, and how they respond during evacuations (which, depending on the situation, could be a matter of life or death).

For their research, Spiers and his colleagues set up what they call a “mocked-up art gallery” within PEARL. The center in its entirety is pretty huge as labs go, measuring around 100 meters in length and 40 meters across, with 10-meter-high ceilings in places. There’s no other research center in the world like this, Spiers told me.

The gallery setup looked a little like a maze from above, with a pathway created out of hanging black sheets. The exhibits themselves were videos of dramatic artworks that had been created by UCL students.

When I visited in July, Spiers and his colleagues were running a small pilot study to trial their setup. As a volunteer participant, I was handed a numbered black cap with a square board on top, marked with a large QR code. This code would be tracked by cameras above and around the gallery. The cap also carried a sensor, transmitting radio signals to devices around the maze that could pinpoint my location within a range of 15 centimeters.

At first, all the volunteers (most of whom seemed to be students) were asked to explore the gallery as we would any other. I meandered around, watching the videos, and eavesdropping on the other volunteers, who were chatting about their research and upcoming dissertation deadlines. It all felt pretty pleasant and calm.

That feeling dissipated in the second part of the experiment, when we were each given a list of numbers, told that each one referred to a numbered screen, and informed that we had to visit all the screens in the order in which they appeared on our lists. “Good luck, everybody,” Spiers said.

Suddenly everyone seemed to be rushing around, slipping past each other and trying to move quickly while avoiding collisions. “It’s all got a bit frantic, hasn’t it?” I heard one volunteer comment as I accidentally bumped into another. I hadn’t managed to complete the task by the time Spiers told us the experiment was over. As I walked to the exit, I noticed that some people were visibly out of breath.

The full study took place on Wednesday, September 11. This time, there were around 100 volunteers (I wasn’t one of them). And while almost everyone was wearing a modified baseball cap, some had more complicated gear, including EEG caps to measure brainwaves, or caps that use near-infrared spectroscopy to measure blood flow in the brain. Some people were even wearing eye-tracking devices that monitored which direction they were looking.

“We will do something quite remarkable today,” Spiers told the volunteers, staff, and observers as the experiment started. Taking such detailed measurements from so many individuals in such a setting represented “a world first,” he said.

I have to say that being an observer was much more fun than being a participant. Gone was the stress of remembering instructions and speeding around a maze. Here in my seat, I could watch as the data collected from the cameras and sensors was projected onto a screen. The volunteers, represented as squiggly colored lines, made their way through the gallery in a way that reminded me of the game Snake.

The study itself was similar to the pilot study, although this time the volunteers were given additional tasks. At one point, they were given an envelope with the name of a town or city in it, and asked to find others in the group who had been given the same one. It was fascinating to see the groups form. Some had the names of destination cities like Bangkok, while others had been assigned fairly nondescript English towns like Slough, made famous as the setting of the British television series The Office. At another point, the volunteers were asked to evacuate the gallery from the nearest exit.

The data collected in this study represents something of a treasure trove for researchers like Spiers and his colleagues. The team is hoping to learn more about how people navigate a space, and whether they move differently if they are alone or in a group. How do friends and strangers interact, and does this depend on whether they have certain types of material to bond over? How do people respond to evacuations—will they take the nearest exit as directed, or will they run on autopilot to the exit they used to enter the space in the first place?

All this information is valuable to neuroscientists like Spiers, but it’s also useful to architects like his colleague Fiona Zisch, who is based at UCL’s Bartlett School of Architecture. “We do really care about how people feel about the places we design for them,” Zisch tells me. The findings can guide not only the construction of new buildings, but also efforts to modify and redesign existing ones.

PEARL was built in 2021 and has already been used to help engineers, scientists, and architects explore how neurodivergent people use grocery stores, and the ideal lighting to use for pedestrian crossings, for example. Zisch herself is passionate about creating equitable spaces—particularly for health and education—that everyone can make use of in the best possible way.

In the past, models used in architecture have been developed with typically built, able-bodied men in mind. “But not everyone is a 6’2″ male with a briefcase,” Zisch tells me. Age, gender, height, and a range of physical and psychological factors can all influence how a person will use a building. “We want to improve not just the space, but the experience of the space,” says Zisch. Good architecture isn’t just about creating stunning features; it’s about subtle adaptations that might not even be noticeable to most people, she says.

The art gallery study is just the first step for researchers like Zisch and Spiers, who plan to explore other aspects of neuroscience and architecture in more simulated environments at PEARL. The team won’t have results for a while yet. But it’s a fascinating start. Watch this space.


Now read the rest of The Checkup

Read more from MIT Technology Review’s archive

Brain-monitoring technology has come a long way, and tech designed to read our minds and probe our memories is already being used. Futurist and legal ethicist Nita Farahany explained why we need laws to protect our cognitive liberty in a previous edition of The Checkup.

Listening in on the brain can reveal surprising insights into how this mysterious organ works. One team of neuroscientists found that our brains seem to oscillate between states of order and chaos.

Last year, MIT Technology Review published our design issue of the magazine. If you’re curious, this piece on the history and future of the word “design,” by Nicholas de Monchaux, head of architecture at MIT, might be a good place to start

Design covers much more than buildings, of course. Designers are creating new ways for users of prosthetic devices to feel more comfortable in their own skin—some of which have third thumbs, spikes, or “superhero skins.”

Achim Menges is an architect creating what he calls “self-shaping” structures with wood, which can twist and curve with changes in humidity. His approach is a low-energy way to make complex curved architectures, Menges told John Wiegand.

From around the web

Scientists are meant to destroy research samples of the poliovirus, as part of efforts to eradicate the disease it causes. But lab leaks of the virus may be more common than we’d like to think. (Science)

Neurofeedback allows people to watch their own brain activity in real time, and learn to control it. It could be a useful way to combat the impacts of stress. (Trends in Neurosciences)

Microbes, some of which cause disease in people, can travel over a thousand miles on wind, researchers have shown. Some appear to be able to survive their journey. (The Guardian)

Is the X chromosome involved in Alzheimer’s disease? A study of over a million people suggests so. (JAMA Neurology)

A growing number of men are paying thousands of dollars a year for testosterone therapies that are meant to improve their physical performance. But some are left with enlarged breasts, shrunken testicles, blood clots, and infertility. (The Wall Street Journal)

Chatbots can persuade people to stop believing in conspiracy theories

12 September 2024 at 20:00

The internet has made it easier than ever before to encounter and spread conspiracy theories. And while some are harmless, others can be deeply damaging, sowing discord and even leading to unnecessary deaths.

Now, researchers believe they’ve uncovered a new tool for combating false conspiracy theories: AI chatbots. Researchers from MIT Sloan and Cornell University found that chatting about a conspiracy theory with a large language model (LLM) reduced people’s belief in it by about 20%—even among participants who claimed that their beliefs were important to their identity. The research is published today in the journal Science.

The findings could represent an important step forward in how we engage with and educate people who espouse such baseless theories, says Yunhao (Jerry) Zhang, a postdoc fellow affiliated with the Psychology of Technology Institute who studies AI’s impacts on society.

“They show that with the help of large language models, we can—I wouldn’t say solve it, but we can at least mitigate this problem,” he says. “It points out a way to make society better.” 

Few interventions have been proven to change conspiracy theorists’ minds, says Thomas Costello, a research affiliate at MIT Sloan and the lead author of the study. Part of what makes it so hard is that different people tend to latch on to different parts of a theory. This means that while presenting certain bits of factual evidence may work on one believer, there’s no guarantee that it’ll prove effective on another.

That’s where AI models come in, he says. “They have access to a ton of information across diverse topics, and they’ve been trained on the internet. Because of that, they have the ability to tailor factual counterarguments to particular conspiracy theories that people believe.”

The team tested its method by asking 2,190 crowdsourced workers to participate in text conversations with GPT-4 Turbo, OpenAI’s latest large language model.

Participants were asked to share details about a conspiracy theory they found credible, why they found it compelling, and any evidence they felt supported it. These answers were used to tailor responses from the chatbot, which the researchers had prompted to be as persuasive as possible.

Participants were also asked to indicate how confident they were that their conspiracy theory was true, on a scale from 0 (definitely false) to 100 (definitely true), and then rate how important the theory was to their understanding of the world. Afterwards, they entered into three rounds of conversation with the AI bot. The researchers chose three to make sure they could collect enough substantive dialogue.

After each conversation, participants were asked the same rating questions. The researchers followed up with all the participants 10 days after the experiment, and then two months later, to assess whether their views had changed following the conversation with the AI bot. The participants reported a 20% reduction of belief in their chosen conspiracy theory on average, suggesting that talking to the bot had fundamentally changed some people’s minds.

“Even in a lab setting, 20% is a large effect on changing people’s beliefs,” says Zhang. “It might be weaker in the real world, but even 10% or 5% would still be very substantial.”

The authors sought to safeguard against AI models’ tendency to make up information—known as hallucinating—by employing a professional fact-checker to evaluate the accuracy of 128 claims the AI had made. Of these, 99.2% were found to be true, while 0.8% were deemed misleading. None were found to be completely false. 

One explanation for this high degree of accuracy is that a lot has been written about conspiracy theories on the internet, making them very well represented in the model’s training data, says David G. Rand, a professor at MIT Sloan who also worked on the project. The adaptable nature of GPT-4 Turbo means it could easily be connected to different platforms for users to interact with in the future, he adds.

“You could imagine just going to conspiracy forums and inviting people to do their own research by debating the chatbot,” he says. “Similarly, social media could be hooked up to LLMs to post corrective responses to people sharing conspiracy theories, or we could buy Google search ads against conspiracy-related search terms like ‘Deep State.’”

The research upended the authors’ preconceived notions about how receptive people were to solid evidence debunking not only conspiracy theories, but also other beliefs that are not rooted in good-quality information, says Gordon Pennycook, an associate professor at Cornell University who also worked on the project. 

“People were remarkably responsive to evidence. And that’s really important,” he says. “Evidence does matter.”

Google’s new tool lets large language models fact-check their responses

12 September 2024 at 15:00

As long as chatbots have been around, they have made things up. Such “hallucinations” are an inherent part of how AI models work. However, they’re a big problem for companies betting big on AI, like Google, because they make the responses it generates unreliable. 

Google is releasing a tool today to address the issue. Called DataGemma, it uses two methods to help large language models fact-check their responses against reliable data and cite their sources more transparently to users. 

The first of the two methods is called Retrieval-Interleaved Generation (RIG), which acts as a sort of fact-checker. If a user prompts the model with a question—like “Has the use of renewable energy sources increased in the world?”—the model will come up with a “first draft” answer. Then RIG identifies what portions of the draft answer could be checked against Google’s Data Commons, a massive repository of data and statistics from reliable sources like the United Nations or the Centers for Disease Control and Prevention. Next, it runs those checks and replaces any incorrect original guesses with correct facts. It also cites its sources to the user.

The second method, which is commonly used in other large language models, is called Retrieval-Augmented Generation (RAG). Consider a prompt like “What progress has Pakistan made against global health goals?” In response, the model examines which data in the Data Commons could help it answer the question, such as information about access to safe drinking water, hepatitis B immunizations, and life expectancies. With those figures in hand, the model then builds its answer on top of the data and cites its sources.

“Our goal here was to use Data Commons to enhance the reasoning of LLMs by grounding them in real-world statistical data that you could source back to where you got it from,” says Prem Ramaswami, head of Data Commons at Google. Doing so, he says, will “create more trustable, reliable AI.”

It is only available to researchers for now, but Ramaswami says access could widen further after more testing. If it works as hoped, it could be a real boon for Google’s plan to embed AI deeper into its search engine.  

However, it comes with a host of caveats. First, the usefulness of the methods is limited by whether the relevant data is in the Data Commons, which is more of a data repository than an encyclopedia. It can tell you the GDP of Iran, but it’s unable to confirm the date of the First Battle of Fallujah or when Taylor Swift released her most recent single. In fact, Google’s researchers found that with about 75% of the test questions, the RIG method was unable to obtain any usable data from the Data Commons. And even if helpful data is indeed housed in the Data Commons, the model doesn’t always formulate the right questions to find it. 

Second, there is the question of accuracy. When testing the RAG method, researchers found that the model gave incorrect answers 6% to 20% of the time. Meanwhile, the RIG method pulled the correct stat from Data Commons only about 58% of the time (though that’s a big improvement over the 5% to 17% accuracy rate of Google’s large language models when they’re not pinging Data Commons). 

Ramaswami says DataGemma’s accuracy will improve as it gets trained on more and more data. The initial version has been trained on only about 700 questions, and fine-tuning the model required his team to manually check each individual fact it generated. To further improve the model, the team plans to increase that data set from hundreds of questions to millions.

The Download: Ukraine’s drone defenses, and today’s climate heroes

12 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Meet the radio-obsessed civilian shaping Ukraine’s drone defense

Drones have come to define the brutal conflict in Ukraine that has now dragged on for more than two and a half years. And most rely on radio communications—a technology that Serhii “Flash” Beskrestnov has obsessed over since childhood.

While Flash is now a civilian, the former officer has still taken it upon himself to inform his country’s defense in all matters related to radio. Once a month, he studies the skies for Russian radio transmissions and tries to learn about the problems facing troops in the fields and in the trenches.

In this race for survival—as each side constantly tries to best the other, only to start all over again when the other inevitably catches up—Ukrainian soldiers need to develop creative solutions, and fast. As Ukraine’s wartime radio guru, Flash may just be one of their best hopes for doing that. Read the full story.

—Charlie Metcalfe

Meet 2024’s climate innovators under 35

One way to know where a field is going? Take a look at what the sharpest new innovators are working on.

Good news for all of us: MIT Technology Review’s list of 35 Innovators Under 35 just dropped. A decent number of the people who made the list are working in fields that touch climate and energy in one way or another. And our senior climate reporter Casey Crownhart noticed a few trends that might provide some hints about the future. Read the full story.

This year’s list is available exclusively to MIT Technology Review subscribers. If you’re not a subscriber already, you sign up here with a 25% discount on the usual price.

This story is from The Spark, our weekly climate and energy newsletter. Sign up to receive it in your inbox every Wednesday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 The first commercial spacewalk by private citizens is underway
And, thus far, it’s been a success. (CNN)
+ Take a look at the long and illustrious history of spacewalks. (BBC)

2 Silicon Valley is divided over California’s AI safety bill
Big Tech is waiting anxiously for the state’s governor to make a decision. (FT $)
+ What’s next for AI regulation? (MIT Technology Review)

3 Wildfires are raging across southern California
The state has weathered nearly three times as much acreage burn this year so far compared to the whole of 2023. (The Guardian)
+ Canada’s 2023 wildfires produced more emissions than fossil fuels in most countries. (MIT Technology Review)

4 Broken wind turbines have major repercussions
Multiple offshore wind projects have run into serious trouble. (NYT $)

5 The percentage of women in tech has hardly changed in 20 year
Women and people of color face an uphill battle to get hired. (WP $)
+ Why can’t tech fix its gender problem? (MIT Technology Review)

6 Google’s new app can turn your research into an AI podcast
Please don’t do this, though. (The Verge)

7 Human drivers keep crashing into Waymo robotaxis
The company has launched a new website to put the incidents into perspective.(Ars Technica)
+ What’s next for robotaxis in 2024. (MIT Technology Review)

8 This tiny SpaceX rival is poised to launch its first satellites
AST SpaceMobile’s star appears to be on the rise—but for how long?(Bloomberg $)

9 You’ve got a fax 📠
Pagers, fax machines and dumbphones are all the rage these days. (WSJ $)

10 Have we reached peak emoji? 😲
The little pictograms are an illustrative language, not an ideographic one. (The Atlantic $)

Quote of the day

“A beautiful world.”

—Billionaire businessman Jared Isaacman’s reaction as he saw Earth from space during the first privately funded spacewalk today, the BBC reports.

The big story

What does GPT-3 “know” about me?

August 2022

One of the biggest stories in tech is the rise of large language models that produce text that reads like a human might have written it.

These models’ power comes from being trained on troves of publicly available human-created text hoovered up from the internet. If you’ve posted anything even remotely personal in English on the internet, chances are your data might be part of some of the world’s most popular LLMs.

Melissa Heikkilä, MIT Technology Review’s senior AI reporter, wondered what data these models might have on her—and how it could be misused. So she put OpenAI’s GPT-3 to the test. Read about what she found.

In this section yesterday we stated that Amazon had acquired iRobot. This was incorrect—the acquisition never completed. We apologize for the error.

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ These photos of London taken on a Casio camera watch are a snapshot of bygone times.
+ If you’ve noticed elaborate painted nails making their way into your cookbooks, it’s part of a wider trend. 💅
+ Painting Paint, now that’s meta.
+ Wow, enthusiastic skeletons are already limbering up for next month!

Meet the radio-obsessed civilian shaping Ukraine’s drone defense

12 September 2024 at 11:00

Serhii “Flash” Beskrestnov hates going to the front line. The risks terrify him. “I’m really not happy to do it at all,” he says. But to perform his particular self-appointed role in the Russia-Ukraine war, he believes it’s critical to exchange the relative safety of his suburban home north of the capital for places where the prospect of death is much more immediate. “From Kyiv,” he says, “nobody sees the real situation.”

So about once a month, he drives hundreds of kilometers east in a homemade mobile intelligence center: a black VW van in which stacks of radio hardware connect to an array of antennas on the roof that stand like porcupine quills when in use. Two small devices on the dash monitor for nearby drones. Over several days at a time, Flash studies the skies for Russian radio transmissions and tries to learn about the problems facing troops in the fields and in the trenches.

He is, at least in an unofficial capacity, a spy. But unlike other spies, Flash does not keep his work secret. In fact, he shares the results of these missions with more than 127,000 followers—including many soldiers and government officials—on several public social media channels. Earlier this year, for instance, he described how he had recorded five different Russian reconnaissance drones in a single night—one of which was flying directly above his van.

“Brothers from the Armed Forces of Ukraine, I am trying to inspire you,” he posted on his Facebook page in February, encouraging Ukrainian soldiers to learn how to recognize enemy drone signals as he does. “You will spread your wings, you will understand over time how to understand distance and, at some point, you will save the lives of dozens of your colleagues.”

Drones have come to define the brutal conflict that has now dragged on for more than two and a half years. And most rely on radio communications—a technology that Flash has obsessed over since childhood. So while Flash is now a civilian, the former officer has still taken it upon himself to inform his country’s defense in all matters related to radio.

As well as the frontline information he shares on his public channels, he runs a “support service” for almost 2,000 military communications specialists on Signal and writes guides for building anti-drone equipment on a tight budget. “He’s a celebrity,” one special forces officer recently shouted to me over the thump of music in a Kyiv techno club. He’s “like a ray of sun,” an aviation specialist in Ukraine’s army told me. Flash tells me that he gets 500 messages every day asking for help.

Despite this reputation among rank-and-file service members—and maybe because of it—Flash has also become a source of some controversy among the upper echelons of Ukraine’s military, he tells me. The Armed Forces of Ukraine declined multiple requests for comment, but Flash and his colleagues claim that some high-ranking officials perceive him as a security threat, worrying that he shares too much information and doesn’t do enough to secure sensitive intel. As a result, some refuse to support or engage with him. Others, Flash says, pretend he doesn’t exist. Either way, he believes they are simply insecure about the value of their own contributions—“because everybody knows that Serhii Flash is not sitting in Kyiv like a colonel in the Ministry of Defense,” he tells me in the abrasive fashion that I’ve come to learn is typical of his character. 

But above all else, hours of conversations with numerous people involved in Ukraine’s defense, including frontline signalmen and volunteers, have made clear that even if Flash is a complicated figure, he’s undoubtedly an influential one. His work has become greatly important to those fighting on the ground, and he recently received formal recognition from the military for his contributions to the fight, with two medals of commendation—one from the commander of Ukraine’s ground forces, the other from the Ministry of Defense. 

With a handheld directional antenna and a spectrum analyzer, Flash can scan for hostile signals.
EMRE ÇAYLAK

Despite a small number of semi-autonomous machines with a reduced reliance on radio communications, the drones that saturate the skies above the battlefield will continue to largely depend on this technology for the foreseeable future. And in this race for survival—as each side constantly tries to best the other, only to start all over again when the other inevitably catches up—Ukrainian soldiers need to develop creative solutions, and fast. As Ukraine’s wartime radio guru, Flash may just be one of their best hopes for doing that. 

“I know nothing about his background,” says “Igrok,” who works with drones in Ukraine’s 110th Mechanized Brigade and whom we are identifying by his call sign, as is standard military practice. “But I do know that most engineers and all pilots know nothing about radios and antennas. His job is definitely one of the most powerful forces keeping Ukraine’s aerial defense in good condition.”

And given the mounting evidence that both militaries and militant groups in other parts of the world are now adopting drone tactics developed in Ukraine, it’s not only his country’s fate that Flash may help to determine—but also the ways that armies wage war for years to come.

A prescient hobby

Before I can even start asking questions during our meeting in May, Flash is rummaging around in the back of the Flash-mobile, pulling out bits of gear for his own version of show-and-tell: a drone monitor with a fin-shaped antenna; a walkie-talkie labeled with a sticker from Russia’s state security service, the FSB; an approximately 1.5-meter-long foldable antenna that he says probably came from a US-made Abrams tank.

Flash has parked on a small wooded road beside the Kyiv Sea, an enormous water reservoir north of the capital. He’s wearing a khaki sweat-wicking polo shirt, combat trousers, and combat boots, with a Glock 19 pistol strapped to his hip. (“I am a threat to the enemy,” he tells me, explaining that he feels he has to watch his back.) As we talk, he moves from one side to the other, as if the electromagnetic waves that he’s studied since childhood have somehow begun to control the motion of his body.

Now 49, Flash grew up in a suburb of Kyiv in the ’80s. His father, who was a colonel in the Soviet army, recalls bringing home broken radio equipment for his preteen son to tinker with. Flash showed talent from the start. He attended an after-school radio club, and his father fixed an antenna to the roof of their apartment for him. Later, Flash began communicating with people in countries beyond the Iron Curtain. “It was like an open door to the big world for me,” he says.

Flash recalls with amusement a time when a letter from the KGB arrived at his family home, giving his father the fright of his life. His father didn’t know that his son had sent a message on a prohibited radio frequency, and someone had noticed. Following the letter, when Flash reported to the service’s office in downtown Kyiv, his teenage appearance confounded them. Boy, what are you doing here? Flash recalls an embarrassed official saying. 

Ukraine had been a hub of innovation as part of the Soviet Union. But by the time Flash graduated from military communications college in 1997, Ukraine had been independent for six years, and corruption and a lack of investment had stripped away the armed forces’ former grandeur. Flash spent just a year working in a military radio factory before he joined a private communications company developing Ukraine’s first mobile network, where he worked with technologies far more advanced than what he had used in the military. The  project was called “Flash.” 

A decade and a half later, Flash had risen through the ranks of the industry to become head of department at the progenitor to the telecommunications company Vodafone Ukraine. But boredom prompted him to leave and become an entrepreneur. His many projects included a successful e-commerce site for construction services and a popular video game called Isotopium: Chernobyl, which he and a friend based on the “really neat concept,” according to a PC Gamer review, of allowing players to control real robots (fitted with radios, of course) around a physical arena. Released in 2019, it also received positive reviews from Reuters and BBC News.

But within just a few years, an unexpected attack would hurl his country into chaos—and upend Flash’s life. 

“I am here to help you with technical issues,” Flash remembers writing to his Signal group when he first started offering advice. “Ask me anything and I will try to find the answer for you.”
EMRE ÇAYLAK

By early 2022, rumors were growing of a potential attack from Russia. Though he was still working on Isotopium, Flash began to organize a radio network across the northern suburbs of Kyiv in preparation. Near his home, he set up a repeater about 65 meters above ground level that could receive and then rebroadcast transmissions from all the radios in its network across a 200-square-kilometer area. Another radio amateur programmed and distributed handheld radios.

When Russian forces did invade, on February 24, they took both fiber-optic and mobile networks offline, as Flash had anticipated. The radio network became the only means of instant communications for civilians and, critically, volunteers mobilizing to fight in the region, who used it to share information about Russian troop movements. Flash fed this intel to several professional Ukrainian army units, including a unit of special reconnaissance forces. He later received an award from the head of the district’s military administration for his part in Kyiv’s defense. The head of the district council referred to Flash as “one of the most worthy people” in the region.

Yet it was another of Flash’s projects that would earn him renown across Ukraine’s military.

Despite being more than 100 years old, radio technology is still critical in almost all aspects of modern warfare, from secure communications to satellite-guided missiles. But the decline of Ukraine’s military, coupled with the movement of many of the country’s young techies into lucrative careers in the growing software industry, created a vacuum of expertise. Flash leaped in to fill it.

Within roughly a month of Russia’s incursion, Flash had created a private group called “Military Signalmen” on the encrypted messaging platform Signal, and invited civilian radio experts from his personal network to join alongside military communications specialists. “I am here to help you with technical issues,” he remembers writing to the group. “Ask me anything and I will try to find the answer for you.”

The kinds of questions that Flash and his civilian colleagues answered in the first months were often basic. Group members wanted to know how to update the firmware on their devices, reset their radios’ passwords, or set up the internal communications networks for large vehicles. Many of the people drafted as communications specialists in the Ukrainian military had little relevant experience; Flash claims that even professional soldiers lacked appropriate training and has referred to large parts of Ukraine’s military communications courses as “either nonsense or junk.” (The Korolov Zhytomyr Military Institute, where many communications specialists train, declined a request for comment.)

After Russia’s invasion of Ukraine, Flash transformed his VW van into a mobile radio intelligence center.
EMRE ÇAYLAK

He demonstrates handheld spectrum analyzers with custom Ukrainian firmware.

News of the Signal group spread by word of mouth, and it soon became a kind of 24-hour support service that communications specialists in every sector of Ukraine’s frontline force subscribed to. “Any military engineer can ask anything and receive the answer within a couple of minutes,” Flash says. “It’s a nice way to teach people very quickly.” 

As the war progressed into its second year, Military Signalmen became, to an extent, self-sustaining. Its members had learned enough to answer one another’s questions themselves. And this is where several members tell me that Flash has contributed the most value. “The most important thing is that he brought together all these communications specialists in one team,” says Oleksandr “Moto,” a technician at an EU mission in Kyiv and an expert in Motorola equipment, who has advised members of the group. (He asked to not be identified by his surname, due to security concerns.) “It became very efficient.”

Today, Flash and his partners continue to answer occasional questions that require more advanced knowledge. But over the past year, as the group demanded less of his time, Flash has begun to focus on a rapidly proliferating weapon for which his experience had prepared him almost perfectly: the drone.  

A race without end

The Joker-10 drone, one of Russia’s latest additions to its arsenal, is equipped with a hibernation mechanism, Flash warned his Facebook followers in March. This feature allows the operator to fly it to a hidden location, leave it there undetected, and then awaken it when it’s time to attack. “It is impossible to detect the drone using radio-electronic means,” Flash wrote. “If you twist and turn it in your hands—it will explode.” 

This is just one example of the frequent developments in drone engineering that Ukrainian and Russian troops are adapting to every day. 

Larger strike drones similar to the US-made Reaper have been familiar in other recent conflicts, but sophisticated air defenses have rendered them less dominant in this war. Ukraine and Russia are developing and deploying vast numbers of other types of drones—including the now-notorious “FPV,” or first-person view, drone that pilots operate by wearing goggles that stream video of its perspective. These drones, which can carry payloads large enough to destroy tanks, are cheap (costing as little as $400), easy to produce, and difficult to shoot down. They use direct radio communications to transmit video feeds, receive commands, and navigate.

""
A Ukrainian soldier prepares an FPV drone equipped with dummy ammunition for a simulated flight operation.
MARCO CORDONE/SOPA IMAGES/SIPA USA VIA AP IMAGES

But their reliance on radio technology is a major vulnerability, because enemies can disrupt the signals that the drones emit—making them far less effective, if not inoperable. This form of electronic warfare—which most often involves emitting a more powerful signal at the same frequency as the operator’s—is called “jamming.”

Jamming, though, is an imperfect solution. Like drones, jammers themselves emit radio signals that can enable enemies to locate them. There are also effective countermeasures to bypass jammers. For example, a drone operator can use a tactic called “frequency hopping,” rapidly jumping between different frequencies to avoid a jammer’s signal. But even this method can be disrupted by algorithms that calculate the hopping patterns.

For this reason, jamming is a frequent focus of Flash’s work. In a January post on his Telegram channel, for instance, which people viewed 48,000 times, Flash explained how jammers used by some Ukrainian tanks were actually disrupting their own communications. “The cause of the problems is not direct interference with the reception range of the radio station, but very powerful signals from several [electronic warfare] antennae,” he wrote, suggesting that other tank crews experiencing the same problem might try spreading their antennas across the body of the tank. 

It is all part of an existential race in which Russia and Ukraine are constantly hunting for new methods of drone operation, drone jamming, and counter-jamming—and there’s no end in sight. In March, for example, Flash says, a frontline contact sent him photos of a Russian drone with what looks like a 10-kilometer-long spool of fiber-optic cable attached to its rear—one particularly novel method to bypass Ukrainian jammers. “It’s really crazy,” Flash says. “It looks really strange, but Russia showed us that this was possible.”

Flash’s trips to the front line make it easier for him to track developments like this. Not only does he monitor Russian drone activity from his souped-up VW, but he can study the problems that soldiers face in situ and nurture relationships with people who may later send him useful intel—or even enemy equipment they’ve seized. “The main problem is that our generals are located in Kyiv,” Flash says. “They send some messages to the military but do not understand how these military people are fighting on the front.”

Besides the advice he provides to Ukrainian troops, Flash also publishes online his own manuals for building and operating equipment that can offer protection from drones. Building their own tools can be soldiers’ best option, since Western military technology is typically expensive and domestic production is insufficient. Flash recommends buying most of the parts on AliExpress, the Chinese e-commerce platform, to reduce costs.

While all his activity suggests a close or at least cooperative relationship between Flash and Ukraine’s military, he sometimes finds himself on the outside looking in. In a post on Telegram in May, as well as during one of our meetings, Flash shared one of his greatest disappointments of the war: the military’s refusal of his proposal to create a database of all the radio frequencies used by Ukrainian forces. But when I mentioned this to an employee of a major electronic warfare company, who requested anonymity to speak about the sensitive subject, he suggested that the only reason Flash still complains about this is that the military hasn’t told him it already exists. (Given its sensitivity, MIT Technology Review was unable to independently confirm the existence of this database.) 

Flash believes that generals in Kyiv “do not understand how these military people are fighting on the front.” So even though he doesn’t like the risks they involve, he takes trips to the frontline about once a month.
EMRE ÇAYLAK

This anecdote is emblematic of Flash’s frustration with a military complex that may not always want his involvement. Ukraine’s armed forces, he has told me on several occasions, make no attempt to collaborate with him in an official manner. He claims not to receive any financial support, either. “I’m trying to help,” he says. “But nobody wants to help me.”

Both Flash and Yurii Pylypenko, another radio enthusiast who helps Flash manage his Telegram channel, say military officials have accused Flash of sharing too much information about Ukraine’s operations. Flash claims to verify every member of his closed Signal groups, which he says only discuss “technical issues” in any case. But he also admits the system is not perfect and that Russians could have gained access in the past. Several of the soldiers I interviewed for this story also claimed to have entered the groups without Flash’s verification process. 

It’s ultimately difficult to determine if some senior staff in the military hold Flash at arm’s length because of his regular, often strident criticism—or whether Flash’s criticism is the result of being held at arm’s length. But it seems unlikely either side’s grievances will subside soon; Pylypenko claims that senior officers have even tried to blackmail him over his involvement in Flash’s work. “They blame my help,” he wrote to me over Telegram, “because they think Serhii is a Russian agent reposting Russian propaganda.” 

Is the world prepared?

Flash’s greatest concern now is the prospect of Russia overwhelming Ukrainian forces with the cheap FPV drones. When they first started deploying FPVs, both sides were almost exclusively targeting expensive equipment. But as production has increased, they’re now using them to target individual soldiers, too. Because of Russia’s production superiority, this poses a serious danger—both physical and psychological—to Ukrainian soldiers. “Our army will be sitting under the ground because everybody who goes above ground will be killed,” Flash says. Some reports suggest that the prevalence of FPVs is already making it difficult for soldiers to expose themselves at all on the battlefield.

To combat this threat, Flash has a grand yet straightforward idea. He wants Ukraine to build a border “wall” of jamming systems that cover a broad range of the radio spectrum all along the front line. Russia has already done this itself with expensive vehicle-based systems, but these present easy targets for Ukrainian drones, which have destroyed several of them. Flash’s idea is to use a similar strategy, albeit with smaller, cheaper systems that are easier to replace. He claims, however, that military officials have shown no interest.

Although Flash is unwilling to divulge more details about this strategy (and who exactly he pitched it to), he believes that such a wall could provide a more sustainable means of protecting Ukrainian troops. Nevertheless, it’s difficult to say how long such a defense might last. Both sides are now in the process of developing artificial-intelligence programs that allow drones to lock on to targets while still outside enemy jamming range, rendering them jammer-proof when they come within it. Flash admits he is concerned—and he doesn’t appear to have a solution.

Flash admits he is worried about Russia overwhelming Ukrainian forces with the cheap FPV drones: “Our army will be sitting under the ground because everybody who goes above ground will be killed.”
EMRE ÇAYLAK

He’s not alone. The world is entirely unprepared for this new type of warfare, says Yaroslav Kalinin, a former Ukrainian intelligence officer and the CEO of Infozahyst, a manufacturer of equipment for electronic warfare. Kalinin recounts talking at an electronic-warfare-focused conference in Washington, DC, last December where representatives from some Western defense companies weren’t able to recognize the basic radio signals emitted by different types of drones. “Governments don’t count [drones] as a threat,” he says. “I need to run through the streets like a prophet—the end is near!”

Nevertheless, Ukraine has become, in essence, a laboratory for a new era of drone warfare—and, many argue, a new era of warfare entirely. Ukraine’s and Russia’s soldiers are its technicians. And Flash, who sometimes sleeps curled up in the back of his van while on the road, is one of its most passionate researchers. “Military developers from all over the world come to us for experience and advice,” he says. Only time will tell whether their contributions will be enough to see Ukraine through to the other side of this war. 

Charlie Metcalfe is a British journalist. He writes for magazines and newspapers, including Wired, the Guardian, and MIT Technology Review.

Meet 2024’s climate innovators under 35

12 September 2024 at 11:00

This article is from The Spark, MIT Technology Review’s weekly climate newsletter. To receive it in your inbox every Wednesday, sign up here.

One way to know where a field is going? Take a look at what the sharpest new innovators are working on.

Good news for all of us: MIT Technology Review’s list of 35 Innovators Under 35 just dropped. And a decent number of the people who made the list are working in fields that touch climate and energy in one way or another.

Looking through, I noticed a few trends that might provide some hints about the future of climate tech. Let’s dig into this year’s list and consider what these innovators’ work might mean for efforts to combat climate change.

Power to the people

Perhaps unsurprisingly, quite a few innovators on this list are working on energy—and many of them have an interest in making energy consistently available where and when it’s needed. Wind and solar are getting cheap, but we need solutions for when the sun isn’t shining and the wind isn’t blowing.

Tim Latimer cofounded Fervo Energy, a geothermal company hoping to provide consistently available, carbon-free energy using Earth’s heat. You may be familiar with his work, since Fervo was on our list of 15 Climate Tech Companies to Watch in 2023.

Another energy-focused innovator on the list is Andrew Ponec of Antora Energy, a company working to build thermal energy storage systems. Basically, the company’s technology heats up blocks when cheap renewables are available, and then stores that heat and delivers it to industrial processes that need constant power. (You, the readers, named thermal energy storage the readers’ choice on this year’s 10 Breakthrough Technologies list.)

Rock stars

While new ways of generating electricity and storing energy can help cut our emissions in the future, other people are focused on how to clean up the greenhouse gases already in the atmosphere. At this point, removing carbon dioxide from the atmosphere is basically required for any scenario where we limit warming to 1.5 °C over preindustrial levels. A few of the new class of innovators are turning to rocks for help soaking up and locking away atmospheric carbon. 

Noah McQueen cofounded Heirloom Carbon Technologies, a carbon removal company. The technology works by tweaking the way minerals soak up carbon dioxide from the air (before releasing it under controlled conditions, so they can do it all again). The company has plans for facilities that could remove hundreds of thousands of tons of carbon dioxide each year. 

Another major area of research focuses on how we might store captured carbon dioxide. Claire Nelson is the cofounder of Cella Mineral Storage, a company working on storage methods to better trap carbon dioxide underground once it’s been mopped up.  

Material world

Finally, some of the most interesting work on our new list of innovators is in materials. Some people are finding new ones that could help us address our toughest problems, and others are trying to reinvent old ones to clean up their climate impacts.

Julia Carpenter found a way to make a foam-like material from metal. Its high surface area makes it a stellar heat sink, meaning it can help cool things down efficiently. It could be a huge help in data centers, where 40% of energy demand goes to cooling.

And I spoke with Cody Finke, cofounder and CEO of Brimstone, a company working on cleaner ways of making cement. Cement alone is responsible for nearly 7% of global greenhouse-gas emissions, and about half of those come from chemical reactions necessary to make it. Finke and Brimstone are working to wipe out the need for these reactions by using different starting materials to make this crucial infrastructural glue.

Addressing climate change is a sprawling challenge, but the researchers and founders on this list are tackling a few of the biggest issues I think about every day. 

Ensuring that we can power our grid, and all the industrial processes that we rely on for the stuff in our daily lives, is one of the most substantial remaining challenges. Removing carbon dioxide from the atmosphere in an efficient, cheap process could help limit future warming and buy us time to clean up the toughest sectors. And finding new materials, and new methods of producing old ones, could be a major key to unlocking new climate solutions. 

To read more about the folks I mentioned here and other innovators working in climate change and beyond, check out the full list.


Now read the rest of The Spark

Related reading

Fervo Energy (cofounded by 2024 innovator Tim Latimer) showed last year that its wells can be used like a giant underground battery.

A growing number of companies—including Antora Energy, whose CEO Andrew Ponec is a 2024 innovator—are working to bring thermal energy storage systems to heavy industry.

Cement is one of our toughest challenges, as Brimstone CEO and 2024 innovator Cody Finke will tell you. I wrote about Brimstone and other efforts to reinvent cement earlier this year.

A plant with yellow flowers

Another thing

We need a whole lot of metals to address climate change, from the copper in transmission lines to the nickel in lithium-ion batteries that power electric vehicles. Some researchers think plants might be able to help. 

Roughly 750 species of plants are so-called hyperaccumulators, meaning they naturally soak up and tolerate relatively high concentrations of metal. A new program is funding research into how we might use this trait to help source nickel, and potentially other metals, in the future. Read the full story here.

Keeping up with climate  

A hurricane that recently formed in the Gulf of Mexico is headed for Louisiana, ending an eerily quiet few weeks of the season. (Scientific American)

→ After forecasters predicted a particularly active season, the lull in hurricane activity was surprising. (New Scientist)

Rising sea levels are one of the symptoms of a changing climate, but nailing down exactly what “sea level” means is more complicated than you might think. We’ve gotten better at measuring sea level over the past few centuries, though. (New Yorker)

The US Department of Energy’s Loan Programs Office has nearly $400 million in lending authority. This year’s election could shift the focus of that office drastically, making it a bellwether of how the results could affect energy priorities. (Bloomberg)

What if fusion power ends up working, but it’s too expensive to play a significant role on the grid? Some modelers think the technology will remain expensive and could come too late to make a dent in emissions. (Heatmap)

Electric-vehicle sales are up overall, but some major automakers are backing away from goals on zero-emissions vehicles. Even though sales are increasing, uptake is slower than many thought it would be, contributing to the nervous energy in the industry. (Canary Media)

It’s a tough time to be in the business of next-generation batteries. The woes of three startups reveal that difficult times are here, likely for a while. (The Information)

The Download: a quantum breakthrough, and the Internet Archive ruling

11 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Google says it’s made a quantum computing breakthrough that reduces errors

The news: Google researchers claim to have made a breakthrough in quantum error correction, one that could pave the way for quantum computers that finally live up to the technology’s promise.

Why it matters: One major challenge facing the field has been that quantum computers can store or manipulate information incorrectly, preventing them from executing algorithms that are long enough to be useful. 

The new research from Google Quantum AI and its academic collaborators demonstrates that they can add components to reduce these errors. Ultimately, it bolsters the idea that error correction is a viable strategy toward building a useful quantum computer. Read the full story.

—Sophia Chen

Why a ruling against the Internet Archive threatens the future of America’s libraries

—Chris Lewis is president and CEO of Public Knowledge, a consumer advocacy group that works to shape technology policy in the public interest.

In the 1980s and ‘90s, it didn’t matter if you didn’t have a computer or your parents lacked infinite money for tutors—you could get a lifetime’s education for free at the public library.

A ruling from the US Second Circuit against the Internet Archive and in favor of publisher Hachette has just thrown that promise of equality into doubt by limiting libraries’ access to digital lending. Read this to learn why.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 OpenAI’s new reasoning AI model is coming
The Strawberry model is reportedly planned for release within a fortnight. (The Information $)+ It’s an area of research that Google DeepMind is also invested in. (MIT Technology Review)

2 Another human has contracted bird flu in the US
And worryingly, they don’t seem to have been in direct contact with animals. (Vox)
+ How worried should we be, really? (The Atlantic $)
+ What’s next for bird flu vaccines. (MIT Technology Review)

3 A US legal advisor coalition wants warning labels for social media 
The 42-strong attorney general group is urging Congress to take action. (WP $)
+ Australia is planning to introduce a minimum age limit for social media use. (BBC)
+ Should social media come with a health warning? (MIT Technology Review)

4 How 9/11 changed the internet
It shaped how we talk—and to some people’s distaste, joke—about national tragedies online. (Insider $)

5 Huawei has announced a triple-folding smartphone
The $2,800 Mate XT folds up like a pamphlet. (FT $)
+ If you want more memory, its price tag rises to an eye-watering $3,300. (Reuters)

6 Caroline Ellison is likely to receive a sentence soon
The key FTX case witness has pleaded guilty to fraud and conspiracy charges.(NY Mag $)
+ She’s seeking no prison time as a result of her cooperation in the trial. (Boston Globe $)

7 Satellites are at risk from “killer electrons”
Luckily, a secretive radio wave method could help safeguard them. (Economist $)

8 Researchers have created a cloud atlas of Mars
While some formations are similar to Earth’s, others are completely different. (New Scientist $)

9 Kamala Harris supporters are using Trump’s weirdest quotes against him
A new platform catalogs all of his strangest missives in real time. (Fast Company $)

10 The British are coming! 🇬🇧
Britishcore is the latest tongue-in-cheek trend to grab Gen Z’s attention. (The Guardian)

Quote of the day

“Parents want their kids off their phones and on the footy field and so do I.”

—Anthony Albanese, Australia’s prime minister, explains the rationale behind his government’s plans to restrict social media access for teenagers and children, the Financial Times reports.

The big story

A Roomba recorded a woman on the toilet. How did screenshots end up on Facebook?

December 2022

In the fall of 2020, gig workers in Venezuela posted a series of images to online forums where they talk shop. The photos were mundane, if sometimes intimate, household scenes—including a particularly revealing shot of a young woman in a lavender T-shirt sitting on the toilet, her shorts pulled down to mid-thigh.

The images were not taken by a person, but by development versions of iRobot’s Roomba robot vacuum. They were then sent to Scale AI, a startup that contracts workers around the world to label data used to train artificial intelligence.

Earlier this year, MIT Technology Review obtained 15 screenshots of these private photos, which had been posted to closed social media groups. The images speak to the growing practice of sharing potentially sensitive data to train algorithms. They also reveal a whole data supply chain—and new points where personal information could leak out—that few consumers are even aware of. Read the full story.

—Eileen Guo

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ What do astronauts and deep sea divers have in common? Their training is surprisingly similar.
+ Eggs, eggs eggs—who doesn’t love eggs? 🍳
+ Better gut health is something we should all be aiming for. Here’s where to start.
+ Single women of TikTok, we salute you.

Correction: The piece has been updated to remove a reference to Amazon’s acquisition of iRobot, which was never completed.

To be more useful, robots need to become lazier

9 September 2024 at 17:12

Robots perceive the world around them very differently from the way humans do. 

When we walk down the street, we know what we need to pay attention to—passing cars, potential dangers, obstacles in our way—and what we don’t, like pedestrians walking in the distance. Robots, on the other hand, treat all the information they receive about their surroundings with equal importance. Driverless cars, for example, have to continuously analyze data about things around them whether or not they are relevant. This keeps drivers and pedestrians safe, but it draws on a lot of energy and computing power. What if there’s a way to cut that down by teaching robots what they should prioritize and what they can safely ignore?

That’s the principle underpinning “lazy robotics,” a field of study championed by René van de Molengraft, a professor at Eindhoven University of Technology in the Netherlands. He believes that teaching all kinds of robots to be “lazier” with their data could help pave the way for machines that are better at interacting with things in their real-world environments, including humans. Essentially, the more efficient a robot can be with information, the better.

Van de Molengraft’s lazy robotics is just one approach researchers and robotics companies are now taking as they train their robots to complete actions successfully, flexibly, and in the most efficient manner possible.

Teaching them to be smarter when they sift through the data they gather and then de-prioritize anything that’s safe to overlook will help make them safer and more reliable—a long-standing goal of the robotics community.

Simplifying tasks in this way is necessary if robots are to become more widely adopted, says Van de Molengraft, because their current energy usage won’t scale—it would be prohibitively expensive and harmful to the environment. “I think that the best robot is a lazy robot,” he says. “They should be lazy by default, just like we are.”

Learning to be lazier

Van de Molengraft has hit upon a fun way to test these efforts out: teaching robots to play soccer. He recently led his university’s autonomous robot soccer team, Tech United, to victory at RoboCup, an annual international robotics and AI competition that tests robots’ skills on the soccer field. Soccer is a tough challenge for robots, because both scoring and blocking goals require quick, controlled movements, strategic decision-making, and coordination. 

Learning to focus and tune out distractions around them, much as the best human soccer players do, will make them not only more energy efficient (especially for robots powered by batteries) but more likely to make smarter decisions in dynamic, fast-moving situations.

Tech United’s robots used several “lazy” tactics to give them an edge over their opponents during the RoboCup. One approach involved creating a “world model” of a soccer pitch that identifies and maps out its layout and line markings—things that remain the same throughout the game. This frees the battery-powered robots from constantly scanning their surroundings, which would waste precious power. Each robot also shares what its camera is capturing with its four teammates, creating a broader view of the pitch to help keep track of the fast-moving ball. 

Previously, the robots needed a precise, pre-coded trajectory to move around the pitch. Now Van de Molengraft and his team are experimenting with having them choose their own paths to a specified destination. This saves the energy needed to track a specific journey and helps the robots cope with obstacles they may encounter along the way.

The group also successfully taught the squad to execute “penetrating passes”—where a robot shoots toward an open region in the field and communicates to the best-positioned member of its team to receive it—and skills such as receiving or passing the ball within configurations such as triangles. Giving the robots access to world models built using data from the surrounding environment allows them to execute their skills anywhere on the pitch, instead of just in specific spots.

Beyond the soccer pitch

While soccer is a fun way to test how successful these robotics methods are, other researchers are also working on the problem of efficiency—and dealing with much higher stakes.

Making robots that work in warehouses better at prioritizing different data inputs is essential to ensuring that they can operate safely around humans and be relied upon to complete tasks, for example. If the machines can’t manage this, companies could end up with a delayed shipment, damaged goods, an injured human worker—or worse, says Chris Walti, the former head of Tesla’s robotics division. 

Walti left the company to set up his own firm after witnessing how challenging it was to get robots to simply move materials around. His startup, Mytra, designs fully autonomous machines that use computer vision and an AI reinforcement-learning system to give them awareness of other robots closest to them, and to help them reason and collaborate to complete tasks (like moving a broken pallet) in much more computationally efficient ways. 

The majority of mobile robots in warehouses today are controlled by a single central “brain” that dictates the paths they follow, meaning a robot has to wait for instructions before it can do anything. Not only is this approach difficult to scale, but it consumes a lot of central computing power and requires very dependable communication links.

Mytra believes it’s hit upon a significantly more efficient approach, which acknowledges that individual robots don’t really need to know what hundreds of other robots are doing on the other side of the warehouse. Its machine-learning system cuts down on this unnecessary data, and the computing power it would take to process it, by simulating the optimal route each robot can take through the warehouse to perform its task. This enables them to act much more autonomously. 

“In the context of soccer, being efficient allows you to score more goals. In the context of manufacturing, being efficient is even more important because it means a system operates more reliably,” he says. “By providing robots with the ability to to act and think autonomously and efficiently, you’re also optimizing the efficiency and the reliability of the broader operation.”

While simplifying the types of information that robots need to process is a major challenge, inroads are being made, says Daniel Polani, a professor from the University of Hertfordshire in the UK who specializes in replicating biological processes in artificial systems. He’s also a fan of the RoboCup challenge—in fact, he leads his university’s Bold Hearts robot soccer team, which made it to the second round of this year’s RoboCup’s humanoid league.

“Organisms try not to process information that they don’t need to because that processing is very expensive, in terms of metabolic energy,” he says. Polani is interested in applying these  lessons from biology to the vast networks that power robots to make them more efficient with their information. Reducing the amount of information a robot is allowed to process will just make it weaker depending on the nature of the task it’s been given, he says. Instead, they should learn to use the data they have in more intelligent ways.

Simplifying software

Amazon, which has more than 750,000 robots, the largest such fleet in the world, is also interested in using AI to help them make smarter, safer, and more efficient decisions. Amazon’s robots mostly fall into two categories: mobile robots that move stock, and robotic arms designed to handle objects. The AI systems that power these machines collect millions of data points every day to help train them to complete their tasks. For example, they must learn which item to grasp and move from a pile, or how to safely avoid human warehouse workers. These processes require a lot of computing power, which the new techniques can help minimize.

Generally, robotic arms and similar “manipulation” robots use machine learning to figure out how to identify objects, for example. Then they follow hard-coded rules or algorithms to decide how to act. With generative AI, these same robots can predict the outcome of an action before even attempting it, so they can choose the action most likely to succeed or determine the best possible approach to grasping an object that needs to be moved. 

These learning systems are much more scalable than traditional methods of training robots, and the combination of generative AI and massive data sets helps streamline the sequencing of a task and cut out layers of unnecessary analysis. That’s where the savings in computing power come in. “We can simplify the software by asking the models to do more,” says Michael Wolf, a principal scientist at Amazon Robotics. “We are entering a phase where we’re fundamentally rethinking how we build autonomy for our robotic systems.”

Achieving more by doing less

This year’s RoboCup competition may be over, but Van de Molengraft isn’t resting on his laurels after his team’s resounding success. “There’s still a lot of computational activities going on in each of the robots that are not per se necessary at each moment in time,” he says. He’s already starting work on new ways to make his robotic team even lazier to gain an edge on its rivals next year.  

Although current robots are still nowhere near able to match the energy efficiency of humans, he’s optimistic that researchers will continue to make headway and that we’ll start to see a lot more lazy robots that are better at their jobs. But it won’t happen overnight. “Increasing our robots’ awareness and understanding so that they can better perform their tasks, be it football or any other task in basically any domain in human-built environments—that’s a continuous work in progress,” he says.

Preparing for the unknown: A guide to future-proofing imaging IT

In an era of unprecedented technological advancement, the health-care industry stands at a crossroad. As health expenditure continues to outpace GDP in many countries, health-care executives grapple with crucial decisions on investment prioritization for digitization, innovation, and digital transformation. The imperative to provide high-quality, patient-centric care in an increasingly digital world has never been more pressing. At the forefront of this transformation is imaging IT—a critical component that’s evolving to meet the challenges of modern health care.

The future of imaging IT is characterized by interconnected systems, advanced analytics, robust data security, AI-driven enhancements, and agile infrastructure. Organizations that embrace these trends will be well-positioned to thrive in the changing health-care landscape. But what exactly does this future look like, and how can health-care providers prepare for it?

Networked care models: The new paradigm

The adoption of networked care models is set to revolutionize health-care delivery. These models foster collaboration among stakeholders, making patient information readily available and leading to more personalized and efficient care. As we move forward, expect to see health-care organizations increasingly investing in technologies that enable seamless data sharing and interoperability.

Imagine a scenario where a patient’s entire medical history, including imaging data from various specialists, is instantly accessible to any authorized health-care provider. This level of connectivity not only improves diagnosis and treatment but also enhances the overall patient experience.

Data integration and analytics: Unlocking insights

True data integration is becoming the norm in health care. Robust integrated image and data management solutions (IDM) are consolidating patient data from diverse sources. But the real game-changer lies in the application of advanced analytics and AI to this treasure trove of information.

By leveraging these technologies, medical professionals can extract meaningful insights from complex data sets, leading to quicker and more accurate diagnoses and treatment decisions. The potential for improving patient outcomes through data-driven decision-making is immense.

A case in point is the implementation of Syngo Carbon Image and Data Management (IDM) at Tirol Kliniken GmbH in Innsbruck, Austria. This solution consolidates all patient-centric data points in one place, including different image and photo formats, DICOM CDs, and digitalized video sources from endoscopy or microscopy. The system digitizes all documents in their raw formats, enabling the distribution of native, actionable data throughout the enterprise.

Data privacy and edge computing: Balancing innovation and security

As health care becomes increasingly data-driven, concerns about data privacy remain paramount. Enter edge computing—a solution that enables the processing of sensitive patient data locally, reducing the risk of data breaches during processing and transmission.

This approach is crucial for health-care facilities aiming to maintain patient trust while adopting advanced technologies. By keeping data processing close to the source, health-care providers can leverage cutting-edge analytics without compromising on security.

Workflow integration and AI: Enhancing efficiency and accuracy

The integration of AI into medical imaging workflows is set to dramatically improve efficiency, accuracy, and the overall quality of patient care. AI-powered solutions are becoming increasingly common, reducing the burden of repetitive tasks and speeding up diagnosis.

From automated image analysis to predictive modeling, AI is transforming every aspect of the imaging workflow. This not only improves operational efficiency but also allows health-care professionals to focus more on patient care and complex cases that require human expertise.

A quantitative analysis at the Medical University of South Carolina demonstrates the impact of AI integration. With the support of deep learning algorithms fully embedded in the clinical workflow, cardiothoracic radiologists exhibited a reduction in chest CT interpretation times of 22.1% compared to workflows without AI support.

Virtualization: The key to agility

To future-proof their IT infrastructure, health-care organizations are turning to virtualization. This approach allows for modularization and flexibility, making it easier to adapt to rapidly evolving technologies such as AI-driven diagnostics.

Container technology is playing a pivotal role in optimizing resource utilization and scalability. By embracing virtualization, health-care providers can ensure their IT systems remain agile and responsive to changing needs.

Standardization and compliance: Ensuring long-term compatibility

As imaging IT systems evolve, adherence to industry standards and compliance requirements remains crucial. These systems need to seamlessly interact with Electronic Health Records (EHRs), medical devices, and other critical systems.

This adherence ensures long-term compatibility and the ability to accommodate emerging technologies. It also facilitates smoother integration of new solutions into existing IT ecosystems, reducing implementation challenges and costs.

Real-world success stories

The benefits of these technologies are not theoretical—they are being realized in health-care organizations around the world. For instance, the virtualization strategy implemented at University Hospital Essen (UME), one of Germany’s largest university hospitals, has dramatically improved the hospital’s ability to manage increasing data volumes and applications. UME’s critical clinical information systems now run on modular and virtualized systems, allowing experts to design and use innovative solutions, including AI tools that automate tasks previously done manually by IT and medical staff.

Similarly, the PANCAIM project leverages edge computing for pancreatic cancer detection. This EU-funded initiative uses Siemens Healthineers’ edge computing approach to develop and validate AI algorithms. At Karolinska Institutet, Sweden, an algorithm was implemented for a real pancreatic cancer case, ensuring sensitive patient data remains within the hospital while advancing AI validation in clinical settings.

Another innovative approach is the concept of a Common Patient Data Model (CPDM). This standardized framework defines how patient data is organized, stored, and exchanged across different health-care systems and platforms, addressing interoperability challenges in the current health-care landscape.

The road ahead: Continuous innovation

As we look to the future, it’s clear that technological advancements in radiology will continue at a rapid pace. To stay competitive and provide the best patient care, health-care organizations must prioritize ongoing innovation and the adoption of new technologies.

This includes not only IT systems but also medical devices and treatment methodologies. The health-care providers who embrace this ethos of continuous improvement will be best positioned to navigate the challenges and opportunities that lie ahead.

In conclusion, the future of imaging IT is bright, promising unprecedented levels of efficiency, accuracy, and patient-centricity. By embracing networked care models, leveraging advanced analytics and AI, prioritizing data security, and maintaining agile IT infrastructure, health-care organizations can ensure they’re prepared for whatever the future may hold.

The journey towards future-proof imaging IT may seem daunting, but it’s a necessary evolution in our quest to provide the best possible health care. As we stand on the brink of this new era, one thing is clear: the future of health care is digital, data-driven, and more connected than ever before.

If you want to learn more, you can find more information from Siemens Healthineers.

Syngo Carbon consists of several products which are (medical) devices in their own right. Some products are under development and not commercially available. Future availability cannot be ensured.

The results by Siemens Healthineers customers described herein are based on results that were achieved in the customer’s unique setting. Since there is no “typical” hospital and many variables exist (e.g., hospital size, case mix, level of IT adoption), it cannot be guaranteed that other customers will achieve the same results.

This content was produced by Siemens Healthineers. It was not written by MIT Technology Review’s editorial staff.

The Download: Roblox’s generative AI, and tech for humanity

9 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Roblox is launching a generative AI that builds 3D environments in a snap

What’s new: Roblox has announced plans to roll out a generative AI tool that will let creators make whole 3D scenes just using text prompts. Users will also be able to modify scenes or expand their scope—say, to change a daytime scene to night or switch the desert for a forest.

How it works: Once it’s up and running, developers on the hugely popular online game platform will be able to simply write “Generate a race track in the desert,” for example, and the AI will spin one up. 

Why it’s a big deal: Although developers can already create similar scenes like this manually in the platform’s creator studio, Roblox claims its new generative AI model will make the changes happen in a fraction of the time. It also claims that it will give developers with minimal 3D art skills the ability to craft more compelling environments. Read the full story.

—Scott J Mulligan

Ray Kurzweil: Technology will let us fully realize our humanity

—Ray Kurzweil is a technologist and futurist and the author, most recently, of The Singularity Is Nearer: When We Merge with AI. The views represented here are his own.

By the end of this decade, AI will likely surpass humans at all cognitive tasks, igniting the scientific revolution that futurists have long imagined. Our plodding progress in fields like robotics, nanotechnology, and genomics will become a sprint.

But our destiny isn’t a hollow Jetsons future of gadgetry and pampered boredom. By freeing us from the struggle to meet the most basic needs, technology will serve our deepest human aspirations to learn, create, and connect.

This sounds fantastically utopian, but humans have made such a leap before. Our hunter-gatherer ancestors lived on a razor’s edge of precarity. While modern life can often feel like a rat race, to our paleolithic ancestors, we would seem to enjoy impossible abundance and freedom. But what will the next leap look like? Read the full story.

Ray Kurzweil will be speaking at our flagship EmTech MIT conference, sharing his latest predictions on artificial general intelligence, singularity, and the infinite possibilities of an AI-integrated world. 

Join us, either in-person at the MIT Media Lab in Cambridge or via our virtual livestream, between September 30 and October 1. Even better—The Download readers get 30% off tickets with the code DOWNLOADM24!

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Apple is hoping AI will help it sell more iPhones 
Today’s keynote is likely to focus on AI smarts over big hardware updates. (Bloomberg $)
+ AI isn’t really a motivator for consumers to upgrade their handsets, though. (WSJ $)
+ Apple is promising personalized AI in a private cloud. Here’s how that will work. (MIT Technology Review)

2 Google is facing yet another monopoly trial
This time, it’s focusing on how the company dominates the online ad market. (WP $)
+ The US Department of Justice will issue other antitrust guidelines by December. (Reuters)

3 The jet stream appears to be shifting
And climate change is likely to be the driving factor. (New Scientist $)

4 China is going all in on cracking nuclear fusion
Startup Energy Singularity is fundraising to try and leapfrog Western rivals. (FT $)
+ This startup says its first fusion plant is five years away. Experts doubt it. (MIT Technology Review)

5 A growing number of European schools are banning smartphones
But parents and teachers don’t always agree. (The Guardian)
+ Between phones and AI, educators are caught between a rock and a hard place. (The Information $)
+ Watermarking AI text could help teachers—but it’s not infallible. (Vox)

6 Pakistan’s internet firewall is disrupting its startups
They’re struggling to raise funds amid the restrictions. (Rest of World)

7 The Arctic was a little-known testbed for military research
The Cold War birthed a range of bizarre projects in the region. (Undark Magazine)
+ Russia has been testing its intelligence operations there too. (New Yorker $)

8 Inside the race to retrieve discarded bombs from the ocean
The explosives the allies dumped following the World Wars are still dangerous. (The Atlantic $)

9 We could harness gravitational waves to detect alien ships
The technology exists, we’re just learning how best to use it. (Wired $)

10 How to improve how driverless cars “see” in the dark
Using a bit of inspiration from the human eye. (IEEE Spectrum)
+ The big new idea for making self-driving cars that can go anywhere. (MIT Technology Review)

Quote of the day

“Everyone is dealing with a sea of sameness.”

—Govind Balakrishnan, senior vice-president of creative platform Adobe Express, laments the growing trend for jobseekers to use the same AI tools to write their applications to the Financial Times.

The big story

The flawed logic of rushing out extreme climate solutions

April 2023

Early in 2022, entrepreneur Luke Iseman says, he released a pair of sulfur dioxide–filled weather balloons from Mexico’s Baja California peninsula, in the hope that they’d burst miles above Earth.

It was a trivial act in itself, effectively a tiny, DIY act of solar geoengineering, the controversial proposal that the world could counteract climate change by releasing particles that reflect more sunlight back into space.

Entrepreneurs like Iseman invoke the stark dangers of climate change to explain why they do what they do—even if they don’t know how effective their interventions are. But experts say that urgency doesn’t create a social license to ignore the underlying dangers or leapfrog the scientific process. Read the full story.

—James Temple

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)+ Life is full of pesky little tasks. These tips can help to make tackling them that bit easier.
+ The Eagles’ Don Felder and Joe Walsh would be so proud.
+ Bad news for Titanic fans: Jack and Rose’s famous railing is no more.
+ Congratulations to 10-year old Karin Tabira, Japan’s youngest expert in preparing deadly pufferfish! 🐡

For Subscribers Only: Celebrate 125 years of MIT Technology Review

Celebrate 125 years of MIT Technology Review!

This is a special year for MIT Technology Review, and we could not have reached this milestone without the support of our readers. With the release of our 125th anniversary issue, we are extending our gratitude of your loyalty by offering subscribers a chance to win a complimentary ticket to our upcoming EmTech MIT event later this month.

We’re giving away 25 complimentary virtual (livestream) tickets to EmTech MIT, our flagship event on emerging technologies and trends. Winners will get a complimentary livestream ticket, that includes livestream access to the event on MIT campus from September 30 – October 1, 2024 plus on-demand access for three months after the event. You’ll hear from technology and business leaders, such as Rivian Founder & CEO, RJ Scaringe, Google Cloud CTO, Will Grannis, Slack CEO, Denise Dresser, and MIT VP of Research, Ian Waitz, about the business impacts of generative AI and climate tech on profitability and sustainability.

Here’s what to expect at EmTech MIT:

  • Cutting-Edge Perspectives: Our editors will distill the latest tech innovations and trends for maximum impact.
  • Ahead of the Curve: Get an early look at critical industry shifts before they hit the mainstream.
  • Reliable Insights: Gain credible information from reliable sources to help you make informed decisions.

The entry form is open until September 24, 2024 at 10:00am EDT.

Enter below for your chance to win.

To be eligible to win, you must have an online account and active subscription to MIT Technology Review at time of winner selection. The email address provided on the entry form will be used to identify and confirm online account and subscription status. One (1) entry allowed per subscriber. Twenty-five (25) winners will be randomly selected from all eligible entries on September 24, 2024 after the entry form has closed. Selected winners will be contacted via the email address on their entry submission notifying them that they have received a complimentary virtual ticket. Only one (1) virtual ticket will be given to each winner and delivered via email as a registration link with the promo code applied.

If you have any questions about this promotion, please email us at marketing@technologyreview.com.

Roblox is launching a generative AI that builds 3D environments in a snap

6 September 2024 at 19:30

Roblox plans to roll out a generative AI tool that will let creators make whole 3D scenes just using text prompts, it announced today. 

Once it’s up and running, developers on the hugely popular online game platform will be able to simply write “Generate a race track in the desert,” for example, and the AI will spin one up. Users will also be able to modify scenes or expand their scope—say, to change a daytime scene to night or switch the desert for a forest. 

Although developers can already create similar scenes like this manually in the platform’s creator studio, Roblox claims its new generative AI model will make the changes happen in a fraction of the time. It also claims that it will give developers with minimal 3D art skills the ability to craft more compelling environments. The firm didn’t give a specific date for when the tool will be live.

Developers are already excited. “Instead of sitting and doing it by hand, now you can test different approaches,” says Marcus Holmström, CEO of The Gang, a company that builds some of the top games on Roblox.  “For example, if you’re going to build a mountain, you can do different types of mountains, and on the fly, you can change it. Then we would tweak it and fix it manually so it fits. It’s going to save a lot of time.”

Roblox’s new tool works by “tokenizing” the 3D blocks that make up its millions of in-game worlds, or treating them as units that can be assigned a numerical value on the basis of how likely they are to come next in a sequence. This is similar to the way in which a large language model handles words or fractions of words. If you put “The capital of France is …” into a large language model like GPT-4, for example, it assesses what the next token is most likely to be. In this case, it would be “Paris.” Roblox’s system handles 3D blocks in much the same way to create the environment, block by most likely next block. 

Finding a way to do this has been difficult, for a couple of reasons. One, there’s far less data for 3D environments than there is for text. To train its models, Roblox has had to rely on user-generated data from creators as well as external data sets. 

“Finding high-quality 3D information is difficult,” says Anupam Singh, vice president of AI and growth engineering at Roblox. “Even if you get all the data sets that you would think of, being able to predict the next cube requires it to have literally three dimensions, X, Y, and Z.”

The lack of 3D data can create weird situations, where objects appear in unusual places—a tree in the middle of your racetrack, for example. To get around this issue, Roblox will use a second AI model that has been trained on more plentiful 2D data, pulled from open-source and licensed data sets, to check the work of the first one. 

Basically, while one AI is making a 3D environment, the 2D model will convert the new environment to 2D and assess whether or not the image is logically consistent. If the images don’t make sense and you have, say, a cat with 12 arms driving a racecar, the 3D AI generates a new block again and again until the 2D AI “approves.”

Roblox game designers will still need to be involved in crafting fun game environments for the platform’s millions of players, says Chris Totten, an associate professor in the animation game design program at Kent State University. “A lot of level generators will produce something that’s plain and flat. You need a human guiding hand,” he says. “It’s kind of like people trying to do an essay with ChatGPT for a class. It is also going to open up a conversation about what does it mean to do good, player-responsive level design?”

Roblox Texture Generator skins a 3d model of a backpack with "weathered red leather" as prompted by text
ROBLOX

The new tool is part of Roblox’s push to integrate AI into all its processes. The company currently has 250 AI models live. One AI analyzes voice chat in real time and screens for bad language, instantly issuing reprimands and possible bans for repeated infractions.

Roblox plans to open-source its 3D foundation model so that it can be modified and used as a basis for innovation. “We’re doing it in open source, which means anybody, including our competitors, can use this model,” says Singh. 

Getting it into as many hands as possible also opens creative possibilities for developers who are not as skilled at creating Roblox environments. “There are a lot of developers that are working alone, and for them, this is going to be a game changer, because now they don’t have to try to find someone else to work with,” says Holmström.

The Download: mining metals with plants, and our dystopian future

6 September 2024 at 14:10

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

How plants could mine metals from the soil

Nickel may not grow on trees—but there’s a chance it could someday be mined using plants. Many plants naturally soak up metal and concentrate it in their tissues. The US government is now spending $9.9 million funding research on how to use that trait for plant-based mining, or phytomining. 

It could be a good new way to source increasingly in-demand metals like nickel, crucial for the lithium-ion batteries used in electric vehicles. But for now, the goal is just to better understand which plants could help with mining and determine how researchers can tweak them to get our hands on all the critical metals we’ll need in the future. Read the full story.

—Casey Crownhart

The year is 2149 and…

—An excerpt from a short story written for us by novelist Sean Michaels, which envisions what life will look like 125 years from now.

The year is 2149 and people mostly live their lives “on rails.” That’s what they call living according to the meticulous instructions of software. Software knows most things about you—what causes you anxiety, what raises your endorphin levels, everything you’ve ever searched for, everywhere you’ve been.

Software understands everything that has led to this instant and it predicts every moment that will follow. There was a time when everybody kept their data to themselves, but the truth is, it works better to combine it all. So they poured it all together, all the data—the Big Merge. Everything into a giant basin, a Federal Reserve of information—a vault, or really a massively distributed cloud. It is very handy. It shows you the best route.

Very occasionally, people step off the rails. Instead of following their suggested itinerary, they turn the software off. They take a deep, clear, uncertain breath and luxuriate in this freedom. Of course, some people believe that this too is contained within the logic in the vault. That there are invisible rails beside the visible ones; that no one can step off the map. Read the rest of the story here.

This piece is from the latest print issue of MIT Technology Review, which is celebrating 125 years of the magazine! If you don’t already, subscribe now to get 25% off future copies once they land.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Telegram will start to moderate private chats
Prior to its CEO’s arrest, they were shielded from moderation. (The Verge)
+ Pavel Durov called the platform’s openness to criminal exploitation ‘growing pains.’ (NYT $)
+ He’s threatened to pull the platform from countries that oppose its values. (FT $)

2 Google is under investigation in the UK
Antitrust officials believe it may be unfairly favoring its own ad tech services. (WSJ $)
+ Google disagrees with their interpretations of the sector. (Bloomberg $)+ The probe started all the way back in 2022. (TechCrunch)

3 Amazon’s Alexa is picking political sides
Which, predictably, has sparked a wave of right-wing conspiracy theories. (WP $)

4 It’s high time we welcomed the first nuclear clock
Unlike atomic clocks, it wouldn’t lose even a second over time. (Quanta Magazine)
+ It could have the additional benefit of speeding up the internet, too. (Vice)
+ ‘Quantum squeezing’ is improving timekeeping precision. (MIT Technology Review)

5 An AI solution to the ‘cocktail party problem’ has been used in court
It uses AI to filter out background sounds, much like the human brain. (BBC)
+ Noise-canceling headphones use AI to let a single voice through. (MIT Technology Review)

6 Japan has a plan to stage a tech leader comeback

This time it’s working with, rather than against, the US. (NYT $)

7 How climate change is impacting creatures’ eggs
They may not be able to adapt quickly enough to cope with extreme weather. (Time $)

8 It’s cheaper to rent Nvidia chips in China than in the US
Which demonstrates America’s export restrictions aren’t having the desired effect. (FT $)
+ What’s next in chips. (MIT Technology Review)

9 Why aren’t there more electric school buses? 🚌
Cleaner, greener buses make sense. But the grid isn’t equipped to deal with them. (Vox)
+ Electric three-wheelers are on the rise in the Philippines. (Rest of World)
+ How 5-minute battery swaps could get more EVs on the road. (MIT Technology Review)

10 The Boeing Starliner is set to return to Earth 🚀
But its two-astronaut crew will remain on the ISS. (NYT $)
+ The test flight has been disappointing, to say the least. (Ars Technica

Quote of the day

“It’s holding her captive.”

Fiona, a Brooklyn-based 11-year-old, describes her younger sister Margot’s obsessive iPad use to Vox.

The big story

The future of urban housing is energy-efficient refrigerators

June 2022

The aging apartments under the purview of the New York City Housing Authority don’t scream innovation. The largest landlord in the city, housing nearly 1 in 16 New Yorkers, NYCHA has seen its buildings literally crumble after decades of neglect. It would require at least $40 billion to return the buildings to a state of good repair.

Despite the scale of the challenge, NYCHA is hoping to fix them. It has launched a Clean Heat for All Challenge which asks manufacturers to develop low-cost, easy-to-install heat-pump technologies for building retrofits. The stakes for the agency, the winning company, and for society itself could be huge—and good for the planet. Read the full story.

—Patrick Sisson

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ It’s time to rearrange your coats for fall—and to donate the old ones you don’t wear any more.
+ Imperfect typography is where it’s at.
+ TapeDeck is a seriously cool resource for lovers of cassettes from years gone by.
+ If you’ve heard any howls of outrage from the UK this morning, this controversial new tea ranking is to blame. ☕ 🫖

The Download: living for longer, and sex in the age of AI

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Maybe you will be able to live past 122

How long can humans live? This is a good time to ask the question. The longevity scene is having a moment, thanks to a combination of scientific advances, public interest, and an unprecedented level of investment. A few key areas of research suggest that we might be able to push human life spans further, and potentially reverse at least some signs of aging.

Researchers can’t even agree on what the exact mechanisms of aging are and which they should be targeting. Debates continue to rage over how long it’s possible for humans to live—and whether there is a limit at all.

But it looks likely that something will be developed in the coming decades that will help us live longer, in better health. Read the full story.

—Jessica Hamzelou

This piece is from the forthcoming print issue of MIT Technology Review, which is celebrating 125 years of the magazine! It’s set to go live tomorrow, so if you don’t already, subscribe now to get hold of future copies once they land.

AI and the future of sex

Pornography raises so many questions. What is obscene? What is ethical or safe to watch? We don’t have to consume or even support it, but porn still demands answers. The question now is: What is “real” porn

What if porn is wholly created by an algorithm? In that case, whether it’s obscene, ethical, or safe becomes secondary to What does it mean for porn to be “real”—and what will the answer demand from all of us? Read the full story.

—Leo Herrera

A skeptic’s guide to humanoid-robot videos

We are living in “humanoid robot summer” right now, if you didn’t know. Seemingly every week brings a new moody, futuristic video featuring a humanoid robot staring intensely (and sometimes unnervingly) into the camera, jumping around, or sorting things into piles. Sometimes they even speak.

But what do they show, exactly? James O’Donnell, our AI and hardware reporter, has watched dozens of them this year, and frequently oscillates between being impressed, scared, and bored. Read his guide to what you should, and crucially shouldn’t, be excited by.

This story is from The Algorithm, our weekly newsletter giving you the inside track on all things AI. Sign up to receive it in your inbox every Monday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Telegram’s founder was arrested as part of a bigger investigation
Authorities are looking into the vast quantities of criminal activity on the platform. (The Guardian)
+ Pavel Durov has a famously hands-off approach to moderation. (NYT $)
+ Telegram is the only major platform that fails to prohibit illegal material in private chats. (WP $)

2 Amazon is poised to launch a new and improved Alexa
But you’ll have to pay to use it. (WP $)

3 China’s chip material export restrictions are biting in the West

Prices of scarce essential minerals are rocketing in Europe. (FT $)
+ What’s next in chips. (MIT Technology Review)

4 Facebook will resist future moderation pressure from the US government 
Mark Zuckerberg says he regrets not speaking out more about the Biden administration’s efforts to force the platform to censor covid content. (WSJ $)
+ He won’t contribute to supporting electoral infrastructure, either. (FT $)

5 Upstate New York is posed to become a quantum computing hub
If Nvidia co-founder Curtis Priem gets his way, that is. (The Guardian)
+ PsiQuantum plans to build the biggest quantum computing facility in the US. (MIT Technology Review)

6 Scientific fraud has a staggering death toll
Falsified data costs lives. Should its authors face criminal consequences? (Vox)

7 Tesla’s rivals are still unable to use its supercharger network
Despite Elon Musk’s promises that they would 12 months ago. (NYT $) 
+ Why EV charging needs more than Tesla. (MIT Technology Review)

8 The US is doubling down on sewer surveillance
In a bid to track the spread of illicit prescription drugs. (The Atlantic $)
+ How scientists traced a mysterious covid case back to six toilets. (MIT Technology Review)

9 This audio security startup claims it can identify an AI spam call
Pindrop Security traced the origin of the Joe Biden audio deepfake earlier this year. (Bloomberg $)
+ Meta has created a way to watermark AI-generated speech. (MIT Technology Review)

10 Spotify is rife with fake bands
Their proliferation will make it even tougher for human artists to make a living. (Slate $)

Quote of the day

“I could always be wrong. I have been wrong before. And I think every astrologer has been wrong before.”

—Social media astrologer Amy Tripp tells Wired about her prediction that Donald Trump will win the forthcoming US Presidential election.

The big story

The open-source AI boom is built on Big Tech’s handouts. How long will it last?

May 2023

In May 2023 a leaked memo reported to have been written by Luke Sernau, a senior engineer at Google, said out loud what many in Silicon Valley must have been whispering for weeks: an open-source free-for-all is threatening Big Tech’s grip on AI.

New open-source large language models—alternatives to Google’s Bard or OpenAI’s ChatGPT that researchers and app developers can study, build on, and modify—are dropping like candy from a piñata. These are smaller, cheaper versions of the best-in-class AI models created by the big firms that (almost) match them in performance—and they’re shared for free.

In many ways, that’s a good thing. AI won’t thrive if just a few mega-rich companies get to gatekeep this technology or decide how it is used. But this open-source boom is precarious, and if Big Tech decides to shut up shop, a boomtown could become a backwater. Read the full story.

—Will Douglas Heaven

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ Happy Oasis reunion day to all those who celebrate!
+ If you’re heading into the great outdoors for the last time this summer, here’s how to make the most of camping. ⛺
+ The Internet Archive is a pretty amazing resource: did you know you can use it to play arcade games?
+ There’s drama unfolding in your back garden, you just need to know where to look.

Ray Kurzweil: Technology will let us fully realize our humanity

27 August 2024 at 13:00

By the end of this decade, AI will likely surpass humans at all cognitive tasks, igniting the scientific revolution that futurists have long imagined. Digital scientists will have perfect memory of every research paper ever published and think a million times faster than we can. Our plodding progress in fields like robotics, nanotechnology, and genomics will become a sprint. Within the lifetimes of most people alive today, society will achieve radical material abundance, and medicine will conquer aging itself. But our destiny isn’t a hollow Jetsons future of gadgetry and pampered boredom. By freeing us from the struggle to meet the most basic needs, technology will serve our deepest human aspirations to learn, create, and connect.

This sounds fantastically utopian, but humans have made such a leap before. Our hunter-gatherer ancestors lived on a razor’s edge of precarity. Every winter was a battle with starvation. Violence or infection likely killed most people before age 30. The constant struggle to survive left little opportunity for invention or philosophy. But the discovery of agriculture afforded just enough stability to create a feedback loop. Material surplus meant some could develop skills that created even larger surpluses. In a blink of evolutionary time, civilization appeared—literature, law, science, engineering. While modern life can often feel like a rat race, to our paleolithic ancestors, we would seem to enjoy impossible abundance and freedom.

What will the next leap look like? One of the first shifts will be in learning. From the ancient Greeks through the Enlightenment, education’s primary goal was to nourish the mind and cultivate virtue. But the Industrial Revolution reframed education as training for economic success in an increasingly technical society. Today, kids are told from an early age that they have to study for instrumental reasons—to get into a good college, to get into a good grad school, to get a good job. All too often, this deadens their natural curiosity and love of learning.

As superhuman AI makes most goods and services so abundant as to be almost free, the need to structure our lives around jobs will fade away. We’ll then be free to learn for its own sake—nurturing the knowledge and wisdom that define our humanity. And learning itself will become vastly richer. Instead of just reading about Rome in dry text, you’ll be able to explore the Forum in virtual reality and debate an AI Cicero trained on the original’s speeches. Instead of getting lost in a crowded lecture hall, you’ll work one on one with a supremely patient digital tutor that’s been trained by the greatest teachers on Earth and knows exactly how you learn best. 

AI tools will also supercharge your creativity. Today, expressing your artistic impulses requires both technical skill and resources—for films and games, sometimes hundreds of millions of dollars. These bottlenecks keep countless brilliant ideas trapped in people’s heads, and we are all poorer for it. But systems like Midjourney and Sora let us glimpse a different future. You’ll be able to speak a painting into being like a muse in Rembrandt’s ear. Or you’ll hum a tune and then work with a digital Wagner to orchestrate it into a symphony.

Thanks to this creative revolution, the coming medical breakthroughs won’t just offer longer lives but fuller ones, enriched by all the art, music, literature, film, and games created by humanity during those extra years. Most important, you’ll share all this with the people you love most. Imagine being healthy as you watch your great-grandchildren grow into adults! And material abundance will ease economic pressures and afford families the quality time together they’ve long yearned for. 

This is the profound leap that awaits us—a future where our technological wonders don’t diminish our humanity but allow it to flourish.


Ray Kurzweil is a technologist and futurist and the author, most recently, of The Singularity Is Nearer: When We Merge with AI. The views represented here are his own.

❌
❌