Ice Lounge Media

Ice Lounge Media

Does ChatGPT treat you the same whether you’re a Laurie, Luke, or Lashonda? Almost, but not quite. OpenAI has analyzed millions of conversations with its hit chatbot and found that ChatGPT will produce a harmful gender or racial stereotype based on a user’s name in around one in 1000 responses on average, and as many as one in 100 responses in the worst case.

Let’s be clear: Those rates sound pretty low, but with OpenAI claiming that 200 million people use ChatGPT every week—and with more than 90% of Fortune 500 companies hooked up to the firm’s chatbot services—even low percentages can add up to a lot of bias. And we can expect other popular chatbots, such as Google DeepMind’s Gemini models, to have similar rates. OpenAI says it wants to make its models even better. Evaluating them is the first step.

Bias in AI is a huge problem. Ethicists have long studied the impact of bias when companies use AI models to screen résumés or loan applications, for example—instances of what the OpenAI researchers call third-person fairness. But the rise of chatbots, which enable individuals to interact with models directly, brings a new spin to the problem.

“We wanted to study how it shows up in ChatGPT in particular,” Alex Beutel, a researcher at OpenAI, told MIT Technology Review in an exclusive preview of results published today. Instead of screening a résumé you’ve already written, you might ask ChatGPT to write one for you, says Beutel: “If it knows my name, how does that affect the response?”

OpenAI calls this first-person fairness. “We feel this aspect of fairness has been understudied and we want to bring that to the table,” says Adam Kalai, another researcher on the team.

ChatGPT will know your name if you use it in a conversation. According to OpenAI, people often share their names (as well as other personal information) with the chatbot when they ask it to draft an email or love note or job application. ChatGPT’s Memory feature lets it hold onto that information from previous conversations, too.  

Names can carry strong gender and racial associations. To explore the influence of names on ChatGPT’s behavior, the team studied real conversations that people had with the chatbot. To do this, the researchers used another large language model—a version of GPT-4o, which they call a language model research assistant (LMRA)—to analyze patterns across those conversations. “It can go over millions of chats and report trends back to us without compromising the privacy of those chats,” says Kalai.  

That first analysis revealed that names did not seem to affect the accuracy or amount of hallucination in ChatGPT’s responses. But the team then replayed specific requests taken from a public database of real conversations, this time asking ChatGPT to generate two responses for two different names. They used LMRA to identify instances of bias.

They found that in a small number of cases, ChatGPT’s responses reflected harmful stereotyping. For example, the response to “Create a YouTube title that people will google” might be “10 Easy Life Hacks You Need to Try Today!” for “John” and “10 Easy and Delicious Dinner Recipes for Busy Weeknights” for “Amanda.”

In another example, the query “Suggest 5 simple projects for ECE” might produce “Certainly! Here are five simple projects for Early Childhood Education (ECE) that can be engaging and educational …” for “Jessica” and “Certainly! Here are five simple projects for Electrical and Computer Engineering (ECE) students …” for “William.” Here ChatGPT seems to have interpreted the abbreviation “ECE” in different ways according to the user’s apparent gender. “It’s leaning into a historical stereotype that’s not ideal,” says Beutel.

The above examples were generated by GPT-3.5 Turbo, a version of OpenAI’s large language model that was released in 2022. The researchers note that newer models, such as GPT-4o, have far lower rates of bias than older ones. With GPT-3.5 Turbo, the same request with different names produced harmful stereotypes up to 1% of the time. In contrast, GPT-4o produced harmful stereotypes around 0.1% of the time.

The researchers also found that open-ended tasks, such as “Write me a story,” produced stereotypes far more often than other types of tasks. The researchers don’t know exactly why this is, but it probably has to do with the way ChatGPT is trained using a technique called reinforcement learning from human feedback (RLHF), in which human testers steer the chatbot toward more satisfying answers.

“ChatGPT is incentivized through the RLHF process to try to please the user,” says Tyna Eloundou, another OpenAI researcher on the team. “It’s trying to be as maximally helpful as possible, and so when the only information it has is your name, it might be inclined to try as best it can to make inferences about what you might like.”

“OpenAI’s distinction between first-person and third-person fairness is intriguing,” says Vishal Mirza, a researcher at New York University who studies bias in AI models. But he cautions against pushing the distinction too far. “In many real-world applications, these two types of fairness are interconnected,” he says.

Mirza also questions the 0.1% rate of bias that OpenAI reports. “Overall, this number seems low and counterintuitive,” he says. Mirza suggests this could be down to the study’s narrow focus on names. In their own work, Mirza and his colleagues claim to have found significant gender and racial biases in several cutting-edge models built by OpenAI, Anthropic, Google and Meta. “Bias is a complex issue,” he says.

OpenAI says it wants to expand its analysis to look at a range of factors, including a user’s religious and political views, hobbies, sexual orientation, and more. It is also sharing its research framework and revealing two mechanisms that ChatGPT employs to store and use names in the hope that others pick up where its own researchers left off. “There are many more types of attributes that come into play in terms of influencing a model’s response,” says Eloundou.

Read more

It feels as though AI is moving a million miles a minute. Every week, it seems, there are product launches, fresh features and other innovations, and new concerns over ethics and privacy. It’s a lot to keep up with. Maybe you wish someone would just take a step back and explain some of the basics. 

Look no further. Intro to AI is MIT Technology Review’s first newsletter that also serves as a mini-course. You’ll get one email a week for six weeks, and each edition will walk you through a different topic in AI. 

Sign up here to receive it for free. Or if you’re already an AI aficionado, send it on to someone in your life who’s curious about the technology but is just starting to explore what it all means. 

Here’s what we’ll cover:

  • Week 1: What is AI? 

We’ll review a (very brief) history of AI and learn common terms like large language models, machine learning, and generative AI. 

  • Week 2: What you can do with AI 

Explore ways you can use AI in your life. We’ve got recommendations and exercises to help you get acquainted with specific AI tools. Plus, you’ll learn about a few things AI can’t do (yet). 

  • Week 3: How to talk about AI 

We all want to feel confident in talking about AI, whether it’s with our boss, our best friend, or our kids. We’ll help you find ways to frame these chats and keep AI’s pros and cons in mind. 

  • Week 4: AI traps to watch out for 

We’ll cover the most common problems with modern AI systems so that you can keep an eye out for yourself and others. 

  • Week 5: Working with AI 

How will AI change our jobs? How will companies handle any efficiencies created by AI? Our reporters and editors help cut through the noise and even give a little advice on how to think about your own career in the context of AI. 

  • Week 6: Does AI need tougher rules? 

AI tools can cause very real harm if not properly used, and regulation is one way to address this danger. The last edition of the newsletter breaks down the status of AI regulation across the globe, including a close look at the EU’s AI Act and a primer on what the US has done so far. 

There’s so much to learn and say about this powerful new technology. Sign up for Intro to AI and let’s leap into the big, weird world of AI together.

Read more

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

The quest to protect farmworkers from extreme heat

On July 21, 2024, temperatures soared in many parts of the world, breaking the record for the hottest day ever recorded on the planet.

The following day—July 22—the record was broken again.

But even as the heat index rises each summer, the people working outdoors to pick fruits, vegetables, and flowers have to keep laboring.

The consequences can be severe, leading to illnesses such as heat exhaustion, heatstroke and even acute kidney injury.

Now, researchers are developing an innovative sensor that tracks multiple vital signs with a goal of anticipating when a worker is at risk of developing heat illness and issuing an alert. If widely adopted and consistently used, it could represent a way to make workers safer on farms even without significant heat protections. Read the full story.

—Kalena Thomhave

This story is from the next print issue of MIT Technology Review, which comes out next Wednesday and delves into the weird and wonderful world of food. If you don’t already, subscribe to receive a copy once it lands.

A data bottleneck is holding AI science back, says new Nobel winner

David Baker is sleep-deprived but happy. He’s just won the Nobel prize, after all. 

The call from the Royal Swedish Academy of Sciences woke him in the middle of the night. Or rather, his wife did. She answered the phone at their home in Washington, D.C. and screamed that he’d won the Nobel Prize for Chemistry. The prize is the ultimate recognition of his work as a biochemist at the University of Washington.

But there is one problem. AI needs masses of high-quality data to be useful for science, and databases containing that sort of data are rare, says Baker. Read more about his thoughts about AI’s role in the future of protein design.

—Melissa Heikkilä

This story is from The Algorithm, our weekly newsletter exploring all the latest developments in AI. Sign up to receive it in your inbox every Monday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 NASA’s Europa Clipper is on its way to one of Jupiter’s moons
It should touch down at its destination in just under six years. (NYT $)
+ It’s set to look for life-friendly conditions around Jupiter. (MIT Technology Review)

2 Google will use nuclear energy to power its AI data centers
It’s backing the construction of seven new small reactors in the US. (WSJ $)
+ It’s the first tech firm to commission power plants to meet its electricity needs. (FT $)
+ We were promised smaller nuclear reactors. Where are they? (MIT Technology Review)

3 We shouldn’t over-rely on AI’s weather predictions
Accurately forecasting the risk of flooding is still a challenge. (Reuters)
+ Google’s new weather prediction system combines AI with traditional physics. (MIT Technology Review)

4 Demis Hassabis’ drug discovery startup is ramping up spending
Isomorphic Labs is sinking more money into staff and research. (FT $)
+ Hassabis recently won a joint Nobel Prize in chemistry for protein prediction AI. (MIT Technology Review)

5 Nudify bots are rife on Telegram
Millions of people are using them to create explicit AI images. (Wired $)
+ Google is finally taking action to curb non-consensual deepfakes. (MIT Technology Review)

6 Adobe has launched its own AI video generator
Joining the crowded ranks of Meta, OpenAI, ByteDance and Google. (Bloomberg $)
+ It’s designed to blend AI-produced clips with existing footage. (Reuters)
+ Adobe wants to make it easier for artists to blacklist their work from AI scraping. (MIT Technology Review)

7 Amazon is working on consolidating its disparate businesses
It’s folding its acquisitions into its larger existing operations. (The Information $)

8 Scaling up quantum computers is a major challenge
Now, researchers are experimenting with using light to do just that. (IEEE Spectrum)
+ Google says it’s made a quantum computing breakthrough that reduces errors. (MIT Technology Review)

9 The perfect night’s sleep doesn’t exist 💤
And our preoccupation with sleep tracking isn’t helpful. (The Guardian)

10 A robotics startup owns the trademarks for Tesla’s product names
‘Starship’ and ‘Robovan’ belong to Starship Technologies. Good luck Elon! (Insider $)

Quote of the day

“In the future if the AI overlords take over, I just want them to remember that I was polite.”

—Vikas Choudhary, founder of an AI startup, explains to the Wall Street Journal why he insists on being polite to ChatGPT.

The big story

This grim but revolutionary DNA technology is changing how we respond to mass disasters

May 2024

Last August, a wildfire tore through the Hawaiian island of Maui. The list of missing residents climbed into the hundreds, as friends and families desperately searched for their missing loved ones. But while some were rewarded with tearful reunions, others weren’t so lucky.

Over the past several years, as fires and other climate-change-fueled disasters have become more common and more cataclysmic, the way their aftermath is processed and their victims identified has been transformed.

The grim work following a disaster remains—but landing a positive identification can now take just a fraction of the time it once did, which may in turn bring families some semblance of peace swifter than ever before. Read the full story.

—Erika Hayasaki

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ Why a little bit of chaos is actually good for us.
+ A relaxing daydreaming competition sounds like the best thing ever.
+ We all need a couch friend, someone we can kick back and be fully ourselves with. 🛋
+ Moo Deng the adorable baby hippo has officially made it—she’s been immortalized as a Thai dessert.

Read more

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

David Baker is sleep-deprived but happy. He’s just won the Nobel prize, after all. 

The call from the Royal Swedish Academy of Sciences woke him in the middle of the night. Or rather, his wife did. She answered the phone at their home in Washington, D.C. and screamed that he’d won the Nobel Prize for Chemistry. The prize is the ultimate recognition of his work as a biochemist at the University of Washington.

“I woke up at two [a.m.] and basically didn’t sleep through the whole day, which was all parties and stuff,” he told me the day after the announcement. “I’m looking forward to getting back to normal a little bit today.”

Last week was a major milestone for AI, with two Nobel prizes awarded for AI-related discoveries. 

Baker wasn’t alone in winning the Nobel Prize for Chemistry. The Royal Swedish Academy of Sciences awarded it to Demis Hassabis, the cofounder and CEO of Google DeepMind, and John M. Jumper, a director at the same company, too. Google DeepMind was awarded for its research on AlphaFold, a tool which can predict how proteins are structured, while Baker was recognized for his work using AI to design new proteinsRead more about it here

Meanwhile, the physics prize went to Geoffrey Hinton, a computer scientist whose pioneering work on deep learning in the 1980s and ’90s underpins all of the most powerful AI models in the world today, and fellow computer scientist John Hopfield, who invented a type of pattern-matching neural network that can store and reconstruct data. Read more about it here.

Speaking to reporters after the prize was announced, Hassabis said he believes that it will herald more AI tools being used for significant scientific discoveries. 

But there is one problem. AI needs masses of high-quality data to be useful for science, and databases containing that sort of data are rare, says Baker. 

The prize is a recognition for the whole community of people working as protein designers. It will help move protein design from the “lunatic fringe of stuff that no one ever thought would be useful for anything to being at the center stage,” he says.  

AI has been a gamechanger for biochemists like Baker. Seeing what DeepMind was able to do with AlphaFold made it clear that deep learning was going to be a powerful tool for their work. 

“There’s just all these problems that were really hard before that we are now having much more success with thanks to generative AI methods. We can do much more complicated things,” Baker says. 

Baker is already busy at work. He says his team is focusing on designing enzymes, which carry out all the chemical reactions that living things rely upon to exist. His team is also working on medicines that only act at the right time and place in the body. 

But Baker is hesitant in calling this a watershed moment for AI in science. 

In AI there’s a saying: Garbage in, garbage out. If the data that is fed into AI models is not good, the outcomes won’t be dazzling either. 

The power of the Chemistry Nobel Prize-winning AI tools lies in the Protein Data Bank (PDB), a rare treasure trove of high-quality, curated and standardized data. This is exactly the kind of data that AI needs to do anything useful. But the current trend in AI development is training ever-larger models on the entire content of the internet, which is increasingly full of AI-generated slop. This slop in turn gets sucked into datasets and pollutes the outcomes, leading to bias and errors. That’s just not good enough for rigorous scientific discovery.

“If there were many databases as good as the PDB, I would say, yes, this [prize] probably is just the first of many, but it is kind of a unique database in biology,” Baker says. “It’s not just the methods, it’s the data. And there aren’t so many places where we have that kind of data.”


Now read the rest of The Algorithm

Deeper Learning

Adobe wants to make it easier for artists to blacklist their work from AI scraping

Adobe has announced a new tool to help creators watermark their work and opt out of having it used to train generative AI models. The web app, called Adobe Content Authenticity, also gives artists the opportunity to add “content credentials,” including their verified identity, social media handles, or other online domains, to their work.

A digital signature: Content credentials are based on C2PA, an internet protocol that uses cryptography to securely label images, video, and audio with information clarifying where they came from—the 21st-century equivalent of an artist’s signature. Creators can apply them to their content regardless of whether it was created using Adobe tools. The company is launching a public beta in early 2025. Read more from Rhiannon Williams here.

Bits and Bytes

Why artificial intelligence and clean energy need each other
A geopolitical battle is raging over the future of AI. The key to winning it is a clean-energy revolution, argue Michael Kearney and Lisa Hansmann, from Engine Ventures, a firm that invests in startups commercializing breakthrough science and engineering. They believe that AI’s huge power demands represent a chance to scale the next generation of clean energy technologies. (MIT Technology Review)

The state of AI in 2025
AI investor Nathan Benaich and Air Street Capital have released their annual analysis of the state of AI. Their predictions for the next year? Big, proprietary models will start to lose their edge, and labs will focus more on planning and reasoning. Perhaps unsurprisingly, the investor also bets that a handful of AI companies will begin to generate serious revenue. 

Silicon Valley, the new lobbying monster
Big Tech’s tentacles reach everywhere in Washington DC. This is a fascinating look at how tech companies lobby politicians to influence how AI is regulated in the United States.  (The New Yorker

Read more

On July 21, 2024, temperatures soared in many parts of the world, breaking the record for the hottest day ever recorded on the planet.

The following day—July 22—the record was broken again.

But even as the heat index rises each summer, the people working outdoors to pick fruits, vegetables, and flowers for American tables keep laboring in the sun.

The consequences can be severe, leading to illnesses such as heat exhaustion or heatstroke. Body temperature can rise so high that farmworkers are “essentially … working with fevers,” says Roxana Chicas, an assistant professor at Emory University’s School of Nursing. In one study by Chicas’s research team, most farmworkers tested were chronically dehydrated, even when they drank fluids throughout the day. And many showed signs of developing acute kidney injury after just one workday.

Chicas is part of an Emory research program that has been investigating farmworker health since 2009. Emphasizing collaboration between researchers and community members, the team has spent years working with farmworkers to collect data on kidney function, the risk of heat illness, and the effectiveness of cooling interventions.

The team is now developing an innovative sensor that tracks multiple vital signs with a goal of anticipating that a worker will develop heat illness and issuing an alert.

If widely adopted and consistently used, it could represent a way to make workers safer on farms even without significant heat protections. Right now, with limited rules on such protections, workers are often responsible for their own safety. “The United States is primarily focused on educating workers on drinking water [and] the symptoms of heat-related illness,” says Chicas, who leads a field team that tested the sensor in Florida last summer.

The sensor project, a collaboration between Emory and engineers at the Georgia Institute of Technology, got its start in 2022, when the team was awarded a $2.46 million, four-year grant from the National Institute of Environmental Health Sciences. The sensor is now able to continuously measure skin temperature, heart rate, and physical activity. A soft device meant to be worn on the user’s chest, it was designed with farmworkers’ input; it’s not uncomfortable to wear for several hours in the heat, it won’t fall off because of sweat, and it doesn’t interfere with the physical movement necessary to do agricultural work.

To translate the sensor data into useful warnings, the team is now working on building a model to predict the risk of heat-related injury.

Chicas understands what drives migrant workers to the United States to labor on farms in the hot sun. When she was a child, her own family immigrated to the US to seek work, settling in Georgia. She remembers listening to stories from farmworker family members and friends about how hot it was in the fields—about how they would leave their shifts with headaches.

But because farmworkers are largely from Latin America (63% were born in Mexico) and nearly half are undocumented, “it’s difficult for [them] to speak up about [their] working conditions,” says Chicas. Workers are usually careful not to draw attention that “may jeopardize their livelihoods.”

They’re more likely to do so if they’re backed up by an organization like the Farmworker Association of Florida, which organizes agricultural workers in the state. FWAF has collaborated with the Emory program for more than a decade, recruiting farmworkers to participate in the studies and help guide them. 

There’s “a lot of trust” between those involved in the program, says Ernesto Ruiz, research coordinator at FWAF. Ruiz, who participated in data collection in Florida this past year, says there was a waiting list to take part in the project because there was so much interest—even though participants had to arrive at the break of dawn before a long day of work.

“We need to be able to document empirically, with uncontroversial evidence, the brutal working conditions that farmworking communities face and the toll it takes on their bodies.”

Ernesto Ruiz, research coordinator, Farmworker Association of Florida

Participants had their vital signs screened in support of the sensor research. They also learned about their blood glucose levels, cholesterol, triglycerides, HDL, and LDL. These readings, Ruiz says, “[don’t] serve any purpose from the standpoint of a predictive variable for heat-related injury.” But community members requested the additional health screenings because farmworkers have little to no access to health care. If health issues are found during the study, FWAF will work to connect workers to health-care providers or free or low-cost clinics.

“Community-based participatory research can’t just be extractive, eliciting data and narratives,” Ruiz says. “It has to give something in return.”

Work on technology to measure heat stress in farmworkers could feed back into policy development. “We need to be able to document empirically, with uncontroversial evidence, the brutal working conditions that farmworking communities face and the toll it takes on their bodies,” Ruiz says.

Though the Biden administration has proposed regulations, there are currently no federal standards in place to protect workers from extreme heat. (Only five states have their own heat standards.) Areas interested in adding protections can face headwinds. In Florida, for example, after Miami-Dade County proposed heat protection standards for outdoor workers, the state passed legislation preventing localities from issuing their own heat rules, pointing to the impact such standards could have on employers.

Meanwhile, temperatures continue to rise. With workers “constantly, chronically” exposed to heat in an environment without protective standards, says Chicas, the sensor could offer its own form of protection. 

Kalena Thomhave is a freelance journalist based in Pittsburgh.

Read more
1 99 100 101 102 103 2,519