Logo

    Are You Overestimating The Algorithm?

    en-usSeptember 03, 2024
    What was the main topic of the podcast episode?
    Summarise the key points discussed in the episode?
    Were there any notable quotes or insights from the speakers?
    Which popular books were mentioned in this episode?
    Were there any points particularly controversial or thought-provoking discussed in the episode?
    Were any current events or trending topics addressed in the episode?

    Podcast Summary

    • Ancient origins of algorithmsAlgorithms are mathematical processes used since ancient times for various purposes, from simple calculations to dividing crops, and have evolved into advanced technologies like Netflix recommendations and Spotify playlists

      Algorithms, while often associated with advanced technology like Netflix recommendations or Spotify playlists, actually have simple roots. They are essentially step-by-step mathematical processes that have been used since ancient times for various purposes, from calculating volumes to dividing crops. Kyle Choka from The New Yorker and author of "Filter World" explained this concept during a panel at the National Book Festival. He was joined by Dr. Joy Bolamwini, a computer scientist and author of "Unmasking AI," who also emphasized the basic nature of algorithms. Despite their integration with modern technology, algorithms have been around for thousands of years, making them a fundamental aspect of human life.

    • Algorithms and their impact on societyAlgorithms have the power to amplify racial bias and lead to harmful consequences, it's crucial to understand their limitations and use them responsibly.

      Algorithms, which can be as simple as a sequence of steps in a video game or as complex as the code behind facial recognition technology, have the power to shape our lives in profound ways. At their best, they can help us solve problems and express ourselves. But at their worst, they can perpetuate biases and lead to harmful consequences. Dr. Joy's book explores the serious side of algorithms, highlighting their potential to amplify racial bias and even lead to false arrests. The algorithms behind artificial intelligence can expand our worldview, but it's important to be aware of their limitations and the potential harm they can cause. As we continue to rely on technology to shape our lives, it's crucial to understand the role of algorithms and how they can be used responsibly.

    • AI shift from calculations to thinkingAI systems rely on high-quality data sets for accurate outputs, as they're trained to recognize patterns and make predictions based on historical data.

      We're witnessing a shift from traditional software that solves problems through calculations to artificial intelligence (AI) that thinks for itself. This transition involves creating pattern recognition systems through algorithms, which are trained on data sets to determine the quality of outputs. Common and Pete Rock discussed their musical collaborations on Bullseye, while at the National Book Festival, the conversation turned to AI. We're currently exploring ways to give machines human-like abilities, such as recognizing faces or making predictions. However, it's crucial to remember that these systems rely on high-quality data sets for accurate outputs. For instance, a machine built in 1769, mentioned in Kyle's book "Filter Bubble," may appear to think for itself, but it's essential to understand the historical context and limitations of such early AI systems.

    • Human influence on AIDespite the appearance of autonomy, AI and algorithmic recommendations are influenced by humans, driven by profit, and should be recognized as such to avoid misunderstanding their capabilities.

      The seemingly intelligent and automated processes of artificial intelligence and algorithmic recommendations, such as those used by Spotify and Netflix, are not as independent as they may appear. The Mechanical Turk, a historical example of a chess-playing machine, serves as a metaphor for this concept. While it appeared to be an autonomous machine, it was actually controlled by a human manipulating pieces from within. Similarly, these technology companies are driven by human engineers and motivated by profit, primarily through advertising. Therefore, it's essential to recognize the human influence behind these systems and understand that our beliefs about their capabilities can have significant consequences.

    • Technology consequencesTechnology's impact on our lives can lead to misidentification, misgendering, wrongful arrests, and shape experiences in dating and social media, potentially affecting freedom and love lives.

      Our interaction with technology, particularly in areas like facial recognition and social media, can have unintended and far-reaching consequences. Dr. Adjo's work highlights how these technologies can lead to misidentification, misgendering, and even wrongful arrests. In our personal lives, algorithms are increasingly shaping our experiences, from dating app matches to the content we see on social media. Even our relationships are being mediated by these systems. The failure of these technologies, as the speaker experienced firsthand, can impact our freedom and love lives. It's important to be aware of these implications and consider the potential consequences of our reliance on technology.

    • Technological HallucinationsTechnology, including social media and AI, can produce inaccurate and misleading information, leading to a 'hallucinatory' experience. It's important to approach technology with a critical mindset to avoid falling victim to these false results, which can have significant consequences.

      Our interactions with technology, such as social media and AI, are not always an accurate reflection of reality. During a discussion on AI and its ability to generate incorrect information, which was referred to as a "hallucination," it was highlighted that our online experiences can be manipulated and misleading. For instance, Facebook uses algorithms to filter and show users content based on their behavior, which may not align with their organic interests. Similarily, AI tools like ChatGPT can generate incorrect information, leading to a hallucinatory experience. In one example, ChatGPT was asked to find a Russian folktale about an automaton, but it kept providing fake sources and non-existent information. This demonstrates how technology can produce fake and misleading results, which can be dangerous when seeking information on important topics like health care or benefits. While the term "hallucination" might seem like an elevated way to describe inaccurate information, it's important to remember that it's just a false or misleading result. The gravity of the situation should not be downplayed, as the consequences of receiving incorrect information can be significant. It's crucial to approach technology with a critical and discerning mindset to avoid falling victim to these technological hallucinations.

    • Addressing biases in AISharing personal experiences can lead to change in AI industry and push companies to address biases, as seen with Microsoft and Amazon stopping sales to law enforcement after Dr. Joy Buolamwini's advocacy.

      Our conversations this week highlighted the importance of addressing biases in artificial intelligence (AI) and the power of sharing personal stories in driving change. Dr. Joy Buolamwini shared her experiences with coded bias in AI and the impact it had on pushing companies like Microsoft and Amazon to stop selling facial recognition technology to law enforcement. This demonstrates the significance of speaking up and sharing experiences, regardless of background or credentials. Additionally, the conversation around book bans in various states was explored on the Code Switch podcast, emphasizing the importance of understanding the reasons behind these bans and the potential consequences. By addressing these limitations, we can reach the aspirations we have for AI and create more inclusive and equitable communities.

    Recent Episodes from Short Wave

    Solutions Week: Reducing The Food Waste Problem

    Solutions Week: Reducing The Food Waste Problem
    We close out Climate Solutions Week with a look at the final step in the food system: waste. Roughly 30-40% of all food produced globally gets thrown out — a huge problem when it comes to climate change. Shuggie's Trash Pie and Natural Wine has answers for what to do with all that waste – be it bruised fruit and wilted greens salad or meatballs made out of beef hearts. The San Francisco restaurant describes itself as a "climate-solutions restaurant" — and they're hoping you find these dishes appealing. Or at least, that you're willing to give them a try. Shuggie's is one of a growing number of eateries trying to address this problem.

    Interested in hearing more climate solutions? Email us at shortwave@npr.org.

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 13, 2024

    Solutions Week: Reinventing Chocolate

    Solutions Week: Reinventing Chocolate
    Climate change is contributing to erratic weather where cocoa beans are grown and threatening the global chocolate supply. Record rainfall last year led to fungal infections among cacao trees and dwindled supply of cocoa beans. Heat is also making it more difficult for cocoa beans to thrive. So, for day three of Climate Solutions Week, we look at one innovation in the food industry: chocolate substitutes.

    As big chocolate manufacturers rush to stockpile cocoa beans, some companies like Planet A Foods are looking for a more sustainable solution: an alternative that looks like chocolate, tastes like chocolate and feels like chocolate... without chocolate.

    You can read more of international correspondent Rob Schmitz's reporting here.

    Interested in hearing more climate solutions? Email us at
    shortwave@npr.org – we'd love to hear your ideas!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 11, 2024

    Solutions Week: The Cost Of Food Delivery

    Solutions Week: The Cost Of Food Delivery
    Since the height of the pandemic, there has been a boom in the use of food delivery services. Day 2 of NPR's Climate Solutions Week is all about the environmental impacts of how we shop for our food. So in this episode, NPR correspondent Scott Neuman reports on a question we've all wanted to know the answer to: What is the impact of getting food delivered on our carbon footprint?

    Interested in hearing more climate solutions? Email us at shortwave@npr.org – we'd love to hear your ideas!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 10, 2024

    Solutions Week: Climate Change Is Coming For Your Wine

    Solutions Week: Climate Change Is Coming For Your Wine
    In California's Napa Valley, the nation's unofficial wine capital, one varietal reigns supreme: cabernet sauvignon. But climate change is threatening the small blue-black grapes for which cabernet sauvignon is named. Increasingly severe heat waves are taking a toll on the grape variety, especially in late summer during ripening.

    To kick off NPR's Climate Solutions Week, climate correspondent Lauren Sommer joins host Regina G. Barber for a deep dive into the innovations wineries are actualizing — and the ways that cabernet farmers and fans alike could learn to adapt.

    Climate change is affecting our food, and our food is affecting the climate. NPR is dedicating a week to stories and conversations about the search for solutions.

    Interested in hearing more climate solutions? Email us at shortwave@npr.org.

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 09, 2024

    Body Electric: How AI Is Changing Our Relationships

    Body Electric: How AI Is Changing Our Relationships
    Hey, Short Wavers! Today, we have a special present for all of you: An episode from our good friends at NPR's Body Electric podcast all a bout artificial intimacy! Thanks to advances in AI, chatbots can act as personalized therapists, companions and romantic partners. The apps offering these services have been downloaded millions of times. If these relationships relieve stress and make us feel better, does it matter that they're not "real"? On this episode of Body Electric, host Manoush Zomorodi talks to MIT sociologist and psychologist Sherry Turkle about her new research into what she calls "artificial intimacy" and its impact on our mental and physical health.

    Binge the whole Body Electric series here. Plus, sign up for the Body Electric Challenge and our newsletter here.

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 07, 2024

    Dogs Go Viral For 'Talking' To Humans — But Can They?

    Dogs Go Viral For 'Talking' To Humans — But Can They?
    Last year, a dog named Bunny went viral on TikTok for pressing buttons with words on them to "communicate" with her owner. But can dogs even understand those words on a soundboard in the first place? A new study in the journal PLOS One seeks answers. Host Regina G. Barber and producer Rachel Carlson break down that story and more of the week's news with the help of All Things Considered's Ari Shapiro.

    Have other viral headlines that you want us to put to the test for its scientific truth? Email us at shortwave@npr.org — we might cover it on a future episode!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 06, 2024

    Feeling Itchy? Air Pollution Might Be Making It Worse

    Feeling Itchy? Air Pollution Might Be Making It Worse
    Short Wave producer Hannah Chinn has adult-onset eczema. They're not the only one. Up to ten percent of people in the United States have it, according to the National Eczema Association — and its prevalence is increasing. Despite its ubiquity, a lot about this skin condition remains a mystery.
    So today, Hannah's getting answers. They sat down with Raj Fadadu, a dermatologist at UC San Diego, to ask: What is eczema? What triggers it in the first place? And might climate change make it worse sometimes?

    If you liked this episode, check out our episode on the science of itchiness. Also, follow us! That way you never miss another Short Wave episode.

    Interested in hearing more about climate change and human health? Email us at shortwave@npr.orgwe'd love to hear your feedback!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 04, 2024

    Are You Overestimating The Algorithm?

    Are You Overestimating The Algorithm?
    Humans hallucinate. Algorithms lie.

    At least, that's one difference that Joy Buolamwini and Kyle Chayka want to make clear. When ChatGPT tells you that a book exists when it doesn't – or professes its undying love – that's often called a "hallucination." Buolamwini, a computer scientist, prefers to call it "spicy autocomplete." But not all algorithmic errors are as innocuous. So today's show, we get into: How do algorithms work? What are their impacts? And how can we speak up about changing them?

    This is a shortened version of Joy and Kyle's live interview, moderated by Regina G. Barber, at this year's Library of Congress National Book Festival.

    If you liked this episode, check out our other episodes on facial recognition in Gaza, why AI is not a silver bullet and tech companies limiting police use of facial recognition.

    Interested in hearing more technology stories? Email us at shortwave@npr.org — we'd love to consider your idea for a future episode!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 03, 2024

    Batteries: The Lemonade Of Life

    Batteries: The Lemonade Of Life
    Just in time for the return of the school year, we're going "Back To School" by revisiting a classic at-home experiment that turns lemons into batteries — powerful enough to turn on a clock or a small lightbulb. But how does the science driving that process show up in household batteries we use daily? Host Emily Kwong and former host Maddie Sofia talk battery 101 with environmental engineer Jenelle Fortunato.

    Want us to cover more science basics? Email us your ideas at shortwave@npr.org — we might feature them on a future episode!

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usSeptember 02, 2024

    Detecting Pests By Eavesdropping On Insects

    Detecting Pests By Eavesdropping On Insects
    From Indonesia to Wisconsin, farmers all over the world struggle with a huge problem: pests. On top of that, it's tough for farmers to identify where exactly they have the pests and when. Reporter Lina Tran from NPR member station WUWM in Milwaukee joins host Emily Kwong to tell the story of how researchers in the Midwest are inventing new forms of pest detection that involve eavesdropping on the world of insects. Plus, hear what aphid slurping sounds like.

    If you liked this episode, check out behind-the-scenes photos of Insect Eavesdropper experiments in Lina's digital story!

    Interested in hearing more insect news? Email us at shortwave@npr.org.

    Learn more about sponsor message choices: podcastchoices.com/adchoices

    NPR Privacy Policy
    Short Wave
    en-usAugust 30, 2024