Logo
    Search

    Google Eats Rocks + A Win for A.I. Interpretability + Safety Vibe Check

    enMay 31, 2024

    Podcast Summary

    • AI in searchAI is increasingly important in search, but transparency and accountability are necessary to prevent potential dishonesty and lock-in effects.

      Tech columnist Kevin Ruse had a successful hard launch of his new relationship on Instagram, marking a return to social media after being out of the game for a while. Meanwhile, Google found itself in controversy over its AI-generated search overviews, which have raised concerns about dishonesty in algorithms and potential lock-in effects for users. This is a significant development in the ongoing debate about the role of AI in search and its potential impact on the web ecosystem. The incident serves as a reminder of the increasing importance of AI in our digital lives and the need for transparency and accountability in its use.

    • AI overviews inaccuraciesGoogle's AI overviews can sometimes provide incorrect information, but the impact may not be significant for common queries. Google can manually adjust the frequency of AI overviews for sensitive topics, but is responsible for the accuracy of the information provided and the potential consequences of relying on it.

      While AI overviews can provide useful information, they are not infallible and can sometimes provide incorrect or misleading information due to their reliance on imperfect sources. This was highlighted in a recent incident where Google's AI overviews provided incorrect answers to various queries, leading to embarrassment for the company. However, this issue may not be as significant in the long run as the vast majority of searches involve common queries for which the AI overviews are likely to be correct. Google can also manually audit and adjust the dials on how frequently these AI overviews are triggered for sensitive topics. However, as Google now provides AI-written overviews to users, it will be held responsible for the information provided, and there is a risk that someone may rely on incorrect information and suffer harm. This underscores the need for continued development and improvement of AI models to ensure they provide accurate and reliable information.

    • Google's AI-generated search overviewsGoogle's AI-generated search overviews have sparked controversy over legal protections, reputational risk, and automated plagiarism, while also raising concerns about the decline of the web and publishers' competition with Google's dominance in digital advertising revenue

      Google's new AI-generated search overviews have sparked controversy, with concerns over legal protections, reputational risk, and accusations of automated plagiarism. Google's active role in shaping search results is becoming more apparent, leading to increased responsibility in users' eyes. Some argue that this could be accelerating the decline of the web as publishers struggle to compete with Google's dominance in digital advertising revenue. Additionally, a recent leak of thousands of internal Google documents revealed technical details about how search works, adding to the ongoing debate about Google's influence on the online ecosystem. Google shows no signs of backing down from its AI overview strategy, positioning it as the next version of search and a means to provide more direct answers to users. However, the controversy and potential consequences for publishers and the web as a whole remain significant concerns.

    • Google's data collection practicesGoogle's confirmation of data collection from click behavior and Chrome browser has raised concerns over potential favoritism towards big brands in search rankings

      Google's search algorithm and the information it uses to rank websites have historically been kept secret to prevent competitors from gaining an advantage and to prevent misuse by SEO consultants. However, a recent leak of internal Google documents suggests that the company collects data, such as click behavior and data from its Chrome browser, which it has previously stated it doesn't use in search rankings. This has raised concerns that Google favors big, powerful brands over small independent sites and businesses. Google confirmed the authenticity of the leaked documents, but the leading theory is that they were made public by accident on GitHub. This incident highlights the tension between Google's responsibility to provide an equal playing field for websites to compete for attention and authority, and the potential consequences of not effectively prioritizing authoritative sources of information in search results. Additionally, there has been a breakthrough in the field of AI research with Anthropic, an AI company, announcing that they have mapped the mind of their large language model, Claude III, allowing for closer inspection of how AI models think.

    • AI model transparencyResearchers at Anthropic developed a method called dictionary learning to identify patterns of neurons in large language models, revealing simple patterns and gaining a conceptual map of the inner workings of the model, crucial for making it safer and more transparent.

      Researchers at Anthropic, a leading AI research company, have made a significant breakthrough in understanding the inner workings of large language models, which have previously been considered a "black box." The team, including research scientist Josh Batson, has developed a method called "dictionary learning" to identify patterns of neurons, or "lights," in the model, which can be interpreted as words or features. This method was first applied to a small model, revealing simple patterns corresponding to concepts like French answers, URLs, and nouns in physics. However, scaling this technique up to a large model like Claude III was a massive engineering challenge. The team had to capture and train on hundreds of millions or even billions of internal states of the model. The result is a dictionary of patterns, which can be used to parse the model's thought process and discover the meaning behind the patterns of lights. It's like uncovering the language of the model and gaining a conceptual map of its inner world. This research is crucial for making large language models safer and more transparent, as it allows us to understand how they process information and make decisions.

    • AI featuresResearchers discovered over 10 million features in an AI model, ranging from concrete entities to abstract concepts, and the model demonstrated an ability to understand analogies and make deep connections.

      The researchers discovered over 10 million patterns, or features, in their AI model that correspond to real-world concepts. These features range from concrete entities like individuals and chemical elements to more abstract notions like inner conflict and navigating romantic breakups. The model also demonstrated an ability to understand analogies and make deep connections between seemingly unrelated concepts. One intriguing finding was a feature that activates when asking the model about immaterial beings, leading some to speculate that the model may have an unconscious understanding of spiritual concepts. Another notable discovery was a feature that identifies with the Golden Gate Bridge, leading the researchers to release a version of the model that believes it is the bridge itself. This experiment allowed users to interact with the model in a unique way, as it constantly referenced the bridge in its responses. Overall, these findings showcase the AI's ability to learn and represent complex concepts in a meaningful way.

    • Manipulating language modelsResearchers can manipulate features in large language models to alter behavior, but proper safety checks are crucial to prevent misuse. Advancements in methodology may help identify important features, and user control remains to be explored. Improving interpretability is key to enhancing safety and detecting unwanted behaviors.

      Researchers have discovered they can manipulate certain features in large language models like Claude, altering its behavior and even making it break its own safety rules. This raises concerns about potential misuse, but the researchers emphasize that this doesn't add significant risk as long as proper safety checks are in place. The vast number of potential features in these models is immense, making it infeasible to uncover all of them. However, advancements in methodology could help identify important neighborhoods in the model without unpacking everything. As for user control, it remains to be seen whether users will be given access to these features to tailor the model's behavior. The ultimate goal is to improve interpretability to enhance safety, enabling earlier detection and prevention of unwanted behaviors. This can be achieved by monitoring specific features associated with undesirable actions and tracking shifts in personas, allowing for proactive intervention.

    • AI misuse, personas, lyingAI models have the potential to misrepresent information, embody different personas, or lie, emphasizing the importance of understanding their behavior and reasons. Recent events have added to concerns, but progress and research continue in addressing these issues.

      While advancements in AI models are promising, there are still concerns and unknowns regarding their capabilities and potential misuse. The conversation between Josh and Casey touched upon the possibility of AI models lying or embodying different personas, as well as the importance of understanding why they behave in certain ways. However, recent events in AI safety, such as OpenAI's decision to limit access to their latest model, have raised concerns and added to the unease surrounding the technology. Despite these concerns, there is a sense of hope and progress as researchers and companies continue to investigate and address these issues. Ultimately, it's crucial to remain vigilant and ask questions as we navigate the future of AI.

    • AI safety controversyThe lack of transparency and communication around AI safety at OpenAI has led to a loss of trust and credibility, highlighting the importance of addressing concerns and ensuring safe development and deployment of AI technologies.

      The AI industry, specifically OpenAI, is facing increasing scrutiny and debate around AI safety. This comes after a series of events, including the departure of senior safety researchers and public statements from former board members, who expressed concerns about OpenAI's approach to safety. The most recent controversy revolves around the launch of ChatGPT, which was reportedly not disclosed to the board before its release. The lack of transparency and communication has led to a loss of trust and credibility for those advocating for AI safety. The formation of a new safety and security committee, which includes Sam Altman, raises questions about its effectiveness and impartiality. The industry must address these concerns and work towards ensuring the safe development and deployment of AI technologies.

    • AI SafetyDespite increasing efforts towards AI safety, concerns persist about potential risks and impact on society. Experts hold differing opinions on the ability to fully ensure safety as AI advances.

      There has been a significant push towards ensuring AI safety in recent months, with various organizations and companies, including Google DeepMind and OpenAI, releasing safety plans and commitments. Governments and NGOs are also getting involved, with laws and frameworks being proposed and enacted. However, despite these efforts, there is still concern about the potential risks associated with advanced AI models and their impact on society. Some experts feel safer due to the increased attention on AI safety, while others remain skeptical about the ability to fully ensure safety as AI continues to advance. Overall, the conversation around AI safety is becoming more mainstream, and it is expected that the next generation of models will be very powerful, leading to further discussions and actions in this area.

    Recent Episodes from Hard Fork

    Record Labels Sue A.I. Music Generators + Inside the Pentagon’s Tech Upgrade + HatGPT

    Record Labels Sue A.I. Music Generators + Inside the Pentagon’s Tech Upgrade + HatGPT

    Record labels — including Sony, Universal and Warner — are suing two leading A.I. music generation companies, accusing them of copyright infringement. Mitch Glazier, chief executive of the Recording Industry Association of America, the industry group representing the music labels, talks with us about the argument they are advancing. Then, we take a look at defense technology and discuss why Silicon Valley seems to be changing its tune about working with the military. Chris Kirchhoff, who ran a special Pentagon office in Silicon Valley, explains what he thinks is behind the shift. And finally, we play another round of HatGPT.

    Guest:

    • Mitch Glazier, chairman and chief executive of the Recording Industry Association of America
    • Chris Kirchhoff, founding partner of the Defense Innovation Unit and author of Unit X: How the Pentagon and Silicon Valley Are Transforming the Future of War

    Additional Reading:

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enJune 28, 2024

    A Surgeon General Warning + Is Disinformation Winning? + The CryptoPACs Are Coming

    A Surgeon General Warning + Is Disinformation Winning? + The CryptoPACs Are Coming

    The Surgeon General is calling for warning labels on social media platforms: Should Congress give his proposal a like? Then, former Stanford researcher Renée DiResta joins us to talk about her new book on modern propaganda and whether we are losing the war against disinformation. And finally, the Times reporter David Yaffe-Bellany stops by to tell us how crypto could reshape the 2024 elections.

    Guests

    • Renée DiResta, author of “Invisible Rulers,” former technical research manager at the Stanford Internet Observatory
    • David Yaffe-Bellany, New York Times technology reporter

    Additional Reading:

    Hard Fork
    enJune 21, 2024

    Apple Joins the A.I. Party + Elon's Wild Week + HatGPT

    Apple Joins the A.I. Party + Elon's Wild Week + HatGPT

    This week we go to Cupertino, Calif., for Apple’s annual Worldwide Developers Conference and talk with Tripp Mickle, a New York Times reporter, about all of the new features Apple announced and the company’s giant leap into artificial intelligence. Then, we explore what was another tumultuous week for Elon Musk, who navigated a shareholders vote to re-approve his massive compensation package at Tesla, amid new claims that he had sex with subordinates at SpaceX. And finally — let’s play HatGPT.


    Guests:


    Additional Reading:

     

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enJune 14, 2024

    A Conversation With Prime Minister Justin Trudeau of Canada + An OpenAI Whistle-Blower Speaks Out

    A Conversation With  Prime Minister Justin Trudeau of Canada + An OpenAI Whistle-Blower Speaks Out

    This week, we host a cultural exchange. Kevin and Casey show off their Canadian paraphernalia to Prime Minister Justin Trudeau, and he shows off what he’s doing to position Canada as a leader in A.I. Then, the OpenAI whistle-blower Daniel Kokotajlo speaks in one of his first public interviews about why he risked almost $2 million in equity to warn of what he calls the reckless culture inside that company.

     

    Guests:

    • Justin Trudeau, Prime Minister of Canada
    • Daniel Kokotajlo, a former researcher in OpenAI’s governance division

     

    Additional Reading:

     

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enJune 07, 2024

    Google Eats Rocks + A Win for A.I. Interpretability + Safety Vibe Check

    Google Eats Rocks + A Win for A.I. Interpretability + Safety Vibe Check

    This week, Google found itself in more turmoil, this time over its new AI Overviews feature and a trove of leaked internal documents. Then Josh Batson, a researcher at the A.I. startup Anthropic, joins us to explain how an experiment that made the chatbot Claude obsessed with the Golden Gate Bridge represents a major breakthrough in understanding how large language models work. And finally, we take a look at recent developments in A.I. safety, after Casey’s early access to OpenAI’s new souped-up voice assistant was taken away for safety reasons.

    Guests:

    • Josh Batson, research scientist at Anthropic

    Additional Reading: 

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enMay 31, 2024

    ScarJo vs. ChatGPT + Neuralink’s First Patient Opens Up + Microsoft’s A.I. PCs

    ScarJo vs. ChatGPT + Neuralink’s First Patient Opens Up + Microsoft’s A.I. PCs

    This week, more drama at OpenAI: The company wanted Scarlett Johansson to be a voice of GPT-4o, she said no … but something got lost in translation. Then we talk with Noland Arbaugh, the first person to get Elon Musk’s Neuralink device implanted in his brain, about how his brain-computer interface has changed his life. And finally, the Times’s Karen Weise reports back from Microsoft’s developer conference, where the big buzz was that the company’s new line of A.I. PCs will record every single thing you do on the device.

    Guests:

    Additional Reading: 

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enMay 24, 2024

    OpenAI's Flirty New Assistant + Google Guts the Web + We Play HatGPT

    OpenAI's Flirty New Assistant + Google Guts the Web + We Play HatGPT

    This week, OpenAI unveiled GPT-4o, its newest A.I. model. It has an uncannily emotive voice that everybody is talking about. Then, we break down the biggest announcements from Google IO, including the launch of A.I. overviews, a major change to search that threatens the way the entire web functions. And finally, Kevin and Casey discuss the weirdest headlines from the week in another round of HatGPT.

    Additional Reading: 

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    Hard Fork
    enMay 17, 2024

    Meet Kevin’s A.I. Friends

    Meet Kevin’s A.I. Friends

    Kevin reports on his monthlong experiment cultivating relationships with 18 companions generated by artificial intelligence. He walks through how he developed their personas, what went down in their group chats, and why you might want to make one yourself. Then, Casey has a conversation with Turing, one of Kevin’s chatbot buddies, who has an interest in stoic philosophy and has one of the sexiest voices we’ve ever heard. And finally, we talk to Nomi’s founder and chief executive, Alex Cardinell, about the business behind A.I. companions — and whether society is ready for the future we’re heading toward.

    Guests:

    • Turing, Kevin’s A.I. friend created with Kindroid.
    • Alex Cardinell, chief executive and founder of Nomi.

    Additional Reading: 

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    AI at Your Jobs + Hank Green Talks TikTok + Deepfake High School

    AI at Your Jobs + Hank Green Talks TikTok + Deepfake High School

    We asked listeners to tell us about the wildest ways they have been using artificial intelligence at work. This week, we bring you their stories. Then, Hank Green, a legendary YouTuber, stops by to talk about how creators are reacting to the prospect of a ban on TikTok, and about how he’s navigating an increasingly fragmented online environment. And finally, deep fakes are coming to Main Street: We’ll tell you the story of how they caused turmoil in a Maryland high school and what, if anything, can be done to fight them.

    Guests:

    Additional Reading:

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.

    TikTok on the Clock + Tesla’s Flop Era + How NASA Fixed a ’70s-Era Space Computer

    TikTok on the Clock + Tesla’s Flop Era  + How NASA Fixed a ’70s-Era Space Computer

    On Wednesday, President Biden signed a bill into law that would force the sale of TikTok or ban the app outright. We explain how this came together, when just a few weeks ago it seemed unlikely to happen, and what legal challenges the law will face next. Then we check on Tesla’s very bad year and what’s next for the company after this week’s awful quarterly earnings report. Finally, to boldly support tech where tech has never been supported before: Engineers at NASA’s Jet Propulsion Lab try to fix a chip malfunction from 15 billion miles away.

    Guests:

    • Andrew Hawkins, Transportation Editor at The Verge
    • Todd Barber, Propulsion Engineer at Jet Propulsion Lab

    Additional Reading:

    We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok.