Logo
    Search

    Demis Hassabis - Scaling, Superhuman AIs, AlphaZero atop LLMs, Rogue Nations Threat

    enFebruary 28, 2024

    Podcast Summary

    • Advancements in large language models and their surprising effectsCEO of DeepMind, Demis Hassabis, discusses the unexpected improvements in specific domains leading to advancements in other areas for large language models, and encourages computational neuroscience researchers to analyze these models to understand the underlying principles and mechanisms

      The advancements in large language models (LLMs) are surprising and their effectiveness goes beyond what was initially expected. Demis Hassabis, the CEO of DeepMind, believes that there might be underlying principles that govern intelligence, both human and artificial, despite its broad applicability and the existence of specialized skills. He also mentioned that improvements in specific domains can lead to surprising advancements in other areas, as seen in human learning. The most surprising example of this transfer, according to Hassabis, is the potential for LLMs to get better at coding and mathematics, which could lead to a general improvement in reasoning. However, the mechanistic analysis of these systems is not yet sophisticated enough to identify the specific areas of improvement within the neural networks. Hassabis encourages computational neuroscience researchers to apply their knowledge to analyzing these large models, proposing the term "virtual brain analytics." From his neuroscience background, Hassabis believes that neuroscience has added valuable insights to understanding intelligence, but there is still much to learn about the underlying principles and mechanisms.

    • Neuroscience-inspired advancements in AINeuroscientific concepts like reinforcement learning, deep learning, and attention have guided AI development, and understanding how the brain constructs world models and creating efficient planning mechanisms could improve LLMs' reliability and goal achievement.

      Neuroscience has played a significant role in inspiring advancements in artificial intelligence (AI) over the past few decades. Neuroscientific concepts like reinforcement learning, deep learning, experience replay, and attention have provided directional clues for AI development. The discovery that general intelligence is possible due to the existence of the human brain has made researchers believe that it's just a matter of time and effort to make significant progress. Moving forward, there are still challenges to address, such as understanding how the brain constructs world models and creating more efficient planning mechanisms. Large language models (LLMs) could benefit from these advancements by incorporating tree search or planning mechanisms on top of them, making them more reliable world models and enabling them to make concrete plans to achieve goals. However, making these approaches more efficient remains a challenge, with researchers focusing on sample-efficient methods, reusing existing data, and developing better world models to make search more efficient.

    • The importance of world models in AI search efficiencyAdvanced world models improve search efficiency in AI, while brute force systems rely heavily on search. Defining the right objective function for real-world problems is a challenge, but the future of RL may hold the answer with synthetic data generation.

      The sophistication of the world models we build plays a crucial role in the efficiency of search in artificial intelligence. Brute force systems, which lack advanced models, rely heavily on search to make decisions. However, human intelligence, with its rich and accurate models, allows for more efficient decision-making with less search. The challenge in real-world systems is defining the right objective function or reward function for the AI. Games, like Go, provide a concrete win condition, making it easier to specify a reward function. In contrast, real-world problems and scientific discoveries require a more nuanced approach. Humans use their intuition, knowledge, and experience to build accurate models and make efficient decisions. The future of reinforcement learning (RL) may hold the answer to generating synthetic data to overcome data bottlenecks. I'm optimistic about this potential development, as there is still a vast amount of data to be utilized and society continues to generate more data daily. The ability to generate synthetic data could significantly improve the efficiency and effectiveness of AI systems.

    • Expanding machine learning capabilities with synthetic dataSynthetic data generated through simulation and self-play is a promising approach to expand machine learning models, but ensuring it's new and representative is complex. Data curation and analysis are essential for fairness, bias, and representation.

      Creating synthetic data through simulation and self-play is a promising approach to expand the capabilities of machine learning models. However, ensuring that synthetic data is not just a repetition of what's already in the dataset but something new and representative of the distribution being learned is a complex challenge. Data curation and analysis are essential to address issues of fairness, bias, and representation. Older ideas, such as scaling reinforcement learning and combining it with deep learning, may hold significant potential when combined with new advances in large models. While there is theoretically no reason why AGI couldn't come from a pure RL approach, starting with existing knowledge and using it as a prior to make predictions and bootstrap learning seems a more plausible and efficient way to AGI.

    • The importance of grounding in AGI systemsAGI systems with large multimodal models may need additional planning and search capabilities for abstract concepts. Grounding, which comes from human feedback or language data, is crucial for understanding and interpreting concepts. Continuous scaling and innovation are necessary to tackle philosophical questions and increasing complexity.

      The final Artificial General Intelligence (AGI) system is expected to include large multimodal models as part of the solution, but they may not be sufficient on their own. Additional planning and search capabilities will likely be required. This idea, known as the scaling hypothesis, has been surprising in its effectiveness, but it may not be enough for explicitly abstract concepts or neatly defined concepts. However, these systems have shown implicit learning of concepts and some form of grounding, which is unexpected given their reliance on language. The grounding may come from RLHF (Reinforcement Learning with Human Feedback) systems, as human raters provide grounded feedback, or from the vast amount of information contained in language. The advances made in this field have only scratched the surface of philosophical questions, and it's important to continue pushing scaling as hard as possible while also inventing new architectures and algorithms. The grounding may become more difficult as models get smarter and operate in domains where human labels are insufficient, and as more compute is used for tasks beyond next token prediction. Overall, both scaling and innovation should be pursued aggressively.

    • Connecting AI to the physical world and human thoughtGrounding AI systems enables them to understand and interact with the environment, improving real-world applications. Ensuring alignment with human values and understanding is crucial as AI becomes more autonomous.

      Grounding is crucial for AI systems to achieve their goals effectively in the real world. Grounding refers to the connection of AI systems to the physical world and human thought, enabling them to understand and interact with the environment. As AI systems become more multimodal and ingest various types of data, they will start to understand the physics of the real world better. This understanding will enable feedback loops and active learning, which are essential for robotics and other real-world applications. However, as these systems become smarter and more autonomous, ensuring their alignment with human values and understanding becomes increasingly challenging. Ideas for addressing this include developing more stringent evaluation systems, creating hardened sandboxes or simulations, and using narrow AI tools to help analyze and understand the concepts and representations the system is building. Ultimately, the goal is to create AI systems that are both powerful and controllable, capable of transformative scientific discoveries while minimizing potential negative consequences.

    • Making Progress Towards AGI: Balancing Optimism and CautionWe're optimistic about developing AGI within the next decade, but we need to ensure safety and understand its capabilities before proceeding. AGI could lead to faster progress, but its timeline is uncertain. Safety implications, transparency, and explanation are crucial. Sandbox simulations can help identify potential issues.

      We're making significant progress towards developing Artificial General Intelligence (AGI) within the next decade, but it's essential to ensure safety and understand the systems better before continuing development. Demis Hassabis, the CEO of DeepMind, shares this optimistic view and believes that once we have an AGI system, it could lead to faster progress in AI research. However, the outcome's timeline is uncertain due to numerous unknowns and human ingenuity's unpredictability. Additionally, the safety implications of AGI systems designing future versions of themselves must be considered. To proceed with confidence, we need to develop the right evaluations, metrics, and possibly formal proofs to understand the capabilities and limitations of these systems. Transparency and explanation are crucial, as they could open up possibilities of using AGI systems to explain their thought processes and even design future versions. Conversely, there could be unexpected observations that might require halting the training of AGI systems, such as observations related to deception or unintended behaviors. Sandbox simulations could help us identify these potential issues. In summary, while we're making progress towards AGI, it's crucial to be prepared for the challenges and opportunities that come with it.

    • Exploring the unexpected capabilities and challenges of scaling up AI modelsScaling up AI models brings unexpected capabilities and challenges, requiring continuous optimization, practical limitations management, and understanding of reasons behind occurrences.

      Scaling up AI models like Gemini comes with unexpected capabilities and challenges. While it's possible to predict some metrics like training loss, the actual capabilities may not follow linearly. The development process involves adjusting recipes and obtaining new data points to optimize hyperparameters. Practical limitations include compute capacity, distributed computing challenges, and the need for hardware innovations. Despite perceptions, DMG's Gemini used roughly the same amount of compute as rumored for other labs' models. The process requires efficient use of compute and continuous innovation. Unexpected capabilities and challenges make it essential to pause, investigate, and understand the reasons behind these occurrences before continuing the development.

    • Exploring new ideas in AI requires substantial compute resources and a focus on generality and learningNew ideas in AI may not work at small scales but can be valuable at larger scales, making exploration compute-intensive. Google's investment in generality and learning led to breakthroughs like reinforcement learning, search, and deep learning, which have scaled and not required extensive human priors.

      Significant innovation and progress in AI require both substantial compute resources and a focus on generality and learning. The speaker, who co-founded DeepMind and is now at Google, notes that new ideas may not work at small scales but can be valuable at larger scales, making the exploration process compute-intensive. He also mentions that Google, as a research lab, is fortunate to have the most compute this year. Looking back to 2010 when DeepMind was founded, the speaker admits they didn't anticipate the need for such massive investments in models. However, they did bet on generality and learning, which led them to reinforcement learning, search, and deep learning. These algorithms have scaled and not required extensive human priors, as opposed to the rigid logic-based systems of the past. The speaker believes they were on the right track and that the success of DeepMind and subsequent advancements, like AlphaGo and transformers, have inspired others. As we look to the future, the speaker emphasizes the importance and potential consequences of superhuman intelligence and the need for appropriate governance.

    • The role of stakeholders in advancing AI and ethical considerationsAI's integration into society requires collaboration from various stakeholders, and ethical considerations are crucial as AI's capabilities expand beyond memorization to generalization and imagination.

      The advancement of AI and its integration into our society is a collaborative effort that requires involvement from various stakeholders including civil society, academia, and government. The recent popularity of chatbot systems has opened up conversations about the potential uses and ethical considerations of AI. These systems have immense power but are still at the beginning of their capabilities, and we need new advancements in areas like planning, search, personalization, and memory for them to be truly useful in our everyday lives. The potential benefits of AI are vast, from curing diseases to addressing climate change, but it's crucial that we reach international consensus on how to deploy these models responsibly and ensure that the benefits are accessible to everyone. The claim that these models are just memorizing is a simplistic view, and while memorization is a form of compression, it's not the only capability we need. The new era of AI is capable of generalizing to new constructs and going beyond memorization. The link between memory and imagination, as discussed in a 2007 paper, highlights the importance of understanding the nuanced capabilities of AI and recognizing that it's not just about memorization.

    • The importance of imagination in AI and its absence in current systemsGoogle DeepMind recognizes the need for AI to possess imagination for planning and plans to publish more about it, while implementing safeguards to secure their models from potential misuse or theft.

      Imagination in AI, as in human cognition, is a reconstructive process using familiar components to create something novel for specific purposes, such as planning. This idea, which started in neuroscience, is still missing from current AI systems, which primarily simulate new experiences by combining different parts of a world model. Google DeepMind, a leading AI lab, recognizes the importance of this concept and plans to publish more publicly about it, as well as implementing safeguards to secure their models from potential misuse or theft. They already have robust cybersecurity measures in place, but acknowledge the need for continued improvement and potential future solutions like air gaps and secure data centers. The open-source nature of AI technology raises additional questions about preventing bad actors from repurposing it for harmful ends.

    • Balancing open sourcing and security concerns for advanced AI modelsExperts from cybersecurity and safety institutions are crucial in addressing security concerns for advanced AI models. Google, with its resources and expertise, is well-positioned to lead in multimodal interactions. Collaboration with independent researchers and safety-focused institutions is essential for safe development and deployment.

      As AI models become more advanced and capable, there is a need to balance open sourcing for collaboration and progress with security concerns. While tech deserves credit for funding R&D, the access to these advanced systems raises questions about securing them from potential threats. Expertise from cybersecurity and safety institutions will be crucial in addressing these concerns. Additionally, multimodal interactions with AI systems, such as through chat, video, and voice, are expected to evolve and become more fluid in the coming years. Google, with its advanced model and resources, is well-positioned to lead in this area, leveraging its history in simulations, games environments, and cybersecurity expertise. However, the vast surface area of these general systems necessitates automated testing and boundary condition exploration. Collaboration with independent researchers and safety-focused institutions will also be essential in ensuring the safe development and deployment of these advanced AI systems.

    • The Future of Technology is MultimodalTrue multimodality, enabling machines to understand and interact with the world through multiple senses, is becoming a reality, revolutionizing robotics and other fields. AI can assist humans by narrowing down search spaces and finding solutions to complex problems, but it's not yet capable of asking questions or creating hypotheses.

      We are on the brink of a new era in technology, where true multimodality – the ability for machines to understand and interact with the world through multiple senses, including touch and various types of sensors – is becoming a reality. This is an exciting development, but it presents challenges, particularly in the field of robotics, where data is scarce. However, progress is being made in areas like robot transformer and other multimodal systems, which can learn from data in various domains and then apply that knowledge to new tasks. These systems are still in development, but they have the potential to revolutionize robotics and other fields. Additionally, while AI is making strides in areas like math and coding, it is not yet capable of asking the right questions or creating hypotheses, which are crucial aspects of scientific discovery. Instead, AI can help humans by narrowing down the search space and finding solutions to complex problems. Overall, the future of technology is multimodal, and while there are challenges to overcome, the potential benefits are significant.

    • DeepMind focuses on domain-specific solutions while waiting for AGIDeepMind prioritizes practical benefits in AI research, collaborating with Google Brain to develop advanced systems and maintain a safety-conscious approach.

      DeepMind, despite its potential to bring about AGI in the future, chooses to focus on domain-specific solutions now due to the uncertainty of when AGI will arrive and the immediate benefits these solutions can bring to various fields such as science, health, and everyday life. The integration of DeepMind and Google Brain, known as Gemini, has proven successful in bringing together world-class organizations, resources, and expertise to build advanced AI systems and keep research on the right path. The collaboration also ensures a responsible and safety-conscious approach to AI development, as both DeepMind and Google have taken these concerns seriously.

    • Approaching new technologies with caution and responsibilityBe thoughtful, humble, and use the scientific method when developing and deploying new technologies. Focus on responsible scaling and address potential risks before deployment.

      As we continue to develop and deploy advanced technologies, it's crucial to approach them with caution and responsibility. These technologies, while bringing immense potential for positive change, also come with significant uncertainties and potential risks. We must be thoughtful, humble, and use the scientific method to understand these systems and their consequences before widespread deployment. This means moving away from a "move fast and break things" attitude and instead focusing on responsible scaling. When evaluating new technologies, ensuring security and detecting potential risks ahead of time are essential. If a capability could be misused, it should be addressed before deployment. The recent interest in and advancements of AI and related technologies may have been a surprise to some, but for those who have anticipated these developments, it's essential to continue to approach them with caution and responsibility.

    • Navigating the chaos of AI developmentAI development is growing rapidly, but it's crucial to approach it responsibly and ethically to ensure careful progression

      The current state of AI development is experiencing significant growth and chaos, with a surge in VC funding and numerous assistant-type systems emerging. This has created a new environment for the field, but it also poses challenges. It's essential for the AI community to approach this development responsibly, thoughtfully, and scientifically, using the scientific method to ensure a careful and optimistic progression. The risk is that the importance of this approach could be lost amidst the excitement and rush of the current AI landscape. Overall, it's crucial to remember the importance of ethical and responsible development as AI continues to evolve. Thank you, Demos, for sharing your insights on this topic. We hope you enjoyed this episode, and please remember to share the podcast with others who might find it interesting. Stay tuned for more thought-provoking discussions on AI and related topics.

    Recent Episodes from Dwarkesh Podcast

    Tony Blair - Life of a PM, The Deep State, Lee Kuan Yew, & AI's 1914 Moment

    Tony Blair - Life of a PM, The Deep State, Lee Kuan Yew, & AI's 1914 Moment

    I chatted with Tony Blair about:

    - What he learned from Lee Kuan Yew

    - Intelligence agencies track record on Iraq & Ukraine

    - What he tells the dozens of world leaders who come seek advice from him

    - How much of a PM’s time is actually spent governing

    - What will AI’s July 1914 moment look like from inside the Cabinet?

    Enjoy!

    Watch the video on YouTube. Read the full transcript here.

    Follow me on Twitter for updates on future episodes.

    Sponsors

    - Prelude Security is the world’s leading cyber threat management automation platform. Prelude Detect quickly transforms threat intelligence into validated protections so organizations can know with certainty that their defenses will protect them against the latest threats. Prelude is backed by Sequoia Capital, Insight Partners, The MITRE Corporation, CrowdStrike, and other leading investors. Learn more here.

    - This episode is brought to you by Stripe, financial infrastructure for the internet. Millions of companies from Anthropic to Amazon use Stripe to accept payments, automate financial processes and grow their revenue.

    If you’re interested in advertising on the podcast, check out this page.

    Timestamps

    (00:00:00) – A prime minister’s constraints

    (00:04:12) – CEOs vs. politicians

    (00:10:31) – COVID, AI, & how government deals with crisis

    (00:21:24) – Learning from Lee Kuan Yew

    (00:27:37) – Foreign policy & intelligence

    (00:31:12) – How much leadership actually matters

    (00:35:34) – Private vs. public tech

    (00:39:14) – Advising global leaders

    (00:46:45) – The unipolar moment in the 90s



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
    Dwarkesh Podcast
    enJune 26, 2024

    Francois Chollet, Mike Knoop - LLMs won’t lead to AGI - $1,000,000 Prize to find true solution

    Francois Chollet, Mike Knoop - LLMs won’t lead to AGI - $1,000,000 Prize to find true solution

    Here is my conversation with Francois Chollet and Mike Knoop on the $1 million ARC-AGI Prize they're launching today.

    I did a bunch of socratic grilling throughout, but Francois’s arguments about why LLMs won’t lead to AGI are very interesting and worth thinking through.

    It was really fun discussing/debating the cruxes. Enjoy!

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here.

    Timestamps

    (00:00:00) – The ARC benchmark

    (00:11:10) – Why LLMs struggle with ARC

    (00:19:00) – Skill vs intelligence

    (00:27:55) - Do we need “AGI” to automate most jobs?

    (00:48:28) – Future of AI progress: deep learning + program synthesis

    (01:00:40) – How Mike Knoop got nerd-sniped by ARC

    (01:08:37) – Million $ ARC Prize

    (01:10:33) – Resisting benchmark saturation

    (01:18:08) – ARC scores on frontier vs open source models

    (01:26:19) – Possible solutions to ARC Prize



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
    Dwarkesh Podcast
    enJune 11, 2024

    Leopold Aschenbrenner - China/US Super Intelligence Race, 2027 AGI, & The Return of History

    Leopold Aschenbrenner - China/US Super Intelligence Race, 2027 AGI, & The Return of History

    Chatted with my friend Leopold Aschenbrenner on the trillion dollar nationalized cluster, CCP espionage at AI labs, how unhobblings and scaling can lead to 2027 AGI, dangers of outsourcing clusters to Middle East, leaving OpenAI, and situational awareness.

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here.

    Follow me on Twitter for updates on future episodes. Follow Leopold on Twitter.

    Timestamps

    (00:00:00) – The trillion-dollar cluster and unhobbling

    (00:20:31) – AI 2028: The return of history

    (00:40:26) – Espionage & American AI superiority

    (01:08:20) – Geopolitical implications of AI

    (01:31:23) – State-led vs. private-led AI

    (02:12:23) – Becoming Valedictorian of Columbia at 19

    (02:30:35) – What happened at OpenAI

    (02:45:11) – Accelerating AI research progress

    (03:25:58) – Alignment

    (03:41:26) – On Germany, and understanding foreign perspectives

    (03:57:04) – Dwarkesh’s immigration story and path to the podcast

    (04:07:58) – Launching an AGI hedge fund

    (04:19:14) – Lessons from WWII

    (04:29:08) – Coda: Frederick the Great



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
    Dwarkesh Podcast
    enJune 04, 2024

    John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

    John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

    Chatted with John Schulman (cofounded OpenAI and led ChatGPT creation) on how posttraining tames the shoggoth, and the nature of the progress to come...

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

    Timestamps

    (00:00:00) - Pre-training, post-training, and future capabilities

    (00:16:57) - Plan for AGI 2025

    (00:29:19) - Teaching models to reason

    (00:40:50) - The Road to ChatGPT

    (00:52:13) - What makes for a good RL researcher?

    (01:00:58) - Keeping humans in the loop

    (01:15:15) - State of research, plateaus, and moats

    Sponsors

    If you’re interested in advertising on the podcast, fill out this form.

    * Your DNA shapes everything about you. Want to know how? Take 10% off our Premium DNA kit with code DWARKESH at mynucleus.com.

    * CommandBar is an AI user assistant that any software product can embed to non-annoyingly assist, support, and unleash their users. Used by forward-thinking CX, product, growth, and marketing teams. Learn more at commandbar.com.



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
    Dwarkesh Podcast
    enMay 15, 2024

    Mark Zuckerberg - Llama 3, Open Sourcing $10b Models, & Caesar Augustus

    Mark Zuckerberg - Llama 3, Open Sourcing $10b Models, & Caesar Augustus

    Mark Zuckerberg on:

    - Llama 3

    - open sourcing towards AGI

    - custom silicon, synthetic data, & energy constraints on scaling

    - Caesar Augustus, intelligence explosion, bioweapons, $10b models, & much more

    Enjoy!

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Human edited transcript with helpful links here.

    Timestamps

    (00:00:00) - Llama 3

    (00:08:32) - Coding on path to AGI

    (00:25:24) - Energy bottlenecks

    (00:33:20) - Is AI the most important technology ever?

    (00:37:21) - Dangers of open source

    (00:53:57) - Caesar Augustus and metaverse

    (01:04:53) - Open sourcing the $10b model & custom silicon

    (01:15:19) - Zuck as CEO of Google+

    Sponsors

    If you’re interested in advertising on the podcast, fill out this form.

    * This episode is brought to you by Stripe, financial infrastructure for the internet. Millions of companies from Anthropic to Amazon use Stripe to accept payments, automate financial processes and grow their revenue. Learn more at stripe.com.

    * V7 Go is a tool to automate multimodal tasks using GenAI, reliably and at scale. Use code DWARKESH20 for 20% off on the pro plan. Learn more here.

    * CommandBar is an AI user assistant that any software product can embed to non-annoyingly assist, support, and unleash their users. Used by forward-thinking CX, product, growth, and marketing teams. Learn more at commandbar.com.



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Sholto Douglas & Trenton Bricken - How to Build & Understand GPT-7's Mind

    Sholto Douglas & Trenton Bricken - How to Build & Understand GPT-7's Mind

    Had so much fun chatting with my good friends Trenton Bricken and Sholto Douglas on the podcast.

    No way to summarize it, except: 

    This is the best context dump out there on how LLMs are trained, what capabilities they're likely to soon have, and what exactly is going on inside them.

    You would be shocked how much of what I know about this field, I've learned just from talking with them.

    To the extent that you've enjoyed my other AI interviews, now you know why.

    So excited to put this out. Enjoy! I certainly did :)

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. 

    There's a transcript with links to all the papers the boys were throwing down - may help you follow along.

    Follow Trenton and Sholto on Twitter.

    Timestamps

    (00:00:00) - Long contexts

    (00:16:12) - Intelligence is just associations

    (00:32:35) - Intelligence explosion & great researchers

    (01:06:52) - Superposition & secret communication

    (01:22:34) - Agents & true reasoning

    (01:34:40) - How Sholto & Trenton got into AI research

    (02:07:16) - Are feature spaces the wrong way to think about intelligence?

    (02:21:12) - Will interp actually work on superhuman models

    (02:45:05) - Sholto’s technical challenge for the audience

    (03:03:57) - Rapid fire



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Demis Hassabis - Scaling, Superhuman AIs, AlphaZero atop LLMs, Rogue Nations Threat

    Demis Hassabis - Scaling, Superhuman AIs, AlphaZero atop LLMs, Rogue Nations Threat

    Here is my episode with Demis Hassabis, CEO of Google DeepMind

    We discuss:

    * Why scaling is an artform

    * Adding search, planning, & AlphaZero type training atop LLMs

    * Making sure rogue nations can't steal weights

    * The right way to align superhuman AIs and do an intelligence explosion

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here.

    Timestamps

    (0:00:00) - Nature of intelligence

    (0:05:56) - RL atop LLMs

    (0:16:31) - Scaling and alignment

    (0:24:13) - Timelines and intelligence explosion

    (0:28:42) - Gemini training

    (0:35:30) - Governance of superhuman AIs

    (0:40:42) - Safety, open source, and security of weights

    (0:47:00) - Multimodal and further progress

    (0:54:18) - Inside Google DeepMind



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Patrick Collison (Stripe CEO) - Craft, Beauty, & The Future of Payments

    Patrick Collison (Stripe CEO) - Craft, Beauty, & The Future of Payments

    We discuss:

    * what it takes to process $1 trillion/year

    * how to build multi-decade APIs, companies, and relationships

    * what's next for Stripe (increasing the GDP of the internet is quite an open ended prompt, and the Collison brothers are just getting started).

    Plus the amazing stuff they're doing at Arc Institute, the financial infrastructure for AI agents, playing devil's advocate against progress studies, and much more.

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

    Timestamps

    (00:00:00) - Advice for 20-30 year olds

    (00:12:12) - Progress studies

    (00:22:21) - Arc Institute

    (00:34:27) - AI & Fast Grants

    (00:43:46) - Stripe history

    (00:55:44) - Stripe Climate

    (01:01:39) - Beauty & APIs

    (01:11:51) - Financial innards

    (01:28:16) - Stripe culture & future

    (01:41:56) - Virtues of big businesses

    (01:51:41) - John



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Tyler Cowen - Hayek, Keynes, & Smith on AI, Animal Spirits, Anarchy, & Growth

    Tyler Cowen - Hayek, Keynes, & Smith on AI, Animal Spirits, Anarchy, & Growth

    It was a great pleasure speaking with Tyler Cowen for the 3rd time.

    We discussed GOAT: Who is the Greatest Economist of all Time and Why Does it Matter?, especially in the context of how the insights of Hayek, Keynes, Smith, and other great economists help us make sense of AI, growth, animal spirits, prediction markets, alignment, central planning, and much more.

    The topics covered in this episode are too many to summarize. Hope you enjoy!

    Watch on YouTube. Listen on Apple PodcastsSpotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

    Timestamps

    (0:00:00) - John Maynard Keynes

    (00:17:16) - Controversy

    (00:25:02) - Fredrick von Hayek

    (00:47:41) - John Stuart Mill

    (00:52:41) - Adam Smith

    (00:58:31) - Coase, Schelling, & George

    (01:08:07) - Anarchy

    (01:13:16) - Cheap WMDs

    (01:23:18) - Technocracy & political philosophy

    (01:34:16) - AI & Scaling



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Lessons from The Years of Lyndon Johnson by Robert Caro [Narration]

    Lessons from The Years of Lyndon Johnson by Robert Caro [Narration]

    This is a narration of my blog post, Lessons from The Years of Lyndon Johnson by Robert Caro.

    You read the full post here: https://www.dwarkeshpatel.com/p/lyndon-johnson

    Listen on Apple Podcasts, Spotify, or any other podcast platform. Follow me on Twitter for updates on future posts and episodes.



    Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

    Related Episodes

    How to decode a thought

    How to decode a thought
    Can researchers decipher what people are thinking about just by looking at brain scans? With AI, they're getting closer. How far can they go, and what does it mean for privacy? To buy tickets to our upcoming live show in New York, go to http://vox.com/unexplainablelive For more, go to http://vox.com/unexplainable It’s a great place to view show transcripts and read more about the topics on our show. Also, email us! unexplainable@vox.com We read every email. Support Unexplainable by making a financial contribution to Vox! bit.ly/givepodcasts Learn more about your ad choices. Visit podcastchoices.com/adchoices

    A.I. Vibe Check With Ezra Klein + Kevin Tries Phone Positivity

    A.I. Vibe Check With Ezra Klein + Kevin Tries Phone Positivity

    The New York Times Opinion columnist Ezra Klein has spent years talking to artificial intelligence researchers. Many of them feel the prospect of A.I. discovery is too sweet to ignore, regardless of the technology’s risks.

    Today, Mr. Klein discusses the profound changes that an A.I.-powered world will create, how current business models are failing to meet the A.I. moment, and the steps government can take to achieve a positive A.I. future.

    Also, radical acceptance of your phone addiction may just help your phone addiction.

    On today’s episode:

    Additional reading:

    #120 - GigaChat + HuggingChat, a LOT of research, EU Act passed, #promptography

    #120 - GigaChat + HuggingChat, a LOT of research, EU Act passed, #promptography

    Our 120th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter at https://lastweekin.ai/

    Check out Jeremie's new book Quantum Physics Made Me Do It

    Quantum Physics Made Me Do It tells the story of human self-understanding through the lens of physics. It explores what we can and can’t know about reality, and how tiny tweaks to quantum theory can reshape our entire picture of the universe. And because I couldn't resist, it explains what that story means for AI and the future of sentience   

    You can find it on Amazon in the UK, Canada, and the US — here are the links:

    UK version | Canadian version | US version 

     

    Outline:

    (00:00) Intro / Banter (04:35) Episode Preview (06:00) Russia's Sberbank releases ChatGPT rival GigaChat + Hugging Face releases its own version of ChatGPT + Stability AI launches StableLM, an open source ChatGPT alternative (14:30) Stack Overflow joins Reddit and Twitter in charging AI companies for training data + Inside the secret list of websites that make AI like ChatGPT sound smart (24:45) Big Tech is racing to claim its share of the generative AI market (27:42) Microsoft Building Its Own AI Chip on TSMC's 5nm Process (30:45) Snapchat’s getting review-bombed after pinning its new AI chatbot to the top of users’ feeds (33:30) Create generative AI video-to-video right from your phone with Runway’s iOS app (35:50) Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models (40:30) Autonomous Agents & Agent Simulations (46:13) Scaling Transformer to 1M tokens and beyond with RMT (49:05) Meet MiniGPT-4: An Open-Source AI Model That Performs Complex Vision-Language Tasks Like GPT-4 (50:50) Visual Instruction Tuning (52:25) AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head (54:05) Performance of ChatGPT on the US Fundamentals of Engineering Exam: Comprehensive Assessment of Proficiency and Potential Implications for Professional Environmental Engineering Practice (58:20) ChatGPT is still no match for humans when it comes to accounting (01:01:13) Large Language Models Are Human-Level Prompt Engineers (01:05:00) RedPajama, a project to create leading open-source models, starts by reproducing LLaMA training dataset of over 1.2 trillion tokens (01:05:55) Do Embodied Agents Dream of Pixelated Sheep: Embodied Decision Making using Language Guided World Modelling (01:08:45) Fundamental Limitations of Alignment in Large Language Models (01:11:35) Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond (01:15:40) Tool Learning with Foundation Models (01:17:20) With AI Watermarking, Creators Strike Back (01:22:02) EU lawmakers pass draft of AI Act, includes copyright rules for generative AI (01:26:44) How can we build human values into AI? (01:32:20) How prompt injection can hijack autonomous AI agents like Auto-GPT (01:34:30) AI Simply Needs a Kill Switch (01:39:35) Anthropic calls for $15 million in funding to boost the government’s AI risk assessment work (01:41:48) ‘AI isn’t a threat’ – Boris Eldagsen, whose fake photo duped the Sony judges, hits back (01:45:20) AI Art Sites Censor Prompts About Abortion (01:48:15) Outro

    #113 - Nvidia’s 10k GPU, Toolformer, AI alignment, John Oliver

    #113 - Nvidia’s 10k GPU, Toolformer, AI alignment, John Oliver

    Our 113th episode with a summary and discussion of last week's big AI news!

    Check out our text newsletter at https://lastweekin.ai/

    Stories this week:

    The rise of AutoGPTs and AI anxieties with Sunny Madra and Vinny Lingham | E1720

    The rise of AutoGPTs and AI anxieties with Sunny Madra and Vinny Lingham | E1720

    Vinny and Sunny join Jason to discuss AI’s blistering pace and compare its development to other product launches (2:47). They also break down the anxiety artists and developers feel from AI automation (14:47), the rise of AutoGPTs, Twitter’s reported LLM project, and more (37:01).

    (0:00) Jason kicks off the show

    (2:47) The blistering pace of AI

    (9:44) Developer builds Flappy Bird in 1-hour

    (11:24) Squarespace - Use offer code TWIST to save 10% off your first purchase of a website or domain at https://Squarespace.com/TWIST

    (12:53) Developers leveraging gains in AI

    (14:47) Ai anxiety

    (23:55) Instacart’s ChatGPT plugin 

    (26:06) Preserving your advantage against AI

    (29:10) Vanta - Get $1000 off your SOC 2 at https://vanta.com/twist

    (30:14) AI artists

    (35:54) Crowdbotics - Get a free scoping session for your next big app idea at http://crowdbotics.com/twist

    (37:01) Automating with AutoGPT

    (47:09) LLMs vs. Knowledge retrieval systems 

    (57:06) Is Google’s Bard behind? 

    (1:02:42) Twitter’s alleged generative AI project

    (1:08:30) Ethereum and Bitcoin updates


    FOLLOW Sunny: https://twitter.com/sundeep

    FOLLOW Vinny: https://twitter.com/VinnyLingham

    FOLLOW Jason: https://linktr.ee/calacanis


    Subscribe to our YouTube to watch all full episodes:

    https://www.youtube.com/channel/UCkkhmBWfS7pILYIk0izkc3A?sub_confirmation=1


    FOUNDERS! Subscribe to the Founder University podcast:

    https://podcasts.apple.com/au/podcast/founder-university/id1648407190