Logo

    280 | François Chollet on Deep Learning and the Meaning of Intelligence

    enJune 24, 2024
    What limitations do large language models have according to Francois Chollet?
    How does the Archathon competition aim to improve AI systems?
    What are practical applications of large language models in business?
    Why should LLMs not be solely relied upon for decision-making?
    What characterizes true creativity in algorithms according to the text?

    Podcast Summary

    • AI memorization vs. intelligenceLarge language models, while impressive in mimicking human speech and behavior, lack the ability to abstract and reason beyond what they have learned, and are primarily memorizing and interpolating information.

      While large language models, such as those developed using software libraries like Keras, have made significant strides in mimicking human speech and behavior, they are not truly intelligent in the conventional sense. Francois Chollet, a deep learning researcher at Google, argues that these models have merely memorized a vast amount of information and can interpolate between it, allowing them to perform well on various intelligence tests. However, they lack the ability to abstract and reason beyond what they have learned. Chollet's perspective is not unique, but it is a reminder that while large language models are impressive, they are not yet capable of true intelligence. The Archathon competition, which focuses on logic puzzles not already in the training data, aims to push researchers to develop AI systems that can go beyond memorization and interpolation. The debate about the definition and achievement of intelligence in AI is ongoing, but it is clear that significant progress has been made in the connectionist tradition of machine learning, and the field continues to evolve.

    • Transformer architectures in deep learningTransformers surpassed RNNs for sequence processing but have limitations in modeling complex situations and providing exact quantitative answers, requiring refinement using symbolic search or human intervention.

      The advancements in deep learning have been marked by significant improvements in image and sequence processing, driven by the development of new GPU-based models and architectures. One of the most notable shifts was the rise of transformer architectures, which surpassed the capabilities of earlier recurrent neural networks for sequence processing. The transformers were trained on vast amounts of data, often collected and annotated by thousands of people, enabling them to prefer the correct answers across a wide range of queries. However, these models have limitations. They operate by mapping queries onto a high-dimensional curve, interpolating across data points but unable to model complex situations or understand queries in quantitative terms. As a result, they can provide directionally accurate answers but are not reliable for exact answers, especially for quantitative problems. It's essential to use these models as a stepping stone and refine the answers using a symbolic search system or human intervention. Overall, the progress in deep learning can be understood as a shift from models where the programmer builds in a structure to models where the model learns a structure from data, with the intelligence residing in the model's ability to learn rather than in the programmer.

    • AI limitations and creativityAI's curve fitting and symbolic approaches have limitations when it comes to true creativity, as they are both limited to interpolations between existing data points and cannot create new ideas. New methods that can operate in a more unconstrained search space are needed to achieve true creativity.

      Both curve fitting and symbolic approaches in AI have their limitations when it comes to true creativity. Curve fitting, which is a common method for learning patterns from data, is limited to interpolations between existing data points and cannot create anything new. On the other hand, symbolic approaches, which rely on a predefined search space, are limited by the programmer's imagination and ability to anticipate all possible scenarios. To achieve true creativity, an algorithm needs to operate in a relatively unconstrained search space with an open-ended search process. Genetic algorithms, for example, can demonstrate creativity by generating new combinations and mutations, leading to inventions that humans could not anticipate. LLMs, which have received much attention lately, work by mapping words or tokens into a vector space and computing pairwise distances between them. The idea is that the distance between points represents a semantically similar area. While there is a connection between this approach and the way the brain learns, the result is a high-dimensional manifold that is limited to interpolations between existing data points and cannot create truly novel ideas. To summarize, the limitations of current AI approaches, whether it's curve fitting or symbolic, highlight the need for new methods that can operate in a more unconstrained search space and demonstrate true creativity.

    • Large Language Models and ManifoldsLarge Language Models (LLMs) function by organizing information in a high-dimensional manifold, enabling efficient encoding of semantically rich relationships and discovery of useful semantic programs. However, they are limited in their ability to compose different programs and can only generate results based on memorized patterns from training data.

      Large language models (LLMs) function by organizing information in a high-dimensional space called a manifold. This manifold is differentiable and smooth, allowing for efficient and scalable fitting of curves and encoding of semantically rich relationships between tokens. The organization of tokens on this manifold results in the discovery of useful semantic programs, or patterns of data transformation. These programs help compress the data and express it in a more concise fashion. Essentially, LLMs are vast stores of vector functions, or programs, which can interpolate between different transformations. These vector functions enable the encoding of semantically meaningful relationships and can even transform the style of text, such as turning a paragraph into poetry. The magic of deep learning lies in expressing relationships between things as a distance function in vector space and fitting curves to a wide range of data. However, despite the large size and complexity of these models, they are limited in their ability to compose different programs and can only generate results based on patterns they have memorized from their training data. This makes them sensitive to the way queries are phrased and prone to failure when presented with unfamiliar information. Data annotation and prompt engineering are crucial for improving the performance of LLMs by teaching them new patterns and refining their understanding of existing ones.

    • Large Language ModelsLarge Language Models can generate human-like responses but lack general intelligence and human intuition, and their ability to handle novelty is limited.

      Large Language Models (LLMs) can generate human-like responses by interpolating between known data points and creating new responses based on patterns they've learned. However, they don't have a symbolic model of the world or general intelligence. Instead, they work with models of world and semantic space, which can be interpolated and merged to create new responses. These models are not as generalizable as human intuition and can only handle novelty to a certain extent. While LLMs can do active inference and generate new things, it's data inefficient and they lack the ability to truly adapt and learn new skills like humans can. So, while LLMs have some level of intelligence, they are vastly less intuitive and adaptable than humans. It's important to remember that LLMs are tools that can be useful for specific tasks, but they don't possess human-like understanding or consciousness.

    • LLM limitationsLarge Language Models excel at memorization but lack the ability to understand concepts from first principles and struggle with novel problems

      While Large Language Models (LLMs) excel at memorization and can mimic human-like responses, they lack the ability to understand concepts from first principles and struggle with novel problems. Schools often focus on memorization, but true intelligence requires understanding and the ability to solve novel problems. The Arc-AGI benchmark, which includes unique, never-before-seen puzzles, highlights this limitation of LLMs. The competition, which incentivizes new ideas, shows that LLMs cannot adapt to novelty in the way that human intelligence can. To test true intelligence, we need to move beyond memorization and focus on problems that require understanding and the ability to adapt to novel situations.

    • Large Language Models and General IntelligenceLarge Language Models are not truly intelligent beings, they function as interpretive databases, lacking the ability to reason, plan, adapt, or invent, and the idea of them being on the path to Human-Level General Intelligence is misguided.

      Large Language Models (LLMs) are not truly intelligent beings as they lack the ability to reason or plan. Instead, they function as interpretive databases of programs, memorizing and reapplying patterns. LLMs cannot adapt to new situations or synthesize new programs on the fly, which is a key aspect of general intelligence. The idea that LLMs are on the path to Human-Level General Intelligence (HGI) based on their current performance levels is misguided. The field is still in the process of understanding how LLMs work and interpreting their behavior. While there have been some insights, there is no evidence that LLMs can synthesize new programs or possess the ability to invent. The difference between LLMs and general intelligence lies in the programmer's ability to adapt and invent, while LLMs are limited to their stored programs. The focus should be on improving the quality of training data and better curation rather than increasing the scale.

    • LLMs and human intelligenceLLMs mimic human-like intelligence through memorization but lack the ability to truly learn and create beyond their extrapolated knowledge. Human intelligence and symbolic search systems are needed for invention.

      Current large language models (LLMs) excel at mimicking human-like intelligence through memorization but lack the ability to truly learn and create beyond their extrapolated knowledge. This limitation is akin to a GitHub, which stores and retrieves patterns, but cannot truly understand or invent new concepts. LLMs struggle with novel problems that fall outside their memorized solutions, and while they can be used as memory extensions for scientists, they do not replace the need for human intelligence and creativity. The dream of having an AI program a smarter AI is not feasible with current LLMs, as they are limited to interpolating solutions from their trained data. Instead, the focus should be on combining LLMs with human intelligence and symbolic search systems to create hybrid systems capable of invention.

    • AI takeoverThe human brain's complexity makes it unlikely for AI to take over the world without new ideas and inventions, and creating dangerous AI requires more than just intelligence, including autonomy, goal setting, and a value system.

      Relying on brute force methods or algorithms to discover new technologies, such as a General Intelligence (GI), is not effective. The human brain is incredibly complex, and we don't fully understand how it works. Therefore, we cannot assign a large probability to the idea that AI will take over the world, as we need new ideas and inventions to create such technology. Intelligence is just a conversion ratio between the information we have and our ability to operate in novel situations. However, creating dangerous AI requires more than just intelligence; it requires autonomy, goal setting, and a value system. These elements should be approached with caution and deliberation. Furthermore, there is a lot of positive potential in developing deep learning models and large language models. These tools should be accessible to everyone, not just a fixed set of companies. Keras, a deep learning library, is designed to be accessible and approachable, allowing software developers to tackle their own problems using these technologies. It's essential to remember that intelligence is just a tool and that the potential risks associated with it depend on how it is engineered and used.

    • Large Language ModelsLarge Language Models can generate new content, answer questions, and automate tasks, but they should not be relied upon for decision-making or important content without human oversight. They are accessible and easy to use with free GPU services or personal computers.

      Large Language Models (LLMs) can be trained and used by individuals or businesses to generate new content, answer questions, or even automate specific tasks. For instance, using models like the JAMA 8 billion model, one can generate new podcast transcripts or train an LLM on all the works of a philosopher like Daniel Dennett to answer questions about his work. This not only can be fun and educational but also has practical applications. Businesses can use LLMs to automate tasks, such as turning spreadsheet data into emails, or even develop their own private repository of programs tailored to their specific needs. However, it's important to note that while LLMs can be powerful tools, they should not be relied upon for decision-making or generating important content, such as emails, without human oversight. Instead, they can be used as a shortcut towards the general area of interest and help speed up tasks or fix typos. The ease of getting started with LLMs is also notable, as one can use free GPU notebook services like Colab or use their own MacBook Pro with a GPU to begin training models. With the increasing accessibility and capabilities of LLMs, their impact on our lives is transformative and will continue to grow.

    Recent Episodes from Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

    AMA | September 2024

    AMA | September 2024

    Welcome to the September 2024 Ask Me Anything episode of Mindscape! These monthly excursions are funded by Patreon supporters (who are also the ones asking the questions). We take questions asked by Patreons, whittle them down to a more manageable number -- based primarily on whether I have anything interesting to say about them, not whether the questions themselves are good -- and sometimes group them together if they are about a similar topic. Enjoy!

    Blog post with AMA questions and transcript: https://www.preposterousuniverse.com/podcast/2024/09/02/ama-september-2024/

    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    287 | Jean-Paul Faguet on Institutions and the Legacy of History

    287 | Jean-Paul Faguet on Institutions and the Legacy of History

    One common feature of complex systems is sensitive dependence on initial conditions: a small change in how systems begin evolving can lead to large differences in their later behavior. In the social sphere, this is a way of saying that history matters. But it can be hard to quantify how much certain specific historical events have affected contemporary conditions, because the number of variables is so large and their impacts are so interdependent. Political economist Jean-Paul Faguet and collaborators have examined one case where we can closely measure the impact today of events from centuries ago: how Colombian communities are still affected by 16th-century encomienda, a colonial forced-labor institution. We talk about this and other examples of the legacy of history.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/08/26/287-jean-paul-faguet-on-institutions-and-the-legacy-of-history/

    Jean-Paul Faguet received a Ph.D. in Political Economy and an M.Sc. in Economics from the London School of Economics, and an Master of Public Policy from the Kennedy School of Government at Harvard. He is currently Professor of the Political Economy of Development at LSE. He serves as the Chair of the Decentralization Task Force for the Initiative for Policy Dialogue. Among his awards are the W.J.M. Mackenzie Prize for best political science book.


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    286 | Blaise Agüera y Arcas on the Emergence of Replication and Computation

    286 | Blaise Agüera y Arcas on the Emergence of Replication and Computation

    Understanding how life began on Earth involves questions of chemistry, geology, planetary science, physics, and more. But the question of how random processes lead to organized, self-replicating, information-bearing systems is a more general one. That question can be addressed in an idealized world of computer code, initialized with random sequences and left to run. Starting with many such random systems, and allowing them to mutate and interact, will we end up with "lifelike," self-replicating programs? A new paper by Blaise Agüera y Arcas and collaborators suggests that the answer is yes. This raises interesting questions about whether computation is an attractor in the space of relevant dynamical processes, with implications for the origin and ubiquity of life.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/08/19/286-blaise-aguera-y-arcas-on-the-emergence-of-replication-and-computation/

    Blaise Agüera y Arcas received a B.A. in physics from Princeton University. He is currently a vice-president of engineering at Google, leader of the Cerebra team, and a member of the Paradigms of Intelligence team. He is the author of the books Ubi Sunt and Who Are We Now?, and the upcoming What Is Intelligence?


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    285 | Nate Silver on Prediction, Risk, and Rationality

    285 | Nate Silver on Prediction, Risk, and Rationality

    Being rational necessarily involves engagement with probability. Given two possible courses of action, it can be rational to prefer the one that could possibly result in a worse outcome, if there's also a substantial probability for an even better outcome. But one's attitude toward risk -- averse, tolerant, or even seeking -- also matters. Do we work to avoid the worse possible outcome, even if there is potential for enormous reward? Nate Silver has long thought about probability and prediction, from sports to politics to professional poker. In his his new book On The Edge: The Art of Risking Everything, Silver examines a set of traits characterizing people who welcome risks.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/08/12/285-nate-silver-on-prediction-risk-and-rationality/

    Nate Silver received a B.A. in economics from the University of Chicago. He worked as a baseball analyst, developing the PECOTA statistical system (Player Empirical Comparison and Optimization Test Algorithm). He later founded the FiveThirtyEight political polling analysis site. His first book, The Signal and the Noise, was awarded the Phi Beta Kappa Society Book Award in Science. He is the co-host (with Maria Konnikova) of the Risky Business podcast.


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    AMA | August 2024

    AMA | August 2024

    Welcome to the August 2024 Ask Me Anything episode of Mindscape! These monthly excursions are funded by Patreon supporters (who are also the ones asking the questions). We take questions asked by Patreons, whittle them down to a more manageable number -- based primarily on whether I have anything interesting to say about them, not whether the questions themselves are good -- and sometimes group them together if they are about a similar topic. Enjoy!

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/08/05/ama-august-2024/

    Support Mindscape on Patreon: https://www.patreon.com/seanmcarroll

    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    284 | Doris Tsao on How the Brain Turns Vision Into the World

    284 | Doris Tsao on How the Brain Turns Vision Into the World

    The human brain does a pretty amazing job of taking in a huge amount of data from multiple sensory modalities -- vision, hearing, smell, etc. -- and constructing a coherent picture of the world, constantly being updated in real time. (Although perhaps in discrete moments, rather than continuously, as we learn in this podcast...) We're a long way from completely understanding how that works, but amazing progress has been made in identifying specific parts of the brain with specific functions in this process. Today we talk to leading neuroscientist Doris Tsao about the specific workings of vision, from how we recognize faces to how we construct a model of the world around us.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/07/29/284-doris-tsao-on-how-the-brain-turns-vision-into-the-world/

    Doris Tsao received her Ph.D. in neurobiology from Harvard University. She is currently a professor of molecular and cell biology, and a member of the Helen Wills Neuroscience Institute, at the University of California, Berkeley. Among her awards are a MacArthur Fellowship, membership in the National Academy of Sciences, the Eppendorf and Science International Prize in Neurobiology, the National Institutes of Health Director’s Pioneer Award, the Golden Brain Award from the Minerva Foundation, the Perl-UNC Neuroscience Prize, and the Kavli Prize in Neuroscience.

    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    283 | Daron Acemoglu on Technology, Inequality, and Power

    283 | Daron Acemoglu on Technology, Inequality, and Power

    Change is scary. But sometimes it can all work out for the best. There's no guarantee of that, however, even when the change in question involves the introduction of a powerful new technology. Today's guest, Daron Acemoglu, is a political economist who has long thought about the relationship between economics and political institutions. In his most recent book (with Simon Johnson), Power and Progress: Our Thousand-Year Struggle Over Technology and Prosperity, he looks at how technological innovations affect the economic lives of ordinary people. We talk about how such effects are often for the worse, at least to start out, until better institutions are able to eventually spread the benefits more broadly.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/07/22/283-daron-acemoglu-on-technology-inequality-and-power/

    Daron Acemoglu received a Ph.D. in economics from the London School of Economics. He is currently Institute Professor at the Massachusetts Institute of Technology. He is a fellow of the National Academy of Sciences, the American Academy of Arts and Sciences, and the Econometric Society. Among his awards are the John Bates Clark Medal and the Nemmers Prize in Economics. In 2015, he was named the most cited economist of the past 10 years.


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    282 | Joel David Hamkins on Puzzles of Reality and Infinity

    282 | Joel David Hamkins on Puzzles of Reality and Infinity

    The philosophy of mathematics would be so much easier if it weren't for infinity. The concept seems natural, but taking it seriously opens the door to counterintuitive results. As mathematician and philosopher Joel David Hamkins says in this conversation, when we say that the natural numbers are "0, 1, 2, 3, and so on," that "and so on" is hopelessly vague. We talk about different ways to think about the puzzles of infinity, how they might be resolved, and implications for mathematical realism.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/07/15/282-joel-david-hamkins-on-puzzles-of-reality-and-infinity/

    Support Mindscape on Patreon.

    Joel David Hamkins received his Ph.D. in mathematics from the University of California, Berkeley. He is currently the John Cardinal O'Hara Professor of Logic at the University of Notre Dame. He is a pioneer of the idea of the set theory multiverse. He is the top-rated user by reputation score on MathOverflow. He is currently working on The Book of Infinity, to be published by MIT Press.


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    Ask Me Anything | July 2024

    Ask Me Anything | July 2024

    Welcome to the July 2024 Ask Me Anything episode of Mindscape! These monthly excursions are funded by Patreon supporters (who are also the ones asking the questions). We take questions asked by Patreons, whittle them down to a more manageable number -- based primarily on whether I have anything interesting to say about them, not whether the questions themselves are good -- and sometimes group them together if they are about a similar topic. Enjoy!

    Blog post with questions and transcript: https://www.preposterousuniverse.com/podcast/2024/07/08/ama-july-2024/

    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    281 | Samir Okasha on the Philosophy of Agency and Evolution

    281 | Samir Okasha on the Philosophy of Agency and Evolution

    Just like with physics, in biology it is perfectly possible to do most respectable work without thinking much about philosophy, but there are unmistakably foundational questions where philosophy becomes crucial. When do we say that a collection of matter (or bits) is alive? When does it become an agent, capable of making decisions? What are the origins of morality and altruistic behavior? We talk with one of the world's leading experts, Samir Okasha, about the biggest issues in modern philosophy of biology.

    Support Mindscape on Patreon.

    Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/07/01/281-samir-okasha-on-the-philosophy-of-agency-and-evolution/

    Samir Okasha received his D.Phil. in Philosophy from the University of Oxford. He is currently Professor of the Philosophy of Science at the University of Bristol. He is a winner of the Lakatos Award for his book Evolution and the Levels of Selection, and is a Fellow of the British Academy.


    See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

    Related Podcasts

    AI Unpacked

    AI Unpacked

    AI Unpacked is a biweekly podcast exploring AI's impact on our lives and society. Join us as we take a deep dive into the latest AI trends and topics, unpacking the complex issues and considering the global implications of this evolutionary technology. We talk to the experts, academics, and innovators who are making an impact in the field, and discuss how AI is shaping our future and what it means for us as individuals and as a society.


    By: James Cogan

    Total Episodes: 9

    Topics:society & culturetechnologyscience

    Agora Entendi!

    Agora Entendi!
    Todos somos um pouco cientistas quando crianças. É a curiosidade que nos faz aprender a andar, conhecer sabores e resolver problemas. Mas só alguns, quando adultos, mantém esse espírito vivo e continuam a se questionar, permanecem movidos a aprender. Se você mantém esse espírito ou mesmo se é um cético em relação a alguns fatos científicos esse podcast vai te levar ao encontro de cientistas que explicarão sobre assuntos atuais e que te ajudarão a tirar muitas de suas dúvidas em relação às principais conclusões da ciência. Mas não nos interessa somente que você entenda. Queremos receber seus questionamentos, seus argumentos, suas opiniões. Vamos debater ciência! Só estaremos satisfeitos quando juntos dissermos: Agora entendi! - Duração média dos episódios: 15 a 20 minutos - Criação, roteiro e entrevistas: Klena Sarges (Pesquisadora em Saúde Pública / FIOCRUZ) - Apresentação/locução: Henrique Sarges (estudante de Artes Cênicas / Casa das Artes de Laranjeiras/RJ) - Edição: Alcyr de Morisson (arquiteto/doutorando em Habitação de Interesse Social/FAU/UFRJ) Trilha sonora das vinhetas: Club Seamus (Kevin MacLeod) (Free License)

    By: podcastagoraentendi

    Total Episodes: 8

    Topics:technologyscience