Logo

    Google DeepMind's Vision for AI, Search and Gemini with Oriol Vinyals from Google DeepMind

    enAugust 01, 2024
    What is the goal of the Gemini project?
    How are chat-based and search-based models expected to interact?
    What challenges do current large language models face?
    What is the significance of long context windows in AI?
    What potential applications are anticipated for Gemini technology?

    Podcast Summary

    • Google's AI research directionGoogle's AI research is focusing on creating a powerful core model for language technology through the Gemini project, aiming to enhance both chat-based and search-based models for better user experience.

      Google and DeepMind's research efforts have undergone significant changes in the past year, leading to the formation of the Gemini project and the merging of various AI research organizations under Google DeepMind. The goal of Gemini is to create a powerful core model to power the technology used in large language models worldwide. The team interacts with the rest of the company by focusing on building state-of-the-art technology, addressing the needs of different product areas, and integrating AI into products. While chat-based models and traditional search-based models have different use cases, both are expected to enhance each other and play important roles in the future. The integration of AI into search products is expected to significantly improve the user experience.

    • Long context windows in AIThe integration of long context windows into AI models is a significant step forward in natural language processing, enabling the processing and answering of questions from long videos or texts, and is expected to be widely adopted within the next one to two years due to research developments and hardware capabilities.

      We are on the brink of a new era in AI and language models, where long context windows will become the norm. Google, among other companies, is exploring this technology through projects like Gemini, which allows for infinite context length. This capability, which seems trivial now, has already shown surprising results, such as the ability to process and answer questions from long videos or texts. The use cases for this technology are still emerging, but it's expected that it will be widely adopted in both enterprise and consumer applications within the next one to two years. This will be driven by both the research developments and the availability of hardware capable of handling the increased memory requirements. The potential applications are vast, from allowing companies to access all of a consumer's context to enabling enterprise users to upload and query large datasets. Despite some limitations, the motivation to find compelling use cases for this technology is strong, and the technological challenges will be addressed in due time. Overall, the integration of long context windows into AI models marks a significant step forward in the field of natural language processing.

    • Reasoning in large language modelsThe next frontier for large language models is to perfect reasoning capabilities, making them crisp and accurate, and expanding multimodal capabilities to bring us closer to AGI.

      As we move towards an era of infinite context, the relevance of retrieval architectures and hierarchical memory systems remains significant, especially from an efficiency perspective. The ability to effectively contextualize and reason about complex information will continue to be important, even as we combine retrieval-based methods with neural-based ones. The field of large language models (LLMs) has seen tremendous growth in recent years, with an influx of new researchers and advances in data, compute, and algorithms. However, current state-of-the-art LLMs still have limitations, such as the inability to perfectly reason crisply and accurately. The next frontier is to push the boundaries of these models and perfect the reasoning step, making them more crisp and accurate, and expanding their multimodal capabilities. This will bring us closer to achieving artificial general intelligence (AGI).

    • AI development balanceThe future of AI development involves balancing computational resources for pre-training, reinforcement learning, and inference. While pre-training is important, inference time requires attention for making AI systems as good as humans. Accurately assigning rewards remains a challenge, and new methods are needed as AI surpasses human performance.

      The future of AI development lies in balancing the computational resources dedicated to training and inference. While large-scale pre-training is crucial for achieving high performance, the use of reinforcement learning and search algorithms at inference time are essential for making AI systems as good as humans. However, assigning accurate rewards to these systems remains a significant challenge. The speaker suggests that the current trend of skewing computational resources towards pre-training may shift in the future, with a more even distribution between pre-training, reinforcement learning, and inference. The exact percentage of compute dedicated to each stage is not clear but is expected to be less than 90% for inference time and more for pre-training. The speaker also emphasizes the importance of research in scaling reward functions beyond traditional games and applications. While supervised learning can be used to scale rewards, human annotation and labeling have been crucial in advancing deep learning. However, as AI systems surpass human performance, new methods for assigning rewards will be necessary. The speaker also touches on the potential of making AI systems more explainable and logical, which could lead to faster inference times and fewer errors. The ultimate goal is to create AI systems that can learn new skills and adapt to new situations as effectively as humans, while minimizing errors and maximizing efficiency.

    • Self-assessment in language modelsLanguage models may be able to evaluate their own outputs more accurately, leading to a reinforcement learning loop for improvement. This involves using the model itself as a reward, but specific task annotations remain a challenge. The future involves enabling self-assessment and progressing towards increasingly general models, while addressing unique challenges in various domains.

      There's an exciting potential for language models to evaluate their own outputs more accurately than generating them, leading to a reinforcement learning loop where the model gets better based on its own assessments. This idea involves using the model itself as a reward, which can be further refined through the development of generative reward models. However, the need for specific task annotations remains a question, and the hope is that in the limit, the user may provide the system with as many labels as needed. This concept can be likened to the Nyquist-Shannon sampling theorem, which suggests that to reconstruct a wave accurately, you need to sample it at a certain rate. Similarly, a model needs to be smart enough to evaluate its own intelligence. The future of this field involves enabling these capabilities and continuing to progress from general algorithms to increasingly general models, while also addressing unique challenges in various domains through specialization. Current models may not yet be able to fully solve complex problems like protein folding or nuclear fusion, but focusing on these areas could lead to significant advancements.

    • AI development beyond AGIThe future of AI research may shift from solely pursuing Artificial General Intelligence to focusing on improving models' ability to distinguish truth from falsehood and recognizing their limitations.

      The pursuit of Artificial General Intelligence (AGI) may not be the sole focus in the future of AI research. The speaker suggests that while we may still see a hybrid model of generalist and specialized AI, the definition and achievement of AGI might not be as clear-cut as some predict. Instead, the emphasis could shift towards understanding and improving the models' abilities to distinguish truth from falsehood, rather than solely focusing on achieving a singular, definitive AGI. The speaker also expresses the importance of recognizing the limitations and potential pitfalls of current AI models and continuing to address their egregious errors. Furthermore, the speaker believes that the use of AI in various fields, such as research and science, is an exciting prospect regardless of whether or not AGI is achieved in the traditional sense. The speaker's perspective encourages a more nuanced approach to AI development, emphasizing the importance of continuous improvement and a distribution of capabilities rather than a singular, definitive goal.

    • AI and EducationDiscover passions, embrace tools, language important in AI, unexplored areas in climate modeling, LLM a promising field for next decade, balance passions and opportunities

      As technology continues to evolve, particularly in the realm of AI, it presents both challenges and opportunities, especially for parents raising children who will soon enter the workforce. While there's no clear answer on what specific field of study will be most valuable in the future, it's important for children to discover their passions and embrace the tools and technologies available to them. The speaker emphasized the importance of language in understanding and utilizing AI, making it a potentially fruitful area of study for those not as technologically inclined. Additionally, there are still many unexplored areas in AI, such as climate modeling, where specialized knowledge and innovation could lead to significant advancements. For those interested in technology and research, the field of LLM (Large Language Models) is expected to be a worthwhile area of investigation for at least the next decade. Overall, the key takeaway is to find a balance between following one's passions and staying open to the opportunities presented by technology.

    Recent Episodes from No Priors: Artificial Intelligence | Machine Learning | Technology | Startups

    Future of LLM Markets, Consolidation, and Small Models with Sarah and Elad

    Future of LLM Markets, Consolidation, and Small Models with Sarah and Elad
    In this episode of No Priors, Sarah and Elad go deep into what's on everyone’s mind. They break down new partnerships and consolidation in the LLM market, specialization of AI models, and AMD’s strategic moves. Plus, Elad is looking for a humanoid robot.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil  Show Notes: (0:00) Introduction (0:24) LLM market consolidation  (2:18) Competition and decreasing API costs (3:58) Innovation in LLM productization  (8:20) Comparing  the LLM and social network market (11:40) Increasing competition in image generation (13:21) Trend in smaller models with higher performance (14:43) Areas of innovation (17:33) Legacy of AirBnB and Uber pushing boundaries (24:19) AMD Acquires ZT  (25:49) Elad’s looking for a Robot

    The Road to Autonomous Intelligence with Andrej Karpathy

    The Road to Autonomous Intelligence with Andrej Karpathy
    Andrej Karpathy joins Sarah and Elad in this week of No Priors. Andrej, who was a founding team member of OpenAI and former Senior Director of AI at Tesla, needs no introduction. In this episode, Andrej discusses the evolution of self-driving cars, comparing Tesla and Waymo’s approaches, and the technical challenges ahead. They also cover Tesla’s Optimus humanoid robot, the bottlenecks of AI development today, and  how AI capabilities could be further integrated with human cognition.  Andrej shares more about his new company Eureka Labs and his insights into AI-driven education, peer networks, and what young people should study to prepare for the reality ahead. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @Karpathy Show Notes:  (0:00) Introduction (0:33) Evolution of self-driving cars (2:23) The Tesla  vs. Waymo approach to self-driving  (6:32) Training Optimus  with automotive models (10:26) Reasoning behind the humanoid form factor (13:22) Existing challenges in robotics (16:12) Bottlenecks of AI progress  (20:27) Parallels between human cognition and AI models (22:12) Merging human cognition with AI capabilities (27:10) Building high performance small models (30:33) Andrej’s current work in AI-enabled education (36:17) How AI-driven education reshapes knowledge networks and status (41:26) Eureka Labs (42:25) What young people study to prepare for the future

    Building toward a bright post-AGI future with Eric Steinberger from Magic.dev

    Building toward a bright post-AGI  future with Eric Steinberger from Magic.dev
    Today on No Priors, Sarah Guo and Elad Gil are joined by Eric Steinberger, the co-founder and CEO of Magic.dev. His team is developing a software engineer co-pilot that will act more like a colleague than a tool. They discussed what makes Magic stand out from the crowd of AI co-pilots, the evaluation bar for a truly great AI assistant, and their predictions on what a post-AGI world could look like if the transition is managed with care.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @EricSteinb Show Notes:  (0:00) Introduction (0:45) Eric’s journey to founding Magic.dev (4:01) Long context windows for more accurate outcomes (10:53) Building a path toward AGI (15:18) Defining what is enough compute for AGI (17:34) Achieving Magic’s final UX (20:03) What makes a good AI assistant (22:09) Hiring at Magic (27:10) Impact of AGI (32:44) Eric’s north star for Magic (36:09) How Magic will interact in other tools

    Cloud Strategy in the AI Era with Matt Garman, CEO of AWS

    Cloud Strategy in the AI Era with Matt Garman, CEO of AWS
    In this episode of No Priors, hosts Sarah and Elad are joined by Matt Garman, the CEO of Amazon Web Services. They talk about the evolution of Amazon Web Services (AWS) from its inception to its current position as a major player in cloud computing and AI infrastructure. In this episode they touch on AI commuting hardware,  partnerships with AI startups, and the challenges of scaling for AI workloads. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil  Show Notes:  (00:00) Introduction  (00:23) Matt’s early days at Amazon (02:53) Early conception of AWS (06:36) Understanding the full opportunity of cloud compute (12:21) Blockers to cloud migration (14:19) AWS reaction to Gen AI (18:04) First-party models at hyperscalers (20:18) AWS point of view on open source (22:46) Grounding and knowledge bases (26:07) Semiconductors and data center capacity for AI workloads (31:15) Infrastructure investment for AI startups (33:18) Value creation in the AI ecosystem (36:22) Enterprise adoption  (38:48) Near-future predictions for AWS usage (41:25) AWS’s role for startups

    The marketplace for AI compute with Jared Quincy Davis from Foundry

    The marketplace for AI compute with Jared Quincy Davis from Foundry
    In this episode of No Priors, hosts Sarah and Elad are joined by Jared Quincy Davis, former DeepMind researcher and the Founder and CEO of Foundry, a new AI cloud computing service provider. They discuss the research problems that led him to starting Foundry, the current state of GPU cloud utilization, and Foundry's approach to improving cloud economics for AI workloads. Jared also touches on his predictions for the GPU market and the thinking behind his recent paper on designing compound AI systems. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @jaredq_ Show Notes:  (00:00) Introduction  (02:42) Foundry background (03:57) GPU utilization for large models (07:29) Systems to run a large model (09:54) Historical value proposition of the cloud (14:45) Sharing cloud compute to increase efficiency  (19:17) Foundry’s new releases (23:54) The current state of GPU capacity (29:50) GPU market dynamics (36:28) Compound systems design (40:27) Improving open-ended tasks

    How AI can help build smarter systems for every team with Eric Glyman and Karim Atiyeh of Ramp

    How AI can help build  smarter systems for every team  with Eric Glyman and Karim Atiyeh of Ramp
    In this episode of No Priors, hosts Sarah and Elad are joined by Ramp co-founders Eric Glyman and Karim Atiyeh of Ramp. The pair has been working to build one of the fastest growing fintechs since they were teenagers. This conversation focuses on how Ramp engineers have been building new systems to help every team from sales and marketing to product. They’re building best-in-class SaaS solutions just for internal use to make sure their company remains competitive. They also get into how AI will augment marketing and creative fields, the challenges of selling productivity, and how they’re using LLMs to create internal podcasts using sales calls to share what customers are saying with the whole team.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @eglyman l @karimatiyeh Show Notes:  (0:00) Introduction to Ramp (3:17) Working with startups (8:13) Ramp’s implementation of AI (14:10) Resourcing and staffing (17:20) Deciding when to build vs buy (21:20) Selling productivity (25:01) Risk mitigation when using AI (28:48) What the AI stack is missing (30:50) Marketing with AI (37:26) Designing a modern marketing team (40:00) Giving creative freedom to marketing teams (42:12) Augmenting bookkeeping (47:00) AI-generated podcasts

    Innovating Spend Management through AI with Pedro Franceschi from Brex

    Innovating Spend Management through AI with Pedro Franceschi from Brex
    Hunting down receipts and manually filling out invoices kills productivity. This week on No Priors, Sarah Guo and Elad Gil sit down with Pedro Franceschi, co-founder and CEO of Brex. Pedro discusses how Brex is harnessing AI to optimize spend management and automate tedious accounting and compliance tasks for teams. The conversation covers the reliability challenges in AI today, Pedro’s insights on the future of fintech in an AI-driven world, and the major transitions Brex has navigated in recent years. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @Pedroh96 Show Notes:  (0:00) Introduction (0:32) Brex’s business and transitioning to solo CEO (3:04) Building AI into Brex  (7:09) Solving for risk and reliability in AI-enabled financial products (11:41) Allocating resources toward AI investment (14:00) Innovating data use in marketing  (20:00) Building durable businesses in the face of AI (25:36) AI’s impact on finance (29:15) Brex’s decision to focus on startups and enterprises

    Google DeepMind's Vision for AI, Search and Gemini with Oriol Vinyals from Google DeepMind

    Google DeepMind's Vision for AI, Search and Gemini with Oriol Vinyals from Google DeepMind
    In this episode of No Priors, hosts Sarah and Elad are joined by Oriol Vinyals, VP of Research, Deep Learning Team Lead, at Google DeepMind and Technical Co-lead of the Gemini project. Oriol shares insights from his career in machine learning, including leading the AlphaStar team and building competitive StarCraft agents. We talk about Google DeepMind, forming the Gemini project, and integrating AI technology throughout Google products. Oriol also discusses the advancements and challenges in long context LLMs, reasoning capabilities of models, and the future direction of AI research and applications. The episode concludes with a reflection on AGI timelines, the importance of specialized research, and advice for future generations in navigating the evolving landscape of AI. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @oriolvinyalsml Show Notes:  (00:00) Introduction to Oriol Vinyals (00:55) The Gemini Project and Its Impact (02:04) AI in Google Search and Chat Models (08:29) Infinite Context Length and Its Applications (14:42) Scaling AI and Reward Functions (31:55) The Future of General Models and Specialization (38:14) Reflections on AGI and Personal Insights (43:09) Will the Next Generation Study Computer Science? (45:37) Closing thoughts

    Low-Code in the Age of AI and Going Enterprise, with Howie Liu from Airtable

    Low-Code in the Age of AI and Going Enterprise, with Howie Liu from Airtable
    This week on No Priors, Sarah Guo and Elad Gil are joined by Howie Liu, the co-founder and CEO of Airtable. Howie discusses their Cobuilder launch, the evolution of Airtable from a simple productivity tool to an enterprise app platform with integrated AI capabilities. They talk about why the conventional wisdom of “app not platform” can be wrong,  why there’s a future for low-code in the age of AI and code generation, and where enterprises need help adopting AI. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @Howietl Show Notes:  (00:00) Introduction (00:29) The Origin and Evolution of Airtable (02:31) Challenges and Successes in Building Airtable (06:09) Airtable's Transition to Enterprise Solutions (09:44) Insights on Product Management (16:23) Integrating AI into Airtable (21:55) The Future of No Code and AI (30:30) Workshops and Training for AI Adoption (36:28) The Role of Code Generation in No Code Platforms

    How AI is opening up new markets and impacting the startup status quo with Sarah Guo and Elad Gil

    How AI is opening up new markets and impacting the startup status quo with Sarah Guo and Elad Gil
    This week on No Priors, we have a host-only episode. Sarah and Elad catch up to discuss how tech history may be repeating itself. Much like in the early days of the internet, every company is clamoring to incorporate AI into their products or operations while some legacy players are skeptical that investment in AI will pay off. They also get into new opportunities and capabilities that AI is opening up, whether or not incubators are actually effective, and what companies are poised to stand the test of time in the changing tech landscape. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil Show Notes:  (0:00) Introduction (0:16) Old school operators AI misunderstandings (5:10) Tech history is repeating itself with slow AI adoption (6:09) New AI Markets (8:48) AI-backed buyouts (13:03) AI incubation (17:18) Exciting incubating applications (18:26) AI and the public markets (22:20) Staffing AI companies  (25:14) Competition and shrinking head count