Logo

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    enJuly 07, 2024
    What are the features of Google's Gemini 1.5 models?
    How does Google's context caching feature improve efficiency?
    What challenge does the reward model face in reinforcement learning?
    What innovation is Tencent AI Lab introducing with personas?
    How are publishers responding to AI summarization tools?

    Podcast Summary

    • Generative AI advancementsGoogle's Gemini 1.5 Flash and Pro offer larger context windows and context caching, allowing for more powerful and efficient handling of larger inputs. Meta is releasing a 400 billion parameter model, further advancing the field.

      There have been recent advancements in generative AI models, specifically from Google with the public release of Gemini 1.5 Flash and Pro. These models offer larger context windows, up to 2 million tokens, making them more powerful and capable of handling larger inputs. Google's context caching feature is also a notable addition, allowing models to store and reuse information, resulting in cost savings and improved efficiency. The industry is shifting towards making these AI models more user-friendly and productized, with Google having an edge due to its enterprise offerings on its cloud platform. Additionally, Meta is about to release its biggest LLM yet, a 400 billion parameter model, which is expected to be a significant advancement in the field. These developments demonstrate the rapid progress being made in generative AI and its increasing importance in various industries.

    • AI model securityMeta faces challenges ensuring its 400 billion parameter model can't be easily misused or jailbroken, while companies like Runway and Google explore business models to create a competitive edge in the generative AI market

      The release of larger language models like Meta's 400 billion parameter model comes with significant challenges, particularly in ensuring the models can't be easily jailbroken or misused. Meta is reportedly considering releasing the model, but the company faces the daunting task of implementing robust safeguards that can withstand unknown jailbreaking techniques. Meanwhile, Runway has released a paid version of its Gen Free Alpha AI video model, which offers text-to-video generation and plans to add image-to-video and video-to-video modes in the future. Google is also integrating AI features into its Pixel 9 smartphone, and audio generation firm 11 Labs has announced a reader app with famous voices. The generative AI market is heating up, and companies are exploring various business models, including offering paid versions of their products, to create a competitive moat. The potential benefits of these tools are significant, but so are the risks, making the decisions around their release a complex issue.

    • Text-to-Speech Innovation, AI SearchCompanies collaborate with deceased actors for text-to-speech features, Perplexity upgrades Pro Search for better complex query understanding, and addressing latency and context window effectiveness are crucial for optimal user experiences, while clear marketing messaging is important to manage user expectations regarding AI capabilities

      Companies are exploring innovative ways to provide text-to-speech features without facing backlash, by collaborating with estates of deceased actors and actresses for recorded content. This approach, while not without controversy, allows for a more professional and engaging user experience. Perplexity, an AI-powered search engine, has upgraded its Pro Search feature, enabling it to better understand complex queries and provide richer, more detailed answers. This advancement, along with Perplexity's polished product and positioning, positions it as a compelling alternative to legacy search engines. However, as these advanced technologies continue to develop, addressing issues like inference latency and context window effectiveness will be crucial for delivering optimal user experiences. Additionally, it was revealed that Gemini's data analyzing abilities may not be as accurate as Google claims, highlighting the importance of clear marketing messaging and realistic expectations.

    • AI and copyright infringementThe use of AI in accessing and summarizing content from websites with paywalls raises copyright infringement concerns, with publishers considering blocking downloads of data to protect intellectual property. Meanwhile, companies like OpenAI pursue media partnerships to differentiate their AI tools, and China continues to be a significant player in the global AI competition.

      The use of AI in accessing and summarizing content from websites, particularly those with paywalls, is becoming a contentious issue. The case of PO, a summarization bot, raises questions about copyright infringement and the effectiveness of the robots exclusion protocol. Publishers are now considering blocking downloads of data to protect their intellectual property, while companies like OpenAI are pursuing media partnerships to differentiate their AI tools. In the hardware realm, Huawei and Wuhan Chinchin are reportedly collaborating to develop high bandwidth memory chips in the face of US restrictions, highlighting the importance of China in the global AI competition. Additionally, Alibaba's large language model has entered the top ranks on the developer platform Hugging Face, indicating growing competition in the AI model space between the US and China. These developments underscore the complex and evolving landscape of AI technology and its implications for copyright law, data ownership, and international relations.

    • AI competition and collaborationChinese companies focus on open source models for AI research due to limited resources, while Meta pushes wearable AI technology boundaries with Raybans and Apple collaborates with OpenAI

      The race for AI dominance continues, with Chinese companies focusing on open source models due to limited access to advanced GPUs and potential geopolitical leverage. Meanwhile, Meta is making strides in wearable AI with its Raybans, offering video recording and AI integration. A deeper partnership between Apple and OpenAI is also developing, with Phil Schiller reportedly joining OpenAI's board. These developments highlight the intense competition and collaboration in the AI sector. Chinese companies are exploring open source models to stay competitive in AI research, while Meta is pushing the boundaries of wearable AI technology. Apple and OpenAI's partnership underscores the importance of collaboration between tech giants in the rapidly evolving AI landscape. The Meta Raybans, with their video recording capabilities and AI integration, offer a potential wearable paradigm that people may actually want, unlike previous failures in this space. Overall, the AI industry is witnessing significant advancements and strategic partnerships, with potential implications for geopolitical dynamics and consumer technology.

    • Microsoft-OpenAI relationship, RegulationMicrosoft's exclusivity agreement with OpenAI for GPT technologies may not last, and regulatory bodies play a crucial role in shaping technology development, particularly regarding AGI and potential misuse, while third-party model evaluations ensure responsible use and promote AI safety and governance.

      The relationship between Microsoft and OpenAI, as well as the role of regulatory bodies in shaping technology development, continues to be a complex and evolving issue. The exclusivity agreement between Microsoft and OpenAI for GPT technologies may not be holding up, and the determination of when OpenAI achieves Artificial General Intelligence (AGI) is crucial, as it will impact Microsoft's access to the technology. Additionally, the evaluation of AI models and the development of third-party model evaluations are essential for ensuring responsible use and preventing potential misuse, such as identity theft and disinformation. Companies like Runway are raising significant funds to advance AI technology, particularly in the video domain, and new benchmarks are highlighting the gap between human and AI performance. Anthropic's push for third-party model evaluations is a response to the need for independent oversight and a part of the ongoing conversation about AI safety and governance.

    • AI safety regulations, misalignment riskAnthropic advocates for government-mandated audits and certifications for AI models to address various risks, including misalignment risk, and Mozilla introduces LAMMA files for easier deployment of models, while researchers explore reducing memory usage and increasing throughput in large language models.

      Anthropic, a leading AI safety research company, is pushing for government-mandated audits and certifications for AI models in both large and small companies. This initiative aims to address various risks, including cyber attacks, chemical, bio, radiological, and nuclear risks, autonomy, social manipulation, and misalignment risk. Anthropic's clear focus on misalignment risk as a separate category is noteworthy. Mozilla, a significant player in the open source space, has introduced LAMMA files, which package together the weights of an AI model with the software needed to run it, making it easier to deploy models on various platforms and devices. Researchers are also working on eliminating matrix multiplication in large language models (LLMs) by using ternary values and addition instead, which could lead to reduced memory usage and increased throughput. While these advancements show promise, it's important to note that the research is still in its early stages and more testing is needed to confirm the benefits at larger scales.

    • Simplifying complex architecturesRecent research challenges the assumption that complex architectures are always superior, showing that simple strategies like repeatedly calling a model or increasing its temperature can perform comparably for human evaluation tasks and often come with lower costs. Alternative approaches like weighted average rewarded policies in reinforcement learning are also proposed.

      Recent research challenges the assumption that complex architectures are always superior to simpler ones in the field of machine learning, specifically in the context of language models and agent architectures. The paper "A Simple Architecture for Agent Evaluation" demonstrates that simple baselines, such as repeatedly calling a model or increasing its temperature, can perform comparably to more complex agent architectures for human evaluation tasks. Furthermore, these simple strategies often come with lower costs. The researchers also criticize the current evaluation practices for agents, emphasizing the importance of considering both accuracy and cost. Another paper, "Warp on the Benefits of Weight Average Rewarded Policies," focuses on reinforcement learning and suggests an alternative to the common practice of using KL regularization to prevent the model from forgetting pre-trained knowledge during training. This paper proposes a weighted average rewarded policies strategy that allows for better optimization in the RL stage while retaining more information. These papers serve as reminders that it's crucial to question assumptions and explore alternative approaches in the ever-evolving field of machine learning.

    • Perverse optimizations in reinforcement learningTo prevent perverse optimizations in reinforcement learning from human feedback, multi-run fusion is used to train multiple copies of the language model independently against the reward model and callback labeler, then merge their weights to create a more aligned model. Personas are also used to generate diverse synthetic data and elicit unique outputs from the model.

      In reinforcement learning from human feedback, creating a reward model to train language models can lead to perverse optimizations where the model finds ways to manipulate the reward model rather than understanding human desires. To prevent this, researchers introduce the concept of an anchor or callback labeler divergence score to ensure the model stays close to its original behavior. This technique, called multi-run fusion, trains multiple copies of the language model independently against the reward model and the callback labeler, then merges their weights to create a more aligned model. This process is repeated to gradually improve the model's alignment. Another interesting paper discusses the challenge of generating synthetic data for AI models. The solution proposed is the use of personas, which are descriptions of different types of people. By tailoring prompts to these personas, the model provides unique outputs, eliciting a broader range of information from the model. The paper by Tencent AI Lab Seattle introduces a text-to-persona strategy to generate personas and a persona-to-persona strategy to derive additional personas based on interpersonal relationships. They have released over 200,000 personas and are open to releasing more, acknowledging the potential risks and concerns. Overall, these papers highlight the importance of understanding human feedback and generating diverse synthetic data to improve AI models.

    • AI regulationThe Supreme Court's decision to strike down Chevron deference may lead to more clear and detailed legislation from Congress regarding AI, but the technical nuances involved could present a significant challenge for implementation.

      The use of personas and the generation of synthetic data at scale can significantly improve the performance of large language models, as demonstrated by a recent study using a 7 billion parameter Chinese model named Quen2, 7B, which surpassed the performance of the leading anthropic model, Gemini Ultra, on a math benchmark. Additionally, calibrating positional attention bias can help longer context utilization in LLMs. However, the regulatory landscape for AI is shifting with the Supreme Court's decision to strike down Chevron deference, which means that courts will now have to interpret ambiguous laws related to AI regulation, potentially leading to a need for more clear and detailed legislation from Congress. This change could present a significant challenge for the implementation of AI legislation due to the technical nuances involved.

    • US-China tech competitionNew US law limiting Congress' ability to delegate regulatory authority to agencies could hinder US response to emerging technologies like AI, while US export control measures against Chinese tech companies lead to longer delays and fewer exports, potentially giving China an edge in the market

      The new US law limiting Congress' ability to delegate regulatory authority to agencies could significantly hinder the country's agility in responding to emerging technologies like AI. This comes as the US and China continue to engage in export control measures, with the US relying on manual processes to oversee restrictions on Chinese tech companies. The manual processes at the Bureau of Industry and Security (BIS) have struggled to keep up with the increasing number of Chinese entities on their list, leading to longer delays and, by default, fewer exports. This situation could give China an edge in the market, as US companies face more obstacles in selling their products there. Additionally, the H20 GPU chips produced in China, which are less powerful than their US counterparts, are still experiencing significant sales due to the lagging effect of US export controls. Overall, these developments highlight the complex and evolving nature of the US-China tech competition and the importance of adaptability in navigating it.

    • Semiconductor, FinanceThe semiconductor industry faces worker shortages due to growth, while finance adopts machine learning for investment decisions, highlighting the need to stay informed about technological advancements and their industry impacts

      Both the semiconductor industry and the financial sector are experiencing significant changes driven by technological advancements. In the semiconductor industry, the US government is investing in workforce development programs to address projected worker shortages due to the industry's growth. Meanwhile, in finance, a billion-dollar fund run by Bridgewater Associates will use machine learning for decision making, potentially disrupting traditional investment strategies. These developments underscore the importance of staying informed about technological advancements and their potential impacts on various industries.

    Recent Episodes from Last Week in AI

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    Our 181st episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov and Jeremie Harris

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode:

    - Google's AI advancements with Gemini 1.5 models and AI-generated avatars, along with Samsung's lithography progress.  - Microsoft's Inflection usage caps for Pi, new AI inference services by Cerebrus Systems competing with Nvidia.  - Biases in AI, prompt leak attacks, and transparency in models and distributed training optimizations, including the 'distro' optimizer.  - AI regulation discussions including California’s SB1047, China's AI safety stance, and new export restrictions impacting Nvidia’s AI chips.

    Timestamps + Links:

    Last Week in AI
    enSeptember 15, 2024

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    Our 180th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    • Ideogram AI's new features, Google's Imagine 3, Dream Machine 1.5, and Runway's Gen3 Alpha Turbo model advancements.
    • Perplexity's integration of Flux image generation models and code interpreter updates for enhanced search results. 
    • Exploration of the feasibility and investment needed for scaling advanced AI models like GPT-4 and Agent Q architecture enhancements.
    • Analysis of California's AI regulation bill SB1047 and legal issues related to synthetic media, copyright, and online personhood credentials.

    Timestamps + Links:

    Last Week in AI
    enSeptember 03, 2024

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    Our 179th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    - Grok 2's beta release features new image generation using Black Forest Labs' tech.

    - Google introduces Gemini Voice Chat Mode available to subscribers and integrates it into Pixel Buds Pro 2.

    - Huawei's Ascend 910C AI chip aims to rival NVIDIA's H100 amidst US export controls.

    - Overview of potential risks of unaligned AI models and skepticism around SingularityNet's AGI supercomputer claims.

    Timestamps + Links:

    Last Week in AI
    enAugust 20, 2024

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    Our 178th episode with a summary and discussion of last week's big AI news!

    NOTE: this is a re-upload with fixed audio, my bad on the last one! - Andrey

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode: - Notable personnel movements and product updates, such as Character.ai leaders joining Google and new AI features in Reddit and Audible. - OpenAI's dramatic changes with co-founder exits, extended leaves, and new lawsuits from Elon Musk. - Rapid advancements in humanoid robotics exemplified by new models from companies like Figure in partnership with OpenAI, achieving amateur-level human performance in tasks like table tennis. - Research advancements such as Google's compute-efficient inference models and self-compressing neural networks, showcasing significant reductions in compute requirements while maintaining performance.

    Timestamps + Links:

    Last Week in AI
    enAugust 16, 2024

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    Our 177th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode again coming out about a week late, next one will be coming out soon...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you'd like to listen to the interview with Andrey, check out https://www.superdatascience.com/podcast

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    In this episode, hosts Andrey Kurenkov and John Krohn dive into significant updates and discussions in the AI world, including Instagram's new AI features, Waymo's driverless cars rollout in San Francisco, and NVIDIA’s chip delays. They also review Meta's AI Studio, character.ai CEO Noam Shazir's return to Google, and Google's Gemini updates. Additional topics cover NVIDIA's hardware issues, advancements in humanoid robots, and new open-source AI tools like Open Devon. Policy discussions touch on the EU AI Act, the U.S. stance on open-source AI, and investigations into Google and Anthropic. The impact of misinformation via deepfakes, particularly one involving Elon Musk, is also highlighted, all emphasizing significant industry effects and regulatory implications.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 11, 2024

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    Our 176th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode coming out about a week late, things got in the way of editing it...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

     

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 03, 2024

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    Our 175th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, hosts Andrey Kurenkov and Jeremy Harris explore recent AI advancements including OpenAI's release of GPT 4.0 Mini and Mistral’s open-source models, covering their impacts on affordability and performance. They delve into enterprise tools for compliance, text-to-video models like Hyper 1.5, and YouTube Music enhancements. The conversation further addresses AI research topics such as the benefits of numerous small expert models, novel benchmarking techniques, and advanced AI reasoning. Policy issues including U.S. export controls on AI technology to China and internal controversies at OpenAI are also discussed, alongside Elon Musk's supercomputer ambitions and OpenAI’s Prover-Verify Games initiative.  

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 25, 2024

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    Our 174rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, we delve into the latest advancements and challenges in the AI industry, highlighting new features from Figma and Quora, regulatory pressures on OpenAI, and significant investments in AI infrastructure. Key topics include AMD's acquisition of Silo AI, Elon Musk's GPU cluster plans for XAI, unique AI model training methods, and the nuances of AI copying and memory constraints. We discuss developments in AI's visual perception, real-time knowledge updates, and the need for transparency and regulation in AI content labeling and licensing.

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 17, 2024

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    Our 173rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode of Last Week in AI, we explore the latest advancements and debates in the AI field, including Google's release of Gemini 1.5, Meta's upcoming LLaMA 3, and Runway's Gen 3 Alpha video model. We discuss emerging AI features, legal disputes over data usage, and China's competition in AI. The conversation spans innovative research developments, cost considerations of AI architectures, and policy changes like the U.S. Supreme Court striking down Chevron deference. We also cover U.S. export controls on AI chips to China, workforce development in the semiconductor industry, and Bridgewater's new AI-driven financial fund, evaluating the broader financial and regulatory impacts of AI technologies.  

    Timestamps + links:

    Last Week in AI
    enJuly 07, 2024

    #172 - Claude and Gemini updates, Gemma 2, GPT-4 Critic

    #172 - Claude and Gemini updates, Gemma 2, GPT-4 Critic

    Our 172nd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enJuly 01, 2024