Logo
    Search

    #169 - Google's Search Errors, OpenAI news & DRAMA, new leaderboards

    enJune 03, 2024

    Podcast Summary

    • AI misinformationGoogle's new AI search feature faced backlash due to inaccurate and misleading information it provided, highlighting the need for careful implementation and consideration when adding AI features.

      Last week saw significant developments in AI, but also raised concerns over potential misinformation and errors in AI-generated content. Google's new AI search feature, which generates summaries of sources at the top of search results, faced backlash due to inaccurate and misleading information it provided. This included incorrect information about Muslim presidents in the US and suggestions to eat rocks. Google's response was defensive, stating that the errors only affected a minority of queries and that they were manually removing bad results. However, this incident highlights the need for careful implementation and consideration when adding AI features, especially as the pressure to innovate and ship faster increases. Additionally, the conversation touched on the importance of addressing critics and skeptics in AI discussions, as well as the EU's decentralized approach to AI regulations compared to the US. Overall, the episode emphasized the importance of balancing the benefits and risks of AI technology.

    • AI chatbot integrationsRecent developments include Microsoft's Co-Pilot on Telegram, Google's Gemini on Oprah's browser, and Amazon's plan to overhaul Alexa, signifying a trend towards making AI chatbots ubiquitous but raising strategic considerations

      There have been several recent developments in the integration of AI chatbots into various platforms, with Microsoft's Co-Pilot bot being added to Telegram and Google's Gemini being integrated into Oprah's browser. These integrations signify a trend towards making AI chatbots ubiquitous, as they are being added to multiple platforms for easy access. However, these integrations also raise strategic considerations, as they could potentially create an extra layer between users and the AI providers, leading to data sharing and potential competition. Another notable development is Amazon's plan to give Alexa an AI overhaul and introduce a monthly subscription price for its enhanced features. These advancements demonstrate the growing importance of AI in everyday life and the ongoing competition among tech companies to offer the best AI experiences to users.

    • Amazon's push into generative AIAmazon is forming a new team to compete in generative AI, facing challenges in hardware and not bundling new AI service ChatGPT with Prime membership, while Microsoft is making strides in real-time translation and PwC becomes the first enterprise user of OpenAI's ChatGPT, expanding its reach in the business world.

      Amazon is making a significant push into the generative AI space, forming a new team to help them compete with industry leaders like Microsoft. While they already have a strong brand and distribution with Alexa, they are facing challenges in hardware and need to make up for lost time. Amazon's new AI service, ChatGPT, is not bundled with their Prime membership, indicating that inference may still be too expensive to include. Microsoft, on the other hand, is making strides in real-time translation, making language barriers less of an issue. This could be a game-changer for video content and communication. Another notable development is EO's new AI earbuds, which aim to succeed where other attempts have failed. These earbuds offer intrinsic value as high-quality headphones and have a reasonable price point, making them a safer bet than some previous failures in the market. PwC, a large consulting firm, has also entered the scene, becoming the first reseller and largest enterprise user of OpenAI's ChatGPT. This deal will give 100,000 PwC employees access to the enterprise version of ChatGPT, further expanding its reach in the business world. Overall, these developments highlight the growing importance of generative AI and its potential applications in various industries.

    • OpenAI, PwC dealOpenAI's deal with PwC marks its first foray into a resale model, granting 100,000 PwC employees access to advanced AI technology, generating substantial revenue and boosting PwC's reputation as an AI-driven consulting firm, potentially shifting the balance of power in the media landscape

      OpenAI, a leading AI research lab, has recently signed a significant deal with PwC, one of the world's largest professional services network, granting 100,000 PwC employees access to OpenAI's advanced AI technology. This marks OpenAI's first foray into a resale model and a strategic move towards greater independence from tech giants like Microsoft, with whom they have a complex relationship. The deal is expected to bring substantial revenue for OpenAI and boost PwC's reputation as a forward-thinking, AI-driven consulting firm. This trend of corporate media outlets partnering with AI companies for data licensing and enhanced content generation is likely to continue, potentially shifting the balance of power in the media landscape. As more revenues flow through these AI platforms, the implications for unbiased coverage and the role of social media platforms in shaping media narratives become crucial questions to consider.

    • AI industry partnerships with media outletsTechnology companies are partnering with media outlets for access to updated information and training data, offering incentives such as discounted rates for schools and nonprofits, and China's advancements in nanometer-class process technology could position it as a long-term competitor in the AI industry.

      Technology companies, including those in the AI sector, are seeking access to updated information and training data, leading to potential partnerships and incentives with media outlets. This trend is important from an economic and business perspective as AI continues to advance and become a commodity, with multiple options available for users. OpenAI, for instance, is offering discounted rates for schools and nonprofits to use its chatbot technology. Meanwhile, China is making strides in nanometer-class process technology, despite US sanctions, through multi-patterning and quadruple patterning lithography methods. These developments could potentially position China to compete long-term in the AI industry. However, it's important to note that the relationship between technology companies and media outlets, as well as the timing of certain announcements, may raise questions and concerns.

    • Semiconductor and AI competitionHuawei and SMIC are pushing for smaller node sizes with multi-patterning, while NVIDIA dominates with record revenue and effective use of market power. XAI joins high-valued AI companies, but long-term sustainability is uncertain. Scale AI introduces reliable private leaderboards for AI model evaluation.

      Huawei and SMIC are making significant strides in semiconductor technology with their multi-patterning technique, aiming to reach the three nanometer node size. NVIDIA, on the other hand, continues to dominate the tech industry with record-breaking revenue and profits, accelerating shipping velocity, and effective use of market dominance to secure fabrication capacity. Elon Musk's XAI has raised $6 billion in funding, joining the ranks of high-valued AI companies, but the long-term sustainability of their position remains uncertain. Scale AI's new private leaderboards offer a more reliable evaluation of AI model performance, as they are harder to game and involve human evaluators. These developments demonstrate the ongoing competition and innovation in the AI and semiconductor industries.

    • AI language models evaluationSignificant strides have been made in evaluating AI language models using elo-scale rankings and human evaluations. Multilingual models like Cohere's AIA 23 are expanding the state of the art to nearly half of the world's population, while efforts continue to make advanced models open source.

      There have been significant strides made in evaluating and comparing AI language models, with models like Cloud 3 Opus and GPT-4 leading in different areas. The use of elo-scale rankings and human evaluations has proven to be the most robust method for evaluating these models. Recently, Cohere for AI launched AIA 23, a multilingual model with 8 and 35 billion parameter versions, expanding the state of the art language model to nearly half of the world's population. The focus on multilingual models is a key differentiator for Cohere, as smaller models don't always perform as well on foreign languages. Additionally, there are ongoing efforts to make advanced models like AlphaFold open source, but there are concerns about the specific use cases and capabilities that will be made available. Overall, the field of AI language models is becoming increasingly crowded, with advances being made in specific areas like low resource languages.

    • AI advancementsMistral's Stral model generates code, completes functions, and answers questions about codebase in English, while the new optimization method, 'The Road Less Scheduled,' eliminates the need for learning rate schedules, making training more efficient.

      The recent advancements in AI, specifically Mistral's release of its first generative AI model for code and the introduction of a new optimization method, are making significant strides in the field. The accessibility and potential impact of these developments are impressive, with potential implications for various industries, including drug discovery and technology. Mistral's model, Stral, can generate code, complete functions, write and test code, and answer questions about a codebase in English. Although it's open-source, its large size (22 billion parameters) and long context window (32,000 tokens) limit its accessibility to larger companies with substantial compute infrastructure. The new optimization method, "The Road Less Scheduled," eliminates the need for learning rate schedules, making the training process more efficient and less reliant on trial-and-error hyperparameter tuning. This approach has shown strong performance on various benchmarks, generating excitement in the engineering and machine learning communities. While these advancements are promising, there are concerns about the fairness of comparisons made by Mistral and the potential limitations of their models. As the field continues to evolve, it's crucial to evaluate these developments critically and consider their implications for the future of AI research and application.

    • Learning rate adjustments and compute scalingHistorically, consistent learning rates during training have not been optimal. Strategies like adjusting learning rates during different stages of training have shown better results. Compute scaling for advanced AI models has been increasing significantly over the years, with an estimated growth of around 4-5x per year.

      The learning rate in machine learning models plays a crucial role in determining how quickly and effectively the model adapts to new data. A larger learning rate means bigger changes to the model, while a smaller learning rate implies more cautious adjustments. Historically, a consistent learning rate throughout training has not been optimal, and strategies like adjusting learning rates during different stages of training have shown better results. Additionally, the trend of compute scaling for advanced AI models has grown significantly over the years, with an estimated increase of around 4-5x per year. This trend, which started around 2012 with models like AlexNet, has been driven by companies like Google, OpenAI, and Meta, with Meta showing a slightly more aggressive scaling rate. However, it's important to note that the optimal learning rate and compute scaling trends are subject to change as new research and technologies emerge.

    • Model size vs data sizeMore complex, less compressible data requires more data to train a model effectively, challenging the belief that model size and data size should scale one-to-one.

      The quality and density of data play a significant role in determining the optimal ratio of model size versus data size when scaling up machine learning models. This was highlighted in a research paper that found more complex, less compressible data requires more data to train a model effectively, as opposed to a larger model with more parameters. This discovery challenges the long-held belief that model size and data size should scale one-to-one. The findings have implications for theories of intelligence and learning, suggesting that the more data and compute invested in a model, the more performant it becomes. Additionally, in the context of robotics, scaling laws have been identified, revealing that more data and compute lead to better results, but deployment constraints may necessitate data-optimal training or other approaches. Another intriguing topic in language models is contextual positioning, which allows models to attend to specific tokens based on their context, enabling better performance on tasks like selective copying, counting, and flipping. This approach improves perplexity on language model encoding tasks. Overall, these findings underscore the importance of understanding the relationship between model size, data size, and data quality in machine learning and robotics research.

    • AI limitations, usage gapDespite advancements in AI research, limitations persist and usage gap between hype and reality remains. Meta's CoPE addresses positional encoding challenges, but awareness and usage of advanced AI models like GPT-3 is low.

      While Meta's AI research is making significant strides, there are limitations to transformer models that make advanced reasoning challenging. Meta's Contextual Positional Encoding (CoPE) addresses this issue by providing more contextual positional information, improving performance on tasks requiring precise counting and understanding of context. Despite the hype surrounding new AI products, a recent survey revealed that only a small percentage of people in various countries use these tools daily. This includes popular models like GPT-3, with only 58% awareness in the US and 53% in the UK. This gap between hype and usage is not surprising, as it has been observed in previous technological trends. However, the true impact of AI may not be measured by usage numbers alone, but by the value it creates, such as Google's AI-driven search summaries. In the world of AI governance, the ongoing drama at OpenAI took a new turn with Helen Toner, a board member, revealing more details about Sam Altman's ousting. The board allegedly learned of CHA-GPT on Twitter, and there were concerns over a lack of transparency and misleading information regarding safety processes. Additionally, accusations of a toxic culture and psychological abuse have surfaced, with Altman not being the only person accused. These revelations add to the ongoing narrative of the OpenAI leadership change.

    • OpenAI leadership concernsConcerns about OpenAI's leadership, specifically Sam Altman, include inaccurate information given to the board, silencing of whistleblowers, and a culture of secrecy contradicting OpenAI's public messaging, raising questions about governance and potential impact on humanity.

      There have been concerns raised about OpenAI and its leadership, specifically Sam Altman, regarding transparency, accountability, and safety practices. These concerns include instances of inaccurate information given to the board, silencing of whistleblowers, and a culture of secrecy that contradicts OpenAI's public messaging. The revelations have raised questions about the governance and leadership of the organization, which is working on advanced artificial intelligence technology. The lack of transparency and accountability is particularly concerning given the potential impact of OpenAI's work on humanity. The resignation of a researcher over safety concerns and his move to a rival company further highlights these issues. OpenAI has responded with statements denying concerns regarding product safety or security, but the lack of clear communication and transparency leaves many questioning the validity of these statements. The ongoing debate highlights the importance of openness and accountability in organizations, especially those working on advanced technologies with significant implications for society.

    • OpenAI governanceOpenAI's use of non-disparagement clauses and lack of transparency in addressing criticisms has raised concerns about governance and potential impact on public trust and AI safety regulation.

      OpenAI, a leading AI research lab, has faced numerous criticisms regarding its governance and employment practices. These include the use of strict non-disparagement clauses in employment agreements, which could prevent former employees from speaking negatively about the company and potentially result in the clawback of their equity. The CEO, Sam Altman, has been criticized for being aware of these practices but not addressing them publicly until faced with significant scrutiny. The formation of a new safety and security committee, led by OpenAI insiders, has been met with skepticism. The training of a new advanced AI model, GPD-5, was also announced as a response to these criticisms. These events raise concerns about transparency and oversight within OpenAI, potentially impacting public trust and the broader debate on AI safety and regulation.

    • AI governance and accountabilityOngoing concerns about OpenAI's governance and accountability persist, with fines for deepfake creators and hacked AI models highlighting the need for ethical and regulatory oversight in the industry

      There are ongoing concerns about governance and accountability at OpenAI, despite their expertise and policy leaders. Sam Altman's influence and the composition of the committee raising objections have raised questions. In other news, a person who created deepfake Biden robocalls was fined $6 million by the FCC, setting a precedent. A hacker released a jailbroken version of ChatGPT, highlighting the challenges of aligning AI models. China announced a $47.5 billion chip fund, a significant investment in semiconductor industry, and Alphabet and Meta are partnering with Hollywood on AI, potentially democratizing video production. These developments underscore the importance of addressing ethical and regulatory issues in AI and technology.

    Recent Episodes from Last Week in AI

    #171 - - Apple Intelligence, Dream Machine, SSI Inc

    #171 - - Apple Intelligence, Dream Machine, SSI Inc

    Our 171st episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 24, 2024

    #170 - new Sora rival, OpenAI robotics, understanding GPT4, AGI by 2027?

    #170 - new Sora rival, OpenAI robotics, understanding GPT4, AGI by 2027?

    Our 170th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 09, 2024

    #169 - Google's Search Errors, OpenAI news & DRAMA, new leaderboards

    #169 - Google's Search Errors, OpenAI news & DRAMA, new leaderboards

    Our 168th episode with a summary and discussion of last week's big AI news!

    Feel free to leave us feedback here: https://forms.gle/ngXvXZpNJxaAprDv6

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 03, 2024

    #168 - OpenAI vs Scar Jo + safety researchers, MS AI updates, cool Anthropic research

    #168 - OpenAI vs Scar Jo + safety researchers, MS AI updates, cool Anthropic research

    Our 168th episode with a summary and discussion of last week's big AI news!

    With guest host Gavin Purcell from AI for Humans podcast!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enMay 28, 2024

    #167 - GPT-4o, Project Astra, Veo, OpenAI Departures, Interview with Andrey

    #167 - GPT-4o, Project Astra, Veo, OpenAI Departures, Interview with Andrey

    Our 167th episode with a summary and discussion of last week's big AI news!

    With guest host Daliana Liu (https://www.linkedin.com/in/dalianaliu/) from The Data Scientist Show!

    And a special one-time interview with Andrey in the latter part of the podcast.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 19, 2024

    #166 - new AI song generator, Microsoft's GPT4 efforts, AlphaFold3, xLSTM, OpenAI Model Spec

    #166 - new AI song generator, Microsoft's GPT4 efforts, AlphaFold3, xLSTM, OpenAI Model Spec

    Our 166th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 12, 2024

    #165 - Sora challenger, Astribot's S1, Med-Gemini, Refusal in LLMs

    #165 - Sora challenger, Astribot's S1, Med-Gemini, Refusal in LLMs

    Our 165th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 05, 2024

    #164 - Meta AI, Phi-3, OpenELM, Bollywood Deepfakes

    #164 - Meta AI, Phi-3, OpenELM, Bollywood Deepfakes

    Our 164th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 30, 2024

    #163 - Llama 3, Grok-1.5 Vision, new Atlas robot, RHO-1, Medium ban

    #163 - Llama 3, Grok-1.5 Vision, new Atlas robot, RHO-1, Medium ban

    Our 163rd episode with a summary and discussion of last week's big AI news!

    Note: apology for this one coming out a few days late, got delayed in editing it -Andrey

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 24, 2024

    #162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter

    #162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter

    Our 162nd episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 15, 2024