Logo

    #156 - OpenAI's Sora, Gemini 1.5, BioMistral, V-JEPA, AI Task Force, Fun!

    enFebruary 25, 2024
    What new capabilities does OpenAI's Sora model showcase?
    What limitations does the new LPU chip face?
    How does Scenario's tool benefit video game developers?
    What role does Jeremy Harris play in AI policy announcements?
    What is the purpose of the new 'fun section' introduced?

    • New text-to-video model, Sora, steals the showOpenAI's new text-to-video model, Sora, generates high-resolution, clear videos from text inputs, marking a significant leap forward in text-to-video AI. Policy discussions included an announcement for an experienced AI team leader position at Gladstone AI.

      Last week saw significant advancements in the field of AI, specifically in the areas of text-to-video models and policy discussions. OpenAI's new text-to-video model, Sora, stole the spotlight with its impressive capabilities, showcasing high-resolution, clear videos generated from text inputs. Although some may argue that Gemini 1.9 was also a notable development, the general consensus seems divided. Sora's technology, which uses a transformer model to embed meaning from images, marks a significant leap forward in text-to-video AI. Despite the excitement, it's important to note that OpenAI did not release the specifics of their model's setup in their report. On the policy front, Jeremy Harris, one of the hosts, made an announcement about a potential career opportunity for an experienced AI team leader. The industry veteran, Ben, has experience working at a Fortune 100 company and holds a PhD in physics. Interested parties can reach out to Gladstone AI for more information. Lastly, the hosts introduced a new segment called the "fun section," where they'll share less serious stories that don't quite fit into the other sections. They closed the episode with this new addition, aiming to add some levity to the discussion. Overall, the past week demonstrated the rapid progress being made in AI, from impressive technological breakthroughs to important policy discussions. Stay tuned for more updates on Last Week in AI.

    • OpenAI's new model, Sora, generates physically accurate videos from text descriptionsOpenAI's Sora model uses 'spacetime patches' to understand physical world models and generate meaningful and accurate videos from text descriptions, marking a significant breakthrough in AI-generated videos.

      OpenAI has developed a new model called Sora, which can generate meaningful and physically accurate videos based on text descriptions. This model goes beyond language models by also incorporating video and text data, creating what OpenAI calls "spacetime patches" – atomic units of meaning in the context of video generation. These spacetime patches enable the model to learn and understand physical world models, including laws of physics, as evidenced by its ability to accurately portray real-world phenomena like glasses shattering or balls falling. Sora is not just a language model; it's a diffusion transformer, which generates videos end-to-end, and it can perform various tasks like text-to-video, image-to-video, and video-to-video editing. The impressive results are a significant breakthrough in the field of AI-generated videos and demonstrate OpenAI's ongoing belief in the robustness of transformers to learn world models. However, Sora is currently only available to a select group of red teamers, artists, designers, and filmmakers for assessment and feedback.

    • OpenAI's Sora demonstrates object permanence, Google introduces mid-sized model Gemini 1.5 ProOpenAI's Sora shows understanding of object permanence, Google's Gemini 1.5 Pro offers a large context window and potentially competitive pricing, advancing AI research and potentially pressuring OpenAI.

      OpenAI's latest model, Sora, has demonstrated the ability to understand and track objects over long time horizons, showcasing an emergent understanding of object permanence. This breakthrough aligns with OpenAI's mission to build AGI and their consistent theme of chunking up data and looking at it from the right perspective before applying massive scale. Google's Gemini 1.5 Pro, a mid-sized multimodal model, was also announced, boasting a large context window and reportedly being as good as, if not more efficient than, their higher-tier model. This new model, with its impressive context window and potential competitive pricing, could add pressure on OpenAI. Sora's object permanence understanding and OpenAI's data chunking approach are significant advancements, offering a glimpse into the future of AI research.

    • New AI model with improved recall in large context windowsA new AI model can process and recall info from context windows over a million tokens long, outperforming all previous models in recall tests.

      A new AI model has been developed which can process and recall information from context windows that are over a million tokens (approximately 750,000 words) long. This is a significant improvement over previous models, as they often forgot details mentioned early in their prompts when dealing with such large context windows. This new model, which is incredibly powerful, achieves near-perfect recall in needle and haystack tests, outperforming all previous models including GPT-4 and Gemini Ultra. The exact mechanism by which this recall is achieved is not yet clear, but it may involve some form of stateful memory or algorithmic modification. This breakthrough allows the system to learn and absorb context incredibly quickly, even picking up obscure languages like Kalamang, which has fewer than 200 speakers worldwide. It's likely that this improvement has been achieved through tweaks to the decoding and prompting processes, but further research is needed to understand the full extent of this impressive advancement in AI technology.

    • Gemini 1.5 Pro outperforms expectations, GrogQ's new language processing unit impressesGemini 1.5 Pro surpasses power law fit with potential algorithmic improvement or scaling effect. GrogQ's language processing unit boasts 500 tokens per second throughput but may be limited by onboard RAM.

      The Gemini 1.5 Pro model outperforms expectations in making qualitatively better predictions than what was previously thought possible, even surpassing the power law fit. This could be due to a fundamental algorithmic improvement or an unexpected scaling effect. On a different note, GrogQ Inc's new language processing unit, GrogQ, has gained attention for its impressive 500 tokens per second throughput, which is four times faster than other inference services. However, its limited onboard RAM might limit the number of customers it can serve at a given time. These advancements highlight the ongoing innovation in AI and language models.

    • New LPU chip for LLM inference and game dev tool for consistent character generationA new LPU chip for inference in LLMs and a game dev tool for consistent character generation mark significant advancements in AI, but come with limitations and raise questions about IP protections.

      The new language processing unit (LPU) chip, which is specifically designed for inference in large language models (LLMs), is a significant breakthrough in the field of AI. However, it also comes with limitations, such as the high cost and the fact that it only does inference and not training. This means that companies will need to invest heavily in infrastructure to use this chip effectively. Furthermore, the trend towards models doing more of their thinking during inference rather than training suggests that we can expect to see more specialized chips for LLM use cases in the future. Another interesting development is Scenario's new tool that allows video game developers to create consistent assets of a character from a single reference image. This is a significant step forward in addressing the challenge of creating consistent character generation, which has been a long-standing issue in industries like animation, webcomics, and video games. The tool, which generates IP for commercial applications, raises questions about copyright and IP protections. OpenAI, which already has a web crawler (gpt bot), is also reportedly working on a web search product using its language model technology. This is not surprising given Microsoft's use of gpt for its customized search product and OpenAI's partnership with Microsoft. Overall, these developments highlight the rapid pace of innovation in the field of AI and the increasing importance of specialized hardware for inference in LLMs.

    • Adobe's New AI Assistant in Acrobat and OpenAI's Unique VC FundAdobe introduces AI chat feature in Acrobat, while OpenAI has a unique VC fund structure, raising governance concerns, but the AI market's rapid growth and intense competition indicate ongoing innovation.

      The AI landscape is becoming increasingly crowded as more companies, including Google and Adobe, introduce AI-enabled products. The latest addition is Adobe Acrobat's AI Assistant, which allows users to interact with documents in a chat-like manner. Meanwhile, OpenAI, a leading player in the AI space, has a peculiar ownership structure for its venture capital fund, with Sam Altman personally owning it despite OpenAI's involvement in its operations. This arrangement raises questions about governance and potential risks, especially considering the significant investments made by the fund. The market share imbalance between Google and OpenAI, along with competition from other players, adds to the uncertainty of OpenAI's ability to make a substantial impact. However, the rapid advancement of AI technology and the growing demand for AI-integrated tools suggest that the competition will only intensify.

    • Companies like OpenAI, Reddit, and NVIDIA are making significant strides in AIOpenAI keeps a secret fund for lawyers, Reddit signs a $60M annual AI content deal, and NVIDIA reveals a massive AI supercomputer

      The intersection of technology and business continues to evolve rapidly, with companies like OpenAI, Reddit, and NVIDIA pushing the boundaries of what's possible in the realm of AI. OpenAI, an organization known for its advanced AI research, has been keeping a secret fund for lawyers, adding to the intrigue surrounding the organization's unusual structure. Reddit, meanwhile, has signed a significant AI content licensing deal, reportedly worth $60 million annually, which could set a precedent for future deals. And NVIDIA has revealed its EOS supercomputer, a massive AI processing machine with 4,608 H100 GPUs, marking a notable achievement in the world of individual supercomputers for AI applications. These developments underscore the growing importance of AI and the significant investments being made in this area. The future is sure to bring more advancements and innovations as these companies and others continue to push the boundaries of what's possible.

    • Recent advancements in AI technologyGoogle's Goose AI and EOS supercomputer's language model showcase rapid progress in AI, with Goose trained on decades of engineering expertise and EOS completing a billion token task in minutes.

      The advancements in AI technology are progressing at an unprecedented pace. EOS, a supercomputer, was able to train a large language model on one billion tokens in just four minutes, which was previously a three and a half year long process. Google also launched an internal AI model named Goose, which is aimed at helping employees write code faster by utilizing the internal text tags of the company. This model is reportedly trained on 25 years of engineering expertise at Google. Chinese startup Moonshot AI raised a billion dollars in funding for its open AI play, having launched a smart chat bot, KimiChat, in October. Moonshot AI's large language model, moonshot LLM, is capable of processing up to 200,000 Chinese characters in its context window. These advancements demonstrate the significant investments being made in AI technology and the rapid progress being made in the field. Additionally, it's noteworthy that these developments involve collaboration between various parts of tech giants like Google and the involvement of military-affiliated institutions in China. These context windows, while impressive, should be evaluated not just by their size, but also by the quality of the AI's handling of the context.

    • Chinese AI Market: Significant Investment and InnovationChinese AI companies are making headlines with large fundraises and innovative developments in compute and generative AI training, open source language models for medical domains, and fully open source long context text embedding models.

      The AI industry, particularly in the Chinese market, is seeing significant investment and innovation. Chinese models are making headlines with large fundraises, but a cautious approach is recommended due to the potential focus on vanity metrics. Lambda, a competing firm, raised 320 million and specializes in compute and generative AI training, making them a notable player. Another company, ex sales, raised 110 million for AI agents for businesses, which are enabling various interactions between businesses and customers. Open source developments include Bio Mistral, a collection of pre-trained large language models for medical domains, which outperforms other models on various medical question answering tasks. The model was trained using the French National Center for Scientific Research's high-performance computer, demonstrating the impact of giving researchers access to powerful hardware. NOMIC AI released the first fully open source long context text embedding model, Norney Symbed text V1, which surpasses OpenAI's performance on various benchmarks. This model can handle sequence lengths of 8,000 tokens and is available under an Apache 2 license. These developments showcase the growing importance and potential of AI in various industries and the ongoing advancements in the field. The open-source nature of many of these projects allows for collaboration and further innovation, leading to more advanced models and applications.

    • Advancements in AI: Longformer for Text and V-Jepa for VideoResearchers are developing new AI models and techniques, like Longformer for text and V-Jepa for video, to improve data understanding and push the boundaries of AI capabilities. These models prioritize efficiency and accessibility, making the most of available resources and enabling advanced text and video understanding.

      Researchers are continuously pushing the boundaries of artificial intelligence (AI) by developing advanced models and techniques to improve training and understanding of various forms of data, such as text and video. The first example discussed a paper on a new text model called "Longformer," which uses long context windows and a small model size to achieve superior performance on short and long context benchmarks, surpassing OpenAI's text embedding models. The researchers emphasized the importance of having a small, efficient model that can effectively process long contexts, which was a challenge until now. They also noted the openness and accessibility of the project, making the code, data, and everything else readily available to the public. The second story featured Meta's new V-Jepa AI model, which aims to predict patches of video based on the idea that closely related information often appears in neighboring patches or frames. This model, like OpenAI's Sora, operates on the embedding space and takes advantage of the fact that meaning is similar in closely related parts of the video. The ultimate goal is to train an encoder that can extract meaningful embeddings from patches of video or images. Both projects demonstrate the ongoing efforts to create more advanced and versatile AI models, with a focus on making the most of available resources and pushing the boundaries of what's possible in the realm of text and video understanding.

    • Exploring approaches to AGI: VJAPA and language modelsResearchers are experimenting with different methods for AGI, including unsupervised video learning (VJAPA) and language model reasoning. VJAPA focuses on scalability and handling large data, while language models reveal thought processes and could lead to new applications.

      The research presented in the discussed papers showcases different approaches to achieving artificial general intelligence (AGI) and the importance of both scalability and specialized architectures. The first paper introduces VJAPA, a model for unsupervised learning from video, which has limitations compared to more advanced models like Sora, but reflects a commitment to Jan LeCun's vision of AGI and explores feature prediction as a new objective for unsupervised learning. The model is primarily a research effort, focusing on cell supervised training and can handle large amounts of unlabeled video data, which could lead to more efficient scaling. The second paper discusses the ability of language models to perform chain of thought reasoning without explicit prompting. The researchers show that using the top K alternative tokens during decoding can reveal the chain of thought paths inherent in the sequences, potentially increasing confidence in the final answer when a chain of thought type output is present. This finding highlights the inherent capabilities of language models and could lead to new applications. Both papers demonstrate the ongoing research in the field of AGI and the importance of exploring various approaches, from specialized architectures to scalability, to make progress towards human-level intelligence. The findings in these papers could potentially lead to more efficient methods for training large models and unlocking new capabilities in AI systems.

    • Advancements in AI: Longer Context Windows and Emergent AbilitiesRecent advancements include longer context windows via ring attention mechanisms and emergent abilities in text-to-speech technology, highlighting the ongoing trend towards more sophisticated AI capabilities.

      Recent advancements in AI research, specifically in the areas of longer context windows and emergent abilities, are pushing the boundaries of what's possible in the field. Firstly, the development of ring attention mechanisms in transformer models allows for longer context windows, theoretically with no limit, by passing intermediary keys and values between multiple devices in a ring-like structure. This is a significant step towards achieving more comprehensive understanding and context in AI. Secondly, Amazon's AGI team has made strides in text-to-speech technology with their model, Big Adaptive Streamable TTS, which has shown emergent abilities, handling complex aspects of text, such as foreign words and punctuation, without explicit training. Moreover, the collaboration between OpenAI and Microsoft to document the use of their systems by foreign hackers highlights the importance of safety and security in AI research and application. Hackers have been using AI for relatively mundane tasks, such as drafting emails and debugging code, and OpenAI and Microsoft are working to prevent such unintended uses. These developments underscore the ongoing trend towards longer context and more sophisticated capabilities in AI, with researchers and companies continually pushing the envelope to unlock new possibilities. It's an exciting time for AI research and development, with numerous potential applications and implications for various industries and society as a whole.

    • State-affiliated hacking groups use advanced technologies like OpenAI models to gain insightsState-affiliated hackers use OpenAI models and other advanced tech to expand their operational scope, targeting sectors like gov't, education, comms, oil & gas, and U.S. defense contractors.

      State-affiliated hacking groups, such as Charcoal Typhoon (Chinese), Salmon Typhoon (Chinese), and Forest Blizzard (Russian), are increasingly using advanced technologies, including OpenAI models, to translate technical papers and gain insights into intelligence agencies and potential targets. These groups have a broad operational scope, targeting sectors like government, higher education, communications, oil and gas, and have a history of targeting U.S. defense contractors and cryptographic tech companies. The Russian group Forest Blizzard has been linked to GRU unit 26165 and has been active in the context of Ukraine. In response, the U.S. House of Representatives has launched a bipartisan AI task force, led by Speaker Mike Johnson and Minority Leader Hakeem Jeffries, to support AI innovation and study potential threats. The task force will have 24 members and will be chaired by J. Obernolte, a computer science expert and video game developer, and co-chaired by Ted Lieu, a computer science background and known hawk on AI issues. The task force aims to canvas around for a wide range of opinions on where the field might be going and account for the fact that a lot is unknown, especially regarding the potential speed of technological advancements.

    • Researchers reveal fingerprints from touchscreen soundsResearchers can extract partial fingerprints from touchscreen sounds and complete fingerprints with more effort, highlighting the need to safeguard data. AI ethics and regulations are evolving, with new bills in the US and a legal precedent in Canada.

      Our digital interactions, even seemingly insignificant ones like swiping on a touchscreen, could potentially reveal sensitive information such as fingerprints. Researchers from the University of Colorado and China have developed a side channel attack that can reproduce about 28% of partial fingerprints and 10% of complete fingerprints based on the sounds made while swiping. This highlights the ease with which information can be gathered from our environments and the increasing importance of monitoring and safeguarding data. Another significant development is the increasing number of AI-related bills being introduced in the US, with 15 new bills per week and a total of 407 bills across more than 40 states. New York, California, Tennessee, and Illinois are among the states leading the charge. In the realm of AI ethics, a ruling in Canada set a legal precedent for AI alignment, with Air Canada being found liable for inaccurate information provided by its chatbot. The decision implies a legal requirement for reasonable care in ensuring AI systems give true outputs, opening up interesting discussions about the responsibilities and accountabilities of AI developers and users. Lastly, the FTC issued a warning about potential quiet changes to Terms of Service (ToS) for AI training data, with companies potentially altering their privacy policies to use user data without restriction. The FTC encourages users to stay informed and vigilant about their data and privacy.

    • FTC Warns Firms About Changing Privacy Policies Without ConsentCompanies must be transparent about their privacy policy changes and cannot secretly alter them to monetize user data without consent, according to the FTC. Zoom faced backlash for doing so and a lawsuit was partially dismissed against OpenAI, setting a potential precedent for limiting lawsuits against LLM companies.

      Companies must be transparent about their privacy policies and cannot secretly change them to make money from user data without consent. This was highlighted in the FTC's recent warning to firms, including Zoom, which updated its terms of service in August 2023 to allow the use of user data for AI training without an opt-out option. The lawsuit against OpenAI by Sara Silverman and others was partially dismissed in a California court, leaving only the claims of direct copyright infringement and unfair competition. The judge dismissed the other claims due to a lack of clear economic injury and the speculative nature of the risk of future damage to intellectual property. These rulings could set a precedent for limiting the types of lawsuits that can be brought against LLM companies. The visual guide to Mamba and state space models by Martin Ruten Dorst provides a detailed explanation of the architecture and concepts behind these models, which are built on control theory and hardware optimization. These topics can be quite technical, but the guide offers a comprehensive understanding of the subject matter.

    • AI-generated misinformation in research and creative industriesTwo instances of AI-generated misinformation, in scientific research and creative industries, sparked concerns about the reliability and authenticity of AI-content, highlighting the need for greater oversight and transparency.

      The use of artificial intelligence (AI) in scientific research and creative industries is a topic of growing concern. The discussion highlighted two instances where AI was involved in producing misleading or false information. The first instance involved a retracted scientific paper that contained AI-generated images, and the second instance involved an AI-generated speech read at an awards ceremony by Helen Mirren. Both incidents raised questions about the reliability and authenticity of AI-generated content and sparked a backlash against the use of AI in these fields. The Microsoft Super Bowl ad, which aimed to reassure people that AI is here to help make their dreams come true, seemed to be an attempt to reframe the narrative around AI in a more positive light. However, the incidents discussed serve as a reminder that there is a need for greater oversight and transparency in the use of AI in research and creative industries to prevent the spread of misinformation and maintain the integrity of these fields.

    • Microsoft's Super Bowl ad: AI as an enabler of human achievementMicrosoft's Super Bowl ad challenges the cultural perception of AI as a risk or threat, positioning it as an enabler of human advancement instead.

      Microsoft's recent Super Bowl ad showcases their efforts to reframe Artificial Intelligence (AI) as an enabler of human achievement rather than a replacement or source of fear. The ad, which features Microsoft's co-pilot AI bot, aims to challenge the cultural perception of AI as a risk or threat, much like Apple did in their 1984 Super Bowl ad. Although some viewers found the ad to be underwhelming, its message is clear: AI exists to help us, not replace us. Microsoft's marketing strategy is a reflection of the ongoing conversation about the role of AI in our lives and the potential it holds for human advancement.

    Was this summary helpful?

    Recent Episodes from Last Week in AI

    # 182 - Alexa 2.0, MiniMax, Surskever raises $1B, SB 1047 approved

    # 182 - Alexa 2.0, MiniMax, Surskever raises $1B, SB 1047 approved

    Our 182nd episode with a summary and discussion of last week's big AI news! With hosts Andrey Kurenkov and Jeremie Harris.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/. If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Sponsors:

    - Agent.ai is the global marketplace and network for AI builders and fans. Hire AI agents to run routine tasks, discover new insights, and drive better results. Don't just keep up with the competition—outsmart them. And leave the boring stuff to the robots 🤖

    - Pioneers of AI, is your trusted guide to this emerging technology. Host Rana el Kaliouby (RAH-nuh el Kahl-yoo-bee) is an AI scientist, entrepreneur, author and investor exploring all the opportunities and questions AI brings into our lives. Listen to Pioneers of AI, with new episodes every Wednesday, wherever you tune in.

    In this episode:

    - OpenAI's move into hardware production and Amazon's strategic acquisition in AI robotics. - Advances in training language models with long-context capabilities and California's pending AI regulation bill. - Strategies for safeguarding open weight LLMs against adversarial attacks and China's rise in chip manufacturing. - Sam Altman's infrastructure investment plan and debates on AI-generated art by Ted Chiang.

    Timestamps + Links:

    • (00:00:00) Intro / Banter
    • (00:05:15) Response to listener comments / corrections
    Last Week in AI
    enSeptember 17, 2024

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    Our 181st episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov and Jeremie Harris

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode:

    - Google's AI advancements with Gemini 1.5 models and AI-generated avatars, along with Samsung's lithography progress.  - Microsoft's Inflection usage caps for Pi, new AI inference services by Cerebrus Systems competing with Nvidia.  - Biases in AI, prompt leak attacks, and transparency in models and distributed training optimizations, including the 'distro' optimizer.  - AI regulation discussions including California’s SB1047, China's AI safety stance, and new export restrictions impacting Nvidia’s AI chips.

    Timestamps + Links:

    Last Week in AI
    enSeptember 15, 2024

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    Our 180th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    • Ideogram AI's new features, Google's Imagine 3, Dream Machine 1.5, and Runway's Gen3 Alpha Turbo model advancements.
    • Perplexity's integration of Flux image generation models and code interpreter updates for enhanced search results. 
    • Exploration of the feasibility and investment needed for scaling advanced AI models like GPT-4 and Agent Q architecture enhancements.
    • Analysis of California's AI regulation bill SB1047 and legal issues related to synthetic media, copyright, and online personhood credentials.

    Timestamps + Links:

    Last Week in AI
    enSeptember 03, 2024

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    Our 179th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    - Grok 2's beta release features new image generation using Black Forest Labs' tech.

    - Google introduces Gemini Voice Chat Mode available to subscribers and integrates it into Pixel Buds Pro 2.

    - Huawei's Ascend 910C AI chip aims to rival NVIDIA's H100 amidst US export controls.

    - Overview of potential risks of unaligned AI models and skepticism around SingularityNet's AGI supercomputer claims.

    Timestamps + Links:

    Last Week in AI
    enAugust 20, 2024

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    Our 178th episode with a summary and discussion of last week's big AI news!

    NOTE: this is a re-upload with fixed audio, my bad on the last one! - Andrey

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode: - Notable personnel movements and product updates, such as Character.ai leaders joining Google and new AI features in Reddit and Audible. - OpenAI's dramatic changes with co-founder exits, extended leaves, and new lawsuits from Elon Musk. - Rapid advancements in humanoid robotics exemplified by new models from companies like Figure in partnership with OpenAI, achieving amateur-level human performance in tasks like table tennis. - Research advancements such as Google's compute-efficient inference models and self-compressing neural networks, showcasing significant reductions in compute requirements while maintaining performance.

    Timestamps + Links:

    Last Week in AI
    enAugust 16, 2024

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    Our 177th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode again coming out about a week late, next one will be coming out soon...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you'd like to listen to the interview with Andrey, check out https://www.superdatascience.com/podcast

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    In this episode, hosts Andrey Kurenkov and John Krohn dive into significant updates and discussions in the AI world, including Instagram's new AI features, Waymo's driverless cars rollout in San Francisco, and NVIDIA’s chip delays. They also review Meta's AI Studio, character.ai CEO Noam Shazir's return to Google, and Google's Gemini updates. Additional topics cover NVIDIA's hardware issues, advancements in humanoid robots, and new open-source AI tools like Open Devon. Policy discussions touch on the EU AI Act, the U.S. stance on open-source AI, and investigations into Google and Anthropic. The impact of misinformation via deepfakes, particularly one involving Elon Musk, is also highlighted, all emphasizing significant industry effects and regulatory implications.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 11, 2024

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    Our 176th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode coming out about a week late, things got in the way of editing it...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

     

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 03, 2024

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    Our 175th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, hosts Andrey Kurenkov and Jeremy Harris explore recent AI advancements including OpenAI's release of GPT 4.0 Mini and Mistral’s open-source models, covering their impacts on affordability and performance. They delve into enterprise tools for compliance, text-to-video models like Hyper 1.5, and YouTube Music enhancements. The conversation further addresses AI research topics such as the benefits of numerous small expert models, novel benchmarking techniques, and advanced AI reasoning. Policy issues including U.S. export controls on AI technology to China and internal controversies at OpenAI are also discussed, alongside Elon Musk's supercomputer ambitions and OpenAI’s Prover-Verify Games initiative.  

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 25, 2024

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    Our 174rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, we delve into the latest advancements and challenges in the AI industry, highlighting new features from Figma and Quora, regulatory pressures on OpenAI, and significant investments in AI infrastructure. Key topics include AMD's acquisition of Silo AI, Elon Musk's GPU cluster plans for XAI, unique AI model training methods, and the nuances of AI copying and memory constraints. We discuss developments in AI's visual perception, real-time knowledge updates, and the need for transparency and regulation in AI content labeling and licensing.

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 17, 2024

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    Our 173rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode of Last Week in AI, we explore the latest advancements and debates in the AI field, including Google's release of Gemini 1.5, Meta's upcoming LLaMA 3, and Runway's Gen 3 Alpha video model. We discuss emerging AI features, legal disputes over data usage, and China's competition in AI. The conversation spans innovative research developments, cost considerations of AI architectures, and policy changes like the U.S. Supreme Court striking down Chevron deference. We also cover U.S. export controls on AI chips to China, workforce development in the semiconductor industry, and Bridgewater's new AI-driven financial fund, evaluating the broader financial and regulatory impacts of AI technologies.  

    Timestamps + links:

    Last Week in AI
    enJuly 07, 2024

    Related Episodes

    The Author Behind ‘Arrival’ Doesn’t Fear AI. ‘Look at How We Treat Animals.’

    The Author Behind ‘Arrival’ Doesn’t Fear AI. ‘Look at How We Treat Animals.’

    For years, I’ve kept a list of dream guests for this show. And as long as that list has existed, Ted Chiang has been atop it.

    Chiang is a science fiction writer. But that undersells him. He has released two short story collections over 20 years — 2002’s “Stories of Your Life and Others” and 2019’s “Exhalation.” Those stories have won more awards than I can list, and one of them was turned into the film “Arrival.” They are remarkable pieces of work: Each is built around a profound scientific, philosophical or religious idea, and then the story or the story structure is shaped to represent that idea. They are wonders of precision and craft. But unlike a lot of science fiction, they are never cold. Chiang’s work is deeply, irrepressibly humane.

    I’ve always wondered about the mind that would create Chiang’s stories. And in this conversation I got to watch it in action. Chiang doesn’t like to talk about himself. But he does like to talk about ideas. And so we do: We discuss the difference between magic and technology, why superheroes fight crime but ignore injustice, what it would do to the human psyche if we knew the future is fixed, whether free will exists, whether we’d want to know the exact date of our deaths, why Chiang fears what humans will do to artificial intelligence more than what A.I. will do to humans, the way capitalism turns people against technology, and much more.

    The ideas Chiang offered in this conversation are still ringing in my head, and changing the way I see the world. It’s worth taking your time with this one.

    Recommendations: 

    "Creation" by Steve Grand

    "On the Measure of Intelligence" by Francois Chollet

    "CivilWarLand in Bad Decline" by George Saunders

    "A Visit from the Goon Squad" by Jennifer Egan

    "Royal Space Force: The Wings of Honnêamise" (movie)

    "On Fragile Waves" by Lily Yu

    "Pilgrim at Tinker Creek" by Annie Dillard

    Control (video game)

    Return of the Obra Dinn (video game)

    You can find transcripts (posted midday) and more episodes of "The Ezra Klein Show" at nytimes.com/ezra-klein-podcast, and you can find Ezra on Twitter @ezraklein.

    Thoughts? Guest suggestions? Email us at ezrakleinshow@nytimes.com.

    “The Ezra Klein Show” is produced by Rogé Karma and Jeff Geld; fact-checking by Michelle Harris; original music by Isaac Jones; mixing by Jeff Geld.

    7 Takeaways from the Senate's AI Hearing

    7 Takeaways from the Senate's AI Hearing
    From a surprising lack of skepticism to clear echoes of social media regulatory failures, NLW covers everything that happened in the first Senate AI hearing in the post ChatGPT era.  The AI Breakdown helps you understand the most important news and discussions in AI.  Subscribe to The AI Breakdown newsletter: https://theaibreakdown.beehiiv.com/subscribe Subscribe to The AI Breakdown on YouTube: https://www.youtube.com/@TheAIBreakdown

    EU vs. AI

    EU vs. AI
    The EU has advanced first-of-its-kind AI regulation. The Verge’s Jess Weatherbed tells us whether it will make a difference, and Columbia University’s Anu Bradford explains the Brussels effect. This episode was produced by Amanda Lewellyn, edited by Matt Collette, fact-checked by Laura Bullard, engineered by Patrick Boyd, and hosted by Sean Rameswaram. Transcript at vox.com/todayexplained Support Today, Explained by making a financial contribution to Vox! bit.ly/givepodcasts Learn more about your ad choices. Visit podcastchoices.com/adchoices

    Mayhem at OpenAI + Our Interview With Sam Altman

    Mayhem at OpenAI + Our Interview With Sam Altman

    Last week, we interviewed Sam Altman. Since then, well, everything has changed. The board of OpenAI, maker of ChatGPT, fired Altman as chief executive on Friday. Over the weekend, it looked as if he might return. On Sunday night, Microsoft hired Altman to lead a new A.I. venture. Who knows what will happen next.

    Today, an update on a crazy weekend in tech, and our interview with Sam Altman.

    Today’s Guest:

    • Sam Altman is the former chief executive of OpenAI.

    Additional Reading:

    Logo

    © 2024 Podcastworld. All rights reserved

    Stay up to date

    For any inquiries, please email us at hello@podcastworld.io