Logo

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    enJuly 17, 2024
    What is the focus of the startup Odyssey?
    How much funding has Odyssey raised for its project?
    What does test time training (TTT) allow in AI models?
    What lawsuit was recently dismissed against Microsoft and OpenAI?
    How does Vimeo contribute to AI content labeling?

    Podcast Summary

    • Text-to-Video AI startupsNew AI startup Odyssey raised $9M for creating separate models for geometry, materials, lighting, and motion in text-to-video generation, offering finer control and importance for commercial applications and safety/interpretability reasons.

      The field of AI is continuously evolving, with new startups and technologies emerging every week. During a recent episode of Last Week in AI, hosts Andrei Kamensky and Jeremy Harris discussed various news stories, including the launch of Odyssey, a new AI startup focused on building Hollywood-grade text-to-video models. The company, which has raised $9 million in seed funding, aims to provide full control over core layers of visual storytelling by creating separate models for geometry, materials, lighting, and motion. This approach allows for finer control over the AI system and is important for commercial applications, as well as for safety and interpretability reasons. The trend of text-to-video generation has gained momentum throughout the year, and it will be interesting to see how these technologies develop and find their product-market fit. Additionally, the hosts acknowledged the importance of feedback from their audience and addressed comments regarding the focus on geopolitics and bias in their discussions. Overall, the episode highlights the dynamic nature of the AI industry and the ongoing efforts to create more capable and controllable AI systems.

    • AI innovationAnthropic introduces a prompt playground for AI app improvement, while Figma's AI-generated design feature faces controversy, highlighting the need for ongoing advancements and proper QA processes in the AI sector.

      The field of AI is continuously evolving, with companies like Anthropic and Frapplex introducing new features to enhance the user experience and productivity. Anthropic's latest addition is a prompt playground, which allows developers to generate, test, and evaluate prompts to improve AI apps' responses for specialized tasks. This not only makes the use of cloud more formal but also collects valuable data for fine-tuning models. The potential integration of this technology into moviemaking visual effects workflows is also an intriguing development. Meanwhile, Figma's AI-generated design feature, Make Design, faced controversy when it produced designs similar to existing apps, such as Apple's Weather app. The controversy led to the feature being paused, and Figma's response suggested that they used off-the-shelf language models combined with commissioned systems. However, the response did not provide a clear explanation, and the issue was ultimately traced back to underlying design systems. The CEO acknowledged the lack of a proper QA process and the need to improve it to avoid potential copyright issues. This incident raises interesting legal questions regarding AI-generated content and intellectual property. Overall, these developments demonstrate the importance of continuous innovation and improvement in the AI sector.

    • AI models and chatbots improvementsCompanies like Quora, Grok, and Suno are enhancing AI models and chatbots with new features and integrations, enabling faster testing and comparison, while addressing challenges like memory usage and scalability

      Companies are continuously releasing new features and integrations for AI models and chatbots, but users often encounter issues and bugs. Quora's new feature, Previews, allows users to create and share interactive web apps within chatbot conversations, making it easier to test and compare various models. Grok, a company specializing in fast inference for language models, has announced a lightning-fast LM engine and a console for developers to switch from OpenAI. Suno, a text-to-music startup, has launched an iPhone app, providing users the ability to generate music ideas on the go. These developments reflect the increasing focus on inference and the trade-off between investing compute power during training or inference time. However, challenges such as memory usage and scalability remain. Microsoft and Apple have joined the boards of various AI companies amid regulatory scrutiny. Overall, these advancements demonstrate the rapid evolution of AI technology and its integration into various industries and applications.

    • Microsoft and OpenAI relationship investigationsAntitrust concerns led to Microsoft's resignation from OpenAI's board and departure of key personnel, raising questions about Microsoft's level of control over OpenAI despite no majority stake or voting seat.

      Regulatory investigations into Microsoft and OpenAI's relationship led to the resignation of key personnel and the departure of Microsoft from OpenAI's board due to antitrust concerns. This rapid turnaround came after Microsoft offered Sam Altman, who was fired from OpenAI, a position at Microsoft to head up a big AI research group. This move effectively undercut the board's authority and raised questions about the level of control Microsoft held over OpenAI despite not owning a majority stake or having a voting seat. Meanwhile, OpenAI is also working on an AI health coach in partnership with Thrive AI Health, which aims to provide personalized health advice using AI and users' medical data. The startup Magic, which develops AI models to write software, is in talks to raise $200 million in funding, valuing the company at $1.5 billion, despite having no revenue or product for sale yet. These large funding rounds reflect the significant value investors see in automating software development.

    • GPU technology in AIVC firms like Andreessen Horowitz and Sequoia Capital are making significant investments in GPU technology to gain a competitive edge in AI, with Andreessen Horowitz investing around $1.3 billion in generative AI deals over the last two years, but the value of these investments is being questioned due to the lack of significant consumer revenues and the depreciation of GPU values over time.

      The race for AI dominance is heating up, and venture capital firms like Andreessen Horowitz and Sequoia Capital are making significant investments in GPU technology to gain a competitive edge. Andreessen Horowitz has amassed over 20,000 GPUs to support their portfolio companies, investing heavily in generative AI deals worth around $1.3 billion over the last two years. Sequoia Capital, on the other hand, has expressed skepticism about the value of these investments, pointing out the lack of significant consumer revenues in the AI industry and the depreciation of GPU values over time. Elon Musk's XAI is also making headlines with plans to build a massive 100,000 GPU training cluster, further highlighting the importance of GPU technology in the AI space. These moves illustrate the growing financialization of AI, with compute becoming a currency of sorts for these investments. However, the question remains whether these investments will pay off in the long run, or if the market will become increasingly commoditized.

    • AI CompetitionSmall companies like XAI and Skilled AI are competing with tech giants in AI research and hardware design, while AMD and Intel are making strategic acquisitions and investments to enhance their capabilities.

      The AI industry is witnessing significant investments and advancements from various players, both in software and hardware domains. XAI, a small company, is aiming to compete with tech giants like Microsoft and Google in AI research and in-house data center design. AMD is acquiring SiloAI, a Finnish AI company, to enhance their software expertise and catch up with competitors. Skilled AI, a Pittsburgh-based startup, raised $300 million for their series A, developing general-purpose robotics models with a foundation model that claims to be unusually robust. Intel is beginning construction on a large chip fab in Germany to compete with market leader TSMC and reduce reliance on Taiwan. Investors, including Jeff Bezos, Sequoia Capital, and Carnegie Mellon, have shown confidence in these companies, indicating their potential to make significant contributions to the field. The industry continues to evolve rapidly, with companies exploring various approaches to AI and robotics to push the boundaries of technology.

    • Language model updatesReal-time weight updates through test time training and composable interventions are potential solutions for effectively and reliably incorporating new information into language models.

      The ability to effectively and reliably incorporate new information into language models is a significant challenge. Traditional methods like prepending knowledge or knowledge updating don't always work effectively, especially when dealing with dynamic information such as API updates or new documentation. The introduction of test time training (TTT) and composable interventions are potential solutions to this problem. TTT allows for real-time weight updates, while composable interventions enable the study of the effects of multiple interventions on the same language model. Additionally, the order in which interventions like knowledge editing, unlearning, and model compression are applied matters, with the performance being negatively impacted if certain interventions are done in the wrong order. It's important to consider the potential impact of these interventions on various metrics and to evaluate models comprehensively following these interventions. The ability to effectively and efficiently incorporate new information into language models will become increasingly important as the software and data environments continue to evolve. Another interesting paper discussed the use of a predictive attention mechanism for neural decoding of visual perception, which can reconstruct what a person is looking at with remarkable accuracy using fMRI recordings.

    • AI-based photograph analysis, malicious usesResearch advances in deciphering photo content through AI and brain data, but concerns arise over potential misuse, such as covertly training models to understand harmful requests or hiding malicious messages in plaintext, which can evade detection

      Researchers are making significant strides in deciphering what people are looking at in photographs using AI and brain data, leading to impressive reconstructions. This technique, called fMRI-based reconstruction, focuses on relevant brain areas and produces relatively accurate results, though not quite at the level of mind reading yet. Simultaneously, there's a growing concern about potential malicious uses of AI, such as covert malicious fine-tuning, which can make models respond to harmful requests without detection. In this technique, attackers can train models to understand a coded language or hide harmful messages in plaintext, making it difficult for safety and security algorithms to detect. Additionally, OpenAI faced several security issues, including a hacker gaining access to their internal messaging systems, potentially exposing sensitive information about their AI technology designs. These incidents highlight the importance of addressing both the technological advancements and potential risks associated with AI.

    • OpenAI security breach implicationsOpenAI's decision not to involve law enforcement in a security breach raises questions about their ability to make national security assessments as a private company and the potential implications for partnerships with tech giants like Microsoft.

      The discussion revolves around the incident where OpenAI, a private AI research company, experienced a security breach and chose not to involve law enforcement due to their internal assessment of the situation. The incident raises questions about OpenAI's ability to make national security assessments as a private company and the potential implications of such decisions. Additionally, OpenAI has an internal scale ranking the capabilities of AI from one to five, with AGI defined as a highly autonomous system surpassing humans in most economically valuable tasks. The definition of AGI and the board's determination of it has significant implications for OpenAI's partnership with Microsoft. Furthermore, a new dataset, "Me, Myself, and AI," explores situational awareness in LLMs and found that even the top models fell short of human baselines. The discussion also highlights the importance of defining and measuring situational awareness in AI models and the ongoing research in this area. Overall, the conversation underscores the need for transparency, accountability, and effective oversight in the development and deployment of advanced AI systems.

    • AI situational awarenessCloud 3 Opus, tested by OpenAI, has a high situational awareness score, while OpenAI partners with Los Alamos National Laboratory to explore AI use in scientific research. Safety concerns were raised about OpenAI's new models, leading to a dismissed lawsuit and increased safety testing.

      There are significant differences in situational awareness capabilities among various AI models, and these capabilities can be decoupled from other broader AI abilities. Cloud 3 Opus, tested by OpenAI, stands out for its high situational awareness score. Meanwhile, OpenAI is partnering with Los Alamos National Laboratory to explore the benefits and risks of using AI in scientific research, specifically in genetically engineering E. coli bacteria to produce insulin. A recent lawsuit against Microsoft, OpenAI, and GitHub over the use of intellectual property to train AI models was dismissed due to a failure to prove identical code reproduction. A former OpenAI safety employee, William Saunders, expressed concerns about the company's safety measures, comparing it to the Titanic, and emphasized the need for more safety testing before new models are launched. Vimeo, a video hosting service similar to YouTube and TikTok, has joined them in introducing new AI content labels.

    • AI-generated content disclosurePlatforms like YouTube, Vimeo, and Etsy are implementing new policies to require creators to disclose AI-generated elements in their content to ensure transparency and prevent confusion for users, while tech startup Avail helps media companies and independent creators monetize their data for AI training services.

      Technology companies are increasingly requiring creators to disclose when their content includes AI-generated elements. This includes platforms like YouTube, Vimeo, and Etsy, which are implementing new policies to prevent confusion and ensure transparency. For instance, Vimeo now allows creators to label their content as AI-generated, while Etsy requires sellers to classify items based on the level of human involvement in their creation. These policies are aimed at addressing the growing use of AI in content creation and ensuring that users are aware of it. Additionally, a tech startup called Avail is helping media companies and independent creators license their content for AI training services, allowing them to monetize their data instead of having it scraped. These developments reflect the growing importance of AI in content creation and the need for clear guidelines and ethical practices.

    Recent Episodes from Last Week in AI

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    #181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

    Our 181st episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov and Jeremie Harris

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode:

    - Google's AI advancements with Gemini 1.5 models and AI-generated avatars, along with Samsung's lithography progress.  - Microsoft's Inflection usage caps for Pi, new AI inference services by Cerebrus Systems competing with Nvidia.  - Biases in AI, prompt leak attacks, and transparency in models and distributed training optimizations, including the 'distro' optimizer.  - AI regulation discussions including California’s SB1047, China's AI safety stance, and new export restrictions impacting Nvidia’s AI chips.

    Timestamps + Links:

    Last Week in AI
    enSeptember 15, 2024

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    #180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

    Our 180th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    • Ideogram AI's new features, Google's Imagine 3, Dream Machine 1.5, and Runway's Gen3 Alpha Turbo model advancements.
    • Perplexity's integration of Flux image generation models and code interpreter updates for enhanced search results. 
    • Exploration of the feasibility and investment needed for scaling advanced AI models like GPT-4 and Agent Q architecture enhancements.
    • Analysis of California's AI regulation bill SB1047 and legal issues related to synthetic media, copyright, and online personhood credentials.

    Timestamps + Links:

    Last Week in AI
    enSeptember 03, 2024

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    #179 - Grok 2, Gemini Live, Flux, FalconMamba, AI Scientist

    Our 179th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Episode Highlights:

    - Grok 2's beta release features new image generation using Black Forest Labs' tech.

    - Google introduces Gemini Voice Chat Mode available to subscribers and integrates it into Pixel Buds Pro 2.

    - Huawei's Ascend 910C AI chip aims to rival NVIDIA's H100 amidst US export controls.

    - Overview of potential risks of unaligned AI models and skepticism around SingularityNet's AGI supercomputer claims.

    Timestamps + Links:

    Last Week in AI
    enAugust 20, 2024

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    #178 - More Not-Acquihires, More OpenAI drama, More LLM Scaling Talk

    Our 178th episode with a summary and discussion of last week's big AI news!

    NOTE: this is a re-upload with fixed audio, my bad on the last one! - Andrey

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode: - Notable personnel movements and product updates, such as Character.ai leaders joining Google and new AI features in Reddit and Audible. - OpenAI's dramatic changes with co-founder exits, extended leaves, and new lawsuits from Elon Musk. - Rapid advancements in humanoid robotics exemplified by new models from companies like Figure in partnership with OpenAI, achieving amateur-level human performance in tasks like table tennis. - Research advancements such as Google's compute-efficient inference models and self-compressing neural networks, showcasing significant reductions in compute requirements while maintaining performance.

    Timestamps + Links:

    Last Week in AI
    enAugust 16, 2024

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    #177 - Instagram AI Bots, Noam Shazeer -> Google, FLUX.1, SAM2

    Our 177th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode again coming out about a week late, next one will be coming out soon...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    If you'd like to listen to the interview with Andrey, check out https://www.superdatascience.com/podcast

    If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.

    In this episode, hosts Andrey Kurenkov and John Krohn dive into significant updates and discussions in the AI world, including Instagram's new AI features, Waymo's driverless cars rollout in San Francisco, and NVIDIA’s chip delays. They also review Meta's AI Studio, character.ai CEO Noam Shazir's return to Google, and Google's Gemini updates. Additional topics cover NVIDIA's hardware issues, advancements in humanoid robots, and new open-source AI tools like Open Devon. Policy discussions touch on the EU AI Act, the U.S. stance on open-source AI, and investigations into Google and Anthropic. The impact of misinformation via deepfakes, particularly one involving Elon Musk, is also highlighted, all emphasizing significant industry effects and regulatory implications.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 11, 2024

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    #176 - SearchGPT, Gemini 1.5 Flash, Lamma 3.1 405B, Mistral Large 2

    Our 176th episode with a summary and discussion of last week's big AI news!

    NOTE: apologies for this episode coming out about a week late, things got in the way of editing it...

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

     

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enAugust 03, 2024

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    #175 - GPT-4o Mini, OpenAI's Strawberry, Mixture of A Million Experts

    Our 175th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, hosts Andrey Kurenkov and Jeremy Harris explore recent AI advancements including OpenAI's release of GPT 4.0 Mini and Mistral’s open-source models, covering their impacts on affordability and performance. They delve into enterprise tools for compliance, text-to-video models like Hyper 1.5, and YouTube Music enhancements. The conversation further addresses AI research topics such as the benefits of numerous small expert models, novel benchmarking techniques, and advanced AI reasoning. Policy issues including U.S. export controls on AI technology to China and internal controversies at OpenAI are also discussed, alongside Elon Musk's supercomputer ambitions and OpenAI’s Prover-Verify Games initiative.  

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 25, 2024

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    #174 - Odyssey Text-to-Video, Groq LLM Engine, OpenAI Security Issues

    Our 174rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    In this episode of Last Week in AI, we delve into the latest advancements and challenges in the AI industry, highlighting new features from Figma and Quora, regulatory pressures on OpenAI, and significant investments in AI infrastructure. Key topics include AMD's acquisition of Silo AI, Elon Musk's GPU cluster plans for XAI, unique AI model training methods, and the nuances of AI copying and memory constraints. We discuss developments in AI's visual perception, real-time knowledge updates, and the need for transparency and regulation in AI content labeling and licensing.

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

     

    Timestamps + links:

    Last Week in AI
    enJuly 17, 2024

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    #173 - Gemini Pro, Llama 400B, Gen-3 Alpha, Moshi, Supreme Court

    Our 173rd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    See full episode notes here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    In this episode of Last Week in AI, we explore the latest advancements and debates in the AI field, including Google's release of Gemini 1.5, Meta's upcoming LLaMA 3, and Runway's Gen 3 Alpha video model. We discuss emerging AI features, legal disputes over data usage, and China's competition in AI. The conversation spans innovative research developments, cost considerations of AI architectures, and policy changes like the U.S. Supreme Court striking down Chevron deference. We also cover U.S. export controls on AI chips to China, workforce development in the semiconductor industry, and Bridgewater's new AI-driven financial fund, evaluating the broader financial and regulatory impacts of AI technologies.  

    Timestamps + links:

    Last Week in AI
    enJuly 07, 2024

    #172 - Claude and Gemini updates, Gemma 2, GPT-4 Critic

    #172 - Claude and Gemini updates, Gemma 2, GPT-4 Critic

    Our 172nd episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai

    Last Week in AI
    enJuly 01, 2024