Logo
    Search

    The Data Foundry for AI with Alexandr Wang from Scale

    enMay 22, 2024

    Podcast Summary

    • From Overlooked Data Pillar to Essential AI ComponentScale AI, founded by Alex Wang, emerged as a key player in the AI industry by focusing on the overlooked data pillar and became an essential component of the AI ecosystem by serving as the data backbone for major AI efforts from companies like OpenAI, Meta, and Microsoft.

      Scale AI, founded by Alex Wang in 2016, emerged as a key player in the AI industry by focusing on the data pillar, which was overlooked at the time. Alex recognized that while research labs and companies were working on algorithms and compute, the importance of data would only grow. Starting in his house, Alex dropped out of MIT and went through Y Combinator to build Scale as the data foundry for AI. Initially, the company catered to the autonomous vehicle industry's data needs. However, with AI's versatility as a general-purpose technology, Scale's business has evolved to provide the data infrastructure for a wide range of AI applications. By serving as the data backbone for major AI efforts from companies like OpenAI, Meta, and Microsoft, Scale has become an essential component of the AI ecosystem.

    • Exploring emerging AI use cases for business successAdapting to emerging AI trends and exploring new applications is crucial for business success.

      Anticipating and preparing for the emerging use cases of Artificial Intelligence (AI) is crucial for businesses. When the company was founded in 2016, the focus was on autonomous driving, which became an industry standard with the support of sensor fused data. However, by 2019, the future of AI applications was uncertain, leading the company to explore government applications, specifically geospatial and satellite data, which resulted in the first AI program of record for the US DOD. Around the same time, the company also started working on generative AI with OpenAI, which later became a significant trend in the industry. Today, the company's data foundry fuels the development of major large language models in the industry, including those of OpenAI, Meta, and Microsoft. The company's ability to adapt and focus on emerging AI trends has been essential to its success.

    • Ensuring Data Abundance for AI DevelopmentTo effectively scale large language models, we need high-quality, frontier data for AI to learn effectively, including reasoning chain of thoughts from experts, agent workflows, and multilingual data. The challenge is to produce and collect such data at a large scale.

      We are witnessing an exciting time in the development of AI technology, where it is increasingly being seen as a general-purpose technology with a wide range of business applications. This shift is evident in the emergence of various new types of users, including technology giants, governments, automotive companies, enterprises, and even sovereign AI. The infrastructure required to support this broad industry is substantial, and a major focus is on ensuring data abundance to scale large language models. We have moved beyond easily accessible data on the internet and now need high-quality, frontier data for these models to learn effectively. This includes reasoning chain of thoughts from experts, agent workflows, and multilingual data. The challenge is to produce and collect such data at a large scale. The future of AI development hinges on our ability to ensure data abundance, and our goal is to build the infrastructure to support this need.

    • Producing high-quality data for AI systemsExperts in various fields contribute valuable data for AI systems, small improvements lead to significant impact, and data production is crucial for progress and innovation, despite being an often overlooked resource.

      The production of high-quality data for AI systems is crucial for the future development of the industry. This involves integrating various types of data, including video, audio, and esoteric data, and requires the contributions of experts in various fields. The impact of producing high-quality data can be significant, as even small improvements to models can have a large impact when applied at scale. This is similar to Google's original mission to organize and make accessible all the world's information. The motivation for experts to contribute is not only monetary but also the meaningful impact they can have on fueling the AI movement and progressing human knowledge. A challenge in this field is the lack of adequate data capture for AI use cases, making it a crucial area of focus for companies and investors. In essence, data production for AI is akin to spice production in the game Doom 2, where it is a valuable yet often overlooked resource essential for driving progress and innovation.

    • Combining human talent and proprietary data for powerful AI systemsHuman expertise and machine intelligence will continue to complement each other in creating powerful AI systems, with vast enterprise and government data playing a crucial role.

      The future of AI systems relies on a combination of the best human talent and proprietary data for training. JPMorgan's proprietary dataset, which is 150 petabytes, is just one example of the vast amount of data that exists within enterprises and governments that can be used to create powerful AI systems. However, the question of synthetic data and its role in the future of AI is still open. Hybrid human-AI systems that allow AI to do the heavy lifting while human experts contribute their insight and reasoning capabilities to produce high-quality data are the key to the future. Even when models outperform humans on many dimensions, the combination of human and machine intelligence will continue to produce better results than either alone. The unique qualities and attributes of human intelligence, which are distinct from machine intelligence, will continue to complement each other in practice. When a model produces an answer or response, humans can critique it, identify factual or reasoning errors, and guide the model over a long period to produce correct and deep reasoning chains. This is the focus of our work.

    • Human expertise plays a crucial role in pushing AI capabilities forwardDespite AI's advancements, human intelligence's ability to reason and optimize over long time horizons is still essential. Recent $14B fundraise aims to build an ecosystem around the data foundry to support the future of AI technology.

      While AI models are advancing rapidly and can outperform humans in certain tasks, they still lack the ability to reason and optimize over long time horizons that human intelligence possesses. Human expertise will continue to play a crucial role in pushing the boundaries of AI capabilities, and the fundamental quality of biological intelligence will only be taught to models over time through data transfer. The recent fundraise by the company, valued at $14 billion, aims to serve the entire AI ecosystem and build an ecosystem around their data foundry, bringing along infrastructure providers and key industry players to support the future of the technology.

    • Evaluating and Measuring AI Systems: Ensuring Trust and Responsible DevelopmentEnsuring data quality and measuring AI systems is crucial for building trust in these technologies. Ethical considerations, such as safety, security, and consumer protection, are important for responsible development. Ethals is working on new methods for evaluating AI systems to enable their responsible development and adoption.

      Data is a crucial element in the development and adoption of advanced AI systems, and ensuring data quality and measurement of AI systems is essential for building trust in these technologies. The goal is to leverage data capabilities to empower every layer of the AI stack and invest in data production to enable the technology's growth. However, the evaluation and measurement of AI systems presents challenges, such as the difficulty of automatically grading these systems and the limitations of current academic benchmarks. Ethics and responsibility are also important considerations, as trust in AI systems is necessary for their broader societal adoption. Ethical questions around safety, security, and consumer protection must be addressed. Ethals, a company mentioned in the discussion, is working on developing methods to evaluate AI systems and ensuring their responsible development. The process involves ensuring data abundance and quality, measuring AI systems, and continuously improving them. The evaluation and measurement of AI systems is a critical component of the AI life cycle and is essential for building trust in these technologies. Additionally, the philosophical question of how to measure the intelligence of a system adds complexity to the evaluation process. Ethals is addressing this challenge by developing new methods for evaluating AI systems, which is a significant step towards enabling the responsible development and adoption of advanced AI technologies.

    • Expert evaluations crucial for understanding AI capabilitiesPublic visibility and transparency are essential for safe development and deployment of AI models. Upcoming models will be more powerful, leading to innovative applications. Application builders must focus on self-improvement for future AI development.

      Evaluating the true capabilities of AI models goes beyond just relying on reported performance and requires expert evaluations to understand strengths, weaknesses, and associated risks. Public visibility and transparency through leaderboards and constant evaluations are crucial to ensure safe development and deployment of these models. The current state of AI at the application layer is undergoing a transition, with earlier models like GPD-4 being considered premature for widespread application due to their limitations. However, upcoming models are expected to be significantly more powerful, leading to a surge in innovative applications. Empowering application builders, whether they be enterprises, governments, or startups, to incorporate self-improvement into their applications is a key focus for the future of AI development.

    • Narrowing down data for better model performanceEnterprises need to focus on high-quality data to improve model performance, as not all data is valuable. Platforms like Scale are launching evaluations and leaderboards to ensure consistent progress and improvement in LLMs.

      Enterprises and organizations, regardless of size or industry, need to focus on building applications with self-improving loops to effectively utilize and enhance their data for better model performance. However, not all data is created equal, and filtering down to high-quality data is crucial. JPMorgan, for instance, has a massive amount of data, but not all of it is valuable. Meta's research shows that narrowing the data used can lead to better models, making it essential to identify the data that truly improves the model. Scale, a leading AI platform, is addressing these challenges by launching private held-out evaluations and leaderboards for leading Language Models (LLMs) to consistently benchmark and monitor their performance. They will begin with areas like math, coding, instruction following, and adversarial capabilities, with plans to expand to more domains over time. This creates an "Olympics for LLMs" where models are evaluated every few months, ensuring consistent progress and improvement. Additionally, Scale is collaborating with government customers to leverage LLMs' current capabilities, which can be valuable in various applications.

    • Advancements in AI: Multimodality and Personal AgentsTech companies like Scale and Google are pushing AI forward with multimodal applications, from simple tasks to complex enterprise solutions. However, the lack of good data hampers progress, and competition between similar tech developments signals industry convergence.

      Companies like Scale and tech giants like Google are making significant strides in developing AI applications, particularly in the areas of multimodality and personal agent use cases. These applications can range from mundane tasks like report writing and data transfer to more complex, enterprise-level solutions. However, there is a scarcity of good multimodal data required to fuel these improvements, making data acquisition a key focus. Additionally, both companies are independently developing similar technologies, indicating a convergence in the industry and a shared vision for the future of AI. While there may be competition involved, it's also a sign of the industry's progress and the recognition of multimodality as a major area of development. The industry is also in need of smarter models, like GP5 or Gemini 2, to continue making advancements.

    • The path to AGI is a complex process with numerous individual problems to solveThe development of AGI is more likely to be a gradual process with separate data flywheels for each area of capability

      The path to Artificial General Intelligence (AGI) is likely to be a long, complex process involving the solution of numerous individual problems, rather than a single breakthrough. This perspective contrasts with the belief held by some in the industry that AGI will be achieved through a single, monumental discovery. The speaker draws an analogy between the development of AGI and the process of curing cancer, emphasizing the need to tackle each problem individually before making significant progress towards the ultimate goal. This view has significant implications for how we approach the technology and its societal impact, as it suggests a consistent, gradual progression that will allow society to adapt. Furthermore, the speaker emphasizes the limited generality of current models, suggesting that each area of capability will require separate data flywheels to drive performance. As a CEO leading a scaling organization, the speaker is acutely aware of how early we are in this technology and the importance of a long-term perspective.

    • Staying Nimble in the Early Stages of Emerging TechnologiesThe early stages of emerging technologies offer many opportunities for organizations to adapt and stay ahead. Despite heavy investments and frequent launches, the capabilities of these technologies are only a fraction of what they will be in the future. Organizations must remain nimble to keep up with the rapid advancements and continue learning to succeed.

      While the market for emerging technologies may appear crowded due to heavy investments from tech giants and frequent launches, we are still in the early stages. The technology's capabilities are currently only a fraction of what they will be in the future. Therefore, it's crucial for organizations to remain nimble and adapt alongside the technology's development. This is an exciting time as there are many more chapters to be written in this technological journey. So, stay tuned for new developments and continue to learn and adapt. You can find us on Twitter at @no_priors_pod, subscribe to our YouTube channel, and listen on Apple Podcasts, Spotify, or wherever you prefer. Don't forget to sign up for emails or view transcripts on our website at nodashpriors.com for every episode.

    Recent Episodes from No Priors: Artificial Intelligence | Machine Learning | Technology | Startups

    State Space Models and Real-time Intelligence with Karan Goel and Albert Gu from Cartesia

    State Space Models and Real-time Intelligence with Karan Goel and Albert Gu from Cartesia
    This week on No Priors, Sarah Guo and Elad Gil sit down with Karan Goel and Albert Gu from Cartesia. Karan and Albert first met as Stanford AI Lab PhDs, where their lab invented Space Models or SSMs, a fundamental new primitive for training large-scale foundation models. In 2023, they Founded Cartesia to build real-time intelligence for every device. One year later, Cartesia released Sonic which generates high quality and lifelike speech with a model latency of 135ms—the fastest for a model of this class. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @krandiash | @_albertgu Show Notes:  (0:00) Introduction (0:28) Use Cases for Cartesia and Sonic  (1:32) Karan Goel & Albert Gu’s professional backgrounds (5:06) Steady State Models (SSMs) versus Transformer Based Architectures  (11:51) Domain Applications for Hybrid Approaches  (13:10) Text to Speech and Voice (17:29) Data, Size of Models and Efficiency  (20:34) Recent Launch of Text to Speech Product (25:01) Multimodality & Building Blocks (25:54) What’s Next at Cartesia?  (28:28) Latency in Text to Speech (29:30) Choosing Research Problems Based on Aesthetic  (31:23) Product Demo (32:48) Cartesia Team & Hiring

    Can AI replace the camera? with Joshua Xu from HeyGen

    Can AI replace the camera? with Joshua Xu from HeyGen
    AI video generation models still have a long way to go when it comes to making compelling and complex videos but the HeyGen team are well on their way to streamlining the video creation process by using a combination of language, video, and voice models to create videos featuring personalized avatars, b-roll, and dialogue. This week on No Priors, Joshua Xu the co-founder and CEO of HeyGen,  joins Sarah and Elad to discuss how the HeyGen team broke down the elements of a video and built or found models to use for each one, the commercial applications for these AI videos, and how they’re safeguarding against deep fakes.  Links from episode: HeyGen McDonald’s commercial Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil |  @joshua_xu_ Show Notes:  (0:00) Introduction (3:08) Applications of AI content creation (5:49) Best use cases for Hey Gen (7:34) Building for quality in AI video generation (11:17) The models powering HeyGen (14:49) Research approach (16:39) Safeguarding against deep fakes (18:31) How AI video generation will change video creation (24:02) Challenges in building the model (26:29) HeyGen team and company

    How the ARC Prize is democratizing the race to AGI with Mike Knoop from Zapier

    How the ARC Prize is democratizing  the race to AGI with Mike Knoop from Zapier
    The first step in achieving AGI is nailing down a concise definition and  Mike Knoop, the co-founder and Head of AI at Zapier, believes François Chollet got it right when he defined general intelligence as a system that can efficiently acquire new skills. This week on No Priors, Miked joins Elad to discuss ARC Prize which is a multi-million dollar non-profit public challenge that is looking for someone to beat the Abstraction and Reasoning Corpus (ARC) evaluation. In this episode, they also get into why Mike thinks LLMs will not get us to AGI, how Zapier is incorporating AI into their products and the power of agents, and why it’s dangerous to regulate AGI before discovering its full potential.  Show Links: About the Abstraction and Reasoning Corpus Zapier Central Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @mikeknoop Show Notes:  (0:00) Introduction (1:10) Redefining AGI (2:16) Introducing ARC Prize (3:08) Definition of AGI (5:14) LLMs and AGI (8:20) Promising techniques to developing AGI (11:0) Sentience and intelligence (13:51) Prize model vs investing (16:28) Zapier AI innovations (19:08) Economic value of agents (21:48) Open source to achieve AGI (24:20) Regulating AI and AGI

    The evolution and promise of RAG architecture with Tengyu Ma from Voyage AI

    The evolution and promise of RAG architecture with Tengyu Ma from Voyage AI
    After Tengyu Ma spent years at Stanford researching AI optimization, embedding models, and transformers, he took a break from academia to start Voyage AI which allows enterprise customers to have the most accurate retrieval possible through the most useful foundational data. Tengyu joins Sarah on this week’s episode of No priors to discuss why RAG systems are winning as the dominant architecture in enterprise and the evolution of foundational data that has allowed RAG to flourish. And while fine-tuning is still in the conversation, Tengyu argues that RAG will continue to evolve as the cheapest, quickest, and most accurate system for data retrieval.  They also discuss methods for growing context windows and managing latency budgets, how Tengyu’s research has informed his work at Voyage, and the role academia should play as AI grows as an industry.  Show Links: Tengyu Ma Key Research Papers: Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training Non-convex optimization for machine learning: design, analysis, and understanding Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss Larger language models do in-context learning differently, 2023 Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning On the Optimization Landscape of Tensor Decompositions Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @tengyuma Show Notes:  (0:00) Introduction (1:59) Key points of Tengyu’s research (4:28) Academia compared to industry (6:46) Voyage AI overview (9:44) Enterprise RAG use cases (15:23) LLM long-term memory and token limitations (18:03) Agent chaining and data management (22:01) Improving enterprise RAG  (25:44) Latency budgets (27:48) Advice for building RAG systems (31:06) Learnings as an AI founder (32:55) The role of academia in AI

    How YC fosters AI Innovation with Garry Tan

    How YC fosters AI Innovation with Garry Tan
    Garry Tan is a notorious founder-turned-investor who is now running one of the most prestigious accelerators in the world, Y Combinator. As the president and CEO of YC, Garry has been credited with reinvigorating the program. On this week’s episode of No Priors, Sarah, Elad, and Garry discuss the shifting demographics of YC founders and how AI is encouraging younger founders to launch companies, predicting which early stage startups will have longevity, and making YC a beacon for innovation in AI companies. They also discussed the importance of building companies in person and if San Francisco is, in fact, back.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @garrytan Show Notes:  (0:00) Introduction (0:53) Transitioning from founder to investing (5:10) Early social media startups (7:50) Trend predicting at YC (10:03) Selecting YC founders (12:06) AI trends emerging in YC batch (18:34) Motivating culture at YC (20:39) Choosing the startups with longevity (24:01) Shifting YC found demographics (29:24) Building in San Francisco  (31:01) Making YC a beacon for creators (33:17) Garry Tan is bringing San Francisco back

    The Data Foundry for AI with Alexandr Wang from Scale

    The Data Foundry for AI with Alexandr Wang from Scale
    Alexandr Wang was 19 when he realized that gathering data will be crucial as AI becomes more prevalent, so he dropped out of MIT and started Scale AI. This week on No Priors, Alexandr joins Sarah and Elad to discuss how Scale is providing infrastructure and building a robust data foundry that is crucial to the future of AI. While the company started working with autonomous vehicles, they’ve expanded by partnering with research labs and even the U.S. government.   In this episode, they get into the importance of data quality in building trust in AI systems and a possible future where we can build better self-improvement loops, AI in the enterprise, and where human and AI intelligence will work together to produce better outcomes.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @alexandr_wang (0:00) Introduction (3:01) Data infrastructure for autonomous vehicles (5:51) Data abundance and organization (12:06)  Data quality and collection (15:34) The role of human expertise (20:18) Building trust in AI systems (23:28) Evaluating AI models (29:59) AI and government contracts (32:21) Multi-modality and scaling challenges

    Music consumers are becoming the creators with Suno CEO Mikey Shulman

    Music consumers are becoming the creators with Suno CEO Mikey Shulman
    Mikey Shulman, the CEO and co-founder of Suno, can see a future where the Venn diagram of music creators and consumers becomes one big circle. The AI music generation tool trying to democratize music has been making waves in the AI community ever since they came out of stealth mode last year. Suno users can make a song complete with lyrics, just by entering a text prompt, for example, “koto boom bap lofi intricate beats.” You can hear it in action as Mikey, Sarah, and Elad create a song live in this episode.  In this episode, Elad, Sarah, And Mikey talk about how the Suno team took their experience making at transcription tool and applied it to music generation, how the Suno team evaluates aesthetics and taste because there is no standardized test you can give an AI model for music, and why Mikey doesn’t think AI-generated music will affect people’s consumption of human made music.  Listen to the full songs played and created in this episode: Whispers of Sakura Stone  Statistical Paradise Statistical Paradise 2 Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @MikeyShulman Show Notes:  (0:00) Mikey’s background (3:48) Bark and music generation (5:33) Architecture for music generation AI (6:57) Assessing music quality (8:20) Mikey’s music background as an asset (10:02) Challenges in generative music AI (11:30) Business model (14:38) Surprising use cases of Suno (18:43) Creating a song on Suno live (21:44) Ratio of creators to consumers (25:00) The digitization of music (27:20) Mikey’s favorite song on Suno (29:35) Suno is hiring

    Context windows, computer constraints, and energy consumption with Sarah and Elad

    Context windows, computer constraints, and energy consumption with Sarah and Elad
    This week on No Priors hosts, Sarah and Elad are catching up on the latest AI news. They discuss the recent developments in AI music generation, and if you’re interested in generative AI music, stay tuned for next week’s interview! Sarah and Elad also get into device-resident models, AI hardware, and ask just how smart smaller models can really get. These hardware constraints were compared to the hurdles AI platforms are continuing to face including computing constraints, energy consumption, context windows, and how to best integrate these products in apps that users are familiar with.  Have a question for our next host-only episode or feedback for our team? Reach out to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil  Show Notes:  (0:00) Intro (1:25) Music AI generation (4:02) Apple’s LLM (11:39) The role of AI-specific hardware (15:25) AI platform updates (18:01) Forward thinking in investing in AI (20:33) Unlimited context (23:03) Energy constraints

    Cognition’s Scott Wu on how Devin, the AI software engineer, will work for you

    Cognition’s Scott Wu on how Devin, the AI software engineer, will work for you
    Scott Wu loves code. He grew up competing in the International Olympiad in Informatics (IOI) and is a world class coder, and now he's building an AI agent designed to create more, not fewer, human engineers. This week on No Priors, Sarah and Elad talk to Scott, the co-founder and CEO of Cognition, an AI lab focusing on reasoning. Recently, the Cognition team released a demo of Devin, an AI software engineer that can increasingly handle entire tasks end to end. In this episode, they talk about why the team built Devin with a UI that mimics looking over another engineer’s shoulder as they work and how this transparency makes for a better result. Scott discusses why he thinks Devin will make it possible for there to be more human engineers in the world, and what will be important for software engineers to focus on as these roles evolve. They also get into how Scott thinks about building the Cognition team and that they’re just getting started.  Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @ScottWu46 Show Notes:  (0:00) Introduction (1:12) IOI training and community (6:39) Cognition’s founding team (8:20) Meet Devin (9:17) The discourse around Devin (12:14) Building Devin’s UI (14:28) Devin’s strengths and weakness  (18:44) The evolution of coding agents (22:43) Tips for human engineers (26:48) Hiring at Cognition

    OpenAI’s Sora team thinks we’ve only seen the "GPT-1 of video models"

    OpenAI’s Sora team thinks we’ve only seen the "GPT-1 of video models"
    AI-generated videos are not just leveled-up image generators. But rather, they could be a big step forward on the path to AGI. This week on No Priors, the team from Sora is here to discuss OpenAI’s recently announced generative video model, which can take a text prompt and create realistic, visually coherent, high-definition clips that are up to a minute long. Sora team leads, Aditya Ramesh, Tim Brooks, and Bill Peebles join Elad and Sarah to talk about developing Sora. The generative video model isn’t yet available for public use but the examples of its work are very impressive. However, they believe we’re still in the GPT-1 era of AI video models and are focused on a slow rollout to ensure the model is in the best place possible to offer value to the user and more importantly they’ve applied all the safety measures possible to avoid deep fakes and misinformation. They also discuss what they’re learning from implementing diffusion transformers, why they believe video generation is taking us one step closer to AGI, and why entertainment may not be the main use case for this tool in the future.  Show Links: Bling Zoo video Man eating a burger video Tokyo Walk video Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @_tim_brooks l @billpeeb l @model_mechanic Show Notes:  (0:00) Sora team Introduction (1:05) Simulating the world with Sora (2:25) Building the most valuable consumer product (5:50) Alternative use cases and simulation capabilities (8:41) Diffusion transformers explanation (10:15) Scaling laws for video (13:08) Applying end-to-end deep learning to video (15:30) Tuning the visual aesthetic of Sora (17:08) The road to “desktop Pixar” for everyone (20:12) Safety for visual models (22:34) Limitations of Sora (25:04) Learning from how Sora is learning (29:32) The biggest misconceptions about video models