Logo
    Search

    Podcast Summary

    • Learning AI through TikTok: A New Approach: AI education is evolving, with TikTok emerging as a popular platform for visual learning. Engaging videos and high user engagement make it an effective tool for introducing AI concepts to a wider audience.

      The AI community is expanding its presence on various social media platforms, including TikTok, and it's becoming an engaging and effective way for people to learn about AI and data science. Rajeev Shah, a machine learning engineer at Hugging Face, shared his experience of introducing his son to AI through TikTok videos, which opened his eyes to the potential of visual learning in this field. He also mentioned that he's seen high engagement with his own videos on the platform. This trend highlights the importance of adapting to different learning styles and meeting learners where they are, whether that's through written material, videos, or social media. It's an exciting time for AI education, and the versatility of learning methods is making it more accessible to a wider audience.

    • Exploring new teaching methods and platforms in data science: Adapt to new teaching methods and platforms like TikTok for engaging storytelling, prioritize practical skills, and embrace the current AI innovation wave.

      The way we teach and consume data science information is evolving rapidly, and it's important for educators and practitioners to adapt to new mediums and audiences. The speaker, who creates educational videos on data science, has seen great success with TikTok due to its engaging storytelling format and potential for nuanced conversations. However, not all developments in AI will impact every data scientist immediately, and it's essential to prioritize practical skills for those in enterprise roles. As for reaching different audiences, the speaker acknowledges the influence of technology on how students learn, such as the shift from typing to touchscreens. To stay relevant, we must be open to new teaching methods and platforms. The current period in AI innovation is unprecedented, and it's an exciting time for those interested in the field, but it won't last forever. Embrace the change and enjoy the ride.

    • In context learning with large language models: Large language models can learn from a few examples and continue to answer in the same context without further training, enabling valuable applications

      Data science and AI have become more accessible to a larger community, thanks to advancements in large language models. This transformation was once limited to those with a college education and access to expensive resources. However, now, even teenagers with a GPU can interact with AI, making it a widely engaging and popular field. One specific topic discussed at a conference was in context learning with large language models. In the past, language models could generate text based on statistical probabilities, but they were limited to making weird stories. However, as these models have grown larger and more data has been incorporated, machine learning engineers have noticed an emergent behavior, allowing for in context learning. In context learning, the model is given a few examples of a specific type of question, and it can then continue to answer in that same context without being trained or having its weights changed. For instance, imagine rating movie sentiments. In the traditional data science approach, one would have to label a large number of movies and train the model accordingly. With larger language models, one can give the model a few examples and ask it to rate a new movie review, which it can do without being trained further. This capability of in context learning is significant because it allows the model to understand and respond based on the given context without requiring additional training, making it a valuable tool in various applications.

    • Revolutionizing NLP tasks with prompt engineering: Effective prompting can enable users to accomplish various NLP tasks through a natural language interface, democratizing machine learning for a larger audience.

      The emergence of large language models and the development of prompt engineering have the potential to revolutionize various tasks in Natural Language Processing (NLP) by enabling users to accomplish them through a natural language interface. This new skill in prompting effectively can be applied to practical scenarios within enterprises, such as categorizing unstructured data, by asking the model to clean, summarize, and categorize the information using prompts. Tools like Langchain allow developers to tie together several prompts and create workflows, making it possible to accomplish multiple tasks without the need for separate machine learning models. This approach can lead to significant democratization of machine learning, allowing a larger audience to productively use this technology, including those without a deep understanding of data science concepts. The future of this technology looks promising, with the possibility of integrating it with other APIs and services to accomplish even more complex tasks.

    • Revolutionary moment in AI with large language models: Large language models like Hugging Face's GPT and Auto GPT can answer questions, perform tasks, and even start businesses. They vary in size and data used for training, impacting their reasoning ability and resource requirements.

      We are witnessing a revolutionary moment in artificial intelligence with the advancement of large language models, such as Hugging Face's GPT and Auto GPT, which can interconnect with various models to provide answers and perform tasks. These models can even start a business and find necessary databases or APIs. The landscape of large language models is diverse, with some being proprietary and others open source, like OpenAI and Databricks' DALL-E. The size of these models also varies greatly, from smaller models like the LLAMA models to larger ones like Hugging Face's Bloom, which has 170 billion parameters. The size of the model impacts its reasoning ability and the resources required for inference. Additionally, the data these models were trained on and the amount of code they were trained on are important factors to consider. These advancements, which were unimaginable a year ago, are shaping the future of AI and its applications.

    • Navigating the complex landscape of large language models for text-to-code projects: Start with pre-existing models, understand benefits and limitations, consider community and tooling, leverage Hugging Face's ecosystem, and adapt to the rapidly changing landscape.

      Making sense of the complex and ever-evolving landscape of large language models for text-to-code projects can be overwhelming for organizations. However, starting with a pre-existing model that already understands code and fine-tuning it is a more efficient approach. The characteristics of these models, including their size, licensing, and availability, can be confusing, with open-source and closed-source options available. For organizations looking to navigate this landscape, it's essential to understand the benefits and limitations of different models and consider the community and tooling surrounding them. Hugging Face's ecosystem, with its interoperability and integrated tooling, is a powerful resource for organizations practically implementing these models. Additionally, there's currently a lot of breathing room for enterprises to figure out the best strategy for implementing large language models over the next year. The pace of change can make understanding the landscape challenging, but the power and potential of these models make the effort worthwhile.

    • Making large language models more accessible with Hugging Face: Hugging Face's advancements in parameter efficient fine tuning and reinforcement learning with human feedback make large language models more accessible for fine tuning without loading the entire model or modifying every weight, potentially changing the approach to problem-solving in AI towards more model training for inference

      The Hugging Face ecosystem is leading the way in democratizing machine learning, particularly in the realm of large language models. These models, which can be huge and resource-intensive, are being made more accessible through advancements like parameter efficient fine tuning (Peft) and reinforcement learning with human feedback. This tooling allows for fine tuning without having to load the entire model or modify every weight, making it possible for more people to effectively use these models. This shift towards fine tuning may change the approach to problem-solving in AI, moving away from reliance on APIs and back towards training models, but with a more practical and resource-efficient focus. The Hugging Face ecosystem's infrastructure and tooling are making large language models more accessible, and this could lead to a resurgence in model training for inference, as opposed to relying solely on APIs.

    • Shifting focus towards practical applications and tools for large language models: Enterprises seek control and efficiency with large language models, leading to a focus on building tools for training and use with fewer resources, while addressing ethical and educational concerns.

      The excitement around OpenAI's APIs and large language models in enterprises is being driven by the need for control and efficient use of these models. While there has been a surge in open source developments, there is also a growing focus on building tools to help train and use these models with fewer resources. This shift is due to the desire for models that can fit inside computers or a few GPUs, and can be controlled and tuned by users. However, there are also significant challenges that come with large language models, such as the need for education around their potential for generating inaccurate or hallucinatory output, and the importance of understanding the training data and its origins. In enterprise settings, it's crucial to pair these models with factually-based information retrieval techniques to ensure accuracy and reliability. Overall, the conversation around AI models and large language models is shifting towards practical applications and the development of tools, while also addressing ethical and educational concerns.

    • Navigating legal implications and potential misuse of large language models in education: Ensure equal access, recognize benefits, provide education and training, and foster a collaborative approach to responsibly integrate large language models into education

      As we navigate the integration of large language models into various aspects of our lives, including education, it's crucial to address concerns around legal implications, intellectual property, and potential misuse. These models, which can be incredibly useful and productive, also come with risks such as data leakage and the possibility of skewing academic progress. Pragmatism is key in addressing these challenges. While there will be short-term issues to resolve, such as ensuring equal access to the technology, it's important to recognize the potential benefits for students and coworkers alike. Education and training will be necessary to help people effectively use these tools and understand their limitations. Moreover, the developer community's involvement in this space is promising, as it brings a wider range of perspectives and innovative applications to the table. Ultimately, a collaborative and thoughtful approach will be essential in ensuring that large language models are used responsibly and effectively in education and beyond.

    • Exploring the Rapid Transformation of Our Lives by Technology: Experience AI tools like image generation or chatbots to understand the impact of technology on our lives. The short term brings continued innovation, while the long term holds vast potential for societal change.

      We're living in an unprecedented time where technology, particularly AI, is rapidly transforming various aspects of our lives. The energy around startups and new innovations is palpable, even if most will fail. For those not yet familiar with these advancements, the best way to understand is by experiencing it firsthand. Using tools like image generation or chatbots can help bridge the gap and provide a foundation for exploring the possibilities and limitations. This historical moment is not just significant for the tech industry, but for the world as a whole. In the short term, we can expect continued innovation and breakthroughs. In the long term, the impact on society and our daily lives is yet to be fully realized, but the potential is vast and exciting.

    • Exploring the Future Impact of AI on Humanity: The Hugging Face community is democratizing AI, allowing people without extensive stats training to build code and solve problems using AI. New models and tooling will be released in the coming months, and the potential for AI to enhance productivity and problem-solving capabilities is great, but its impact on larger society must be considered.

      We are at an exciting moment in time where the use of AI is becoming more widespread and accessible to a larger audience. The conversation between the hosts and the representative from Hugging Face touched on the potential future impact of AI on humanity and the importance of figuring out how to integrate it into our daily lives. The Hugging Face community is leading the charge in democratizing AI, making it possible for people without extensive statistics training to build code and solve problems using AI. The representative encouraged listeners to check out the Hugging Face website for resources, including a free online course and community forums, and shared that new models and tooling will be released in the coming months. Overall, the discussion highlighted the potential for AI to greatly enhance productivity and problem-solving capabilities, but also emphasized the importance of considering its impact on larger society.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Stanford's AI Index Report 2024

    Stanford's AI Index Report 2024
    We’ve had representatives from Stanford’s Institute for Human-Centered Artificial Intelligence (HAI) on the show in the past, but we were super excited to talk through their 2024 AI Index Report after such a crazy year in AI! Nestor from HAI joins us in this episode to talk about some of the main takeaways including how AI makes workers more productive, the US is increasing regulations sharply, and industry continues to dominate frontier AI research.

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    First impressions of GPT-4o

    First impressions of GPT-4o
    Daniel & Chris share their first impressions of OpenAI’s newest LLM: GPT-4o and Daniel tries to bring the model into the conversation with humorously mixed results. Together, they explore the implications of Omni’s new feature set - the speed, the voice interface, and the new multimodal capabilities.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.