Logo
    Search

    Podcast Summary

    • Mervha's journey to Hugging Face in NLP: Mervha, a developer advocate at Hugging Face, shares her background in NLP and how she became involved with the organization, emphasizing its role in making machine learning more accessible and practical.

      Hugging Face, a leading organization in the field of open source machine learning, is focused on solving key issues in the realm of machine learning, such as reproducibility and ease of use for various applications. Mervha, a developer advocate engineer at Hugging Face, shared her background in NLP and how she became involved with the organization. She was initially introduced to NLP during her senior year of university and became passionate about it after working on a text mining project. Later, she joined bootcamps and earned a master's degree, eventually working as a machine learning engineer focusing on NLP. During this time, she used Hugging Face in her projects. What caught her attention about Hugging Face was a video by Thomas Wolf discussing the future of NLP and an opportunity to participate in a community sprint on datasets. Merva's experiences highlight the importance of Hugging Face in the machine learning community, particularly in the NLP domain, and its role in making machine learning more accessible and practical for a wide range of use cases.

    • Learning through open source sprints: Beginners can learn and contribute to open source projects by participating in sprints, where they'll receive guidance and actively engage with the community. Understanding the unique challenges of the projects, like text classification for chatbots, can help guide learning and contributions.

      Contributing to open source projects, particularly in data science and machine learning communities, can be an intimidating experience for beginners. However, participating in sprints, where contributors actively engage and guide newcomers, can be an excellent way to start. The speaker shared their personal experience of joining Hugging Face and learning valuable skills like CICD styling, formatting, and contributing to open source through the audio sprint. They emphasized the importance of being aware of community events and sprints, which provide opportunities for active collaboration and learning. The speaker also mentioned their background in building chatbots and how it influenced their perception of NLP tooling needs. They noted that the process of building a chatbot, especially for narrow domains, often involves solving text classification problems. This experience taught them the importance of improving data and iterating on models. In summary, sprints and actively engaging in open source projects can be an effective way for beginners to learn and contribute to the data science and machine learning community. Additionally, understanding the specific challenges of the projects you're interested in, like text classification for chatbots, can help guide your learning and contributions.

    • Creating a conversational agent from scratch: challenges and rewards: Transitioning from math to chatbot development required overcoming challenges in creating conversational agents, adapting to high coding standards, and understanding different approaches to building bots.

      Transitioning from a background in mathematics and operations research to developing a chatbot involved significant challenges. The speaker described the difficulty of creating a conversational agent from scratch, particularly one that could understand and respond appropriately to user queries. They noted that conversational agents can be divided into two categories: those based on generative models, which can discuss any topic, and those based on intent and action, which require defining specific actions for each intent. The speaker found the latter approach to be more challenging due to the need to write extensive training data and the potential biases of language models. Despite these challenges, the speaker was impressed by the use of tools like Hugging Face's sentence transformers and the Pipeline function, which made complex tasks simpler through abstraction. When transitioning into this new skill set, the hardest part for the speaker was adapting to the high standards of code quality and productivity expected in the role. Previously, their code had been functional but not neatly organized or efficient. However, they were determined to improve and became more focused on writing clean, effective code. Overall, the experience of developing a chatbot from scratch was both rewarding and daunting, but the speaker was excited to continue learning and growing in this new field.

    • Leveraging UX design and code improvement in machine learning tools: Hugging Face prioritizes UX design and continuous code improvement to make machine learning models universally accessible and usable for diverse applications, with the Hugging Face Hub as a key component for model accessibility and collaboration.

      Working at Hugging Face has provided valuable insights into the importance of UX design and continuous code improvement in the development of machine learning tools and models. The Hugging Face ecosystem is focused on solving the challenges of open source machine learning, including reproducibility, ease of use, and model accessibility. The Hugging Face Hub is a crucial component, allowing users to declare model limitations and biases, and hosting various libraries such as transformers, Keras, and Stanford NLP models. The ultimate goal is to make machine learning models universally accessible and usable for diverse applications. The speaker's role involves developing tools and creating demos to showcase the capabilities of these libraries. They have realized the significance of UX design and the ongoing process of improving code through collaboration and feedback from colleagues.

    • Effortlessly share and demonstrate machine learning projects with Hugging Face's Spaces and related tools: Streamline sharing of ML projects using Hugging Face's user-friendly tools like Spaces, Streamlit, and Gradio, benefiting beginners and experienced users alike.

      Hugging Face's Spaces and related tools like Streamlit and Gradio offer significant solutions to common pain points for data scientists and machine learning engineers. These tools allow for easy sharing and demonstration of projects without the need for extensive development background or setting up complex environments. The user-friendly interfaces enable users to simply drag and drop their code into Spaces and share the link with clients, teachers, or colleagues. This is particularly beneficial for those in startups or academic settings where time and resources are limited. As the Hugging Face ecosystem continues to grow, it's crucial for experienced users to remember the challenges faced by beginners and communicate effectively to help them navigate the tools and become productive members of the community. Additionally, Hugging Face's recent investment in tabular data further expands the platform's appeal to a wider range of data scientists. Overall, these tools streamline the process of sharing and showcasing machine learning projects, making it an essential resource for the data science community.

    • Explore, train, and share machine learning projects with Hugging Face: Hugging Face simplifies machine learning and data science projects by offering a platform for hosting datasets, training models, and sharing results, promoting reproducibility and collaboration.

      Hugging Face provides a platform for machine learning and data science projects, allowing users to host large datasets, perform exploratory data analysis, train models, and push them to the Hugging Face Hub for collaboration and sharing. This platform caters to various roles, such as machine learning engineers and data scientists, and types of problems, including end-to-end projects and NLP tasks. Hugging Face simplifies the process by offering features like automated model cards, TensorBoard log hosting, and inference widgets, all with minimal code. The primary goal is to promote reproducibility and collaboration in machine learning and data science projects. For NLP tasks, users can train models, push them to the Hugging Face Hub, and build demos using tools like Gradio and its pipe functionality. Overall, Hugging Face offers significant time-saving abstractions for developers, making it an essential resource for the machine learning and data science community.

    • Explore machine learning models with Hugging Face Hub: Hugging Face Hub simplifies machine learning model exploration for developers, offering filtering, assessment, and new collaborative features for improved productivity and adaptability.

      Hugging Face Hub offers software developers an accessible way to explore and utilize machine learning models without needing extensive machine learning knowledge. This platform allows developers to filter models based on their use case, call pipelines or inference APIs, and assess model metrics to determine suitability. The team behind Hugging Face has developed this user-friendly solution, which was released around January or February, to enable software developers to build various products without the need to learn machine learning from scratch. Recently, Hugging Face announced new collaborative features on the hub, including pull requests and community features. These new functionalities allow users to open pull requests on model repositories, which can contain model files, configuration, tokenizers, or application files. This way, people can improve each other's work, similar to how GitHub operates. Hugging Face is focusing on models and infrastructure, and by implementing pull requests, they aim to reduce duplicate work and make it easier for developers to adapt models to different frameworks, such as TensorFlow to PyTorch. Overall, these collaborative features enhance the hub's capabilities and open up new possibilities for the future of machine learning development.

    • Enhancing collaboration and community engagement in machine learning: Integration of discussion sections in machine learning platforms like Hugging Face Spaces fosters a more collaborative and open source-like environment, enabling users to communicate, contribute, and learn from each other, leading to faster innovation and progress in the field.

      The integration of discussion sections in machine learning platforms like Hugging Face Spaces has the potential to significantly enhance collaboration and community engagement within the field of machine learning. This feature allows users to easily communicate with each other about improvements, optimizations, and issues related to specific models or datasets. It enables users to open pull requests, contribute changes, and discuss potential enhancements, fostering a more collaborative and open source-like environment. This social aspect could potentially lead to a massive shift in the machine learning community, similar to the impact GitHub had on the open source world. Users can now follow each other's work, learn from each other, and work together to improve models and datasets, ultimately leading to faster innovation and progress in the field.

    • Hugging Face Hub: Enhancing Collaboration and Optimization in Open Source Machine Learning: The Hugging Face Hub is improving collaboration and optimization in open source machine learning through features like model testing in widgets, evaluation tools, and addressing dataset quality issues.

      The Hugging Face Hub is focusing on improving collaboration and optimization in the open source machine learning community. One way they are doing this is by allowing users to test models in widgets or spaces before cloning the entire repository. They are also investing in evaluation features, such as model cards and leaderboards, to help users identify the best models for their tasks. Additionally, they are addressing issues with dataset quality and encouraging contributions to improve it. The team is constantly evolving and looking for ways to enhance the user experience. A recent example of this is the launch of a feature that allows users to open discussions about dataset quality directly from Twitter. The Hugging Face team is dedicated to promoting ethical practices in the machine learning community and addressing issues as they arise. Overall, the Hugging Face Hub is committed to making open source machine learning more accessible and effective for all users.

    • Hugging Face prioritizes ethical considerations and transparency: Hugging Face emphasizes ethical use of machine learning models, acknowledges potential biases, cares about data privacy, and actively works on projects to improve model production and transparency.

      Hugging Face prioritizes ethical considerations and transparency in their use of machine learning models. They stress the importance of acknowledging potential biases and making declarations, as well as caring about data privacy and ethical restrictions. The team is actively working on various projects within the Hugging Face ecosystem, including improving the production capabilities of scikit-learn and creating model cards for various types of models to provide insight into what the model has learned and how it can be used. The ultimate goal is to make it easy for users to push their models onto the Hugging Face Hub and generate model cards with various information, including visualizations and feature importance. Additionally, they are working on extending Gradio to support tabular data, making it a powerful tool for data analysis and visualization.

    • Exploring the Power of Hugging Face Libraries: Hugging Face libraries like Pandas Profiling and Denim enable efficient data profiling and model training, and their community collaboration leads to innovative tools like auto EDA and AutoML.

      Hugging Face is leading the way in making advanced data analysis and machine learning techniques more accessible through their library and community collaboration. The speaker, Merva, shared her experience of discovering the power of libraries like Pandas Profiling and Denim, which enable efficient data profiling and model training. She is now building tools on the Hugging Face Hub, including an auto EDA (exploratory data analysis) and an AutoML (automated machine learning) tool, which will save time and lower the barrier of entry for data scientists. Merva expressed her admiration for Hugging Face and their impact on the AI community, and the conversation highlighted the importance of community-driven innovation in advancing AI technology.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Stanford's AI Index Report 2024

    Stanford's AI Index Report 2024
    We’ve had representatives from Stanford’s Institute for Human-Centered Artificial Intelligence (HAI) on the show in the past, but we were super excited to talk through their 2024 AI Index Report after such a crazy year in AI! Nestor from HAI joins us in this episode to talk about some of the main takeaways including how AI makes workers more productive, the US is increasing regulations sharply, and industry continues to dominate frontier AI research.

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    First impressions of GPT-4o

    First impressions of GPT-4o
    Daniel & Chris share their first impressions of OpenAI’s newest LLM: GPT-4o and Daniel tries to bring the model into the conversation with humorously mixed results. Together, they explore the implications of Omni’s new feature set - the speed, the voice interface, and the new multimodal capabilities.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.