Logo
    Search

    Podcast Summary

    • LanceDB: A Vector Database for AI ApplicationsLanceDB, founded by Cheng Shi's team, is a standout vector database for AI applications due to its on-disk indexing capabilities, meeting the unique requirements of the field, and commitment to engineering excellence.

      LanceDB, a vector database, was founded by Cheng Shi and his team in response to the growing need for efficient and effective data management for AI applications. With the hype surrounding vector databases, it can be overwhelming for users and developers to determine which tool is best for their needs. Prashant, in a previous episode, highlighted LanceDB as a standout option due to its on-disk indexing capabilities. LanceDB's development was motivated by the increasing demand for AI tooling and applications and the need for a database solution that could keep up with the unique requirements of this field. The company's success is a result of its commitment to engineering excellence and its focus on meeting the specific needs of its users.

    • Creating a unified data infrastructure for tabular and unstructured dataLandCV, initially focused on data infrastructure for computer vision projects, identified the need for handling multimodal data and created an open-source project to improve performance and reduce costs by managing both tabular and unstructured data as a single source of truth.

      LandCV, a company started two years ago, initially focused on building data infrastructure for computer vision projects. The founders, who had extensive experience in data and machine learning, identified a common challenge: handling multimodal data, specifically the integration of tabular and unstructured data like images. Traditional data infrastructure, such as Parquet and Orc, performed poorly with unstructured data. LandCV aimed to create a single source of truth for managing both tabular and unstructured data, improving performance and reducing costs. The team spent a year developing this underlying data infrastructure as an open-source project. When generative AI gained popularity, the community discovered that the vector index built for computer vision users could also be beneficial for generative AI applications. LandCV then separated the vector database component from the main project, making it easier for the community to use. The original motivation for LandCV was to serve companies dealing with large vision datasets, such as those in autonomous driving and recommender systems. The team identified that the underlying data infrastructure was the root cause of longer development times and difficulty in maintaining multimodal AI projects. By focusing on improving the data infrastructure, LandCV aimed to make it easier and more cost-effective for companies to explore and utilize their vision datasets.

    • Managing unstructured data pain points leads to focus on infrastructure for Lance DBLance DB focuses on infrastructure due to challenges of managing unstructured data across multiple systems, and the importance of semantic search and retrieval for generative AI use cases

      The determination to focus on infrastructure for Lance DB was driven by observing the pain points of managing unstructured data across various systems. The interviewees reported issues with data being split across multiple places, making it difficult to maintain and stitch together. This led to subpar tools being used by machine learning engineers and researchers. The transition to generative AI use cases brought a new focus on semantic search and retrieval, making indexing and data management more important than ever. Lance DB has responded by increasing investments in these areas and integrating with relevant frameworks. The company's initial focus on large-scale vision datasets remains relevant, but the generative AI landscape has expanded the potential use cases for Lance DB moving forward.

    • Introducing LanceDB: A User-Friendly Vector Database ToolLanceDB is an easy-to-use, cost-effective, and hyperscalable vector database tool that allows managing all types of data together. Its embedded nature, columnar format, and disk-based vector indices enable efficient data management and scalability.

      LanceDB, a vector database tool, stands out in the market due to its ease of use, hyperscalability, cost effectiveness, and ability to manage all types of data together. The team behind LanceDB was motivated by the need for a user-friendly, easy-to-install package that doesn't require extensive expertise in underlying technologies. LanceDB's embedded nature, which allows it to run in process in Python and JavaScript, sets it apart from other options. Additionally, LanceDB's columnar format and disk-based vector indices enable better data management and scalability. A typical use case for a developer integrating LanceDB into their workflow involves first installing the library, which is a straightforward process due to its ease of use. Next, the developer can load their data, which can include vectors, metadata, and raw assets, into LanceDB. This data can be of various types, such as images, text, or videos. The developer can then create an index using the data, which is disk-based and separates compute and storage for better scalability. Once the index is created, the developer can perform vector searches using LanceDB's efficient algorithms. This workflow allows developers to manage and analyze their data effectively and efficiently, making LanceDB a valuable tool in the field of vector databases.

    • Flexible handling of small and large-scale vector databases with LandCVLandCV offers ease of installation for small projects and supports popular libraries. For large-scale use cases, it processes data using distributed engines and offers flexibility in deployment options.

      LandCV offers flexibility in handling both small and large-scale vector databases. For smaller projects, it's easy to install using PIP or NPM and can be interfaced with popular libraries like pandas and Polaris. It also supports various embedding models and can output results in formats compatible with data frame-based workflows. On the other hand, LandCV stands out for large-scale use cases due to its ability to process data using distributed engines like Spark and GPU acceleration for quick indexing. Its architecture separates compute and storage, allowing for a simple and scalable setup with minimal coordination between query nodes. This architecture is similar to Neon, a database that also follows a "shared nothing" approach. One advantage of using LandCV is its flexibility in deployment. Users can easily test it in a Colab notebook and later deploy it on S3 or other storage options for larger projects. This flexibility makes it a popular choice for various scenarios, from on-premises to cloud-based deployments. When discussing LandCV with clients, it's essential to clarify that deploying it on S3 or other storage options means setting up the database and connecting to it, not just "throwing it up" without any additional effort. However, the ease of deployment and flexibility in storage options is a significant advantage that sets LandCV apart from other vector databases.

    • The separation of compute and storage in data processingThe separation of compute and storage enables efficient querying of large datasets through the use of columnar formats, vector indices, and fast random access columnar formats.

      The separation of compute and storage is a key innovation in data warehousing and data engineering that enables efficient processing of large amounts of data. This concept can be better understood through the analogy of DuckDV processing large data on a laptop. The data architecture that makes this possible includes a columnar format, a vector index based on disk, and a columnar format that can provide fast random access. These three pillars allow for efficient querying of large datasets without the need for massive compute power. Furthermore, the data architecture addressed in the discussion caters to a wide range of needs, including Python and JavaScript. This flexibility is essential as different applications may require different programming languages. In summary, the separation of compute and storage is a game-changer in data processing, allowing for efficient querying of large datasets through the use of columnar formats, vector indices, and fast random access columnar formats. This innovation has been a significant trend in data warehousing and data engineering over the past decade.

    • Switching from C++ to Rust for a vector databaseThe team behind Lance DB successfully transitioned from C++ to Rust, resulting in faster development and improved codebase safety. Rust's popularity in vector databases like Quadrant and Pinecone is due to its productivity and safety features.

      The team behind Lance DB, a vector database, made a productive and safe switch from C++ to Rust, leading to faster development and improved confidence in their codebase. They have developed both a Rust core and embedded databases in JavaScript and other languages based on the Rust crate. This shift to Rust was driven by increased productivity and the safety features it offers, making it a popular choice for vector databases like Quadrant and Pinecone. Looking ahead, the team believes that the trend towards language agnosticism in AI applications and workflows is growing, as the need for diverse tooling and infrastructure becomes more apparent. Generative AI is one area where this trend is particularly noticeable. Overall, the team's experience demonstrates the potential benefits of using Rust for data-intensive projects and the importance of staying adaptable in the ever-evolving landscape of technology.

    • Exploring the Opportunities of Vector Databases and Data Tooling in TypeScript AI ToolsVector databases like LandCV offer versatility and scalability for various industries and applications, including generative AI, productivity tools, and customer success tools. They can handle large single datasets for storing item embeddings and offer time travel queries for tracking changes in large datasets over time.

      The use of vector databases and data tooling, specifically in the TypeScript JavaScript community for building AI tools, presents an opportunity for growth in the open source landscape. This is particularly true for applications that require agile and tightly bundled RAG (Retrieval and Generation) systems. Use cases for LandCV, a vector database and database format, span across various industries and applications, including generative AI, productivity tools, and customer success tools. One unique feature of LandCV is its ability to version tables and perform time travel queries, making it an ideal choice for tracking changes in large datasets over time. Additionally, ecommerce and search applications are traditional use cases for vector databases, which can handle large single datasets for storing item embeddings. Overall, the versatility and scalability of LandCV make it a valuable tool for a wide range of applications.

    • Revolutionizing data management and analysis with LMDB, Land DB, Duck DB, and vector databasesThe integration of various databases and tools enables active learning, deduplication, and high GPU utilization during training and evaluation, with potential applications in edge computing and autonomous vehicles.

      The integration of various databases and tools, such as LMDB, Land DB, Duck DB, and vector databases, is revolutionizing data management and analysis. This combination enables active learning, deduplication, and high GPU utilization during training and evaluation. One exciting application is the use of OpenLM to generate SQL queries based on Lansdv descriptions, which can then be executed using DuckDB. DuckDB's extension mechanism, including the published Rust-based extension framework, allows for more seamless integration and transparency. Furthermore, the discussion touched upon the potential applications of these technologies in the realm of autonomous vehicles and edge computing. The challenges of working with limited resources in these environments make vector databases and other advanced tools particularly valuable. While specific use cases and ideas for the future were not explicitly stated, the potential for improved data management and analysis in edge computing is significant. The goal is to make these vector databases feel like a seamless part of the workflow, rather than a separate entity.

    • Managing Complex Data Integration in Practical AIIn the next 6-12 months, expect an explosion of personalized info retrieval tools & domain-specific agents, as well as the development of low-code/no-code generative AI tools. Exciting developments in practical AI for robotics, vehicles, gaming, and more.

      The integration of complex data from various sources, including visual data, time series data, and manual input, is crucial for companies in the robotics and vehicle industries, as well as any organization generating data in the physical world. Managing and querying this data together will be essential for unlocking the potential of advanced AI capabilities. In the next 6 to 12 months, we can expect an explosion of information retrieval tools, personalized to individual data and cases, as well as domain-specific agents that can deliver magical results in fields like legal and healthcare. Additionally, the development of low-code to no-code tools for building sophisticated applications using generative AI is a promising trend. Personally, I'm excited about the potential of generative AI in gaming, particularly in the context of open-world environments. Overall, the practical AI space is full of exciting developments and opportunities.

    • Exploring the Power of Lance d b with DanielLance d b can save time and effort by generating code and offering endless exploration opportunities. Try it out for productivity gains and stay updated on future improvements.

      The use of AI tools, specifically Lance d b, can significantly enhance productivity and performance in various industries. Daniel, a guest on the Practical AI podcast, shared his personal experience of using Lance d b to generate code and how it saved him time and effort. He emphasized the potential for endless exploration and discovery in the generated world, expressing gratitude towards the Lance d b team for providing excellent tools. Daniel encouraged listeners to check out Lance d b and try it out for themselves, as it only takes a few minutes to get started. He also expressed excitement about the future developments and improvements to come. The podcast concluded with a reminder to subscribe, share Practical AI with others, and thanked Fastly and Fly for their partnership. Additionally, a shoutout was given to Breakmaster Cylinder for providing the best beats for the show. Overall, the discussion highlighted the transformative power of AI tools and their potential to streamline workflows and boost efficiency.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Mamba & Jamba

    Mamba & Jamba
    First there was Mamba… now there is Jamba from AI21. This is a model that combines the best non-transformer goodness of Mamba with good ‘ol attention layers. This results in a highly performant and efficient model that AI21 has open sourced! We hear all about it (along with a variety of other LLM things) from AI21’s co-founder Yoav.

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.