Logo
    Search

    Podcast Summary

    • AI-driven personal devices: Balancing benefits and privacy concernsExploring the benefits of AI-driven personal devices, while acknowledging the importance of privacy and considering potential risks.

      We're seeing an increasing trend of AI-driven personal devices designed to help individuals manage their daily lives and access their personal data more efficiently. These devices, such as the Rabbit r one and the AI pen, have received significant attention recently. While Chris expresses excitement about the potential benefits, including assistance with memory and organization, he also acknowledges feeling torn about the level of integration between AI and personal life. The line between technology and personal privacy is becoming increasingly blurred, and it's important for individuals to consider the implications and potential risks. Daniel shares a personal experience of the challenges that come with technology malfunction, emphasizing the importance of being prepared for unexpected issues. Overall, the discussion highlights the ongoing evolution of AI technology and its potential impact on our daily lives.

    • The perception of privacy shifts with increased automation and agentic nature of AI assistants and devicesAs AI assistants and devices become more automated and agentic, our perception of privacy may change, even if the amount of data collected remains the same. It's important to consider the implications and make informed decisions about what data we share.

      While the use of AI assistants and devices may not significantly change the amount of data being collected compared to what we already give out with our smartphones, the perception of privacy may shift due to the increased automation and agentic nature of these entities. The idea of being watched and having data used without human oversight can be unsettling, even if the data has been freely given for years. The difference lies in the level of automation and the feeling of being interacting with a sentient being, rather than a faceless corporation or entity. However, as we become more accustomed to this level of data collection and analysis, our emotional reactions may lessen. Ultimately, it's important to consider the implications of these technologies on our privacy and make informed decisions about how and what data we share.

    • New level of opt-in data sharing with Rabbit r oneSuccess of Rabbit r one depends on addressing privacy concerns and offering tangible benefits to users

      The perception of privacy invasion and the level of control we have over our data are major factors in how we view and use new technologies. The Rabbit r one, an AI-driven device with a conversational operating system, represents a new level of opt-in data sharing. Its physical design and speech-driven interface set it apart from traditional smartphones and apps. While some may find the idea of a dedicated device appealing, others may prefer the familiarity and convenience of using their existing devices. Ultimately, the success of the Rabbit r one and similar technologies will depend on how well they address privacy concerns and offer tangible benefits to users. The device's design, inspired by hardware synths and sequencers, features a minimal interface with a primary focus on speech commands. The gamble of offering a physical device is that it requires users to place it at the center of their lives, whereas apps are already integrated into our daily routines. The ongoing debate around data privacy and control will continue to shape the adoption and evolution of new technologies.

    • Navigating the Burden of Multiple Apps and the Potential for a New AI OSThe disjointed experience of using multiple apps for simple tasks raises questions about the future of devices, with AI OS like Rabbit potentially becoming the primary hub of our digital lives. Blockchain technology may also provide a solution to the authenticity problem in the digital world as AI-generated content becomes more prevalent.

      While we enjoy the convenience of having an app for everything, the navigation and orchestration of these various apps can be burdensome and not as seamless as we might expect. For instance, performing a simple task like updating a payment method in an Uber ride involves opening multiple apps and copying information between them. This disjointed experience has led some to ponder if we are nearing a turning point in how we think about devices, with the phone no longer being the central hub of our digital lives. Instead, a new device like Rabbit, which packages everything together in an AI OS, could potentially take over as the primary device that runs our lives conversationally. Another related topic that emerged during the discussion was the authenticity problem on the internet, as AI-generated content becomes increasingly prevalent. Chris Dixon's book, "Read, Write, Own," proposes blockchain technology as a possible solution to this issue, allowing us to track the origins of content and ensure transparency and ownership. In the background, there are also technological advancements in multimodal models, like the one discussed in the Ask UI episode, which have the potential to streamline our digital experiences. Overall, these discussions highlight the need for more seamless and transparent technology to keep up with the rapidly evolving digital landscape.

    • Revolutionizing web app interaction with AIRabbit's neurosymbolic large action model revolutionizes web app interaction using AI, performing complex tasks through external system interactions

      Rabbit's new large action model is revolutionizing the way we interact with web applications using AI. This model goes beyond traditional automation tools like Selenium and introduces voice interaction. However, the flexibility and personalization of modern apps present a challenge, as elements can change positions or update frequently. Rabbit's solution is a neurosymbolic large action model, which interacts with external systems like Uber or banks to perform complex tasks. While the model itself may receive the spotlight, the true power lies in the interactions and touchpoints with various systems and services. Large language models, like ChatGPT, also demonstrate this principle through their plugins and external interactions, making the value not only in the model but in how it engages with the world. As we move forward, understanding and harnessing these interactions will be crucial for maximizing the potential impact of AI in our lives.

    • Integrating various types of data and external systems for advanced AI responsesAdvanced AI models can now process multiple types of data and interact with external systems to generate more accurate and contextually relevant responses, enhancing their usefulness for users.

      AI models, like Rabbit, are increasingly capable of integrating various types of data and external systems to generate more accurate and contextually relevant responses. This integration goes beyond simple text-to-text autocomplete and retrieval mechanisms. It includes multimodal models that can process both text and images, and tools that enable AI models to interact with external systems by generating structured outputs. These outputs can then be used to call APIs and receive responses, which can be integrated back into the AI system. For example, an AI model could be used to identify a raccoon in an image and generate a structured output that can be sent to a search API to find related information. This integration of AI with external systems has been happening for some time, but it's an important aspect of the development of advanced AI models that can provide more comprehensive and useful responses to users. The recent advances in multimodal large language models, such as MMLLMs, demonstrate the potential of this technology to process and generate responses based on multiple types of data. Overall, the ability of AI models to integrate various types of data and external systems is a significant step forward in the development of more advanced and useful AI applications.

    • Exploring AI integration in unstructured environmentsRabbit research team is investigating how to enable AI systems to perform arbitrary actions across unstructured applications or systems, beyond APIs and data retrieval.

      While integrating AI models with APIs is a popular approach in enterprises for software-driven environments, not all applications or tools have well-defined and structured APIs. The Rabbit research team is exploring the question of how to reformulate the problem to enable users to trigger an AI system to perform arbitrary actions across applications or systems, regardless of their API structures. The speaker provided an example of using the Shopify API with Prediction Guard, where they create a natural language query to get ShopifyQL or GraphQL queries to interact with the Shopify API. However, not all applications have such well-defined APIs, and sometimes, users may need to perform actions that these APIs don't support. The challenge lies in making human intentions expressed through actions on a computer work in an unstructured world. The Rabbit research team is focusing on this issue, but it remains an open question. While APIs and data retrieval can get us far, they don't cover all the complexities of human interactions with applications. As of now, the Rabbit research team has not published any papers on their research. However, their work highlights the importance of developing AI systems that can adapt to the complexities of an unstructured world and perform arbitrary actions across various applications or systems.

    • Shopify: A Global Commerce Platform for BusinessesShopify simplifies online selling with an easy-to-use platform and powerful features, including the best converting checkout. It supports businesses worldwide and is a significant player in US e-commerce. In AI research, 'large action models' aim to revolutionize AI-UI interaction, creating agentic AI for automated system interaction.

      Shopify is a comprehensive global commerce platform that assists businesses in selling products and services at every stage, from online shops to physical stores. Shopify's ease of use and powerful features, such as the Internet's best converting checkout, help businesses turn browsers into buyers. Shopify powers a significant portion of e-commerce in the US and supports entrepreneurs worldwide. Additionally, the term "large action models" has been used in the context of AI research, specifically in relation to Rabbit's new architecture for observing human interactions with a UI and mapping them onto a flexible, symbolic representation of a program. This architecture aims to create agentic AI that can interact with different systems in an automated way. While the specifics of Rabbit's large action models are not yet fully transparent, their goal is to revolutionize the way AI interacts with user interfaces. Shopify's success and the emergence of large action models demonstrate the exciting potential of AI to transform businesses and industries.

    • Understanding user actions through neurosymbolic approachNeurosymbolic technology interprets user actions into symbols, processes them using logic layers, and executes learned programs within apps.

      The discussed technology uses a neurosymbolic approach to understand user actions on apps like Uber, and generate a series of logical steps to execute the desired function. This approach combines neural networks with symbolic logic processing. The user's action is first interpreted by a neural layer into a set of symbols or representations. Then, these symbols are processed by symbolic logic layers to execute a learned program within the app. This program is not a traditional computer program but a logical one, representing an action sequence. The device's sensors, including camera, microphone, GPS, accelerometer, and gyroscope, may also provide inputs to the large action model, potentially expanding its capabilities beyond speech and camera-based interactions. This neurosymbolic approach allows the system to learn from human demonstrations and execute complex actions within the app.

    • Rabbit device: A new way to interact with technologyThe Rabbit device, without a screen, may change tech interaction through location-based services and fitness tracking using sensors. Potential competition from tech giants, but future of smartphones uncertain.

      The Rabbit device, which resembles a smartphone but lacks a screen, could potentially revolutionize the way we interact with technology, particularly in areas like location-based services and fitness tracking. The device's ability to use sensors to infer user intent and location could make it an intriguing alternative to traditional smartphones. The potential for this new marketplace in the AI space could lead to competition from tech giants like Amazon, Google, and Microsoft, who already have the infrastructure to produce both the back-end and front-end functionality. The future of smartphones could be influenced by this interface rethinking, with devices becoming more like Rabbits and less app-driven. However, it remains to be seen if this is the beginning of a new marketplace opening up, or just a speculative idea.

    • Decentralized AI Platform Rabbit with Minimal ConnectivityRabbit is a decentralized AI platform that operates with minimal connectivity, conserves energy, and is expected to see increased use of action models on low-power devices.

      Rabbit is a decentralized AI platform that utilizes a combination of local and cloud computing for its functions. The platform, which uses Langchain tools, is designed to operate with minimal connectivity and can run on low-power devices. The LAM (Language Model) and Lamb (Local AI Model) powered routines operate on a centralized platform, reducing the workload on individual devices and conserving energy. There's a prediction that a large cloud computing service provider may purchase Rabbit. The platform is expected to see increased use of action models, whether they be tool-using LLMs (Language Learning Models), LAMs (Local AI Models), or SLMs (Symbolic Learning Models). The Rabbit team could have named it "Lamb," but they seem to have mixed up their animals. The platform's documentation and tools, which are not complicated, can be found on Rabbit's research page and Langchain's documentation. It's recommended to check out these resources and try using some of the tools to explore the technology further. Overall, Rabbit offers an interesting direction for AI development, particularly in the area of edge computing and low-power devices.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Mamba & Jamba

    Mamba & Jamba
    First there was Mamba… now there is Jamba from AI21. This is a model that combines the best non-transformer goodness of Mamba with good ‘ol attention layers. This results in a highly performant and efficient model that AI21 has open sourced! We hear all about it (along with a variety of other LLM things) from AI21’s co-founder Yoav.

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.