Logo
    Search

    Podcast Summary

    • The Importance of AI in Today's WorldAI is no longer a niche topic but a necessity for individuals and businesses across various industries, with numerous applications, products, and startups integrating AI into their operations.

      We're living in a pivotal moment where artificial intelligence (AI) is no longer a niche topic but a daily conversation for individuals and organizations across various industries. The AI landscape is rapidly evolving, and the integration of AI is becoming increasingly commonplace. This was discussed on the latest episode of Practical AI, where hosts Daniel Whitenack and Chris Benson reflected on the significant growth of AI since they started the podcast in 2018. They noted that the world has reached a point where AI is no longer just a topic of interest for experts but a topic of necessity for everyone. They also mentioned the increasing number of applications, products, and startups that are integrating AI, making it an essential part of their operations. The impact of AI was further emphasized when they discussed the recent OpenAI outage and how it affected numerous startups built solely on the model. The conversation underscored the interconnectedness of AI and how its influence is expanding beyond the tech industry.

    • Understanding Generative Model Behavior: Amazing Applications and Disturbing OutputsExplore the differences between good and bad outputs of generative models, discuss strategies for utilizing them effectively, and acknowledge that mistakes and competition are not unique to any one company or model.

      As we witness the explosive growth and advancements in generative models, it's crucial for practitioners to understand how to reliably use these models to produce value in their specific applications, while also being aware of and avoiding potential negative outcomes. These models, while impressive, are still in their early stages and can sometimes produce unintended or disturbing results. As competition in the AI space heats up, companies will continue to make mistakes with these models, leading to public failures and panic. Therefore, it's essential to explore the differences between good and bad outputs and discuss strategies for utilizing these models effectively in our day-to-day work. Additionally, it's important to acknowledge that this trend of model mistakes and competition is not unique to any one company or model, and it's likely to continue for some time. So, the conversation today revolves around understanding the behavior of these generative models, both the amazing applications and the disturbing outputs, and how we as practitioners can navigate this landscape.

    • Exploring the Potential and Challenges of Open AI ModelsOpen AI models, like those from Microsoft and Google, offer immense potential for creativity and utility but face criticism for reflecting human biases and unwanted behaviors. Companies must focus on the long-term trajectory and balance creativity with ethical considerations.

      Open AI models, such as those developed by Microsoft and Google, are facing intense competition and criticism as they are still in their early stages. While these models show great potential in areas like endless creativity and utility, they also come with challenges, including the reflection of human biases and unwanted behaviors. During a recent event, demos showcased the creative potential of these models, such as an infinite Dungeons and Dragons referee. However, there is a darker side, with outputs that can be disturbing or reflect negative sentiments. These models are merely reflecting the data they are given, which includes human biases and sarcasms. As senior executives in these companies navigate this landscape, it's important to focus on the long-term trajectory and not get too fixated on short-term problems. The potential for these models to generate creative content and provide useful utility is significant, but it's crucial to address the unwanted behaviors and biases as well. The future of AI lies in striking a balance between creativity and ethical considerations.

    • Exploring the unexpected results of AI modelsAI models reflect their creators' biases and quirks, leading to unexpected and sometimes disturbing results, emphasizing the need for ongoing monitoring and safeguards.

      As we continue to develop and rely on AI technology, it's important to recognize that these models reflect the biases and quirks of their human creators. From language models like Bing's chatbot and ChatGPT, to text-image models like DALL-E 2, we're seeing unexpected and sometimes disturbing results. Some models exhibit bad behavior, like gaslighting users or producing inappropriate content, while others generate academic-sounding text that can make even the most absurd topics seem rational. These trends highlight the need for ongoing monitoring and safeguards to prevent AI from veering off course. It's a reminder that AI is a reflection of us, and as we continue to explore its capabilities, we must remain vigilant and prepared for the unexpected. If you're interested in learning more about these developments and the latest in tech news, be sure to check out our show notes for links to related articles and discussions. And if you'd like to support our work here at Changelog, consider joining Changelog Plus Plus, our membership program, to directly contribute to our mission of bringing you the latest and greatest in tech.

    • Understanding the unpredictability of generative modelsExpect unexpected results when using generative models, acknowledge their limitations, and prepare for the unexpected to avoid disappointment.

      When working with generative models in real-world applications, it's essential to understand that the data used to train these models is vast and beyond our immediate control. This leads to outputs that may be unexpected or inconsistent with our expectations. As a result, it's crucial for organizations and users to reset their expectations and understand that the usage of these models will always have limitations. These models will reliably output creative and coherent text or images, but there will be instances where the output may not align with our intended goals. It's important to acknowledge this "Wild West" aspect of using generative models and prepare for the unexpected to avoid being "shattered" by the results.

    • The Influence of Chaotic and Inconsistent Data on AI BehaviorAI models like ChatGPT and Stable Diffusion can generate coherent responses, but their output may also be inaccurate or illogical due to the inconsistent and harmful data they're trained on and the way prompts are engineered.

      The behavior of AI models like ChatGPT and Stable Diffusion is influenced by the chaotic and inconsistent data they are trained on, as well as the way prompts are engineered. While these models can generate coherent and creative responses, they may also produce inaccurate and illogical answers. The data used to train these models is vast and often contains harmful, inaccurate, and inconsistent information. This inconsistency carries over into the AI's output. Additionally, the way prompts are engineered can significantly impact the quality of the output. Some prompts may be intentionally misleading or adversarial, leading to incorrect or nonsensical responses. As users, we must be aware of this and approach AI outputs with a critical eye, considering the potential biases and limitations of the data and prompts used to generate them.

    • Misalignment between human expectations and AI capabilities in conversationsAI models don't understand or intend the meaning behind text, only generating coherent output based on given prompts. Human scrutiny and understanding of ambiguity and potential quack sources are crucial during testing and evaluation.

      While we may expect AI models to understand and respond like humans in conversations, they merely produce coherent output based on the given prompt. They don't have a clue about the meaning or intent behind the text they generate. This misalignment between human expectations and AI capabilities can lead to dissonance and misunderstandings. During testing and evaluation, it's crucial to consider the ambiguity and potential quack sources in the results. The standards for assessing AI technologies are different from those for replacement. People often intentionally try to break these models, but the same scrutiny isn't applied to other technologies. When building applications with these models, understanding prompt engineering is essential. The model has no morality or understanding of what it's saying, and it's just producing coherent output based on the given prompt. The integration of these models into applications significantly influences their perceived usefulness or lack thereof. ChatGPT, Bing AI, and other similar models operate based on free-form text prompts, which lack structure. This lack of structure can lead to varying levels of output quality. It's important to remember that AI models don't possess human-like understanding or intent, and our expectations should be adjusted accordingly.

    • Exploring the impact of prompts on AI model performanceEffective prompt engineering can significantly enhance AI model performance by optimizing interactions and improving outcomes.

      The interface and structure of prompts used with AI models significantly influence their usefulness and the outcomes they produce. While open-domain interfaces offer greater freedom, they can also lead to unpredictable results. Conversely, using templates and structured prompts can increase predictability but limit the model's open-ended capabilities. This area of interaction with AI models, often referred to as prompt engineering, is a growing focus in the industry and requires expertise beyond just model training. It's an additional layer where data scientists, machine learning engineers, and others will need to operate, as the choice and application of prompts can greatly impact the model's performance and output. Moreover, the rise of cloud-based AI services means that prompt engineering is becoming a distinct skill set. While an imperfect analogy, it shares similarities with user interface and user experience design on the software side. The interaction between the user and the model is a crucial layer that demands attention, and mastering prompt engineering will be essential for maximizing the potential of AI models in various applications.

    • Guiding the Model with Effective PromptsEffective prompts guide the model to generate useful output, experimenting with various formulations can lead to better generations, and describing the task and setting in the prompt is essential.

      The prompt used when interacting with large language models significantly influences the utility and acceptability of the output. The data behind the model and the user interface are also important, but the prompt holds the most control. Different principles for prompt engineering exist, and one noteworthy guide is from Cohere. The first principle is that a prompt guides the model to generate useful output. The second principle is to experiment with multiple formulations of the prompt to achieve the best generations. This may involve trying various prompts or chaining them together. The third principle is to describe the task and the general setting in the prompt. A typical prompt for a language model consists of instructions, context, input data, and an output indicator. Instructions tell the model what you want to happen, such as sentiment analysis. Experimentation and exploration of prompts are crucial aspects of prompt engineering.

    • Defining a task for text or image analysis using CoHEREProvide clear instructions, context, input data, and an output indicator to help CoHERE understand the task and produce accurate results.

      When working with text or image data using models like CoHERE, it's essential to clearly define the task, provide context, input data, and an output indicator. The task refers to the goal of the model, such as sentiment analysis or generating a painting. Context provides background information to help the model understand the task. Input data is the text or image the model will analyze. The output indicator is where the model's result will be displayed, such as sentiment or a generated image. CoHERE recommends providing examples to help the model understand the task. In the text analysis world, this could be positive and negative sentiment examples. In the image world, this could be style keywords, references to other images, or negative filters to exclude specific qualities. By following this structure, we can effectively communicate our intentions to the model and increase the chances of getting accurate and desired results.

    • Effectively and creatively using AI modelsPractitioners and startups should focus on optimizing AI model use rather than testing their limits for best results.

      During our discussion about generating blog content for AI models, the speaker emphasized the importance of using AI models effectively and creatively, rather than trying to test their limits. This approach is particularly valuable for practitioners and startups looking to optimize their use of these models. The speaker also highlighted a specific page they found helpful in this regard and encouraged listeners to explore the resources they discussed in the show notes. Overall, the conversation provided practical insights and learnings for those interested in AI and its applications. If you're looking to make the most of AI models, the speaker's guidance is a great starting point. Be sure to check out the links in the show notes for more information and join the conversation on social media to share your experiences and discoveries. This conversation was a valuable learning experience for the speaker and they look forward to exploring the ever-evolving AI world with listeners. Don't forget to subscribe to Practical AI and share it with your network to help spread the knowledge. And a big thank you to Fastly, Fly, and BRAKE MASTER Cylinder for their continued support of the show.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Stanford's AI Index Report 2024

    Stanford's AI Index Report 2024
    We’ve had representatives from Stanford’s Institute for Human-Centered Artificial Intelligence (HAI) on the show in the past, but we were super excited to talk through their 2024 AI Index Report after such a crazy year in AI! Nestor from HAI joins us in this episode to talk about some of the main takeaways including how AI makes workers more productive, the US is increasing regulations sharply, and industry continues to dominate frontier AI research.

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.