Logo
    Search

    Podcast Summary

    • Innovative use of LoRa technology in a mesh network for disaster reliefLoRa technology's mesh network capabilities can enable real-time communication in critical situations with very low power consumption

      The AI world is continuously evolving, and new advancements are being made in various areas, including hardware and software. At a recent hackathon at Stanford, a project called Meshwork showcased an innovative use of LoRa technology, which involves sets of radio devices communicating in a mesh network over long distances for very low power. The team used this technology for a disaster relief project, allowing for real-time communication between the field and a command center. This demonstrates the potential for AI and low-power technology to make a significant impact in critical situations. Additionally, there are numerous resources available to help individuals level up their AI skills, and staying informed about the latest news and achievements is essential for those working or interested in the field.

    • Exploring advanced technologies at a hackathonParticipants showcased projects using command control interfaces, computer vision, and graph databases, leading to real-time analysis, satellite imaging, and AI face interaction. A new concept, activation hacking, was introduced, involving neural network manipulation for creative results.

      During a recent hackathon, participants showcased impressive projects utilizing advanced technologies like command control interfaces, computer vision, and graph databases. These tools allowed for real-time analysis and tagging of data, satellite imaging, and even interaction with AI faces. The hackathon marked the first time many attendees had seen technologies like Boston Dynamics dogs and Neo4j in person. While the projects ranged from technical to non-technical, one topic that piqued the speaker's interest was activation hacking, a concept mentioned by Quran from Nuse Research during a previous episode. Activation hacking involves manipulating the activations of neural networks to generate new and creative results. The speaker found an informative blog post, "Representation Engineering, Mistral 7 B, an acid trip," by Thea Vogel, which provided insights into this intriguing topic. Overall, the hackathon showcased the power and potential of emerging technologies and sparked curiosity and learning for all involved.

    • Exploring the world of representation engineeringRepresentation engineering goes beyond prompt engineering to control a model's output by influencing its tone, angle, or behavior through specific instructions or biases in the prompt, and recent research is exploring more predictable and controllable outputs using 'activation hacking'.

      Representation engineering, a key aspect of working with AI models, goes beyond prompt engineering. While prompt engineering focuses on crafting the perfect input for the model, representation engineering aims to control the model's output by influencing its tone, angle, or behavior. This is achieved by providing specific instructions or biases in the prompt, but it's not always reliable. Recent research in representation engineering, or "activation hacking," explores the idea of consistently steering a model to give certain types of answers, such as always being happy, confident, or less confident. This approach can be applied to various types of models, not just language models. For instance, it can be used with image generation models to control the mood or style of the generated images. The idea is to create a more predictable and controllable output from the model. An example of this in action is the Mistral model being asked, "What does being an AI feel like?" and the researcher guiding the output to a specific answer. This area of research holds great potential for improving the reliability and effectiveness of AI models in various applications.

    • Leveraging control vectors for emotional control in modelsResearchers propose using control vectors, derived from contrasting datasets, to influence model output without explicit prompts. These vectors represent the difference between hidden states for happy and sad prompts, allowing for subtle emotional control.

      Researchers have proposed a creative method called "control vectors" to influence a model's output without explicitly instructing it through prompts. This approach involves creating contrasting datasets to collect hidden states during the model's processing of happy and sad prompts. The difference between the hidden states for each prompt results in a dataset of vectors representing the difference between the happy and sad hidden states. Through dimensionality reduction techniques like PCA, a single control vector is extracted for each hidden layer. When these control vectors are activated, the model's output corresponds to the respective emotion (happy or sad). This method can be particularly useful in cases where it's challenging to put instructions in prompts or when such instructions become repetitive or boilerplate. By using control vectors, the model can learn to associate specific hidden states with emotions, allowing for more nuanced and subtle control over the output.

    • Streamlining AI model interactions with control vectorsControl vectors simplify prompting by reducing complexity, allowing for more consistent and effective interactions with AI models in various applications.

      When working with AI models, using simple and effective prompts is crucial for achieving desired results. The use of control vectors, a step between basic prompting and fine-tuning, allows for more straightforward and maintainable prompts. This approach simplifies the process by reducing the need for extensive instructions and complex systems. For instance, in a real-world application like a fast food restaurant drive-through, using a model with control vectors can lead to more consistent and positive interactions with customers. By applying retrieval augmented generation and utilizing control vectors, the model can engage in more effective conversations with consumers, enhancing their overall experience. Moreover, having different sets of control vectors allows for flexibility in managing various interactions and flows. This flexibility can be particularly valuable when dealing with complex scenarios where multiple behaviors are required. Overall, the use of control vectors offers a more streamlined and effective approach to prompting AI models, enabling better interactions and improved outcomes.

    • Exploring GPT script, a new natural language-based programming language and its potential use casesGPT script is a new language that simplifies interacting with large language models using natural language syntax. It can be combined with traditional scripts and HTTP calls, and includes a control mechanism for responsible use.

      There's a new scripting language called GPT script that aims to create a fully natural language-based programming experience for interacting with large language models like OpenAI. The script's syntax is largely natural language, making it easy to learn and use. It can be combined with traditional scripts and external HTTP service calls. The central concept is the use of tools, each performing a series of actions, which can be composed to accomplish tasks. Another intriguing topic discussed was the potential use of a control mechanism for AI models. This mechanism could be used to break models out of their intended use or prevent unintended outputs. The application of this technology raises interesting questions about responsible use and potential misuse. The conversation also touched on the concept of a change log news, which provides top developer stories of the week and allows users to subscribe and receive a companion email. The discussion also explored various flavors of responses to hypothetical situations, such as being late for work and pitching a TV show. These examples demonstrated the potential for creativity and divergent thinking. The podcast also encouraged listeners to share their knowledge and experiences related to the topics discussed. Overall, the conversation offered valuable insights into the intersection of AI, programming, and creative expression.

    • AI generates hyperrealistic videos from textOpenAI's Soarer model creates realistic videos, blurring the line between real and generated content, sparking conversations around AI safety.

      OpenAI has made a significant stride in the field of AI by announcing their Soarer model, which can generate hyperrealistic videos from text. This is a notable development as it represents a new level of capability in AI video generation, moving beyond short clips or low-quality videos. The model's release has sparked conversations around AI safety, as the line between real and generated content becomes increasingly blurred. The videos generated by the model showcase a high level of realism, with compelling visuals that mimic human behavior. However, there are concerns about the potential for misuse and the need for safeguards against the creation and dissemination of false or misleading content. Overall, the release of the Soarer model marks an important step forward in the ongoing evolution of AI technology.

    • Google releases open source language model GemmaGoogle's open source language model, Gemma, marks a significant shift towards making AI technology more accessible and practical for a broader audience. Its smaller size and reasonable computational requirements make it ideal for a wide range of applications and practical for local or edge deployments.

      The release of open source language models like Google's Gemma, signifies a significant milestone in the mainstream adoption of AI technology. This comes as a response to the growing popularity and accessibility of models like OpenAI's ChatGPT. Google's decision to release a derivative of their closed source Gemini model as open source is seen as a positive step towards encouraging more organizations to produce and share models. The smaller size and reasonable computational requirements of these models make them practical for local or edge deployments, and ideal for a wide range of applications where large models may not be necessary. The availability of open source models also provides more flexibility for users in terms of deployment strategies, particularly for those with regulatory, security, privacy, or connectivity concerns. As with previous releases of large model families like Llama 2 and Mistral, we can expect to see a surge in fine-tuning efforts off of these models. Overall, the release of open source language models marks a significant shift towards making AI technology more accessible and practical for a broader audience.

    • Discussing Magic, a new large language model for fine-tuning tasks and potential AI developerA new AI model, Magic, is available for fine-tuning tasks and can potentially develop into an AI developer. Chris discussed representation learning and activation hacking, offering insights and practical steps for experimentation.

      A new large language model, which is available under certain licenses, has been released and can be easily used for fine-tuning tasks with popular libraries like Hugging Face's transformers. The model, called Magic, is being framed as a code generation platform, with the ultimate goal of developing an AI developer that can automate coding tasks and potentially contribute to general problem-solving. During the discussion, Chris also shared insights into representation learning and activation hacking, which can be explored further through a blog post and accompanying tutorial. This week has seen a flurry of AI news, and the deep dive into representation learning provided a fascinating perspective on the current state and future potential of AI technology. For those interested in learning more about representation learning and experimenting with the techniques, Chris recommended checking out the blog post and following the steps to perform activation hacking and representation learning using the JEMMA model. Stay tuned for more insights and discussions on practical AI applications in the coming weeks. If you're interested in joining the conversation, be sure to subscribe to Practical AI and join our free Slack community at practicalai.fm/community.

    • Understanding the Role of the Brake Master CylinderRegularly check your brake system, including the Brake Master Cylinder, for signs of failure like a spongy pedal, longer braking distance, or leaks. Prioritize safety by addressing issues promptly and staying informed.

      The BRAKE MASTER Cylinder is a crucial component in a vehicle's braking system, essential for our safety as residents. During our discussion, we learned about its function, which is to convert the force applied to the brake pedal into hydraulic pressure that activates the brakes. We also touched upon the signs of a failing BRAKE MASTER Cylinder, such as a spongy brake pedal, longer braking distance, and brake fluid leaks. Furthermore, we emphasized the importance of regular vehicle maintenance, including checking the brake system, to ensure safety and prevent potential accidents. We also encouraged everyone to pay attention to their vehicles and not ignore any warning signs, no matter how small they may seem. Lastly, we reminded everyone to appreciate the time they spend learning new things, like understanding the importance of the BRAKE MASTER Cylinder, and to look forward to our future discussions. So, in essence, take care of your vehicles, stay informed, and prioritize safety.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Private, open source chat UIs

    Private, open source chat UIs
    We recently gathered some Practical AI listeners for a live webinar with Danny from LibreChat to discuss the future of private, open source chat UIs. During the discussion we hear about the motivations behind LibreChat, why enterprise users are hosting their own chat UIs, and how Danny (and the LibreChat community) is creating amazing features (like RAG and plugins).

    Mamba & Jamba

    Mamba & Jamba
    First there was Mamba… now there is Jamba from AI21. This is a model that combines the best non-transformer goodness of Mamba with good ‘ol attention layers. This results in a highly performant and efficient model that AI21 has open sourced! We hear all about it (along with a variety of other LLM things) from AI21’s co-founder Yoav.

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.