Logo
    Search

    Podcast Summary

    • Impact of NLP technology on society and potential concernsNLP's ability to mimic human language and generate convincing text raises concerns about misinformation and manipulation, despite its applications in various industries.

      The current advancements in Natural Language Processing (NLP) technology, specifically generative models like OpenAI's GPT 3, have the potential to significantly impact society in various ways, raising concerns about misinformation and manipulation. NLP has proven to be one of the most mature branches of Artificial Intelligence (AI) and has clear applications in industries such as customer support, legal, healthcare, and finance. However, its ability to mimic human language and generate convincing text has also made it susceptible to misuse. This can lead to a breakdown of trust in information and even potentially manipulate people's decisions. While the fear of AI becoming sentient and taking over the world may be a common trope, the more immediate concern is the potential for misinformation and manipulation through advanced NLP models. This is a topic that Ivan Lee, the founder and CEO of DataSource, will be speaking about at the ML DataOps Summit in partnership with TechCrunch. The event, which is free and virtual, will gather more than 700 attendees from top AI and ML companies and feature major speakers including Facebook AI, Cruise, Zukes, and GE Healthcare. To learn more and register, visit imerit.net/dataops.

    • Navigating presentations and documents for work: Identifying gaps and discovering new insightsExploring ethical considerations in AI development and the importance of practical approaches to addressing these challenges

      Navigating the process of creating presentations and documents for work can be a challenge, but it can also provide opportunities for self-reflection and improvement. The speaker shared their experience of feeling overwhelmed by the details of design while working for a large company, and how this process can help identify gaps in thinking and lead to new insights. Additionally, they discussed a recent article they found in IEEE Spectrum about a new moral reference guide for AI, called "Machines Learn Good from Common Sense Norm Bank." The approach of this project, which draws from advice columns and ethics message boards, was seen as practical and interesting, as there is currently no standard way to approach AI ethics. Another topic touched upon was the updated version of the AI index report from Stanford University, which they had previously discussed last year. These topics highlight the importance of ethical considerations in AI development and the ongoing exploration of new approaches to address these challenges.

    • Model Delphi makes moral judgments based on common sense ethical dataResearchers developed Delphi AI model to make moral judgments based on ethical dilemmas, achieving high accuracy, but internet-trained GPT-3 performed poorly, suggesting ethical context is less prevalent online.

      Researchers at the Allen Institute for AI have developed a model named Delphi, which makes moral judgment calls based on a large dataset of common sense ethical judgments. This includes examples from various everyday situations, some of which can be quite extreme, like saving a child by killing a bear or exploding a nuclear bomb. Delphi achieved high accuracy compared to other models, reaching 92.1%, as evaluated by crowd workers on Mechanical Turk. However, it's worth noting that GPT-3, which was trained on a massive amount of internet data, performed much lower in this ethical judgment task. This might suggest that the context of ethics is less prevalent or less explicitly presented on the internet compared to other types of content. Overall, this research demonstrates the potential for AI to make moral judgments based on human ethical norms, but also highlights the importance of understanding the limitations and potential biases of such models.

    • Understanding AI language model performanceSpecific models outperform GPT 3 in certain areas, but GPT 3 can still be effective. Performance depends on task, data, and model design. Challenges include understanding why some topics are harder for models to grasp, and addressing ethical implications and potential biases.

      The performance of AI language models like GPT 3 can vary greatly depending on the specific task and the data they are trained on. During a discussion, it was noted that a model specifically trained for a task, like the Delphi model, outperformed GPT 3 in certain areas, such as common sense reasoning and understanding context. However, GPT 3, being a general-purpose model, can still be effective in many scenarios with the right prompts and examples. The discussion also touched on the challenges of understanding why some topics might be easier or harder for AI models to grasp. Researchers are still exploring this area, as the inner workings of these models are not always clear. One factor that can influence a model's performance is the availability and quality of data related to the topic. Adversarial examples, where the model is presented with deliberately misleading information, can also reveal unexpected behaviors. As AI models continue to evolve, it will be essential to consider their ethical implications and potential biases. Some researchers are even exploring the idea of AI models creating their own ethics principles. Overall, the conversation highlighted the importance of ongoing research and development in the field of AI language models and the need to address the complexities and nuances of human language and common sense reasoning.

    • AI investment in drug design and discovery reaches $13.8 billion in 2020Significant investment surge in AI for drug design and discovery, driven by recent health events and ethical discussions, while privacy-focused browsing and ad-blocking features gain attention.

      The latest AI research and investment trends indicate a significant focus on drug design and discovery, with over $13.8 billion invested in 2020, a 4.5x increase from 2019. This shift could be attributed to recent health-related events, such as the COVID-19 pandemic. Another intriguing development is the ongoing discussion about creating AI models that reflect ethical norms for future societies, as our culture adapts to the increasing pervasiveness of automation. The article "Delphi towards Machine Ethics and Norms" touches on this, and the authors plan to expand their dataset to improve transparency and explainability. Meanwhile, the Brave team is working on a better internet by offering privacy-focused browsing and ad-blocking features, allowing users to support content creators through an opt-in reward system. Lastly, the Stanford Institute for Human-Centered Artificial Intelligence released an updated AI Index Report, highlighting these trends and more. It's essential for practitioners to stay informed about these developments and their implications for the future of AI.

    • The Application of AI in Bioinformatics and Its Impact on IndustryPhD graduates in AI are increasingly going into industry for cutting-edge research, higher salaries, and less teaching, leading to a trend of AI expertise in industry and potential competition for academic positions.

      The application of AI in fields like protein folding, genomics, and bioinformatics is increasing, particularly in the commercial sector. This shift is due to the large and complicated nature of the data involved, which is well-suited to AI's capabilities. Additionally, there has been a significant increase in PhD graduates in the AI space going into industry instead of academia, with 65% going into industry in North America compared to 44.4% a decade ago. This trend is reflected in the guests on the show, many of whom have PhDs in related fields or have migrated into the field and are working in industry. The appeal of industry for new PhDs includes the ability to do cutting-edge research, higher salaries, and less teaching. The impact of this trend on academia and the competition for academic positions remains to be seen. Another notable trend mentioned is the rise of generative AI in various applications.

    • The Blurred Line Between Human and AI-Generated ContentAI's ability to generate human-like text, audio, and images raises ethical concerns, particularly in the spread of misinformation. Prioritizing ethical principles, addressing biases, and fostering diversity are crucial for responsible use.

      The distinction between human-generated and AI-generated text, audio, and images has become increasingly blurred, making it difficult to discern the difference. This multimodal capability of AI is a double-edged sword, as it can lead to impressive advancements but also raises concerns about the potential spread of misinformation. Ethical considerations are paramount, as governments, corporations, and individuals grapple with the implications of these tools. In the short term, the generation of convincing misinformation is a significant concern, as it can contribute to the breakdown of trust and understanding among individuals and communities. As we continue to explore the potential of AI, it's crucial to prioritize ethical principles, address biases, and foster a diverse talent pool to ensure that these powerful tools are used responsibly and for the greater good.

    • Advancements in AI technology lead to faster training timesFaster training times enable researchers to experiment with more models and parameters, potentially leading to better outcomes, but also raise concerns about sustainability due to increased power consumption.

      The advancements in AI technology, specifically in the area of faster training times, are opening up new possibilities for researchers and developers. According to a report titled "15 Graphs You Need to See to Understand AI in 2021," the time it takes to train state-of-the-art models on standard datasets has drastically decreased. For instance, training a model on the ImageNet dataset took 6.2 minutes in 2018 and only 47 seconds in 2020. This progress can be attributed to advancements in accelerator chips, distributed training, and specialized hardware. The implication of faster training times is that researchers can experiment with different parameters and models more frequently, potentially leading to better outcomes. However, this also raises concerns about sustainability, as more models may be trained in a given time frame, consuming more power. Overall, the ability to train models faster provides more options for researchers and developers, and this trend is unlikely to slow down. Another interesting point from the report is the idea that we are currently living in an "AI summer," where AI research is experiencing a surge in growth, as evidenced by the increasing number of citations in academic papers related to AI.

    • AI research publications and citations on the riseSince 2000, AI research has seen a steady increase in publications and citations, with a recent surge in 2019 and 2020. China leads in published research, but unpublished R&D in other countries may be significant.

      The number of AI research publications and citations has been on the rise since the early 2000s, with a notable dip around 2015-2018, followed by a sharp increase in 2019 and 2020. This trend suggests that the perception of AI advancements may trail behind the actual research. Furthermore, China has taken the lead in terms of AI research publications due to their emphasis on journal publications, while a significant portion of AI R&D in the US occurs in corporations, which may prioritize trade secrets over publishing. The report also highlights the growing global AI job market. Despite China's dominance in published research, the extent of unpublished research on the non-China side remains a question. Overall, these trends indicate the continued growth and evolution of AI technology and research.

    • AI talent development in diverse countriesGrowing focus on AI talent in Brazil, India, Canada, Singapore, and South Africa, signaling strong representation in the global AI job market. Diversity challenge persists with majority of US AI PhD graduates coming from abroad and staying in the US, possibly due to shift towards data science as post-undergrad career path in US.

      There is a growing focus on developing AI talent in countries like Brazil, India, Canada, Singapore, and South Africa, as indicated by the AI Index. This trend is significant because it signals a strong representation of Asia and Latin America in the AI job market. Additionally, there is a noted diversity challenge in the AI field, with the majority of US AI PhD graduates coming from abroad and staying in the US. The reasons for this trend are complex, but it seems that there may be a shift towards data science as a post-undergrad career path in the US, leading to fewer PhD students in AI. Another important consideration not mentioned in the report is the concern around job security and ethical implications of AI, which is a major concern for many people in the field.

    • Impact of AI on jobs may be more nuanced than expectedAI is automating mundane tasks but not necessarily replacing jobs, instead, roles may change, Hugging Face course offers valuable insights into AI and NLP

      While there is ongoing concern about the impact of AI on jobs and the potential for automation to replace certain positions, the reality may be more nuanced. The availability or competition for positions might not be the primary concern, as deep learning deployment has become cheaper and more common in automating mundane tasks. However, this doesn't necessarily mean that jobs are being completely taken away but rather morphing into something else. Companies introducing automation may still require a similar workforce, albeit with different responsibilities. The Hugging Face course on transformer models is a valuable resource for those interested in learning about AI and natural language processing, and it includes various components like videos, text, and images. This timely and relevant learning opportunity covers topics such as transformer models, bias and limitations, and fine-tuning pre-trained models. Check it out for a deeper understanding of AI and its applications.

    • Access all Changelog podcasts in one place with Changelog MasterSubscribe to Changelog Master to automatically download and manage all Changelog podcast episodes in your preferred podcast app

      Changelog Master is a podcast aggregator where you can access all Changelog podcasts in one place. By subscribing to Changelog Master, your podcast app will automatically download all the episodes produced by Changelog, and you can then select which ones you want to listen to. You can find Changelog Master by searching for "changelogmaster" in your preferred podcast app or visiting changelog.com/master. Additionally, the podcast is brought to you by sponsors Fastly, LaunchDarkly, and Linode. The music for the podcast is provided by BRAKEmaster Cylinder. Tune in next time for another informative episode.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Vectoring in on Pinecone

    Vectoring in on Pinecone
    Daniel & Chris explore the advantages of vector databases with Roie Schwaber-Cohen of Pinecone. Roie starts with a very lucid explanation of why you need a vector database in your machine learning pipeline, and then goes on to discuss Pinecone’s vector database, designed to facilitate efficient storage, retrieval, and management of vector data.

    Stanford's AI Index Report 2024

    Stanford's AI Index Report 2024
    We’ve had representatives from Stanford’s Institute for Human-Centered Artificial Intelligence (HAI) on the show in the past, but we were super excited to talk through their 2024 AI Index Report after such a crazy year in AI! Nestor from HAI joins us in this episode to talk about some of the main takeaways including how AI makes workers more productive, the US is increasing regulations sharply, and industry continues to dominate frontier AI research.

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.