Logo
    Search

    Create AI art with an app, an AI-powered game platform, the year of monster AI models, better images of AI

    enJanuary 01, 2022

    Podcast Summary

    • AI-generated art goes viral with Dream by Wombo.aiText-to-image AI model Dream by Wombo.ai has gone viral, generating over 10 million images from user text prompts, marking a significant step forward in the mainstream adoption of AI technology and opening up new opportunities for innovation and creativity.

      AI technology is increasingly being integrated into various industries and applications, making it more accessible and mainstream. One such application is in the field of art, where a new app called Dream by Wombo.ai allows users to generate artwork from text prompts. This text-to-image AI model, which was previously popular among researchers and hackers, has now gone viral on social media and has already generated over 10 million images. The app's smooth implementation and ability to transform text into art has captured the imagination of users, making AI-generated art a reality for consumers. This is an exciting development for researchers in the field, as it marks a significant step forward in the mainstream adoption of AI technology. Additionally, the app's success highlights the potential of AI to revolutionize various industries and create new opportunities for innovation and creativity.

    • AI-generated content and gaming innovationsNew apps and platforms are using AI to generate content and create games, making it more accessible and convenient for users. These developments could disrupt traditional content creation and game design industries.

      There are exciting developments in the realm of AI-generated content and gaming. The first discussion revolves around a potential new app that allows users to generate their own content, with the possibility of more features and capabilities being added in the future. The convenience of using an app for this purpose was highlighted, and it was suggested that this technology could become more widely adopted in the coming year. The second story introduces Latitude, a company behind AI Dungeon, an interactive text-based game where the AI narrates the story and the user plays a character. Latitude has now launched a closed beta for an experimental AI-powered game platform called Voyage, which aims to create more structured games with defined goals and win conditions. The long-term vision for Voyage is to enable creators to make dynamic and live games using natural language, making game creation more accessible to a wider audience. Both developments showcase the potential of AI to revolutionize content creation and gaming, making it more accessible and convenient for individuals, and potentially disrupting traditional game design platforms.

    • New AI game and journal aim to streamline research processA new AI game, AI Dungeon, and journal, Transactions on Machine Learning Research (TMLR), aim to simplify and improve the process of publishing and reviewing AI research through continuous submissions, faster review turnaround times, and certifications for high-quality papers.

      There are ongoing efforts to simplify and improve the process of publishing and reviewing research in the field of Artificial Intelligence (AI). This was highlighted in a discussion about a new AI-focused game, AI Dungeon, and the announcement of a new journal, Transactions on Machine Learning Research (TMLR). The creators of AI Dungeon expressed their excitement about making a game more suitable for AI interaction and the potential for innovative games. On the research side, TMLR aims to address issues with the current conference-dominated publication process by allowing for continuous submissions and faster review turnaround times. The journal also plans to introduce certifications for high-quality papers and reproducibility, which could lead to valuable insights and improved baselines. Overall, these developments reflect a growing recognition of the need for more efficient and effective methods for sharing and validating AI research.

    • New research platforms and smaller AI models gaining popularitySmaller, more accessible research platforms and AI models are emerging, offering inclusive and efficient solutions, and delivering valuable results with less resources.

      The research community is seeing a shift towards more accessible and less resource-intensive research and publishing platforms, as well as the emergence of smaller, yet effective AI models. The first story discusses the launch of a new conference and journal platform, "arXiv-style for CS," which aims to make the research process more inclusive and efficient by accepting shorter papers and utilizing a team of reviewers throughout the year. This approach is seen as a solution to the challenges of traditional conferences, where only larger, more comprehensive studies are typically published due to the lengthy review process. The second story highlights DeepMind's new language AI model, called "retro," which is smaller in size, with only 7 billion parameters, compared to the gigantic models with 178 or even 280 billion parameters that have been released this year. Despite its smaller size, retro outperforms larger models by utilizing a retrieval database of two trillion words. This allows the model to access and use related information from the database when creating outputs, making it a viable alternative to larger models with significant data requirements. These developments in the research community show that smaller, more accessible platforms and models can still yield valuable results and make a significant impact in the field. It remains to be seen how these new initiatives will be received and how they will shape the research landscape moving forward.

    • DeepMind's progress in language models and UN's stalled efforts on autonomous weaponsDeepMind advanced language models with a massive 280 billion parameter model, while UN discussions on regulating autonomous weapons stalled due to opposition from major powers.

      DeepMind made significant strides in the field of language models this year, with the training of a gigantic 280 billion parameter model and the release of papers discussing the benefits and risks of such large models. However, progress towards regulating autonomous "killer robots" stalled at the UN Convention on Certain Conventional Weapons, with the US, Russia, and India opposing the creation of regulations for these weapons. Despite the lack of progress, it's predicted that public interest and media attention will be necessary for any meaningful regulations to be implemented. The trend of large language models and the debate over autonomous weapons mark important inflection points in the fields of AI and robotics.

    • Concerns over privacy and ethics with increasing use of facial recognition and autonomous weapon systemsAs facial recognition and autonomous weapon systems become more prevalent, ethical concerns and potential privacy violations need to be addressed to prevent misuse and negative consequences in the future.

      While there are efforts to ban facial recognition technology on a local level, its use is rapidly increasing, particularly in airports and for identity verification. The concern is that as it becomes more ubiquitous, it may be harder to regulate bad uses and potential privacy violations. Autonomous weapon systems, on the other hand, are not yet in widespread use, but the potential for their use in future conflicts is a cause for concern. The human tendency to react to crises rather than take preventative measures applies to both technologies. It's important to consider the ethical implications and potential consequences of these technologies as they continue to develop and be implemented.

    • Promoting accurate representations of AI through stock photosThe Better Images of AI project provides free, realistic stock photos to challenge misconceptions and promote accurate understanding of AI as object recognition technology

      The Better Images of AI project aims to provide more accurate and realistic images of AI for journalists and the general public. This project, led by the BBC Research and Development Group, VB and AI UK non-profit, and the Hume Center for the Future of Intelligence, intends to challenge the common misconceptions of AI as being related to robots and brains, instead showcasing it as object recognition technology. The project's goal is to offer Creative Commons licensed stock photos, featuring images like bounding boxes around everyday objects, to create a more accurate representation of AI. This initiative is an essential step towards correcting the public's perception of AI, as misconceptions can lead to confusion and misunderstanding of the technology's capabilities. Overall, the Better Images of AI project is a significant step forward in promoting accurate and realistic portrayals of AI in media and society.

    • Engage with listeners for podcast successSubscribe, leave reviews, provide feedback, and share podcasts to support creators and build a community

      Listener engagement is crucial for a podcast's success. The speakers in this discussion encouraged listeners to subscribe, leave reviews, and provide feedback. They acknowledged that even negative feedback is valuable. By engaging with listeners, podcasters can build a community and grow their audience. Additionally, they emphasized the importance of consistency, with plans to continue producing episodes in the new year. So, if you enjoy a podcast, make sure to subscribe, leave a review, and share it with others. Your feedback matters and helps support the creators.

    Recent Episodes from Last Week in AI

    #171 - - Apple Intelligence, Dream Machine, SSI Inc

    #171 - - Apple Intelligence, Dream Machine, SSI Inc

    Our 171st episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 24, 2024

    #170 - new Sora rival, OpenAI robotics, understanding GPT4, AGI by 2027?

    #170 - new Sora rival, OpenAI robotics, understanding GPT4, AGI by 2027?

    Our 170th episode with a summary and discussion of last week's big AI news!

    With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)

    Feel free to leave us feedback here.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 09, 2024

    #169 - Google's Search Errors, OpenAI news & DRAMA, new leaderboards

    #169 - Google's Search Errors, OpenAI news & DRAMA, new leaderboards

    Our 168th episode with a summary and discussion of last week's big AI news!

    Feel free to leave us feedback here: https://forms.gle/ngXvXZpNJxaAprDv6

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enJune 03, 2024

    #168 - OpenAI vs Scar Jo + safety researchers, MS AI updates, cool Anthropic research

    #168 - OpenAI vs Scar Jo + safety researchers, MS AI updates, cool Anthropic research

    Our 168th episode with a summary and discussion of last week's big AI news!

    With guest host Gavin Purcell from AI for Humans podcast!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + Links:

    Last Week in AI
    enMay 28, 2024

    #167 - GPT-4o, Project Astra, Veo, OpenAI Departures, Interview with Andrey

    #167 - GPT-4o, Project Astra, Veo, OpenAI Departures, Interview with Andrey

    Our 167th episode with a summary and discussion of last week's big AI news!

    With guest host Daliana Liu (https://www.linkedin.com/in/dalianaliu/) from The Data Scientist Show!

    And a special one-time interview with Andrey in the latter part of the podcast.

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 19, 2024

    #166 - new AI song generator, Microsoft's GPT4 efforts, AlphaFold3, xLSTM, OpenAI Model Spec

    #166 - new AI song generator, Microsoft's GPT4 efforts, AlphaFold3, xLSTM, OpenAI Model Spec

    Our 166th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 12, 2024

    #165 - Sora challenger, Astribot's S1, Med-Gemini, Refusal in LLMs

    #165 - Sora challenger, Astribot's S1, Med-Gemini, Refusal in LLMs

    Our 165th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enMay 05, 2024

    #164 - Meta AI, Phi-3, OpenELM, Bollywood Deepfakes

    #164 - Meta AI, Phi-3, OpenELM, Bollywood Deepfakes

    Our 164th episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 30, 2024

    #163 - Llama 3, Grok-1.5 Vision, new Atlas robot, RHO-1, Medium ban

    #163 - Llama 3, Grok-1.5 Vision, new Atlas robot, RHO-1, Medium ban

    Our 163rd episode with a summary and discussion of last week's big AI news!

    Note: apology for this one coming out a few days late, got delayed in editing it -Andrey

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 24, 2024

    #162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter

    #162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter

    Our 162nd episode with a summary and discussion of last week's big AI news!

    Read out our text newsletter and comment on the podcast at https://lastweekin.ai/

    Email us your questions and feedback at contact@lastweekin.ai and/or hello@gladstone.ai

    Timestamps + links:

    Last Week in AI
    enApril 15, 2024

    Related Episodes

    Toni Townes-Whitley (Microsoft) - The Ethics of Innovation

    Toni Townes-Whitley (Microsoft) - The Ethics of Innovation
    How often do entrepreneurs and corporate leaders think about issues like fairness, accessibility or unseen biases in the technologies they invent and advance? That’s the challenge for companies leading the digital transformation that’s disrupting every aspect of society, says Toni Townes-Whitley, Corporate Vice President of Worldwide Public Sector and Industry at Microsoft, in this talk about innovating strategically and responsibly.

    Will.i.am: On Power of AI

    Will.i.am: On Power of AI

    He’s a rapper, a producer, an entrepreneur, and a founding member of the Black Eyed Peas.

    Will.I.Am has worked with some of the biggest names in the business - from Michael Jackson and Britney to Kanye and Rihanna.

    He’s dipped his toes in the shark-infested waters of politics, too. Backing Obama, Biden and Hillary… while lampooning Donald Trump in an infamous music video.

    These days his eyes are firmly on the future. He’s an A.I evangelist, and the co-presenter of his new radio show is a robot!

    Now Will.I.Am is Uncensored.



    YouTube: @PiersMorganUncensored

    X: @PiersUncensored

    TikTok: @piersmorganuncensored

    Insta: @piersmorganuncensored



    Hosted on Acast. See acast.com/privacy for more information.


    Joe DeSimone (Carbon) - The Case for Convergence

    Joe DeSimone (Carbon) - The Case for Convergence

    Joe DeSimone is the founder and executive chairman of Carbon, a global company that is driving the evolution of 3D printing from a prototyping tool into a scalable manufacturing technology. As a professor at the University of North Carolina, DeSimone made scientific breakthroughs in areas including green chemistry, medical devices, and nanotechnology, also co-founding several companies based on his research. In 2016 President Obama awarded him the National Medal of Technology and Innovation, the highest honor in the U.S. for achievement and leadership in advancing technological progress. In this talk, he explores how diverse teams, perspectives and specialties can drive innovations in both technologies and business models.

    Steven Johnson: How We Got to Now

    Steven Johnson: How We Got to Now

    Glass, Cold, Sound, Clean, Time and Light. Just a random collection of items in the physical world, right? Not in the hands of science journalist Steven Johnson, who weaves a revelatory tale of technological wonders in his new book and PBS series, "How We Got to Now: Six Innovations That Made the Modern World." Join Julie and Robert for an interview with the man behind the book and series.

    Learn more about your ad-choices at https://www.iheartpodcastnetwork.com

    See omnystudio.com/listener for privacy information.

    CL219: SuperIntelligence And The Ethics Of AI - Interview with Greg Keiser

    CL219: SuperIntelligence And The Ethics Of AI - Interview with Greg Keiser

    Greg Kieser is the founder of Supersystemic.ly, a Brooklyn-based think-tank and angel investment firm dedicated to increasing humanity’s readiness for the emergence of superintelligent entities through the study and spread of “supersystemic” perspectives and innovations. Kieser previously oversaw a portfolio of technology initiatives at an NYC-based poverty-fighting foundation. His new book, Dear Machine: A Letter to a Super-Aware/Intelligent Machine sets a course for symbiosis with superintelligent machines. Learn more at www.supersystemic.ly/dear-machine/

    In this episode, we cover:

    1:57 – Complex systems
    2:55 – An optimistic view of superintelligent machines
    4:14 – Letters to a machine
    7:44 – UBI
    11:50 – AI and a period of strife
    15:59 – The ethics of AI
    16:54 – Write what you know

    Resources mentioned in this episode:

    supersystemic.ly
    Tools: Gmail
    Book: The Selfish Gene by Richard Dawkins
    Album: Spotify Discover

    If you enjoyed the show, please rate it on YouTube, iTunes or Stitcher and write a brief review. That would really help get the word out and raise the visibility of the Creative Life show.

    The post CL219: SuperIntelligence And The Ethics Of AI - Interview with Greg Keiser appeared first on James Taylor.