Logo
    Search

    Podcast Summary

    • Photonic Computing: A Solution to the Energy Consumption Challenge in AIPhotonic computing, using light instead of electricity for computation, is a promising solution to address the energy consumption issue in the rapidly growing AI market.

      The rapid advancement of AI technology, driven by the development of increasingly complex neural networks, has led to a significant increase in computational power requirements. Traditional transistor-based computers are reaching their energy efficiency limits, and the Department of Energy predicts that by 2040, most of the world's energy will be used for compute and interconnect. Nick Harris, CEO of Lightmatter, is working on a solution to this challenge: photonic computing. He got his start as an engineer at Micron, where he became familiar with the challenges of shrinking transistors. He then went to MIT to study quantum computing, though Lightmatter's work isn't quantum computing per se. Instead, they're using photonics for computation, a unique approach to tackling the energy consumption issue in the growing AI market. The motivation behind photonic computing is to build more energy-efficient AI supercomputers, addressing the capital-intensive nature of current AI infrastructure.

    • MIT researchers discover method to use quantum processors for neural networks with lasersMIT researchers found a way to use quantum processors for neural networks using lasers, resulting in more efficient and scalable AI solutions

      Researchers at MIT have discovered a way to use the same processors used for quantum computing to process neural networks using lasers, leading to significant benefits. This approach, known as photonic computing, leverages the same infrastructure used in traditional communications, such as lasers and silicon photonics, to perform the core computations in deep learning. Unlike quantum computing, which uses single photons, photonic computing for neural networks uses lasers. Deep learning, which is the foundation of many AI models, relies on multiplication and addition, making it a good fit for photonic computing. The form factor of photonic computers is more similar to standard computer chips, making them a better fit for hyperscaler and cloud environments compared to the large and unusual setups required for quantum computing. This development could lead to more efficient and scalable AI solutions.

    • Challenges in Scaling Up AI Systems: Power Consumption and Thermal IssuesThe scaling of AI systems faces significant challenges due to power consumption and thermal issues, limiting our ability to increase compute power while maintaining energy efficiency and reducing environmental impact. Potential solutions include photonics, but energy consumption remains a concern.

      As we strive to power larger and more complex artificial intelligence (AI) systems, such as 500 billion weight neural networks, the primary challenge lies in the immense computing power required and the resulting heat generation. Traditional computer chips are becoming increasingly power-hungry and large in size, leading to significant thermal issues. This issue, combined with Ohmdel's law scaling, which states that adding more compute units does not result in linear performance gains, limits our ability to scale up and contributes to environmental concerns due to the high energy consumption. Photonics, or laser-driven computing, offers a potential solution as it doesn't generate heat like traditional chips. However, it's important to note that lasers still require power. The energy scaling problem, driven by Moore's law, which dictates that transistors should be shrunk every 18 months and use less energy, has been a challenge since around 2005. Dennard scaling, which aimed to maintain constant power per transistor as transistors shrank, has not been achieved, leading to overheating issues. These challenges limit our ability to scale up AI systems while maintaining energy efficiency and reducing environmental impact.

    • Managing the Energy Consumption Challenge in AIAI's energy consumption is projected to consume up to 10% of the planet's energy by 2030, prompting the need for new solutions like photonic computing to eliminate energy scaling issues and reduce overall consumption.

      The current state of AI technology is facing a significant challenge in managing the enormous amounts of power required for computation. Traditional methods, such as water and air cooling, are reaching their limits, and new solutions like immersion cooling are emerging. However, these methods still rely on transistors, which are facing a fundamental limit in energy efficiency due to Dennard scaling. This issue is expected to consume up to 10% of the entire planet's energy by 2030, making it a pressing concern for businesses and researchers alike. Photonic computing, which uses optics instead of transistors, is seen as a potential solution to this problem as it eliminates the energy scaling issue. While progress is being made, it's still an active area of research and development. The shift towards photonics could revolutionize the field of AI by reducing energy consumption and enabling more efficient computations. However, it's important to note that this is a complex issue with many moving parts, and a complete transition to photonic computing is likely to take time.

    • Leading company in photonic chips for AI applicationsLightmatter's silicon photonics technology powers trillion-parameter neural networks, offering a valuable asset for major companies, despite challenges in architecture, team building, and supply chain.

      Lightmatter is a company leading the way in developing and commercializing photonic chips for energy-intensive AI applications. This technology, which is based on silicon photonics, has been in development for over a decade and has already shown promising results in running state-of-the-art neural networks. The processors are capable of powering trillion-parameter neural networks and beyond, making them a valuable asset for major companies. The journey to this point has not been easy, with challenges including figuring out the photonic compute architecture, building teams, and establishing a supply chain. The chips, which are similar in some ways to Google's TPUs, are matrix processors that primarily focus on linear algebra. They consist of multiple processor cores and offer additional capabilities beyond just matrix processing. Despite the complexity and challenges, Lightmatter is confident in the future of this technology and is gearing up to deliver their processors to customers.

    • Light-based quad core computer with unique challengesA light-based quad core computer with intriguing properties includes challenges like preventing interference, designing analog and digital circuits, and getting light into optical wires.

      The discussed product is a unique type of computer that utilizes a quad core design with each core dedicated to linear algebra, but instead of using electrical signals like traditional CPUs, it uses light. The chip appears similar to a Tensor Processing Unit (TPU) with a two-dimensional array of multiply-accumulate units, but with the added visual aspect of light being distributed to each component. However, despite using light, there are still challenges in designing and building the physical unit. Interference from unwanted signals through antennas is a concern, but since light signals don't interfere with each other, it's an intriguing property. The chip also consists of analog and digital circuits, as well as photonic circuits, which require careful consideration to prevent unwanted signals. The optical wires on the chip, though tiny, are challenging to get light into, which can be both fortunate and unfortunate. As technology continues to shrink, there are limitations to consider, such as the difficulty of getting light into increasingly small optical wires, and the potential for new challenges as we explore shorter wavelengths.

    • Developing software for photonic computersLight Matters' idiom SDK enables users to import neural networks from PyTorch and TensorFlow for use on their photonic computer, Envise, simplifying the process of software development in this emerging technology field.

      The development of photonic computers, which use light instead of electricity to process information, holds great promise due to their potential for high clock frequencies and the ability to process multiple colors of light at once. However, the physical production of these computers is challenging, and the integration of software and tooling for use with popular machine learning frameworks like PyTorch and TensorFlow is a significant undertaking. Light Matters, a company working on this technology, has developed a software development kit called idiom, which allows users to import neural networks built in PyTorch and TensorFlow and compile them for use on their photonic computer, Envise. Despite the complexities, the trend in the industry is towards prioritizing the delivery of fast, reliable, and user-friendly technology, with a growing focus on software development.

    • Real-world application and deployment of AI models through inferenceInference is where AI models interact with the world and provide results to users, making it the stage for real-world impact and scale of AI, while the development of advanced multiplexers or demultiplexers is crucial for expanding the color and frequency range in photonic processors.

      While the process of training neural networks is essential for creating AI models, the real-world application and deployment of these models through inference is where most of the energy consumption and practical use cases lie. For data scientists, software developers, and deep learning engineers, the workflow remains largely the same when using frameworks like PyTorch or TensorFlow, but the integration of hardware like Onyx and photonic processors can significantly improve productivity and scalability without thermal limitations. However, the detection of various light wave frequencies and colors using receivers is the current limiting factor, and the development of more advanced multiplexers or demultiplexers will be crucial for expanding the color and frequency range. Inference, as the deployment of trained models, is where most of the energy footprint and economic value of AI will be found, as it's the stage where AI models interact with the world and provide results to users. Training is an essential R&D phase, but inference is where the real-world impact and scale of AI occur.

    • The Future of Computing: Photonic ComputingPhotonic computing, using light for processing and interconnects, offers power efficiency and speed improvements, but faces challenges in scaling beyond AI applications. It may not replace traditional computing entirely, but will coexist with it.

      The future of computing, specifically in the field of AI, is heading towards photonic computing, which uses light for processing and interconnects. This technology, represented by the chip, Passage, and Idiom software integration, offers significant improvements in power efficiency and speed compared to current solutions. For instance, the Envise server, which uses this technology, consumes only a fraction of the power and delivers faster results than NVIDIA's DGX A100 server. However, the challenge lies in scaling this technology beyond AI applications, as general-purpose computing, like running Windows or playing video games, requires non-linear operations that are difficult to achieve with optics. As a result, we can expect a future with multiple competing technologies, each excelling in specific areas. This shift in technology might call for new skills and adaptations for professionals in the industry. While photonic computing is currently targeted at AI, it may not replace traditional computing entirely, as each technology will have its strengths and limitations.

    • Insights from a Pioneer in Optoelectronics on the Future of ComputingDr. Nick Holonyak discusses the future of computing, expressing his belief in the dominance of photonic computers for AI and his enthusiasm for quantum computing, while encouraging everyone to learn more about Light Matters processors.

      The future of computing will involve a variety of technologies each suited to different types of problems. Dr. Nick Holonyak, a pioneer in the field of optoelectronics, shared his insights on the upcoming dominance of photonic computers in AI due to their compatibility with deep learning mathematics. He also expressed his enthusiasm for quantum computing, albeit with a more cautious timeline. Dr. Holonyak's ultimate goal is to have Google run on Light Matters processors, and he encourages everyone to learn more about this innovative technology by checking out their website. As transistors reach their limits, new technologies like analog electronics, digital electronics, and photonic compute units will compete to solve various computational problems. The Changelog team was impressed by the advancements in this field and invited Dr. Holonyak to potentially discuss progress next year. For more information, listeners can find links to Lightmatter and related benchmarks in the show notes. Special thanks to BRAKEmaster Cylinder for the music and to Fastly, LaunchDarkly, and Linode for their long-term sponsorship. Stay tuned for more Changelog podcasts.

    Recent Episodes from Practical AI: Machine Learning, Data Science

    Vectoring in on Pinecone

    Vectoring in on Pinecone
    Daniel & Chris explore the advantages of vector databases with Roie Schwaber-Cohen of Pinecone. Roie starts with a very lucid explanation of why you need a vector database in your machine learning pipeline, and then goes on to discuss Pinecone’s vector database, designed to facilitate efficient storage, retrieval, and management of vector data.

    Stanford's AI Index Report 2024

    Stanford's AI Index Report 2024
    We’ve had representatives from Stanford’s Institute for Human-Centered Artificial Intelligence (HAI) on the show in the past, but we were super excited to talk through their 2024 AI Index Report after such a crazy year in AI! Nestor from HAI joins us in this episode to talk about some of the main takeaways including how AI makes workers more productive, the US is increasing regulations sharply, and industry continues to dominate frontier AI research.

    Apple Intelligence & Advanced RAG

    Apple Intelligence & Advanced RAG
    Daniel & Chris engage in an impromptu discussion of the state of AI in the enterprise. Then they dive into the recent Apple Intelligence announcement to explore its implications. Finally, Daniel leads a deep dive into a new topic - Advanced RAG - covering everything you need to know to be practical & productive.

    The perplexities of information retrieval

    The perplexities of information retrieval
    Daniel & Chris sit down with Denis Yarats, Co-founder & CTO at Perplexity, to discuss Perplexity’s sophisticated AI-driven answer engine. Denis outlines some of the deficiencies in search engines, and how Perplexity’s approach to information retrieval improves on traditional search engine systems, with a focus on accuracy and validation of the information provided.

    Using edge models to find sensitive data

    Using edge models to find sensitive data
    We’ve all heard about breaches of privacy and leaks of private health information (PHI). For healthcare providers and those storing this data, knowing where all the sensitive data is stored is non-trivial. Ramin, from Tausight, joins us to discuss how they have deploy edge AI models to help company search through billions of records for PHI.

    Rise of the AI PC & local LLMs

    Rise of the AI PC & local LLMs
    We’ve seen a rise in interest recently and a number of major announcements related to local LLMs and AI PCs. NVIDIA, Apple, and Intel are getting into this along with models like the Phi family from Microsoft. In this episode, we dig into local AI tooling, frameworks, and optimizations to help you navigate this AI niche, and we talk about how this might impact AI adoption in the longer term.

    AI in the U.S. Congress

    AI in the U.S. Congress
    At the age of 72, U.S. Representative Don Beyer of Virginia enrolled at GMU to pursue a Master’s degree in C.S. with a concentration in Machine Learning. Rep. Beyer is Vice Chair of the bipartisan Artificial Intelligence Caucus & Vice Chair of the NDC’s AI Working Group. He is the author of the AI Foundation Model Transparency Act & a lead cosponsor of the CREATE AI Act, the Federal Artificial Intelligence Risk Management Act & the Artificial Intelligence Environmental Impacts Act. We hope you tune into this inspiring, nonpartisan conversation with Rep. Beyer about his decision to dive into the deep end of the AI pool & his leadership in bringing that expertise to Capitol Hill.

    Full-stack approach for effective AI agents

    Full-stack approach for effective AI agents
    There’s a lot of hype about AI agents right now, but developing robust agents isn’t yet a reality in general. Imbue is leading the way towards more robust agents by taking a full-stack approach; from hardware innovations through to user interface. In this episode, Josh, Imbue’s CTO, tell us more about their approach and some of what they have learned along the way.

    Related Episodes

    When data leakage turns into a flood of trouble

    When data leakage turns into a flood of trouble
    Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. It’s the kind of topic that we don’t often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.

    Stable Diffusion (Practical AI #193)

    Stable Diffusion (Practical AI #193)
    The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL·E 2). (Image from stability.ai)

    AlphaFold is revolutionizing biology

    AlphaFold is revolutionizing biology
    AlphaFold is an AI system developed by DeepMind that predicts a protein’s 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.

    Zero-shot multitask learning (Practical AI #158)

    Zero-shot multitask learning (Practical AI #158)
    In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.