Logo
    Search

    If everyone is building AI, why aren't more projects in production?

    enApril 17, 2024

    Podcast Summary

    • Large Enterprise Adoption of LLMs and GNAIEarly enterprise adoption of LLMs and GNAI focuses on productivity gains and cost optimization. Future goals include improving customer interactions and creating new product offerings. Talent with MLOps or LLM instrumentation expertise is in high demand.

      While the adoption of large language models (LLMs) and generative AI (GNAI) is growing, the business case for investing in operationalizing these technologies within large enterprise organizations is still being explored. The early stages of adoption are focused on productivity gains through simple solutions and optimizing cost bases. However, as we move further, the goal is to improve customer interactions through applications like customer support, HR automation, and legal automation. The aspirational goal is to create new product offerings and monetization opportunities, but we are still at least 18 to 24 months away from seeing these trends. From a talent perspective, there is a clear gap in the market for developers with expertise in MLOps or LLM instrumentation, making it an attractive area for companies looking to acquire and retain top talent. Additionally, developers are drawn to the fresh and challenging nature of this field.

    • AI tool implementation challengesDespite advancements in AI tools, large-scale implementation remains a challenge due to issues like talent gap, data quality, availability, and extensive prompt engineering and integration. Approx. 25-30% of applications don't make it to production due to non-technical reasons.

      While AI tools are making significant strides in areas like prompt engineering, image and text generation, and code generation, reaching production at scale remains a significant challenge. Talent gap is a top issue identified by experts, and the developer audience is particularly ripe for AI tools targeted at developer tooling. However, achieving large-scale implementation requires addressing challenges such as data quality and availability, as well as the need for extensive prompt engineering and integration. Companies like MongoDB are working to help AI tools better understand their products by providing large amounts of data and natural language prompts. But this requires a shift in mindset and brings its own challenges of integration, scale, and maintenance. Data foundation and data maturity are crucial for taking even simple use cases into production, and the challenges become significantly different when considering enterprise-focused solutions. Approximately 25-30% of applications do not make it from ideation to deployment at scale due to non-technical reasons.

    • Model implementation challenges in enterprisesEnterprises face challenges in implementing generative AI at scale due to data governance, model life cycle management, and model selection. Standardization and experimentation in a safe environment are necessary for effective model selection.

      While the use of generative AI via cloud providers offers many benefits, such as easy access to advanced models and standardization, there are still significant challenges for enterprises in implementing it at scale. These challenges include ensuring data governance, managing the life cycle of models, and making the right model selection for specific use cases. The need for standardization arises from the fact that enterprises typically invest in a large number of applications across various functions, requiring a platform that offers choice and optionality in terms of models. The process of model design and selection is currently more art than science, with a growing number of models available and varying in types such as text, image, audio, and video. The ability to experiment with models in a safe space, like a model garden, is crucial for making the best choice for a particular use case. However, the lack of determinism in these models and the need for trial and error add to the complexity of implementation.

    • Machine learning model productionTransitioning a machine learning model to production involves estimating workload, deciding on the most cost-effective and suitable model, dealing with obsolescence risk, ongoing maintenance, and benchmarking different models.

      Implementing and managing machine learning models involves a significant amount of experimentation, iteration, and ongoing effort. While choosing a model for experimentation requires considering factors like fit and business value, transitioning that model to production brings new challenges such as estimating workload, deciding on the most cost-effective and suitable model for the long term, and dealing with the obsolescence risk of new models. The training and fine-tuning process can be time-consuming and require manual intervention, with windows for tuning provided by companies offering these models. After deployment, ongoing maintenance and upkeep are necessary considerations, including monitoring performance and addressing any potential issues. Additionally, the output from machine learning models can vary greatly, making it essential to benchmark and compare different models for specific use cases. The process of implementing machine learning models is an ongoing journey, requiring continuous evaluation and adaptation to new technologies and improvements.

    • AI model selection automationAI is now used to benchmark and train AI models for automating the selection process. Companies increasingly rely on AI Managed Service Providers for help in model evaluation, data foundation, and fine-tuning.

      The landscape of AI model selection, customization, and maintenance has evolved significantly. Early methods involved checking the similarity of models and making small tweaks for best practices. However, with the vast number of models available, automation is necessary. Now, AI is used to benchmark and train AI models. This process is complex and requires ongoing effort, but models today are more powerful, reducing the need for extensive optimization. Fine-tuning and customization are still required, but the cycle is faster than before. Companies are also turning to AI Managed Service Providers (MSPs) for help in model evaluation, data foundation, and fine-tuning. This trend is expected to continue as businesses aim to accelerate the adoption of AI solutions.

    • AI advancementsExpect enhancements in customization, control, cost optimization, and database integration leading to faster testing-production gap and new use cases, but regulatory and safety considerations and costs are important factors to consider.

      The Gen AI space is expected to see significant advancements in the next year, particularly in areas of enhanced customization and control, cost optimization, and database integration. These developments could lead to a decrease in the gap between testing and production, and the emergence of new use cases. However, the regulatory and safety aspects of AI are not to be ignored, and the costs involved will become increasingly important. The adoption of AI is inevitable for most companies, and the challenge lies in balancing the potential benefits against the risks and costs. The managed service providers are expected to play a crucial role in this process by offering end-to-end solutions and orchestration capabilities. Overall, the Gen AI space is in a period of rapid growth, and the next year is expected to bring significant progress and innovation.

    • MongoDB, Google Cloud resources for Gen AIExplore MongoDB's developer center for articles, how-tos, and podcasts, or experiment with Google Cloud's Gen AI models and connect with their team for discussions.

      Both MongoDB and Google Cloud offer valuable resources for those interested in Gen AI. MongoDB, represented by Toni Mirovaeva, invites you to explore their developer center filled with articles, how-tos, and even a podcast, the MortgageB podcast, where you can learn from their team, including Shane. Google Cloud, represented by Miku Ja, encourages you to experiment with their Gen AI models and connect with her on LinkedIn for further discussions. Both teams are dedicated to fostering a community of learning and exploration in the field of Gen AI. So, whether you're just starting out or looking to deepen your understanding, be sure to check out the resources from MongoDB and Google Cloud.

    Recent Episodes from The Stack Overflow Podcast

    How to build open source apps in a highly regulated industry

    How to build open source apps in a highly regulated industry

    Before Medplum, Reshma founded and exited two startups in the healthcare space – MedXT (managing medical images online acquired by Box) and Droplet (at-home diagnostics company acquired by Ro). Reshma has a B.S. in computer science and a Masters of Engineering from MIT.

    You can learn more about Medplum here and check out their Github, which has over 1,200 stars, here.

    You can learn more about Khilnani on her website, GitHub, and on LinkedIn.

    Congrats to Stack Overflow user Kvam for earning a Lifeboat Badge with an answer to the question: 

    What is the advantage of using a Bitarray when you can store your bool values in a bool[]?

    A very special 5-year-anniversary edition of the Stack Overflow podcast!

    A very special 5-year-anniversary edition of the Stack Overflow podcast!

    Cassidy reflect on her time as a CTO of a startup and how the shifting environment for funding has created new pressures and incentives for founders, developers, and venture capitalists.

    Ben tries to get a bead on a new Moore’s law for the GenAI era: when will we start to see diminishing returns and fewer step factor jumps? 

    Ben and Cassidy remember the time they made a viral joke of a keyboard!

    Ryan sees how things goes in cycles. A Stack Overflow job board is back! And what do we make of the trend of AI assisted job interviews where cover letters and even technical interviews have a bot in the background helping out.

    Congrats to Erwin Brandstetter for winning a lifeboat badge with an answer to this question:  How do I convert a simple select query like select * from customers into a stored procedure / function in pg?

    Say goodbye to "junior" engineering roles

    Say goodbye to "junior" engineering roles

    How would all this work in practice? Of course, any metric you set out can easily become a target that developers look to game. With Snapshot Reviews, the goal is to get a high level overview of a software team’s total activity and then use AI to measure the complexity of the tasks and output.

    If a pull request attached to a Jira ticket is evaluated as simple by the system, for example, and a programmer takes weeks to finish it, then their productivity would be scored poorly. If a coder pushes code changes only once or twice a week, but the system rates them as complex and useful, then a high score would be awarded. 

    You can learn more about Snapshot Reviews here.

    You can learn more about Flatiron Software here.

    Connect with Kirim on LinkedIn here.

    Congrats to Stack Overflow user Cherry who earned a great question badge for asking: Is it safe to use ALGORITHM=INPLACE for MySQL?

    Making ETL pipelines a thing of the past

    Making ETL pipelines a thing of the past

    RelationalAI’s first big partner is Snowflake, meaning customers can now start using their data with GenAI without worrying about the privacy, security, and governance hassle that would come with porting their data to a new cloud provider. The company promises it can also add metadata and a knowledge graph to existing data without pushing it through an ETL pipeline.

    You can learn more about the company’s services here.

    You can catch up with Cassie on LinkedIn.

    Congrats to Stack Overflow user antimirov for earning a lifeboat badge by providing a great answer to the question: 

    How do you efficiently compare two sets in Python?

    The world’s most popular web framework is going AI native

    The world’s most popular web framework is going AI native

    Palmer says that a huge percentage of today’s top websites, including apps like ChartGPT, Perplexity, and Claude, were built with Vercel’s Next.JS. 

    For the second goal, you can see what Vercel is up to with its v0 project, which lets developers use text prompts and images to generate code. 

    Third, the Vercel AI SDK, which aims to to help developers build conversational, streaming, and chat user interfaces in JavaScript and TypeScript. You can learn more here.

    If you want to catch Jared posting memes, check him out on Twitter. If you want to learn more abiout the AI SDK, check it out 

    here.

    A big thanks to Pierce Darragh for providing a great answer and earning a lifeboat badge by saving a question from the dustinbin of history. Pierce explained: How you can split documents into training set and test set

    Can software startups that need $$$ avoid venture captial?

    Can software startups that need $$$ avoid venture captial?

    You can find Shestakofsky on his website or check him out on X.

    Grab a copy of his new book: Behind the Startup: How Venture Capital Shapes Work, Innovation, and Inequality. 

    As he writes on his website, the book:

    Draws on 19 months of participant-observation research to examine how investors’ demand for rapid growth created organizational problems that managers solved by combining high-tech systems with low-wage human labor. The book shows how the burdens imposed on startups by venture capital—as well as the benefits and costs of “moving fast and breaking things”—are unevenly distributed across a company’s workforce and customers. With its focus on the financialization of innovation, Behind the Startup explains how the gains generated by tech startups are funneled into the pockets of a small cadre of elite investors and entrepreneurs. To promote innovation that benefits the many rather than the few, Shestakofsky argues that we should focus less on fixing the technology and more on changing the financial infrastructure that supports it.

    A big thanks to our user of the week, Parusnik, who was awarded a Great Question badge for asking: How to run a .NET Core console application on Linux?

    An open-source development paradigm

    An open-source development paradigm

    Temporal is an open-source implementation of durable execution, a development paradigm that preserves complete application state so that upon host or software failure it can seamlessly migrate execution to another machine. Learn how it works or dive into the docs. 

    Temporal’s SaaS offering is Temporal Cloud.

    Replay is a three-day conference focused on durable execution. Replay 2024 is September 18-20 in Seattle, Washington, USA. Get your early bird tickets or submit a talk proposal!

    Connect with Maxim on LinkedIn.

    User Honda hoda earned a Famous Question badge for SQLSTATE[01000]: Warning: 1265 Data truncated for column.