Episodios

  • Expert Insights On Retrieval Augmented Generation And How To Build It
    Jul 28 2024
    SummaryIn this episode we're joined by Matt Zeiler, founder and CEO of Clarifai, as he dives into the technical aspects of retrieval augmented generation (RAG). From his journey into AI at the University of Toronto to founding one of the first deep learning AI companies, Matt shares his insights on the evolution of neural networks and generative models over the last 15 years. He explains how RAG addresses issues with large language models, including data staleness and hallucinations, by providing dynamic access to information through vector databases and embedding models. Throughout the conversation, Matt and host Tobias Macy discuss everything from architectural requirements to operational considerations, as well as the practical applications of RAG in industries like intelligence, healthcare, and finance. Tune in for a comprehensive look at RAG and its future trends in AI.AnnouncementsHello and welcome to the AI Engineering Podcast, your guide to the fast-moving world of building scalable and maintainable AI systemsYour host is Tobias Macey and today I'm interviewing Matt Zeiler, Founder & CEO of Clarifai, about the technical aspects of RAG, including the architectural requirements, edge cases, and evolutionary characteristicsInterviewIntroductionHow did you get involved in the area of data management?Can you describe what RAG (Retrieval Augmented Generation) is?What are the contexts in which you would want to use RAG?What are the alternatives to RAG?What are the architectural/technical components that are required for production grade RAG?Getting a quick proof-of-concept working for RAG is fairly straightforward. What are the failures modes/edge cases that start to surface as you scale the usage and complexity?The first step of building the corpus for RAG is to generate the embeddings. Can you talk through the planning and design process? (e.g. model selection for embeddings, storage capacity/latency, etc.)How does the modality of the input/output affect this and downstream decisions? (e.g. text vs. image vs. audio, etc.)What are the features of a vector store that are most critical for RAG?The set of available generative models is expanding and changing at breakneck speed. What are the foundational aspects that you look for in selecting which model(s) to use for the output?Vector databases have been gaining ground for search functionality, even without generative AI. What are some of the other ways that elements of RAG can be re-purposed?What are the most interesting, innovative, or unexpected ways that you have seen RAG used?What are the most interesting, unexpected, or challenging lessons that you have learned while working on RAG?When is RAG the wrong choice?What are the main trends that you are following for RAG and its component elements going forward?Contact InfoWebsiteLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. [Podcast.__init__]() covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@aiengineeringpodcast.com with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksClarifaiGeoff HintonYann LecunNeural NetworksDeep LearningRetrieval Augmented GenerationContext WindowVector DatabasePrompt EngineeringMistralLlama 3Embedding QuantizationActive LearningGoogle GeminiAI Model AttentionRecurrent NetworkConvolutional NetworkReranking ModelStop WordsMassive Text Embedding Benchmark (MTEB)Retool State of AI ReportpgvectorMilvusQdrantPineconeOpenLLM LeaderboardSemantic SearchHashicorpThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    1 h y 3 m
  • Barking Up The Wrong GPTree: Building Better AI With A Cognitive Approach
    Jul 28 2024
    SummaryArtificial intelligence has dominated the headlines for several months due to the successes of large language models. This has prompted numerous debates about the possibility of, and timeline for, artificial general intelligence (AGI). Peter Voss has dedicated decades of his life to the pursuit of truly intelligent software through the approach of cognitive AI. In this episode he explains his approach to building AI in a more human-like fashion and the emphasis on learning rather than statistical prediction.AnnouncementsHello and welcome to the AI Engineering Podcast, your guide to the fast-moving world of building scalable and maintainable AI systemsYour host is Tobias Macey and today I'm interviewing Peter Voss about what is involved in making your AI applications more "human"InterviewIntroductionHow did you get involved in machine learning?Can you start by unpacking the idea of "human-like" AI?How does that contrast with the conception of "AGI"?The applications and limitations of GPT/LLM models have been dominating the popular conversation around AI. How do you see that impacting the overrall ecosystem of ML/AI applications and investment?The fundamental/foundational challenge of every AI use case is sourcing appropriate data. What are the strategies that you have found useful to acquire, evaluate, and prepare data at an appropriate scale to build high quality models? What are the opportunities and limitations of causal modeling techniques for generalized AI models?As AI systems gain more sophistication there is a challenge with establishing and maintaining trust. What are the risks involved in deploying more human-level AI systems and monitoring their reliability?What are the practical/architectural methods necessary to build more cognitive AI systems?How would you characterize the ecosystem of tools/frameworks available for creating, evolving, and maintaining these applications?What are the most interesting, innovative, or unexpected ways that you have seen cognitive AI applied?What are the most interesting, unexpected, or challenging lessons that you have learned while working on desiging/developing cognitive AI systems?When is cognitive AI the wrong choice?What do you have planned for the future of cognitive AI applications at Aigo?Contact InfoLinkedInWebsiteParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@aiengineeringpodcast.com with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksAigo.aiArtificial General IntelligenceCognitive AIKnowledge GraphCausal ModelingBayesian StatisticsThinking Fast & Slow by Daniel Kahneman (affiliate link)Agent-Based ModelingReinforcement LearningDARPA 3 Waves of AI presentationWhy Don't We Have AGI Yet? whitepaperConcepts Is All You Need WhitepaperHellen KellerStephen HawkingThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    53 m
  • Build Your Second Brain One Piece At A Time
    Jul 28 2024
    SummaryGenerative AI promises to accelerate the productivity of human collaborators. Currently the primary way of working with these tools is through a conversational prompt, which is often cumbersome and unwieldy. In order to simplify the integration of AI capabilities into developer workflows Tsavo Knott helped create Pieces, a powerful collection of tools that complements the tools that developers already use. In this episode he explains the data collection and preparation process, the collection of model types and sizes that work together to power the experience, and how to incorporate it into your workflow to act as a second brain.AnnouncementsHello and welcome to the AI Engineering Podcast, your guide to the fast-moving world of building scalable and maintainable AI systemsYour host is Tobias Macey and today I'm interviewing Tsavo Knott about Pieces, a personal AI toolkit to improve the efficiency of developersInterviewIntroductionHow did you get involved in machine learning?Can you describe what Pieces is and the story behind it?The past few months have seen an endless series of personalized AI tools launched. What are the features and focus of Pieces that might encourage someone to use it over the alternatives?model selectionsarchitecture of Pieces applicationlocal vs. hybrid vs. online modelsmodel update/delivery processdata preparation/serving for models in context of Pieces appapplication of AI to developer workflowstypes of workflows that people are building with piecesWhat are the most interesting, innovative, or unexpected ways that you have seen Pieces used?What are the most interesting, unexpected, or challenging lessons that you have learned while working on Pieces?When is Pieces the wrong choice?What do you have planned for the future of Pieces?Contact InfoLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@aiengineeringpodcast.com with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksPiecesNPU == Neural Processing UnitTensor ChipLoRA == Low Rank AdaptationGenerative Adversarial NetworksMistralEmacsVimNeoVimDartFlutterTypescriptLuaRetrieval Augmented GenerationONNXLSTM == Long Short-Term MemoryLLama 2GitHub CopilotTabninePodcast EpisodeThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    48 m
  • Strategies For Building A Product Using LLMs At DataChat
    Mar 3 2024
    SummaryLarge Language Models (LLMs) have rapidly captured the attention of the world with their impressive capabilities. Unfortunately, they are often unpredictable and unreliable. This makes building a product based on their capabilities a unique challenge. Jignesh Patel is building DataChat to bring the capabilities of LLMs to organizational analytics, allowing anyone to have conversations with their business data. In this episode he shares the methods that he is using to build a product on top of this constantly shifting set of technologies.AnnouncementsHello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery.Your host is Tobias Macey and today I'm interviewing Jignesh Patel about working with LLMs; understanding how they work and how to build your ownInterviewIntroductionHow did you get involved in machine learning?Can you start by sharing some of the ways that you are working with LLMs currently?What are the business challenges involved in building a product on top of an LLM model that you don't own or control? In the current age of business, your data is often your strategic advantage. How do you avoid losing control of, or leaking that data while interfacing with a hosted LLM API?What are the technical difficulties related to using an LLM as a core element of a product when they are largely a black box? What are some strategies for gaining visibility into the inner workings or decision making rules for these models?What are the factors, whether technical or organizational, that might motivate you to build your own LLM for a business or product? Can you unpack what it means to "build your own" when it comes to an LLM?In your work at DataChat, how has the progression of sophistication in LLM technology impacted your own product strategy?What are the most interesting, innovative, or unexpected ways that you have seen LLMs/DataChat used?What are the most interesting, unexpected, or challenging lessons that you have learned while working with LLMs?When is an LLM the wrong choice?What do you have planned for the future of DataChat?Contact InfoWebsiteLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com) with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksDataChatCMU == Carnegie Mellon UniversitySVM == Support Vector MachineGenerative AIGenomicsProteomicsParquetOpenAI CodexLLamaMistralGoogle VertexLangchainRetrieval Augmented GenerationPrompt EngineeringEnsemble LearningXGBoostCatboostLinear RegressionCOGS == Cost Of Goods SoldBruce Schneier - AI And TrustThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    49 m
  • Improve The Success Rate Of Your Machine Learning Projects With bizML
    Feb 18 2024
    SummaryMachine learning is a powerful set of technologies, holding the potential to dramatically transform businesses across industries. Unfortunately, the implementation of ML projects often fail to achieve their intended goals. This failure is due to a lack of collaboration and investment across technological and organizational boundaries. To help improve the success rate of machine learning projects Eric Siegel developed the six step bizML framework, outlining the process to ensure that everyone understands the whole process of ML deployment. In this episode he shares the principles and promise of that framework and his motivation for encapsulating it in his book "The AI Playbook".AnnouncementsHello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery.Your host is Tobias Macey and today I'm interviewing Eric Siegel about how the bizML approach can help improve the success rate of your ML projectsInterviewIntroductionHow did you get involved in machine learning?Can you describe what bizML is and the story behind it? What are the key aspects of this approach that are different from the "industry standard" lifecycle of an ML project?What are the elements of your personal experience as an ML consultant that helped you develop the tenets of bizML?Who are the personas that need to be involved in an ML project to increase the likelihood of success? Who do you find to be best suited to "own" or "lead" the process?What are the organizational patterns that might hinder the work of delivering on the goals of an ML initiative?What are some of the misconceptions about the work involved in/capabilities of an ML model that you commonly encounter?What is your main goal in writing your book "The AI Playbook"?What are the most interesting, innovative, or unexpected ways that you have seen the bizML process in action?What are the most interesting, unexpected, or challenging lessons that you have learned while working on ML projects and developing the bizML framework?When is bizML the wrong choice?What are the future developments in organizational and technical approaches to ML that will improve the success rate of AI projects?Contact InfoLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com) with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksThe AI Playbook: Mastering the Rare Art of Machine Learning Deployment by Eric SiegelPredictive Analytics: The Power to Predict Who Will Click, Buy, Lie, or Die by Eric SiegelColumbia UniversityMachine Learning Week ConferenceGenerative AI WorldMachine Learning Leadership and Practice CourseRexer AnalyticsKD NuggetsCRISP-DMRandom ForestGradient DescentThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    50 m
  • Using Generative AI To Accelerate Feature Engineering At FeatureByte
    Feb 11 2024
    SummaryOne of the most time consuming aspects of building a machine learning model is feature engineering. Generative AI offers the possibility of accelerating the discovery and creation of feature pipelines. In this episode Colin Priest explains how FeatureByte is applying generative AI models to the challenge of building and maintaining machine learning pipelines.AnnouncementsHello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery.Your host is Tobias Macey and today I'm interviewing Colin Priest about applying generative AI to the task of building and deploying AI pipelinesInterviewIntroductionHow did you get involved in machine learning?Can you start by giving the 30,000 foot view of the steps involved in an AI pipeline? Understand the problemFeature ideationFeature engineeringExperimentOptimizeProductionizeWhat are the stages of that process that are prone to repetition? What are the ways that teams typically try to automate those steps?What are the features of generative AI models that can be brought to bear on the design stage of an AI pipeline? What are the validation/verification processes that engineers need to apply to the generated suggestions?What are the opportunities/limitations for unit/integration style tests?What are the elements of developer experience that need to be addressed to make the gen AI capabilities an enhancement instead of a distraction? What are the interfaces through which the AI functionality can/should be exposed?What are the aspects of pipeline and model deployment that can benefit from generative AI functionality? What are the potential risk factors that need to be considered when evaluating the application of this functionality?What are the most interesting, innovative, or unexpected ways that you have seen generative AI used in the development and maintenance of AI pipelines?What are the most interesting, unexpected, or challenging lessons that you have learned while working on the application of generative AI to the ML workflow?When is generative AI the wrong choice?What do you have planned for the future of FeatureByte's AI copilot capabiliteis?Contact InfoLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com) with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksFeatureByteGenerative AIThe Art of WarOCR == Optical Character RecognitionGenetic AlgorithmSemantic LayerPrompt EngineeringThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0Support The Machine Learning Podcast
    Más Menos
    45 m
  • Learn And Automate Critical Business Workflows With 8Flow
    Jan 28 2024
    SummaryEvery business develops their own specific workflows to address their internal organizational needs. Not all of them are properly documented, or even visible. Workflow automation tools have tried to reduce the manual burden involved, but they are rigid and require substantial investment of time to discover and develop the routines. Boaz Hecht co-founded 8Flow to iteratively discover and automate pieces of workflows, bringing visibility and collaboration to the internal organizational processes that keep the business running.AnnouncementsHello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery.Your host is Tobias Macey and today I'm interviewing Boaz Hecht about using AI to automate customer support at 8FlowInterviewIntroductionHow did you get involved in machine learning?Can you describe what 8Flow is and the story behind it?How does 8Flow compare to RPA tools that companies are using today? What are the opportunities for augmenting or integrating with RPA frameworks?What are the key selling points for the solution that you are building? (does AI sell? Or is it about the realized savings?)What are the sources of signal that you are relying on to build model features?Given the heterogeneity in tools and processes across customers, what are the common focal points that let you address the widest possible range of functionality?Can you describe how 8Flow is implemented? How have the design and goals evolved since you first started working on it?What are the model categories that are most relevant for process automation in your product?How have you approached the design and implementation of your MLOps workflow? (model training, deployment, monitoring, versioning, etc.)What are the open questions around product focus and system design that you are still grappling with?Given the relative recency of ML/AI as a profession and the massive growth in attention and activity, how are you addressing the challenge of obtaining and maximizing human talent?What are the most interesting, innovative, or unexpected ways that you have seen 8Flow used?What are the most interesting, unexpected, or challenging lessons that you have learned while working on 8Flow?When is 8Flow the wrong choice?What do you have planned for the future of 8Flow?Contact InfoLinkedInPersonal WebsiteParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com) with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.Links8FlowRobotic Process AutomationThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0Support The Machine Learning Podcast
    Más Menos
    43 m
  • Considering The Ethical Responsibilities Of ML And AI Engineers
    Jan 28 2024
    SummaryMachine learning and AI applications hold the promise of drastically impacting every aspect of modern life. With that potential for profound change comes a responsibility for the creators of the technology to account for the ramifications of their work. In this episode Nicholas Cifuentes-Goodbody guides us through the minefields of social, technical, and ethical considerations that are necessary to ensure that this next generation of technical and economic systems are equitable and beneficial for the people that they impact.AnnouncementsHello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery.Your host is Tobias Macey and today I'm interviewing Nicholas Cifuentes-Goodbody about the different elements of the machine learning workflow where ethics need to be consideredInterviewIntroductionHow did you get involved in machine learning?To start with, who is responsible for addressing the ethical concerns around AI?What are the different ways that AI can have positive or negative outcomes from an ethical perspective? What is the role of practitioners/individual contributors in the identification and evaluation of ethical impacts of their work?What are some utilities that are helpful in identifying and addressing bias in training data?How can practitioners address challenges of equity and accessibility in the delivery of AI products?What are some of the options for reducing the energy consumption for training and serving AI?What are the most interesting, innovative, or unexpected ways that you have seen ML teams incorporate ethics into their work?What are the most interesting, unexpected, or challenging lessons that you have learned while working on ethical implications of ML?What are some of the resources that you recommend for people who want to invest in their knowledge and application of ethics in the realm of ML?Contact InfoWorldQuant University's Applied Data Science LabLinkedInParting QuestionFrom your perspective, what is the biggest barrier to adoption of machine learning today?Closing AnnouncementsThank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com) with your story.To help other people find the show please leave a review on iTunes and tell your friends and co-workers.LinksUNESCO Recommendation on the Ethics of Artificial IntelligenceEuropean Union AI ActHow machine learning helps advance access to human rights informationDisinformation, Team JorgeChina, AI, and Human RightsHow China Is Using A.I. to Profile a MinorityWeapons of Math DestructionFairlearnAI Fairness 360Allen Institute for AI NYTAllen Institute for AITransformersAI4ALLWorldQuant UniversityHow to Make Generative AI GreenerMachine Learning Emissions CalculatorPracticing Trustworthy Machine LearningEnergy and Policy Considerations for Deep LearningNatural Language ProcessingTrolley ProblemProtected Classesfairlearn (scikit-learn)BERT ModelThe intro and outro music is from Hitman's Lovesong feat. Paola Graziano by The Freak Fandango Orchestra/CC BY-SA 3.0
    Más Menos
    39 m