The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.

https://twimlai.com

Eine durchschnittliche Folge dieses Podcasts dauert 43m. Bisher sind 699 Folge(n) erschienen. Alle 4 Tage erscheint eine Folge dieses Podcasts.

Gesamtlänge aller Episoden: 21 days 15 hours 16 minutes

subscribe
share






episode 681: GraphRAG: Knowledge Graphs for AI Applications with Kirk Marple


Today we're joined by Kirk Marple, CEO and founder of Graphlit, to explore the emerging paradigm of "GraphRAG," or Graph Retrieval Augmented Generation. In our conversation, Kirk digs into the GraphRAG architecture and how Graphlit uses it to offer a multi-stage workflow for ingesting, processing, retrieving, and generating content using LLMs (like GPT-4) and other Generative AI tech...


share








   47m
 
 

episode 680: Teaching Large Language Models to Reason with Reinforcement Learning with Alex Havrilla


Today we're joined by Alex Havrilla, a PhD student at Georgia Tech, to discuss "Teaching Large Language Models to Reason with Reinforcement Learning." Alex discusses the role of creativity and exploration in problem solving and explores the opportunities presented by applying reinforcement learning algorithms to the challenge of improving reasoning in large language models...


share








   46m
 
 

episode 679: Localizing and Editing Knowledge in LLMs with Peter Hase


Today we're joined by Peter Hase, a fifth-year PhD student at the University of North Carolina NLP lab. We discuss "scalable oversight", and the importance of developing a deeper understanding of how large neural networks make decisions. We learn how matrices are probed by interpretability researchers, and explore the two schools of thought regarding how LLMs store knowledge...


share








   49m
 
 

episode 678: Coercing LLMs to Do and Reveal (Almost) Anything with Jonas Geiping


Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks...


share








   48m
 
 

episode 677: V-JEPA, AI Reasoning from a Non-Generative Architecture with Mido Assran


Today we’re joined by Mido Assran, a research scientist at Meta’s Fundamental AI Research (FAIR). In this conversation, we discuss V-JEPA, a new model being billed as “the next step in Yann LeCun's vision” for true artificial reasoning. V-JEPA, the video version of Meta’s Joint Embedding Predictive Architecture, aims to bridge the gap between human and machine intelligence by training models to learn abstract concepts in a more efficient predictive manner than generative models...


share








 March 25, 2024  47m
 
 

episode 676: Video as a Universal Interface for AI Reasoning with Sherry Yang


Today we’re joined by Sherry Yang, senior research scientist at Google DeepMind and a PhD student at UC Berkeley. In this interview, we discuss her new paper, "Video as the New Language for Real-World Decision Making,” which explores how generative video models can play a role similar to language models as a way to solve tasks in the real world...


share








 March 18, 2024  49m
 
 

episode 675: Assessing the Risks of Open AI Models with Sayash Kapoor


Today we’re joined by Sayash Kapoor, a Ph.D. student in the Department of Computer Science at Princeton University. Sayash walks us through his paper: "On the Societal Impact of Open Foundation Models.” We dig into the controversy around AI safety, the risks and benefits of releasing open model weights, and how we can establish common ground for assessing the threats posed by AI...


share








 March 11, 2024  40m
 
 

episode 674: OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia


Today we’re joined by Akshita Bhagia, a senior research engineer at the Allen Institute for AI. Akshita joins us to discuss OLMo, a new open source language model with 7 billion and 1 billion variants, but with a key difference compared to similar models offered by Meta, Mistral, and others. Namely, the fact that AI2 has also published the dataset and key tools used to train the model...


share








 March 4, 2024  32m
 
 

episode 673: Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski


Today we’re joined by Ben Prystawski, a PhD student in the Department of Psychology at Stanford University working at the intersection of cognitive science and machine learning. Our conversation centers on Ben’s recent paper, “Why think step by step? Reasoning emerges from the locality of experience,” which he recently presented at NeurIPS 2023...


share








 February 26, 2024  24m
 
 

episode 672: Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh


Today we're joined by Armineh Nourbakhsh of JP Morgan AI Research to discuss the development and capabilities of DocLLM, a layout-aware large language model for multimodal document understanding. Armineh provides a historical overview of the challenges of document AI and an introduction to the DocLLM model...


share








 February 19, 2024  45m