GPT Reviews – Details, episodes & analysis
Podcast details
Technical and general information from the podcast's RSS feed.

GPT Reviews
Earkind
Frequency: 1 episode/2d. Total Eps: 301

Recent rankings
Latest chart positions across Apple Podcasts and Spotify rankings.
Apple Podcasts
🇬🇧 Great Britain - dailyNews
13/11/2024#87
Spotify
No recent rankings available
Shared links between episodes and podcasts
Links found in episode descriptions and other podcasts that share them.
See all- https://cloud.withgoogle.com/next
183 shares
- https://openai.com/sora
92 shares
RSS feed quality and score
Technical evaluation of the podcast's RSS feed quality and structure.
See allScore global : 53%
Publication history
Monthly episode publishing history over the past years.
OpenAI's Strawberry Revolution 🍓 // Nvidia's Lucrative Paychecks 💸 // Google Pipe SQL Simplification 📊
jeudi 29 août 2024 • Duration 14:01
This episode dives into OpenAI's promising new model, Strawberry, which could revolutionize interactions in ChatGPT. We explore the financial envy Nvidia employees inspire in their Google and Meta counterparts due to lucrative stock options. Google’s new Pipe SQL syntax aims to simplify data querying, while concerns about research accessibility are raised. Finally, we discuss BaichuanSEED and Dolphin models, which highlight advancements in extensible data collection and energy-efficient processing, paving the way for enhanced AI capabilities.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:40 OpenAI Races to Launch Strawberry
03:07 Google, Meta workers envy Nvidia staffers’ fat paychecks: ‘Bought a 100K car … all cash’
05:01 Google's New Pipe SQL Syntax
06:12 Fake sponsor
09:20 Dolphin: Long Context as a New Modality for Energy-Efficient On-Device Language Models
11:09 Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders
12:50 Outro
OpenAI's 'Strawberry' AI 🚀 // World's Fastest AI Inference ⚡ // Photo-realistic 3D Avatars 🎨
mercredi 28 août 2024 • Duration 14:14
OpenAI's 'Strawberry' AI tackles complex math and programming with enhanced reasoning, while Cerebras claims to have launched the fastest AI inference, enabling real-time applications at competitive prices. The GenCA model revolutionizes avatar creation with photo-realistic, controllable 3D avatars, and the "Build-A-Scene" paper introduces interactive 3D layout control for text-to-image generation, enhancing creative fields with dynamic object manipulation.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
02:02 OpenAI Shows ‘Strawberry’ AI to the Feds and Uses It to Develop ‘Orion’
03:23 Cerebras Launches the World’s Fastest AI Inference
05:07 Diffusion Models Are Real-Time Game Engines
06:15 Fake sponsor
08:06 The Mamba in the Llama: Distilling and Accelerating Hybrid Models
09:42 GenCA: A Text-conditioned Generative Model for Realistic and Drivable Codec Avatars
11:16 Build-A-Scene: Interactive 3D Layout Control for Diffusion-Based Image Generation
13:04 Outro
Nvidia's Stock Struggles 📉 // Meta's AI Hallucinations 🤖 // Superconducting Microprocessors ⚡
vendredi 2 août 2024 • Duration 14:41
This episode dives into Nvidia's stock struggles amid rising competition, while also unpacking Meta's AI blunders and the implications of "hallucinations" in tech. We explore cutting-edge superconducting microprocessors that promise unprecedented energy efficiency and highlight groundbreaking AI research, including eavesdropping techniques and advancements in reinforcement learning.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:50 Nvidia Sank Again Today -- Time to Buy the Artificial Intelligence (AI) Growth Stock Hand Over Fist?
03:09 Meta blames hallucinations after its AI said Trump rally shooting didn’t happen
04:52 Superconducting Microprocessors? Turns Out They're Ultra-Efficient
06:07 Fake sponsor
09:22 SAPG: Split and Aggregate Policy Gradients
10:45 MindSearch: Mimicking Human Minds Elicits Deep AI Searcher
12:44 Outro
AI Secret Trading in China 💼 // Training Models at Scale 🚀 // Improving User Queries with Backtracing 🔍
vendredi 8 mars 2024 • Duration 14:56
A Google engineer has been indicted for allegedly stealing over 500 confidential files containing AI trade secrets while working for China-based companies seeking an edge in the AI technology race.
A tutorial series explores parallelism strategies for training large deep learning models, making it accessible to everyone regardless of the hardware you have available.
Value functions are a crucial component in deep reinforcement learning, and a new approach using categorical cross-entropy instead of regression can significantly improve performance and scalability in a variety of domains.
Backtracing is the task of retrieving the text segment that most likely caused a user query, and it can help improve content delivery and communication by identifying linguistic triggers that influence user queries.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:33 Google engineer indicted over allegedly stealing AI trade secrets for China
03:57 Training Models at Scale Tutorial
05:24 Autogenerating a Book Series From Three Years of iMessages
06:22 Fake sponsor
08:16 Design2Code: How Far Are We From Automating Front-End Engineering?
10:09 Stop Regressing: Training Value Functions via Classification for Scalable Deep RL
11:43 Backtracing: Retrieving the Cause of the Query
13:27 Outro
Perplexity vs. Google 🔍 // Microsoft vs. NYT ⚖️ // General Computer Control 💻
jeudi 7 mars 2024 • Duration 14:00
Perplexity AI is a search startup that's looking to take on Google by solving the inadequacies of searching the web. They are nearing unicorn status with a valuation of around $1 billion.
Microsoft is being sued by The New York Times for copyright infringement and abusing the newspaper’s intellectual property in training LLMs. Microsoft accuses the Times of "unsubstantiated" claims and compares the lawsuit to Hollywood's resistance to the VCR in the 70s.
A new paper introduces the concept of General Computer Control (GCC), which is the idea of building agents that can master any computer task by taking only screen images and producing keyboard and mouse operations as output. The authors propose a framework called Cradle that has strong reasoning abilities to ensure generalizability and self-improvement across various tasks.
A paper evaluates different tokenizer inference methods and their impact on the performance of downstream NLP tasks. The authors found that for the most commonly used tokenizers, greedy inference performs surprisingly well, and a recently-introduced contextually-informed tokenizer outperforms all others on morphological alignment.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:23 Perplexity Poised To Become Latest AI Startup To Hit Unicorn Status — Report
02:53 Microsoft compares The New York Times’ claims against OpenAI to Hollywood’s early fight against VCR
04:41 Training great LLMs entirely from ground zero in the wilderness as a startup
05:49 Fake sponsor
07:39 Towards General Computer Control: A Multimodal Agent for Red Dead Redemption II as a Case Study
09:15 Design2Code: How Far Are We From Automating Front-End Engineering?
11:01 Greed is All You Need: An Evaluation of Tokenizer Inference Methods
12:49 Outro
OpenAI vs Elon Musk 💻 // Automated Text Embeddings 📊 // Unified Time Series Model 📈
mercredi 6 mars 2024 • Duration 14:34
Groq, an AI chip startup, forms a new business unit and acquires Definitive Intelligence to expand its customer and developer ecosystem.
OpenAI responds to Elon Musk's lawsuit, revealing that Musk himself wanted "absolute control" over the company by merging it with Tesla.
A new Postgres extension called pg_vectorize automates the transformation and orchestration of text to embeddings, providing workflows for vector search and RAG.
UNITS, a unified time series model, achieves superior performance compared to task-specific models and repurposed natural language-based LLMs, demonstrating remarkable zero-shot, few-shot, and prompt learning capabilities.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
02:05 AI chip startup Groq forms new business unit, acquires Definitive Intelligence
03:49 OpenAI says Elon Musk wanted ‘absolute control’ of the company
05:29 pg_vectorize: a VectorDB for Postgres
06:29 Fake sponsor
08:26 Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
09:54 UniTS: Building a Unified Time Series Model
11:39 DéjàVu: KV-cache Streaming for Fast, Fault-tolerant Generative LLM Serving
13:14 Outro
Anthropic's Claude 3 🤖 // Elon Musk sues OpenAI 💥 // Unified Time Series Model 🎧
mardi 5 mars 2024 • Duration 14:58
Anthropic's new and improved Claude 3 model family sets new industry benchmarks across a wide range of cognitive tasks, exhibiting near-human levels of comprehension and fluency on complex tasks.
Elon Musk is suing OpenAI and CEO Sam Altman for allegedly abandoning their original mission to benefit humanity and instead focusing on profits with Microsoft.
Opus 1.5 brings quality improvements, including machine learning-based upgrades, while remaining fully compatible with RFC 6716, and uses deep learning techniques to process or generate signals themselves.
The Multimodal ArXiv dataset represents an important step forward for LVLMs when it comes to interpreting and understanding complex scientific figures, achieving a 10.4% absolute accuracy gain on a multimodal mathematical reasoning benchmark.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:27 Introducing the next generation Claude: Claude 3
03:17 Elon Musk sues Sam Altman and OpenAI
04:59 Opus Gets a Serious Machine Learning Upgrade
06:29 Fake sponsor
08:28 UniTS: Building a Unified Time Series Model
10:10 Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
12:06 Learning and Leveraging World Models in Visual Representation Learning
13:48 Outro
Adobe's GenAI for Audio 🎧 // User Data for AI Backlash 👀 // MOSAIC's Modular Cooking 🍲
lundi 4 mars 2024 • Duration 14:51
Adobe's new generative AI tools for custom audio creation and editing.
Tumblr and WordPress selling user data to train AI tools, sparking backlash.
MOSAIC, a modular system for assistive and interactive cooking using natural language and multiple robots.
A new approach to real-world humanoid control using a causal transformer model trained through autoregressive prediction of sensorimotor trajectories.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:20 Adobe previews new cutting-edge generative AI tools for crafting and editing custom audio
02:39 Tumblr and WordPress to Sell Users’ Data to Train AI Tools
04:14 “AI will cure cancer” misunderstands both AI and medicine
05:56 Fake sponsor
08:12 MOSAIC: A Modular System for Assistive and Interactive Cooking
09:54 Humanoid Locomotion as Next Token Prediction
11:50 In Search of Needles in a 11M Haystack: Recurrent Memory Finds What LLMs Miss
13:22 Outro
Meta's Llama 3 🦙 // Apple's GenAI 🍎 // Unsupervised RL via Reward Encoding 🤖
vendredi 1 mars 2024 • Duration 14:42
Meta Platforms is set to launch its new AI language model, Llama 3, which promises to tackle taboo questions with more grace and respect than its predecessor.
Apple is ramping up its investment in GenAI, with plans to upgrade Siri and iOS’ built-in search tool, Spotlight, with GenAI models to handle more complex queries and multi-turn conversations.
The University of California, Berkeley, has published a paper exploring unsupervised zero-shot reinforcement learning via functional reward encodings, which could enable pre-training of an agent to adapt to any new downstream tasks in a zero-shot manner.
TrustMol, an inverse molecular design method built to be trustworthy, has been proposed by the Max Planck Institute for Informatics, which could make the IMD process more explainable and reliable.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:52 Meta plans launch of new AI language model Llama 3 in July, The Information reports
02:56 Tim Cook says Apple will ‘break new ground’ in GenAI this year
04:35 Things You Should Never Do, Part I
05:46 Fake sponsor
07:28 Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings
09:04 TrustMol: Trustworthy Inverse Molecular Design via Alignment with Molecular Dynamics
10:57 Stochastic Gradient Succeeds for Bandits
13:23 Outro
Pichai on Google Controversy 🤡 // C3.ai's Revenue Surprises AI Market 📈 // 1-bit LLMs for Efficient Language Modeling 💾
jeudi 29 février 2024 • Duration 13:14
Google's image creation tool, Gemini, has been generating offensive and embarrassing results, prompting the company to make structural changes and update product guidelines to avoid bias in AI tools.
C3.ai, a software maker that helps companies build AI applications, reported a narrower-than-expected loss and revenue that topped estimates, causing AI stock to pop more than 14% in extended trading.
A new paper introduces a cost-effective Large Language Model called a 1-bit LLM, which matches the performance of full-precision Transformer LLMs while being significantly more efficient in terms of latency, memory, throughput, and energy consumption.
Another paper proposes a hybrid approach that combines a frozen LLM with a small language model to improve the efficiency of autoregressive decoding for Large Language Models, resulting in substantial speedups of up to 4 times with minor performance penalties. Additionally, a new framework called EMO utilizes a direct audio-to-video synthesis approach to produce highly expressive and lifelike talking head videos.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:38 Google CEO calls AI tool’s controversial responses ‘completely unacceptable’
03:11 Artificial Intelligence Play C3.ai Climbs On Earnings Report, Outlook
04:41 Jason Wei On Sora
06:19 Fake sponsor
08:35 The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
09:19 Think Big, Generate Quick: LLM-to-SLM for Fast Autoregressive Decoding
12:04 Outro