Topik trending
#
Bonk Eco continues to show strength amid $USELESS rally
#
Pump.fun to raise $1B token sale, traders speculating on airdrop
#
Boop.Fun leading the way with a new launchpad on Solana.

Julian Schrittwieser
Anggota Staf Teknis di Anthropic AlphaGo, AlphaZero, MuZero, AlphaCode, AlphaTensor, AlphaProof Gemini RL Prev Insinyur Riset Utama di DeepMind
Fast Opus luar biasa, pertama kali saya menggunakannya, saya tidak bisa berhenti membuat kode selama berjam-jam - sejujurnya terasa seperti kekuatan super, Anda dapat membentuk basis kode Anda secepat yang Anda pikirkan.
Benar-benar luar biasa, tidak ada yang membuat saya lebih merasakan AGI, pasti mencobanya!

Claude8 Feb 2026
Our teams have been building with a 2.5x-faster version of Claude Opus 4.6.
We’re now making it available as an early experiment via Claude Code and our API.
227
Saya bersenang-senang mengobrol dengan @mattturck dari podcast MAD minggu ini! Kami membahas tren dalam AI, RL, dan mengapa AI, dan mengapa membuka kunci Agen, penskalaan, dan banyak lagi:
Tautan ke apa yang kita bicarakan dan bacaan lebih lanjut:

Matt Turck24 Okt 2025
Failing to Understand the Exponential, Again?
My conversation with @Mononofu - Julian Schrittwieser (@AnthropicAI, AlphaGo Zero, MuZero) - on Move 37, Scaling RL, Nobel Prize for AI, and the AI frontier:
00:00 - Cold open: “We’re not seeing any slowdown.”
00:32 - Intro — Meet Julian
01:09 - The “exponential” from inside frontier labs
04:46 - 2026–2027: agents that work a full day; expert-level breadth
08:58 - Benchmarks vs reality: long-horizon work, GDP-Val, user value
10:26 - Move 37 — what actually happened and why it mattered
13:55 - Novel science: AlphaCode/AlphaTensor → when does AI earn a Nobel?
16:25 - Discontinuity vs smooth progress (and warning signs)
19:08 - Does pre-training + RL get us there? (AGI debates aside)
20:55 - Sutton’s “RL from scratch”? Julian’s take
23:03 - Julian’s path: Google → DeepMind → Anthropic
26:45 - AlphaGo (learn + search) in plain English
30:16 - AlphaGo Zero (no human data)
31:00 - AlphaZero (one algorithm: Go, chess, shogi)
31:46 - MuZero (planning with a learned world model)
33:23 -Lessons for today’s agents: search + learning at scale
34:57 - Do LLMs already have implicit world models?
39:02 - Why RL on LLMs took time (stability, feedback loops)
41:43 - Compute & scaling for RL — what we see so far
42:35 - Rewards frontier: human prefs, rubrics, RLVR, process rewards
44:36 - RL training data & the “flywheel” (and why quality matters)
48:02 - RL & Agents 101 — why RL unlocks robustness
50:51 - Should builders use RL-as-a-service? Or just tools + prompts?
52:18 - What’s missing for dependable agents (capability vs engineering)
53:51 - Evals & Goodhart — internal vs external benchmarks
57:35 - Mechanistic interpretability & “Golden Gate Claude”
1:00:03 - Safety & alignment at Anthropic — how it shows up in practice
1:03:48 - Jobs: human–AI complementarity (comparative advantage)
1:06:33 - Inequality, policy, and the case for 10× productivity → abundance
1:09:24 - Closing thoughts
809
Teratas
Peringkat
Favorit
