LessWrong (Curated & Popular)
合計時間:
10 h 09 min
[Linkpost] “Jaan Tallinn’s 2024 Philanthropy Overview” by jaan
LessWrong (Curated & Popular)
01:17
“Impact, agency, and taste” by benkuhn
LessWrong (Curated & Popular)
15:17
[Linkpost] “To Understand History, Keep Former Population Distributions In Mind” by Arjun Panickssery
LessWrong (Curated & Popular)
05:42
“AI-enabled coups: a small group could use AI to seize power” by Tom Davidson, Lukas Finnveden, rosehadshar
LessWrong (Curated & Popular)
15:22
“Accountability Sinks” by Martin Sustrik
LessWrong (Curated & Popular)
28:50
“Training AGI in Secret would be Unsafe and Unethical” by Daniel Kokotajlo
LessWrong (Curated & Popular)
10:46
“Why Should I Assume CCP AGI is Worse Than USG AGI?” by Tomás B.
LessWrong (Curated & Popular)
01:15
“Surprising LLM reasoning failures make me think we still need qualitative breakthroughs for AGI” by Kaj_Sotala
LessWrong (Curated & Popular)
35:51
“Frontier AI Models Still Fail at Basic Physical Tasks: A Manufacturing Case Study” by Adam Karvonen
LessWrong (Curated & Popular)
21:00
“Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research (GDM Mech Interp Team Progress Update #2)” by Neel Nanda, lewis smith, Senthooran Rajamanoharan, Arthur Conmy, Callum McD
LessWrong (Curated & Popular)
57:32
[Linkpost] “Playing in the Creek” by Hastings
LessWrong (Curated & Popular)
04:12
“Thoughts on AI 2027” by Max Harms
LessWrong (Curated & Popular)
40:27
“Short Timelines don’t Devalue Long Horizon Research” by Vladimir_Nesov
LessWrong (Curated & Popular)
02:10
“Alignment Faking Revisited: Improved Classifiers and Open Source Extensions” by John Hughes, abhayesian, Akbir Khan, Fabien Roger
LessWrong (Curated & Popular)
41:04
“METR: Measuring AI Ability to Complete Long Tasks” by Zach Stein-Perlman
LessWrong (Curated & Popular)
11:09
“Why Have Sentence Lengths Decreased?” by Arjun Panickssery
LessWrong (Curated & Popular)
09:08
“AI 2027: What Superintelligence Looks Like” by Daniel Kokotajlo, Thomas Larsen, elifland, Scott Alexander, Jonas V, romeo
LessWrong (Curated & Popular)
54:30
“OpenAI #12: Battle of the Board Redux” by Zvi
LessWrong (Curated & Popular)
18:01
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit
LessWrong (Curated & Popular)
27:39
“OpenAI #12: Battle of the Board Redux” by Zvi
LessWrong (Curated & Popular)
18:01
“You will crash your car in front of my house within the next week” by Richard Korzekwa
LessWrong (Curated & Popular)
01:52
“My ‘infohazards small working group’ Signal Chat may have encountered minor leaks” by Linch
LessWrong (Curated & Popular)
10:33
“Leverage, Exit Costs, and Anger: Re-examining Why We Explode at Home, Not at Work” by at_the_zoo
LessWrong (Curated & Popular)
06:16
“PauseAI and E/Acc Should Switch Sides” by WillPetillo
LessWrong (Curated & Popular)
03:31
“VDT: a solution to decision theory” by L Rudolf L
LessWrong (Curated & Popular)
08:58
“LessWrong has been acquired by EA” by habryka
LessWrong (Curated & Popular)
01:33
“We’re not prepared for an AI market crash” by Remmelt
LessWrong (Curated & Popular)
03:46
“Conceptual Rounding Errors” by Jan_Kulveit
LessWrong (Curated & Popular)
06:21
“Tracing the Thoughts of a Large Language Model” by Adam Jermyn
LessWrong (Curated & Popular)
22:18
“Recent AI model progress feels mostly like bullshit” by lc
LessWrong (Curated & Popular)
14:29
“AI for AI safety” by Joe Carlsmith
LessWrong (Curated & Popular)
34:07
“Policy for LLM Writing on LessWrong” by jimrandomh
LessWrong (Curated & Popular)
04:17
“Will Jesus Christ return in an election year?” by Eric Neyman
LessWrong (Curated & Popular)
07:48
“Good Research Takes are Not Sufficient for Good Strategic Takes” by Neel Nanda
LessWrong (Curated & Popular)
06:58
“Intention to Treat” by Alicorn
LessWrong (Curated & Popular)
03:45
“On the Rationality of Deterring ASI” by Dan H
LessWrong (Curated & Popular)
09:03
[Linkpost] “METR: Measuring AI Ability to Complete Long Tasks” by Zach Stein-Perlman
LessWrong (Curated & Popular)
01:19
“I make several million dollars per year and have hundreds of thousands of followers—what is the straightest line path to utilizing these resources to reduce existential-level AI threats?” by shrimpy
LessWrong (Curated & Popular)
02:17
“Claude Sonnet 3.7 (often) knows when it’s in alignment evaluations” by Nicholas Goldowsky-Dill, Mikita Balesni, Jérémy Scheurer, Marius Hobbhahn
LessWrong (Curated & Popular)
18:05
“Levels of Friction” by Zvi
LessWrong (Curated & Popular)
22:43