Joe Carlsmith Audio
Ukupno trajanje:
30 h 17 min
On attunement
Joe Carlsmith Audio
44:14
On green
Joe Carlsmith Audio
75:13
On the abolition of man
Joe Carlsmith Audio
69:22
Being nicer than Clippy
Joe Carlsmith Audio
47:30
An even deeper atheism
Joe Carlsmith Audio
25:12
Does AI risk "other" the AIs?
Joe Carlsmith Audio
13:15
When "yang" goes wrong
Joe Carlsmith Audio
21:32
Deep atheism and AI risk
Joe Carlsmith Audio
46:59
Gentleness and the artificial Other
Joe Carlsmith Audio
22:39
In search of benevolence (or: what should you get Clippy for Christmas?)
Joe Carlsmith Audio
52:52
Empirical work that might shed light on scheming (Section 6 of "Scheming AIs")
Joe Carlsmith Audio
28:00
Summing up "Scheming AIs" (Section 5)
Joe Carlsmith Audio
15:46
Speed arguments against scheming (Section 4.4-4.7 of "Scheming AIs")
Joe Carlsmith Audio
15:19
Simplicity arguments for scheming (Section 4.3 of "Scheming AIs")
Joe Carlsmith Audio
19:37
The counting argument for scheming (Sections 4.1 and 4.2 of "Scheming AIs")
Joe Carlsmith Audio
10:40
Arguments for/against scheming that focus on the path SGD takes (Section 3 of "Scheming AIs")
Joe Carlsmith Audio
29:03
Non-classic stories about scheming (Section 2.3.2 of "Scheming AIs")
Joe Carlsmith Audio
24:34
Does scheming lead to adequate future empowerment? (Section 2.3.1.2 of "Scheming AIs")
Joe Carlsmith Audio
22:54
The goal-guarding hypothesis (Section 2.3.1.1 of "Scheming AIs")
Joe Carlsmith Audio
19:11
How useful for alignment-relevant work are AIs with short-term goals? (Section 2.2.4.3 of "Scheming AIs")
Joe Carlsmith Audio
09:21
Is scheming more likely if you train models to have long-term goals? (Sections 2.2.4.1-2.2.4.2 of "Scheming AIs")
Joe Carlsmith Audio
09:01
"Clean" vs. "messy" goal-directedness (Section 2.2.3 of "Scheming AIs")
Joe Carlsmith Audio
16:44
Two sources of beyond-episode goals (Section 2.2.2 of "Scheming AIs")
Joe Carlsmith Audio
21:25
Two concepts of an "episode" (Section 2.2.1 of "Scheming AIs")
Joe Carlsmith Audio
12:08
On "slack" in training (Section 1.5 of "Scheming AIs")
Joe Carlsmith Audio
07:12
Situational awareness (Section 2.1 of "Scheming AIs")
Joe Carlsmith Audio
09:27
Why focus on schemers in particular? (Sections 1.3-1.4 of "Scheming AIs")
Joe Carlsmith Audio
31:17
A taxonomy of non-schemer models (Section 1.2 of "Scheming AIs")
Joe Carlsmith Audio
11:20
Varieties of fake alignment (Section 1.1 of "Scheming AIs")
Joe Carlsmith Audio
17:54
Full audio for "Scheming AIs: Will AIs fake alignment during training in order to get power?"
Joe Carlsmith Audio
373:17
Introduction and summary of "Scheming AIs: Will AIs fake alignment during training in order to get power?"
Joe Carlsmith Audio
56:32
In memory of Louise Glück
Joe Carlsmith Audio
21:22
On the limits of idealized values
Joe Carlsmith Audio
60:14
Predictable updating about AI risk
Joe Carlsmith Audio
63:14
Existential Risk from Power-Seeking AI (shorter version)
Joe Carlsmith Audio
55:03
Problems of evil
Joe Carlsmith Audio
35:42
Seeing more whole
Joe Carlsmith Audio
52:26
Why should ethical anti-realists do ethics?
Joe Carlsmith Audio
53:29
Is Power-Seeking AI an Existential Risk?
Joe Carlsmith Audio
201:02
On sincerity
Joe Carlsmith Audio
95:02