Artwork

Το περιεχόμενο παρέχεται από το Turpentine, Erik Torenberg, and Nathan Labenz. Όλο το περιεχόμενο podcast, συμπεριλαμβανομένων των επεισοδίων, των γραφικών και των περιγραφών podcast, μεταφορτώνεται και παρέχεται απευθείας από τον Turpentine, Erik Torenberg, and Nathan Labenz ή τον συνεργάτη της πλατφόρμας podcast. Εάν πιστεύετε ότι κάποιος χρησιμοποιεί το έργο σας που προστατεύεται από πνευματικά δικαιώματα χωρίς την άδειά σας, μπορείτε να ακολουθήσετε τη διαδικασία που περιγράφεται εδώ https://el.player.fm/legal.
Player FM - Εφαρμογή podcast
Πηγαίνετε εκτός σύνδεσης με την εφαρμογή Player FM !

GELU, MMLU, & X-Risk Defense in Depth, with the Great Dan Hendrycks

2:38:31
 
Μοίρασέ το
 

Manage episode 445856065 series 3452589
Το περιεχόμενο παρέχεται από το Turpentine, Erik Torenberg, and Nathan Labenz. Όλο το περιεχόμενο podcast, συμπεριλαμβανομένων των επεισοδίων, των γραφικών και των περιγραφών podcast, μεταφορτώνεται και παρέχεται απευθείας από τον Turpentine, Erik Torenberg, and Nathan Labenz ή τον συνεργάτη της πλατφόρμας podcast. Εάν πιστεύετε ότι κάποιος χρησιμοποιεί το έργο σας που προστατεύεται από πνευματικά δικαιώματα χωρίς την άδειά σας, μπορείτε να ακολουθήσετε τη διαδικασία που περιγράφεται εδώ https://el.player.fm/legal.

Join Nathan for an expansive conversation with Dan Hendrycks, Executive Director of the Center for AI Safety and Advisor to Elon Musk's XAI. In this episode of The Cognitive Revolution, we explore Dan's groundbreaking work in AI safety and alignment, from his early contributions to activation functions to his recent projects on AI robustness and governance. Discover insights on representation engineering, circuit breakers, and tamper-resistant training, as well as Dan's perspectives on AI's impact on society and the future of intelligence. Don't miss this in-depth discussion with one of the most influential figures in AI research and safety.

Check out some of Dan's research papers:

MMLU: https://arxiv.org/abs/2009.03300

GELU: https://arxiv.org/abs/1606.08415

Machiavelli Benchmark: https://arxiv.org/abs/2304.03279

Circuit Breakers: https://arxiv.org/abs/2406.04313

Tamper Resistant Safeguards: https://arxiv.org/abs/2408.00761

Statement on AI Risk: https://www.safe.ai/work/statement-on-ai-risk

Apply to join over 400 Founders and Execs in the Turpentine Network: https://www.turpentinenetwork.co/

SPONSORS:

Shopify: Shopify is the world's leading e-commerce platform, offering a market-leading checkout system and exclusive AI apps like Quikly. Nobody does selling better than Shopify. Get a $1 per month trial at https://shopify.com/cognitive.

LMNT: LMNT is a zero-sugar electrolyte drink mix that's redefining hydration and performance. Ideal for those who fast or anyone looking to optimize their electrolyte intake. Support the show and get a free sample pack with any purchase at https://drinklmnt.com/tcr.

Notion: Notion offers powerful workflow and automation templates, perfect for streamlining processes and laying the groundwork for AI-driven automation. With Notion AI, you can search across thousands of documents from various platforms, generating highly relevant analysis and content tailored just for you - try it for free at https://notion.com/cognitiverevolution

Oracle: Oracle Cloud Infrastructure (OCI) is a single platform for your infrastructure, database, application development, and AI needs. OCI has four to eight times the bandwidth of other clouds; offers one consistent price, and nobody does data better than Oracle. If you want to do more and spend less, take a free test drive of OCI at https://oracle.com/cognitive

CHAPTERS:

(00:00:00) Teaser

(00:00:48) About the Show

(00:02:17) About the Episode

(00:05:41) Intro

(00:07:19) GELU Activation Function

(00:10:48) Signal Filtering

(00:12:46) Scaling Maximalism

(00:18:35) Sponsors: Shopify | LMNT

(00:22:03) New Architectures

(00:25:41) AI as Complex System

(00:32:35) The Machiavelli Benchmark

(00:34:10) Sponsors: Notion | Oracle

(00:37:20) Understanding MMLU Scores

(00:45:23) Reasoning in Language Models

(00:49:18) Multimodal Reasoning

(00:54:53) World Modeling and Sora

(00:57:07) Arc Benchmark and Hypothesis

(01:01:06) Humanity's Last Exam

(01:08:46) Benchmarks and AI Ethics

(01:13:28) Robustness and Jailbreaking

(01:18:36) Representation Engineering

(01:30:08) Convergence of Approaches

(01:34:18) Circuit Breakers

(01:37:52) Tamper Resistance

(01:49:10) Interpretability vs. Robustness

(01:53:53) Open Source and AI Safety

(01:58:16) Computational Irreducibility

(02:06:28) Neglected Approaches

(02:12:47) Truth Maxing and XAI

(02:19:59) AI-Powered Forecasting

(02:24:53) Chip Bans and Geopolitics

(02:33:30) Working at CAIS

(02:35:03) Extinction Risk Statement

(02:37:24) Outro

  continue reading

197 επεισόδια

Artwork
iconΜοίρασέ το
 
Manage episode 445856065 series 3452589
Το περιεχόμενο παρέχεται από το Turpentine, Erik Torenberg, and Nathan Labenz. Όλο το περιεχόμενο podcast, συμπεριλαμβανομένων των επεισοδίων, των γραφικών και των περιγραφών podcast, μεταφορτώνεται και παρέχεται απευθείας από τον Turpentine, Erik Torenberg, and Nathan Labenz ή τον συνεργάτη της πλατφόρμας podcast. Εάν πιστεύετε ότι κάποιος χρησιμοποιεί το έργο σας που προστατεύεται από πνευματικά δικαιώματα χωρίς την άδειά σας, μπορείτε να ακολουθήσετε τη διαδικασία που περιγράφεται εδώ https://el.player.fm/legal.

Join Nathan for an expansive conversation with Dan Hendrycks, Executive Director of the Center for AI Safety and Advisor to Elon Musk's XAI. In this episode of The Cognitive Revolution, we explore Dan's groundbreaking work in AI safety and alignment, from his early contributions to activation functions to his recent projects on AI robustness and governance. Discover insights on representation engineering, circuit breakers, and tamper-resistant training, as well as Dan's perspectives on AI's impact on society and the future of intelligence. Don't miss this in-depth discussion with one of the most influential figures in AI research and safety.

Check out some of Dan's research papers:

MMLU: https://arxiv.org/abs/2009.03300

GELU: https://arxiv.org/abs/1606.08415

Machiavelli Benchmark: https://arxiv.org/abs/2304.03279

Circuit Breakers: https://arxiv.org/abs/2406.04313

Tamper Resistant Safeguards: https://arxiv.org/abs/2408.00761

Statement on AI Risk: https://www.safe.ai/work/statement-on-ai-risk

Apply to join over 400 Founders and Execs in the Turpentine Network: https://www.turpentinenetwork.co/

SPONSORS:

Shopify: Shopify is the world's leading e-commerce platform, offering a market-leading checkout system and exclusive AI apps like Quikly. Nobody does selling better than Shopify. Get a $1 per month trial at https://shopify.com/cognitive.

LMNT: LMNT is a zero-sugar electrolyte drink mix that's redefining hydration and performance. Ideal for those who fast or anyone looking to optimize their electrolyte intake. Support the show and get a free sample pack with any purchase at https://drinklmnt.com/tcr.

Notion: Notion offers powerful workflow and automation templates, perfect for streamlining processes and laying the groundwork for AI-driven automation. With Notion AI, you can search across thousands of documents from various platforms, generating highly relevant analysis and content tailored just for you - try it for free at https://notion.com/cognitiverevolution

Oracle: Oracle Cloud Infrastructure (OCI) is a single platform for your infrastructure, database, application development, and AI needs. OCI has four to eight times the bandwidth of other clouds; offers one consistent price, and nobody does data better than Oracle. If you want to do more and spend less, take a free test drive of OCI at https://oracle.com/cognitive

CHAPTERS:

(00:00:00) Teaser

(00:00:48) About the Show

(00:02:17) About the Episode

(00:05:41) Intro

(00:07:19) GELU Activation Function

(00:10:48) Signal Filtering

(00:12:46) Scaling Maximalism

(00:18:35) Sponsors: Shopify | LMNT

(00:22:03) New Architectures

(00:25:41) AI as Complex System

(00:32:35) The Machiavelli Benchmark

(00:34:10) Sponsors: Notion | Oracle

(00:37:20) Understanding MMLU Scores

(00:45:23) Reasoning in Language Models

(00:49:18) Multimodal Reasoning

(00:54:53) World Modeling and Sora

(00:57:07) Arc Benchmark and Hypothesis

(01:01:06) Humanity's Last Exam

(01:08:46) Benchmarks and AI Ethics

(01:13:28) Robustness and Jailbreaking

(01:18:36) Representation Engineering

(01:30:08) Convergence of Approaches

(01:34:18) Circuit Breakers

(01:37:52) Tamper Resistance

(01:49:10) Interpretability vs. Robustness

(01:53:53) Open Source and AI Safety

(01:58:16) Computational Irreducibility

(02:06:28) Neglected Approaches

(02:12:47) Truth Maxing and XAI

(02:19:59) AI-Powered Forecasting

(02:24:53) Chip Bans and Geopolitics

(02:33:30) Working at CAIS

(02:35:03) Extinction Risk Statement

(02:37:24) Outro

  continue reading

197 επεισόδια

Minden epizód

×
 
Loading …

Καλώς ήλθατε στο Player FM!

Το FM Player σαρώνει τον ιστό για podcasts υψηλής ποιότητας για να απολαύσετε αυτή τη στιγμή. Είναι η καλύτερη εφαρμογή podcast και λειτουργεί σε Android, iPhone και στον ιστό. Εγγραφή για συγχρονισμό συνδρομών σε όλες τις συσκευές.

 

Οδηγός γρήγορης αναφοράς