Remove Benchmark Remove Learning Theory Remove Rate
article thumbnail

Stanford AI Lab Papers and Talks at NeurIPS 2021

Stanford AI Lab Blog

We’re excited to share all the work from SAIL that’s being presented at the main conference , at the Datasets and Benchmarks track and the various workshops , and you’ll find links to papers, videos and blogs below.

Contact 40
article thumbnail

Research directions Open Phil wants to fund in technical AI safety

The AI Alignment Forum

Interpretability benchmarks: Wed like to support more benchmarks for interpretability research. A benchmark should consist of a set of tasks that good interpretability methods should be able to solve. on targeted attack success rate (i.e., This criterion has some commonalities with the focus in Debenedetti et al.

professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Google at NeurIPS 2022

Google Research AI blog

Woodruff * , Fred Zhang * , Qiuyi Zhang Papers From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent Ayush Sekhari, Satyen Kale , Jason D. Goodman Stochastic Online Learning with Feedback Graphs: Finite-Time and Asymptotic Optimality Teodor V.

Google 52
article thumbnail

AXRP Episode 40 - Jason Gross on Compact Proofs and Interpretability

The AI Alignment Forum

In this episode, I speak with Jason Gross about his agenda to benchmark interpretability in this way, and his exploration of the intersection of proofs and modern machine learning. Or according to our singular learning theory friends, the local learning coefficients should be small and that implies this thing about this.

Model 52