Remove Alternative Remove Learning Theory Remove Measure
article thumbnail

Six Tips for Evaluating Your Nonprofit Training Session

Beth's Blog: How Nonprofits Can Use Social Media

Alternately, you may feel so good about it and say my job is done. Use Learning Theory. To guide evaluation, there are several learning theories, including The Four Levels of Evaluation, also referred to as the Kirkpatrick Evaluation Model , was created by Donald Kirkpatrick, Ph.D. The four levels are: Reaction.

article thumbnail

How To Think Like An Instructional Designer for Your Nonprofit Trainings

Beth's Blog: How Nonprofits Can Use Social Media

All of my work these days is focused on designing and delivering effective training for nonprofits -primarily on the topics of social media, strategy, networks, and measurement. As someone who has been designing and delivering training for nonprofits over the past twenty years, the most exciting part is apply theory to your practice.

professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

The Theoretical Reward Learning Research Agenda: Introduction and Motivation

The AI Alignment Forum

A nave answer might be to measure their L 2 -distance. For example, a complete answer to question (2) would be a set of necessary and sufficient conditions on two reward functions R 1 , R 2 which characterise when it would be acceptable (as measured by R 1 ) to maximise R 2 instead of R 1.

article thumbnail

Google at ICLR 2023

Google Research AI blog

If you’re registered for ICLR 2023, we hope you’ll visit the Google booth to learn more about the exciting work we’re doing across topics spanning representation and reinforcement learning, theory and optimization, social impact, safety and privacy, and applications from generative AI to speech and robotics.

Google 105
article thumbnail

Other Papers About the Theory of Reward Learning

The AI Alignment Forum

We also managed to leverage these results to produce a new method for conservative optimisation, that tells you how much (and in what way) you can optimise a proxy reward, based on the quality of that proxy (as measured by a STARC metric ), in order to be guaranteed that the true reward doesnt decrease (and thereby prevent the Goodhart drop).

article thumbnail

Research directions Open Phil wants to fund in technical AI safety

The AI Alignment Forum

Alternatives to adversarial training : Adversarial training (and the rest of todays best alignment techniques) have failed to create LLM agents that reliably avoid misaligned goals. Alternative approaches to mitigating AI risks These research areas lie outside the scope of the clusters above. McAleese et al. Arnesen et al. Khan et al.

article thumbnail

AXRP Episode 40 - Jason Gross on Compact Proofs and Interpretability

The AI Alignment Forum

And the takeaway from this paper is that you can use proofs to measure how much compression you get. And then you prove this, and the measure of compression is how long your proof is. And a technical note: it needs to be in some first-order system or alternatively, you need to measure proof checking time as opposed to proof length.

Model 52