@lukeprog Profile picture

Luke Muehlhauser

@lukeprog

Open Philanthropy Senior Program Officer, AI Governance and Policy

Similar User
Katja Grace 🔍 photo

@KatjaGrace

Scott Alexander photo

@slatestarcodex

Ajeya Cotra photo

@ajeya_cotra

Joe Carlsmith photo

@jkcarlsmith

Kelsey Piper photo

@KelseyTuoc

Rohin Shah photo

@rohinmshah

Michael Aird photo

@michael__aird

Victoria Krakovna photo

@vkrakovna

Amanda Askell photo

@AmandaAskell

Rob Bensinger ⏹️ photo

@robbensinger

Nate Soares ⏹️ photo

@So8res

EA Forum Posts photo

@EAForumPosts

Owain Evans photo

@OwainEvans_UK

Toby Ord photo

@tobyordoxford

Anders Sandberg photo

@anderssandberg

Luke Muehlhauser Reposted

AI skeptics: LLMs are copy-paste engines, incapable of original thought, basically worthless. Professionals who track AI progress: We've worked with 60 mathematicians to build a hard test that modern systems get 2% on. Hope this benchmark lasts more than a couple of years.

Tweet Image 1

.@Miles_Brundage has been doing smart work on AI safety policy for more than a decade and is looking for potential co-founders for a new AI strategy/policy org! milesbrundage.substack.com/p/why-im-leavi…


Luke Muehlhauser Reposted

🚨 🚨 🚨 We are looking for someone great to lead a new program focused on accelerating economic growth in developing countries.🚨 🚨 🚨 If we find the right person, they'll oversee at least $30M of spending over the next few years. Here's why we think this is a great bet:


Luke Muehlhauser Reposted

Our Systemic AI Safety Fast Grants scheme is open for applications. In partnership with @UKRI_News, we’re working to advance this new area of research, building the resilience of our society and infrastructure to AI-related hazards. Find out more: aisi.gov.uk/work/advancing…

Tweet Image 1

Great post from @DarioAmodei on many of the huge benefits we could plausibly get from powerful AI if we mitigate the risks successfully! darioamodei.com/machines-of-lo…


Luke Muehlhauser Reposted

At @open_phil, we see a lot of our work as hits-based giving: high-risk, hopefully high-reward, and built on the understanding that a few exceptional successes can offset a lot of failures. After a decade of grantmaking, we’re beginning to see some of our early bets pay off. 🧵


Luke Muehlhauser Reposted

Repurposing generic drugs to treat additional diseases has great potential, but there's no incentive for the private sector to fund that research or clinical trials. We @Arnold_Ventures are thrilled to join other funders to raise $60 mil for Every Cure to advance these efforts.


Luke Muehlhauser Reposted

Update: we’re hiring for multiple positions! Join GDM to shape the frontier of AI safety, governance, and strategy. Priority areas: forecasting AI, geopolitics and AGI efforts, FSF risk management, agents, global governance. More details below: 🧵

We are hiring! Google DeepMind's Frontier Safety and Governance team is dedicated to mitigating frontier AI risks; we work closely with technical safety, policy, responsibility, security, and GDM leadership. Please encourage great people to apply! 1/ boards.greenhouse.io/deepmind/jobs/…



Luke Muehlhauser Reposted

It's been fun watching papers start to come out of this RFP on benchmarking LLM agents on consequential real world tasks that @ajeya_cotra ran: openphilanthropy.org/rfp-llm-benchm… Bunch of cool ones so far: 🧵


Luke Muehlhauser Reposted

Today, we're excited to announce ForecastBench: a new benchmark for evaluating AI and human forecasting capabilities. Our research indicates that AI remains worse at forecasting than expert forecasters. 🧵 Arxiv: arxiv.org/abs/2409.19839 Website: forecastbench.org


New RFP for LLM-focused AI safety work, application deadline is Nov. 8th! schmidtsciences.org/safe-ai/


Luke Muehlhauser Reposted

Love this call to action -- think it's a very high priority to figure out "if-then" policies that people with widely varying views on the timeline to dangerous capabilities can agree on, and in parallel to develop a rigorous science to assess whether those conditions are met.

Tweet Image 1

Debates over AI Policy like CA SB-1047 highlight fragmentation in the AI community. How can we develop AI policies that help foster innovation while mitigating risks? We propose a path for science- and evidence-based AI policy: understanding-ai-safety.org

Tweet Image 1


Luke Muehlhauser Reposted

Joe Biden tells the UN that we will see more technological change in the next 2-10 years than we have seen in the last 50 and AI will change our ways of life, work and war so urgent efforts are needed on AI safety


Luke Muehlhauser Reposted

Before vs after the vaccines introduction

Tweet Image 1

An interesting case study in reward hacking. Would like to see what happens with bigger models. Paper: arxiv.org/abs/2409.12822

RLHF is a popular method. It makes your human eval score better and Elo rating 🚀🚀. But really❓Your model might be “cheating” you! 😈😈 We show that LLMs can learn to mislead human evaluators via RLHF. 🧵below

Tweet Image 1


Luke Muehlhauser Reposted

Very exciting news on lead from @PowerUSAID & @albrgr ! washingtonpost.com/opinions/2024/…

Tweet Image 1

Loading...

Something went wrong.


Something went wrong.