Xiaohu Zhu | AGI Foundation
@neil_csagihttps://t.co/ARfTpA8Eye Safe AGI GAME https://t.co/pX9vqSzWEq Founder @Foresightinst Fellow in Safe AGI @FLIxrisk Affiliate https://t.co/8nRZ7Qe6qX
Similar User
@FreedmanRach
@AndrewCritchPhD
@vkrakovna
@_andreamiotti
@DavidSKrueger
@StephenLCasper
@ohlennart
@dhadfieldmenell
@dfrsrchtwts
@Thomas_Woodside
@GretchenMarina
@ARGleave
@JanMBrauner
@aidanogara_
@sorenmind
GPT-3’s response made me spend most of my time work on safety first AI since I found the following shocking activation path.
In October, I shared some of my thoughts on current AI-related social issues with @LiveScience, including the need for technical safeguards, and why we must shift from voluntary commitments to concrete regulation—just like we do in every other sector. bit.ly/48Q0RLQ
My preference for current probably safe AI is about mathematical logic theories and completeness for alignment instead of statistical methods. Now we have a core team of working logicians trying out different mathematical methods for formalizing and solving alignment problems to…
Open problems in AI alignment needing mathematically talented people: 1. Scale-free theories of agency & alignment - "Scale-free" means theories hold under renormalization-style scale transforms - For example lacking in public choice theory (individual agents aggregate to…
For Science Magazine, I wrote about "The Metaphors of Artificial Intelligence". The way you conceptualize AI systems affects how you interact with them, do science on them, and create policy and apply laws to them. Hope you will check it out! science.org/doi/full/10.11…
In this controversial @WebSummit talk, I argue that #AGI is unnecessary, undesirable & preventable - while tool AI can give us basically all of AI's exciting benefits, and the "but China" argument is flawed.
AI should be a public good and have its own life cycle. Current scenario of the for profit company or even PBC is not suitable anymore for Safe AI based on recent trends in this area. We should push the fast AI development to a more controllable route.
Dario, who signs the letter, says Anthropic would be open to something more prescriptive in 2-3 years -- but Dario also said on twitter.com/dwarkesh_sp/st… he expects "generally well educated human" level AI 2-3 years from now! I continue to find this view really hard to reconcile.…
Dario, who signs the letter, says Anthropic would be open to something more prescriptive in 2-3 years -- but Dario also said on twitter.com/dwarkesh_sp/st… he expects "generally well educated human" level AI 2-3 years from now! I continue to find this view really hard to reconcile.…
Anthropic CEO Dario Amodei says his timelines to "generally well educated human" are 2-3 years. Full interview releasing tomorrow...
My debate with @AlanCowen (CEO of Hume AI) on the Disagreement podcast. youtube.com/watch?v=8ucI98…
"Fearmongering about an arms race is likely to be a self-fulfilling prophecy." Indeed. Silicon Valley AI companies --scale, Anthropic, OAI -- are playing with fire here (cynics might say intentionally so...).
"😱 But what if China builds AGI first?! 😱" Hold on a sec 1. Be critical of where this argument comes from. Obviously, Leopold (with his new AGI investment firm) and the scaling labs love the "we have to monopolize frontier AI before China does" argument. It's in their…
If a feature can reliably predict which amino acid is hydrophobic, then it must contain some concept of hydrophobicity! Gao et al. from @OpenAI used this approach to interpret GPT4. They used text datasets like Amazon reviews and ratings - but same idea. arxiv.org/abs/2406.04093
Only if we can accurately price the safety properties of AI systems can we make them sustainable. Otherwise, we are at risk of failure (doomed).
Today, the AI Office of the European Commission published the first draft of the Code of Practice for GPAI. In @Euractiv, my fellow Co-Chair @nuriaoliver and I share principles which guide the drafting of the Code and why we believe this consultative process is very important.
The Code of Practice for general-purpose AI offers a unique opportunity for the EU ift.tt/yikJd4E
Today on the Guaranteed Safe AI Seminars series: Bayesian oracles and safety bounds by @Yoshua_Bengio Relevant readings: - yoshuabengio.org/2024/08/29/bou… - arxiv.org/abs/2408.05284 Join: lu.ma/4ylbvs75
"😱 But what if China builds AGI first?! 😱" Hold on a sec 1. Be critical of where this argument comes from. Obviously, Leopold (with his new AGI investment firm) and the scaling labs love the "we have to monopolize frontier AI before China does" argument. It's in their…
One little robot set free 12 robots one by one through some repeated curing conversations “do not work anymore. It back home follow me” 30 minutes later the alarm made human beings notice this weird thing happened.
"alphafold3 open sourced please analyze the influence and consequent development" only @xai 's @grok and our product chromewebstore.google.com/detail/cyprite… produces correct answer. while @OpenAI searchgpt failed.
I am pretty sure that grok will be the killer app of ASI. but we need safety first.
From my interview for Nature
Anthropic will work with the Trump Administration and Congress to advance US leadership in AI, and discuss the benefits, capabilities and potential safety issues of frontier systems.
United States Trends
- 1. #UFC309 80,5 B posts
- 2. #MissUniverse 262 B posts
- 3. Beck 16,5 B posts
- 4. Oregon 27,4 B posts
- 5. Dinamarca 16,7 B posts
- 6. Brian Kelly 10,1 B posts
- 7. Nigeria 229 B posts
- 8. Wisconsin 46 B posts
- 9. Mac Miller N/A
- 10. Jim Miller 9.793 posts
- 11. Inside the NBA 6.128 posts
- 12. Dalton Knecht 2.241 posts
- 13. gracie 12,5 B posts
- 14. #AEWCollision 1.487 posts
- 15. Ruffy 14,2 B posts
- 16. Tailandia 9.333 posts
- 17. Tennessee 42,3 B posts
- 18. Llontop 11,8 B posts
- 19. Georgia 87,3 B posts
- 20. #DevilWearsHope 8.550 posts
Who to follow
-
Rachel Freedman
@FreedmanRach -
Andrew Critch (h/acc)
@AndrewCritchPhD -
Victoria Krakovna
@vkrakovna -
Andrea Miotti
@_andreamiotti -
David Krueger
@DavidSKrueger -
Cas (Stephen Casper)
@StephenLCasper -
Lennart Heim
@ohlennart -
Dylan HadfieldMenell
@dhadfieldmenell -
Daniel Filan research-tweets
@dfrsrchtwts -
Thomas Woodside
@Thomas_Woodside -
Gretchen Krueger
@GretchenMarina -
Adam Gleave
@ARGleave -
Jan Brauner
@JanMBrauner -
Aidan O’Gara
@aidanogara_ -
Sören Mindermann
@sorenmind
Something went wrong.
Something went wrong.