Archana Ahlawat
@archanaahlawatproduct, AI guardrails @ Dynamo AI || prev: research & eng @PrincetonCITP, @Microsoft, @JustFuturesLaw, @DukeU.
Similar User
@reboot_hq
@CharacterHub
@divyasiddarth
@MatthewDSun
@spencerc99
@jasminewsun
@saffronhuang
@will__ye
@mslkmp
@GretchenMarina
@Manderljung
@AbramovichShira
@BenDLaufer
@gucaslelfond
@jessicadai_
personal news: I quit my job! I'm going to write full-time!
Most real-world AI applications involve human-model interaction, yet most current safety evaluations do not. In a new paper with @saffronhuang @_lamaahmad @Manderljung, we argue that we need evaluations which assess human-model interactions for more accurate safety assessments 🧵
👋Hello World! We’re Digital Witness Lab. We build privacy-respecting tools to access hard-to-reach data and empower public interest investigations.
Independent evaluation of the risks of foundation models is a vital means for transparency and instrument for accountability. While possible today, uncertainties and the lack of protections reduces its efficacy. New work led by @ShayneRedford to address sites.mit.edu/ai-safe-harbor/
🚨New paper🚨 Black-Box Access is Insufficient for Rigorous AI Audits AI audits are increasingly seen as key for governing powerful AI systems. But to be effective, audits need to be high-quality, and to produce high-quality audits, auditors need access.🧵 arxiv.org/abs/2401.14446
the thing about AI that people don't understand is that it's got all these risks. but also ! all these opportunities. not to mention the risks. but ! think of the opportunities. but the risks :( but the opportunit
If you are a public figure and tell your followers that “big new risks from advanced AI are fake”, you are wrong. Not only that, you’ll be seen to be wrong *publicly & soon*. This is not an “EA thing”, it is an oncoming train and it is going to hit you, either help out or shut up
A brilliant paper from my @PrincetonCITP colleagues analyzing the knowledge creation and sharing practices (forecasting, EA-funded prize competitions, forums) in the AI safety community that enabled this niche and contested set of ideas to go mainstream: drive.google.com/file/d/1HIwKMn…
The UK AI Safety Summit kicks off this week, highlighting the rapid rise of the field of AI safety and its influence on industry, academia and policy. In a new paper, Shazeda Ahmed, @archanaahlawat, @aawinecoff, @m0namon & I explain how this once-niche field became mainstream.🧵
🧵Announcing GPQA, a graduate-level “Google-proof” Q&A benchmark designed for scalable oversight! w/ @_julianmichael_, @sleepinyourhat GPQA is a dataset of *really hard* questions that PhDs with full access to Google can’t answer. Paper: arxiv.org/abs/2311.12022
Now is probably the time to announce that I've been writing a book about @OpenAI, the AI industry & its impacts. Here is a slice of my book reporting, combined with reporting from the inimitable @cwarzel Inside the year of chaos that led to this weekend. theatlantic.com/technology/arc…
United States Trends
- 1. Joe Douglas 6.882 posts
- 2. Maxey 9.191 posts
- 3. Jaguar 33,7 B posts
- 4. Embiid 17,5 B posts
- 5. Rodgers 10,3 B posts
- 6. #OnlyKash 10,7 B posts
- 7. Woody 13,7 B posts
- 8. Russia 924 B posts
- 9. $CUTO 8.219 posts
- 10. Nancy Mace 45,6 B posts
- 11. Ukraine 993 B posts
- 12. #HMGxCODsweeps N/A
- 13. WWIII 163 B posts
- 14. Merchan 22,2 B posts
- 15. Sony 69,8 B posts
- 16. Hacker 23,9 B posts
- 17. The Jets 26,4 B posts
- 18. SPLC 17,5 B posts
- 19. #InternationalMensDay 82,8 B posts
- 20. Sarah McBride 43,9 B posts
Who to follow
-
⚡ reboot ⚡
@reboot_hq -
CharacterHub
@CharacterHub -
Divya Siddarth
@divyasiddarth -
Matthew Sun
@MatthewDSun -
spencer chang
@spencerc99 -
jasmine sun
@jasminewsun -
Saffron Huang
@saffronhuang -
will
@will__ye -
Max Langenkamp
@mslkmp -
Gretchen Krueger
@GretchenMarina -
Markus Anderljung
@Manderljung -
shira
@AbramovichShira -
Benjamin Laufer
@BenDLaufer -
lucas gelfond
@gucaslelfond -
jessica dai
@jessicadai_
Something went wrong.
Something went wrong.