@_kevinlu Profile picture

Kevin Lu

@_kevinlu

@openai. ex-@berkeley_ai, advised by @pabbeel & @imordatch

Similar User
Igor Mordatch photo

@IMordatch

Abhishek Gupta photo

@abhishekunique7

Dhruv Shah photo

@shahdhruv_

Hao Liu photo

@haoliuhl

Jason Weston photo

@jaseweston

Stefano Ermon photo

@StefanoErmon

Clémentine Dominé 🍊 photo

@ClementineDomi6

Kimin photo

@kimin_le2

Zhuohan Li photo

@zhuohan123

Lerrel Pinto photo

@LerrelPinto

Yu Bai photo

@yubai01

Siddharth Karamcheti photo

@siddkaramcheti

Archit Sharma photo

@archit_sharma97

Denis Yarats photo

@denisyarats

Kuan Fang photo

@KuanFang

Pinned

Come check out o1-mini: SoTA math reasoning in a small package openai.com/index/openai-o… with @ren_hongyu @shengjia_zhao @Eric_Wallace_ & the rest of the OpenAI team

Tweet Image 1

Kevin Lu Reposted

A Codeforces contestant used o1-mini in a live contest today codeforces.com/blog/entry/133… and achieved near master-level performance! Agree with the decision to restrict AI in competition going forwards, and it'll be interesting to see how the contest scene evolves.


Kevin Lu Reposted

Thrilled to release o1-mini, a model near and dear to my heart 💙. o1-mini is an efficient model in the o1 series that’s super performant in STEM reasoning, especially math and coding. I can’t wait to see what you all build with o1-mini!! openai.com/index/openai-o…


Kevin Lu Reposted

“OpenAI says that more than 200 million people use ChatGPT each week […] while API usage has doubled following the release of the company’s cheaper and smarter model GPT-4o mini” Has @OpenAI API usage really doubled in the past five weeks since 4o-mini? theverge.com/2024/8/29/2423…


Kevin Lu Reposted

Exciting Chatbot Arena Update -- GPT-4o mini's result is out! With 4K+ user votes, GPT-4o mini climbs to the top of the leaderboard, now joint #1 with GPT-4o while being 20x cheaper! Significantly better than its early version ("upcoming-gpt-mini") in Arena across the boards.…

Tweet Image 1
Tweet Image 2

Kevin Lu Reposted

Excited to release our new small model, developed by a killer crew of team players. Intelligence per $ is very strong with GPT-4o mini. Your turn, developers! omniminiomniminiomnimini (say it 5 times fast)

We’re continuing to make advanced AI accessible to all with the launch of GPT-4o mini, now available in the API and rolling out in ChatGPT today.



I recently joined OpenAI! Come check out our new model: 82% MMLU at 60 cents per 1M output tokens! openai.com/index/gpt-4o-m…

Tweet Image 1

Kevin Lu Reposted

In our new work - Algorithm Distillation - we show that transformers can improve themselves autonomously through trial and error without ever updating their weights. No prompting, no finetuning. A single transformer collects its own data and maximizes rewards on new tasks. 1/N


Kevin Lu Reposted

LLMs like GPT-3 and Codex contain rich world knowledge. In this fun study, we ask if GPT like models can plan actions for embodied agents. Turns out, with apt sanity checks, even vanilla LLMs without any finetuning can generate good high-level plans given a low-level controller.


Come chat with us about sequence modeling for reinforcement learning @NeurIPSConf tomorrow (Thurs 12/9) at 8:30-10am PT! gather.town/app/XRWlik7kvt…

Tweet Image 1

Can RL algorithms be replaced with transformer-based language models? We’ve looked at this question with our work on Decision Transformer: Website: sites.google.com/corp/berkeley.… Code: github.com/kzl/decision-t… 1/8



Kevin Lu Reposted

Currently It is challenging to measure progress in Unsupervised RL w/o having common tasks & protocol. To take a step in addressing this issue we release our #NeurIPS2021 paper: (URLB) Unsupervised RL Benchmark! Paper: bit.ly/3bwHhY8 Code: bit.ly/3bAvI1S 1/N

Tweet Image 1

Really exciting work looking at how to utilize frozen language models for multimodal tasks! Great to see more successes in this direction.

Our new paper shows how to prompt a pre-trained text language model with a combination of text AND images (🖼️,🔤, 🖼️,🔤, 🖼️,🔤). Keep the language model 🧊 frozen 🧊 and train a vision encoder to embed images into the same space as word sequences. arxiv.org/abs/2106.13884 (1/12)



Excited to share our new work on applying language modeling ideas to RL policy optimization! Jointly led with @lchen915 and other amazing collaborators.

Can RL algorithms be replaced with transformer-based language models? We’ve looked at this question with our work on Decision Transformer: Website: sites.google.com/corp/berkeley.… Code: github.com/kzl/decision-t… 1/8



Complementary blog post to our paper, Pretrained Transformers as Universal Computation Engines, has been released! bair.berkeley.edu/blog/2021/03/2…


Kevin Lu Reposted

What are the limits to the generalization of large pretrained transformer models? We find minimal fine-tuning (~0.1% of params) performs as well as training from scratch on a completely new modality! with @_kevinlu, @adityagrover_, @pabbeel paper: arxiv.org/abs/2103.05247 1/8


Kevin Lu Reposted

sites.google.com/corp/berkeley.… Excited to share work led by amazing Kevin Lu in collaboration with @adityagrover_ and @pabbeel! What’s holding us back from agents that learn in reset-free, lifelong settings?


Loading...

Something went wrong.


Something went wrong.