@ebugliarello Profile picture

Emanuele Bugliarello

@ebugliarello

Multimodal researcher @GoogleDeepMind. He/him

Similar User
Desmond Elliott photo

@delliott

EdinburghNLP photo

@EdinburghNLP

CopeNLU photo

@CopeNLU

Roy Schwartz photo

@royschwartzNLP

Yonatan Bisk photo

@ybisk

Yonatan Belinkov photo

@boknilev

Antonis Anastasopoulos @EMNLP photo

@anas_ant

Edoardo Ponti photo

@PontiEdoardo

Anna Rogers photo

@annargrs

Ivan Vulić photo

@licwu

Sarah Wiegreffe (on faculty job market!) photo

@sarahwiegreffe

Daniel Fried photo

@dan_fried

Julia Kreutzer photo

@KreutzerJulia

Ivan Titov photo

@iatitov

Shiyue Zhang photo

@byryuer

Pinned

Wouldn’t it be cool if AI could help us generate movies?🎬 We built a new benchmark to measure progress in this direction🍿 “StoryBench: A Multifaceted Benchmark for Continuous Story Visualization” 📄 arxiv.org/abs/2308.11606 👩‍💻 github.com/google/storybe… 📈 paperswithcode.com/dataset/storyb…

Tweet Image 1

Emanuele Bugliarello Reposted

Our team at Google DeepMind is seeking a Research Scientist with a strong publication record (multiple first-author papers) on multi-modal LLMs in top ML venues like NeurIPS, ICLR, CVPR. Email me at af_hiring@google.com @CordeliaSchmid


Emanuele Bugliarello Reposted

Day 2 starts in a few hours, let's go! #cardiffnlpworkshop

Tweet Image 1

Embrace cultural diversity in your large-scale data! 🌎🌍🌏 @angelinepouget’s study shows that (quantitatively) you have no reason not to 🌸

PSA: Stop pretraining your VLMs on EN-filtered data, even if it improves ImageNet and COCO‼️ Doing so impairs the model's understanding of non-English cultures❗️ I argued for years, now finally publish concrete results for this (imo) intuitively obvious recommendation A🧾🧶



Emanuele Bugliarello Reposted

PSA: Stop pretraining your VLMs on EN-filtered data, even if it improves ImageNet and COCO‼️ Doing so impairs the model's understanding of non-English cultures❗️ I argued for years, now finally publish concrete results for this (imo) intuitively obvious recommendation A🧾🧶

Want your VLM to reflect the world's rich diversity 🌍? We’re very excited to share our recent research on this topic. TLDR: to build truly inclusive models that work for everyone, don’t filter by English, and check out our recommended evaluation benchmarks. (1/7)

Tweet Image 1


Emanuele Bugliarello Reposted

PaliGemma - Open Vision Model from Google! 💎 > 3B parameter model - SigLiP + Gemma 2B > Supports images upto 896 x 896 resolution > Capable of Document understanding, Image detection, visual question answering, captioning and more > In addition to general purpose checkpoints…

Tweet Image 1

And of course, it's multilingual! 🗺️

We release PaliGemma. I'll keep it short, still on vacation: - sota open base VLM designed to transfer quickly, easily, and strongly to a wide range of tasks - Also does detection and segmentation - We provide lots of examples - Meaty tech report later! ai.google.dev/gemma/docs/pal…

Tweet Image 1


Emanuele Bugliarello Reposted

About a year ago we put "A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision" on arxiv. We call it "LiT-decoder". It's been rejected (NoT sOtA!!1) but the lessons learned have guided us, and we've use it as a benchmark in many works. A🧶about the lessons

Tweet Image 1

Emanuele Bugliarello Reposted

Google presents Revisiting Text-to-Image Evaluation with Gecko On Metrics, Prompts, and Human Ratings While text-to-image (T2I) generative models have become ubiquitous, they do not necessarily generate images that align with a given prompt. While previous work has

Tweet Image 1

Emanuele Bugliarello Reposted

"MULAN 🥷: A Study of Fact Mutability in Language Models" a benchmark to evaluate the ability of English language models to anticipate time-contingency, containing 35 relations extracted from @Wikidata, each with up to 1,500 queries. (Fierro et al, 2024) arxiv.org/pdf/2404.03036…

Tweet Image 1

Are you interested in video generation? Come intern with @laoreja001, @alirezafathi and me @GoogleAI 🇺🇸! 📨 CV + brief intro to: video-gen-internship@google.com 👀 Preferably PhD candidates with papers in computer vision 🗺️ Candidates from underrepresented groups, do apply!

Our team at Google Research is hiring a research intern working on video generation, please email xiuyegu@google.com if you are interested.



Emanuele Bugliarello Reposted

Are you a PhD student interested in memorisation, generalisation and the role of data in the era of LLMs? Come do an internship with me at @AIatMeta! metacareers.com/jobs/704140718… (Send me a ping if you apply)


Emanuele Bugliarello Reposted

✴ Hiring a Postdoctoral Researcher ✴ I am hiring a postdoc with a background in *vision and language processing*, on a 2/3 year contract. Application deadline: 15 Feb 2024 Start: ASAP Apply here: mbzuai.ac.ae/vacancy/postdo… and contact me here or via email. #NLProc #hiring


Fascinated by video generation? Come check out StoryBench: A benchmark aimed at pushing the capabilities of text-to-video models to generate stories 🎞️ Today, Poster 211 at #NeurIPS2023's 10:45-12:45 session 🎷

Wouldn’t it be cool if AI could help us generate movies?🎬 We built a new benchmark to measure progress in this direction🍿 “StoryBench: A Multifaceted Benchmark for Continuous Story Visualization” 📄 arxiv.org/abs/2308.11606 👩‍💻 github.com/google/storybe… 📈 paperswithcode.com/dataset/storyb…

Tweet Image 1


Emanuele Bugliarello Reposted

Student researcher position applications are open at Google Deepmind! I'm hosting a SR in the intersection of bias and generative models. If you're an interested PhD student please reach out! google.com/about/careers/


Loading...

Something went wrong.


Something went wrong.