Emilio Villa Cueva
@evllcv
Followers
72
Following
788
Media
0
Statuses
24
PhD @ mbzuai. social intelligence, multimodality, & multilinguality
Abu Dhabi
Joined May 2023
I'll be in *SEM today, presenting efforts led by @AtnafuLambebo on developing technology and data for African languages. The talk also summarizes our recent survey on what we mean by low-resource in NLP. I hope to amplify the message from our group and others in our field,
aclanthology.org
Israel Abebe Azime, Atnafu Lambebo Tonja, Tadesse Destaw Belay, Yonas Chanie, Bontu Fufa Balcha, Negasi Haile Abadi, Henok Biadglign Ademtew, Mulubrhan Abebe Nerea, Debela Desalegn Yadeta, Derartu...
0
3
10
Imagine losing first authorship because you got hit by a blue shell on the last lap 💀
LLMs are injective and invertible. In our new paper, we show that different prompts always map to different embeddings, and this property can be used to recover input tokens from individual embeddings in latent space. (1/6)
87
4K
52K
Welcome back lunch for RiTUAL lab: a new semester started and we have some new faces and some members completing their appointment with us. I'm thankful for the contributions and connections that the researchers in my group bring. I'm still hiring, visiting students, postdocs,
1
4
21
This wouldn’t have been possible without the amazing team behind it. Huge thanks to all the co-authors @jcblaisecruz
@AlhamFikri
@skylrwang
@radamihalcea
@thamar_solorio
0
0
4
Multimodal input helps, but models still struggle to integrate visual and auditory cues effectively, especially over longer contexts. MOMENTS reveals the limitations in current systems and offers a path toward building more socially intelligent AI
1
0
3
Visual and auditory cues (like gaze, facial expressions, prosody) often play a significant role in understanding the interaction to answer the question correctly. We also reduce answer-set bias using an LLM-in-the-loop process, so models must reason from context, not just guess.
1
0
3
Each question is grounded in social situations within self-contained stories, allowing a deeper understanding of characters and their mental states. We go beyond beliefs/goals, covering emotions, intentions, sensory perceptions, non-literal communication, and more.
1
0
2
Excited to finally share MOMENTS!! A new human-annotated benchmark to evaluate Theory of Mind in multimodal LLMs using long-form videos with real human actors. 📽️ 2.3K+ MCQA items from 168 short films 🧠 Tests 7 different ToM abilities 🔗
1
8
22
We put 5 VLMs to the test, and we found that visual input improves CSI preservation, gender marking, and lexical disambiguation, with multimodal translations often preferred by native speakers even when automatic metrics show modest gains.
1
0
0
CAMMT features: 5.8k+ image–L1–L2 caption triples 1.5k samples focused on Cultural Specific Item (CSI) Data in 👉
huggingface.co
1
0
0
We’re excited to introduce CaMMT: a human-curated benchmark for culturally aware multimodal translation. Covering 23 regions, it shows how images can help preserve cultural nuance in translation. 📷+📝=🌍 📄 https://t.co/iNxK5nSxFT
@AtnafuLambebo, @injy_hamed, @thamar_solorio
arxiv.org
Translating cultural content poses challenges for machine translation systems due to the differences in conceptualizations between cultures, where language alone may fail to convey sufficient...
1
5
17
🔥 Llega la IV edición del #HackathonSomosNLP 2025. Vamos a alinear los #LLMs con la cultura de los países de LATAM y la Península Ibérica. Anímate a participar, a dar una keynote, a patrocinar el evento o a apoyarnos con difusión. ¡Contamos contigo! 🚀 https://t.co/ov4GQpdrBU
0
17
32
It's today! Two chances to hear @Davidromogr present CVQA @NeurIPSConf: oral session 4A at 3:50 in East meeting rooms 1-3 And poster session 4 West, West Ballroom A-D, 4:30-7:30pm. I'll be there and happy to talk about this and any other research going on in my group. CU there!
2
3
8
Very happy to share this!🙌🏼
🎉 @evllcv former @nlp_cimat student, won 2nd place in the 2024 Best MSc Thesis in AI by @smia_mx! Supervised by @apastorlm & @FerdotSV (pictured), his NLP thesis led to 5 publications, including a regular one at #NAACL2024 and top places at PAN@CLEF2023 & PoliticEs@IberLEF2022.
1
1
14
🔥 Presentamos #LaLeadeboard, la primera leaderboard open-source para evaluar automáticamente #LLM en las variedades del español y lenguas oficiales de España y LATAM. https://t.co/EklRbCex8m
6
76
245
🎉Happy to share our recent collaborative effort on building a culturally diverse, multilingual visual QA dataset! CVQA consists of over 9,000 questions across 28 countries, covering 26 languages (with more to be added!) 🌐 https://t.co/2JzSqhyMmp 📜 https://t.co/UbtRGvFvJ2
8
44
199
so proud to have been part of this lab and really grateful for the awesome mentors i got to learn from! 🤘🏽
🎉 Today, Emilio Villa successfully defended his master's thesis in CS at CIMAT. This thesis is the first at CIMAT in the format of a compendium of articles and includes three papers from prestigious forums in NLP: IberLef@SEPLN, PAN@CLEF, and NAACL. Congratulations, Emilio!👏📚
1
0
11