Emilio Villa Cueva Profile
Emilio Villa Cueva

@evllcv

Followers
72
Following
788
Media
0
Statuses
24

PhD @ mbzuai. social intelligence, multimodality, & multilinguality

Abu Dhabi
Joined May 2023
Don't wanna be here? Send us removal request.
@thamar_solorio
thamar |
14 days
I'll be in *SEM today, presenting efforts led by @AtnafuLambebo on developing technology and data for African languages. The talk also summarizes our recent survey on what we mean by low-resource in NLP. I hope to amplify the message from our group and others in our field,
Tweet card summary image
aclanthology.org
Israel Abebe Azime, Atnafu Lambebo Tonja, Tadesse Destaw Belay, Yonas Chanie, Bontu Fufa Balcha, Negasi Haile Abadi, Henok Biadglign Ademtew, Mulubrhan Abebe Nerea, Debela Desalegn Yadeta, Derartu...
0
3
10
@luismbat
Luis Batalha
21 days
Imagine losing first authorship because you got hit by a blue shell on the last lap 💀
@GladiaLab
GLADIA Research Lab
26 days
LLMs are injective and invertible. In our new paper, we show that different prompts always map to different embeddings, and this property can be used to recover input tokens from individual embeddings in latent space. (1/6)
87
4K
52K
@ericzelikman
Eric Zelikman
3 months
16
67
446
@thamar_solorio
thamar |
3 months
Welcome back lunch for RiTUAL lab: a new semester started and we have some new faces and some members completing their appointment with us. I'm thankful for the contributions and connections that the researchers in my group bring. I'm still hiring, visiting students, postdocs,
1
4
21
@evllcv
Emilio Villa Cueva
5 months
This wouldn’t have been possible without the amazing team behind it. Huge thanks to all the co-authors @jcblaisecruz @AlhamFikri @skylrwang @radamihalcea @thamar_solorio
0
0
4
@evllcv
Emilio Villa Cueva
5 months
Multimodal input helps, but models still struggle to integrate visual and auditory cues effectively, especially over longer contexts. MOMENTS reveals the limitations in current systems and offers a path toward building more socially intelligent AI
1
0
3
@evllcv
Emilio Villa Cueva
5 months
Visual and auditory cues (like gaze, facial expressions, prosody) often play a significant role in understanding the interaction to answer the question correctly. We also reduce answer-set bias using an LLM-in-the-loop process, so models must reason from context, not just guess.
1
0
3
@evllcv
Emilio Villa Cueva
5 months
Each question is grounded in social situations within self-contained stories, allowing a deeper understanding of characters and their mental states. We go beyond beliefs/goals, covering emotions, intentions, sensory perceptions, non-literal communication, and more.
1
0
2
@evllcv
Emilio Villa Cueva
5 months
Excited to finally share MOMENTS!! A new human-annotated benchmark to evaluate Theory of Mind in multimodal LLMs using long-form videos with real human actors. 📽️ 2.3K+ MCQA items from 168 short films 🧠 Tests 7 different ToM abilities 🔗
1
8
22
@evllcv
Emilio Villa Cueva
6 months
Feel free to reach out if you are working in this topic!
0
0
0
@evllcv
Emilio Villa Cueva
6 months
We put 5 VLMs to the test, and we found that visual input improves CSI preservation, gender marking, and lexical disambiguation, with multimodal translations often preferred by native speakers even when automatic metrics show modest gains.
1
0
0
@evllcv
Emilio Villa Cueva
6 months
CAMMT features: 5.8k+ image–L1–L2 caption triples 1.5k samples focused on Cultural Specific Item (CSI) Data in 👉
huggingface.co
1
0
0
@evllcv
Emilio Villa Cueva
6 months
We’re excited to introduce CaMMT: a human-curated benchmark for culturally aware multimodal translation. Covering 23 regions, it shows how images can help preserve cultural nuance in translation. 📷+📝=🌍 📄 https://t.co/iNxK5nSxFT @AtnafuLambebo, @injy_hamed, @thamar_solorio
Tweet card summary image
arxiv.org
Translating cultural content poses challenges for machine translation systems due to the differences in conceptualizations between cultures, where language alone may fail to convey sufficient...
1
5
17
@SomosNLP_
SomosNLP
8 months
🔥 Llega la IV edición del #HackathonSomosNLP 2025. Vamos a alinear los #LLMs con la cultura de los países de LATAM y la Península Ibérica. Anímate a participar, a dar una keynote, a patrocinar el evento o a apoyarnos con difusión. ¡Contamos contigo! 🚀 https://t.co/ov4GQpdrBU
0
17
32
@thamar_solorio
thamar |
11 months
It's today! Two chances to hear @Davidromogr present CVQA @NeurIPSConf: oral session 4A at 3:50 in East meeting rooms 1-3 And poster session 4 West, West Ballroom A-D, 4:30-7:30pm. I'll be there and happy to talk about this and any other research going on in my group. CU there!
2
3
8
@evllcv
Emilio Villa Cueva
1 year
Very happy to share this!🙌🏼
@nlp_cimat
NLP-CIMAT
1 year
🎉 @evllcv former @nlp_cimat student, won 2nd place in the 2024 Best MSc Thesis in AI by @smia_mx! Supervised by @apastorlm & @FerdotSV (pictured), his NLP thesis led to 5 publications, including a regular one at #NAACL2024 and top places at PAN@CLEF2023 & PoliticEs@IberLEF2022.
1
1
14
@SomosNLP_
SomosNLP
1 year
🔥 Presentamos #LaLeadeboard, la primera leaderboard open-source para evaluar automáticamente #LLM en las variedades del español y lenguas oficiales de España y LATAM. https://t.co/EklRbCex8m
6
76
245
@AlhamFikri
Alham Fikri Aji
1 year
🎉Happy to share our recent collaborative effort on building a culturally diverse, multilingual visual QA dataset! CVQA consists of over 9,000 questions across 28 countries, covering 26 languages (with more to be added!) 🌐 https://t.co/2JzSqhyMmp 📜 https://t.co/UbtRGvFvJ2
8
44
199
@evllcv
Emilio Villa Cueva
1 year
so proud to have been part of this lab and really grateful for the awesome mentors i got to learn from! 🤘🏽
@nlp_cimat
NLP-CIMAT
1 year
🎉 Today, Emilio Villa successfully defended his master's thesis in CS at CIMAT. This thesis is the first at CIMAT in the format of a compendium of articles and includes three papers from prestigious forums in NLP: IberLef@SEPLN, PAN@CLEF, and NAACL. Congratulations, Emilio!👏📚
1
0
11
@XimGutierrez
Ximena Gutiérrez
1 year
Yesterday at the #NAACL2024 social event #MexicoCity
0
9
33