
Zeeshan khan
@zeeshank95
Followers
68
Following
108
Media
8
Statuses
21
PhD @Inria Willow and @ENS_ULM
Paris, France
Joined May 2021
Below 👇 are some examples of complex prompts, the LLM generated composite object priors, and the corresponding image generation. The work is done with @CordeliaSchmid and @chen_shizhe in the Willow team of @Inria Paris and @ENS_ULM . Arxiv:
0
2
3
RT @MakarandTapaswi: 🔔New @CVPR paper evaluating compositional reasoning of Video-LLMs on 10s, action-packed clips!. 🥁 VELOCITI features 7….
0
11
0
RT @gaur_manu: Can RL fine-tuning endow MLLMs with fine-grained visual understanding?. Using our training recipe, we outperform SOTA open-s….
0
30
0
RT @gaur_manu: 🚨 Introducing Detect, Describe, Discriminate: Moving Beyond VQA for MLLM Evaluation. Given an image pair, it is easier for….
0
17
0
RT @MakarandTapaswi: Thanks to the organizers (@davmoltisanti +) for an opportunity to share my thoughts at the amazing @CVPR Workshop "Wha….
0
12
0
RT @phillip_isola: Our computer vision textbook is released!. Foundations of Computer Vision.with Antonio Torralba and Bill Freeman.https:/….
0
405
0
RT @FuteralMatthieu: Announcing mOSCAR, multilingual interleaved text-image corpus as part of @oscarnlp project. Paper: .
0
26
0
RT @MakarandTapaswi: Given multiple short movie clips, can models generate coherent identity-aware descriptions? 🤔 Turns out, this is a com….
0
14
0
RT @FuteralMatthieu: Excited to introduce MAD Speech: a new set of metrics to measure acoustic diversity in speech. Work done @GoogleDeepM….
0
13
0
RT @MakarandTapaswi: 📢Happy to announce two #CVPR2024 papers from our Katha AI group @iiit_hyderabad! 🎉🎞️🔥. 1. On 📺TV episode story summari….
0
9
0
RT @MakarandTapaswi: Excited to receive Google's India Faculty Research Award 2022, my first Indian research grant 🙂 .
0
4
0
RT @MakarandTapaswi: Given a short movie clip, can we identify who is doing what to/with whom, where, how & why? Our latest paper at @NeurI….
0
8
0