Explore tweets tagged as #Multimodal
@yisongyue
Yisong Yue
2 days
My student @SaberaTalukder and I are creating a new startup that deeply rethinks how we architect and engage with multimodal models. 🚀 We are chatting with investors at #NeurIPS2025, and if you want to get on our radar, DM Sabera.
12
11
179
@grass
Grass
13 hours
2025 has been a breakthrough year. Grass has evolved into one of the largest multimodal data providers, with accelerating commercial demand from AI labs and enterprise clients. Every dollar generated is being reinvested to expand network capacity and support the Grass ecosystem.
149
182
1K
@samruddhi_mokal
Samruddhi Mokal
2 days
Gemini 3 has a capability most people don't even know exists. it's not the 1M tokens. it's not the multimodal processing. it's something else entirely. And it's the reason I built 3,000+ prompts specifically for Gemini 3. Everyone talks about Gemini's specs: → 1 million
2K
1K
2K
@samruddhi_mokal
Samruddhi Mokal
7 days
Gemini 3 + Claude + N8N is absolutely INSANE This combo just replaced the entire operations team. No manual work. No $200K/year salaries. No coordination chaos. Just three AIs working together to run business 24/7. Here's how it works: → Gemini 3 handles multimodal inputs
590
428
1K
@gemsofbabus_
Gems
5 days
Rate it out of 10. 📍Multimodal Transport Hub, Ahmedabad 🇮🇳
69
71
2K
@aaxsh18
Aamir Shakir
2 days
grep is multimodal now. performs better then apple photos and gives you and your agent perfect search. just run npm install -g @mixedbread/mgrep
20
28
352
@KaichenZhang358
Kaichen Zhang
3 days
🚀 Introducing OpenMMReasoner — a transparent, reproducible recipe for multimodal reasoning. We present a 2-stage pipeline uses 874K SFT samples with step-by-step validation and 74K high-quality RL samples. Paper: https://t.co/87o8IwI26Y More in thread:
3
32
139
@elena__belli
Elena Belli
10 days
I am looking for an intern to support my PhD project on multimodal communication in common marmosets! 🐒🌞 You will be part of @EvoCogGroup at the @UZH_en (@NCCR_Language). Minimum commitment is 3 months full-time. This is an unpaid position and remote options are not available.
1
4
6
@AntGroup
Ant Group
5 days
1M downloads in just 4 days! Ant Group’s LingGuang, the multimodal AI assistant with code-driven outputs, is enabling users to build flash programs, animate complex concepts and get to know the world around them in new ways. Want to have a try? Download LingGuang from the
20
82
2K
@AntGroup
Ant Group
9 days
Ant Group today launched LingGuang, a next-generation multimodal AI assistant and the first of its kind in China that interacts with users through code-driven outputs. Equipped with the capability to understand and produce language, image, voice and data, LingGuang delivers
3
1
17
@humphrey_shi
Humphrey Shi
5 days
🚀 I’m hiring across multiple levels — AI engineers, senior IC/leads, and interns — to join me at @nvidia . We’re building the next generation of high-performance, multimodal, and agentic AI systems across the full stack: models, kernels, compilers, and hardware. You don’t need
80
197
3K
@Imagen_Network
Imagen AI
31 minutes
Imagen Network implements secure vision interpreter to enhance multimodal asset validation. Read more: https://t.co/oQQYEOrvCq
20
106
109
@melissaloumd
Melissa Lou Silva, MD
12 hours
Thrilled to see our work published in JAHA! Huge thanks to my amazing mentor Dr. Bruno Lima for this opportunity. This study provides important insights into how multimodal imaging can enhance detection and monitoring of post heart transplantation complications. 🚀 #radiology
0
1
6
@sundarpichai
Sundar Pichai
9 days
Introducing Gemini 3 ✨ It’s the best model in the world for multimodal understanding, and our most powerful agentic + vibe coding model yet. Gemini 3 can bring any idea to life, quickly grasping context and intent so you can get what you need with less prompting.  Find Gemini
878
3K
22K
@razoralign
antisense.
8 days
Evo2HiC: a multimodal foundation model for integrative analysis of genome sequence and architecture https://t.co/JHcmse7NcC
1
18
79
@acontext_io
Acontext
7 days
🚀 Introduce Acontext: the #opensource Context Data Platform for self-learning AI #agents. 💾Multimodal context storage via a unified API 📊Real-time task status observability 📚Automatic skill learning GitHub: https://t.co/uQp4EWOcqV 🥑 #AI #developer #AIAgent #LLM #llmstack
6
8
32
@HuggingPapers
DailyPapers
11 hours
Unveiling the Understanding-Generation Gap in Multimodal Models! New research introduces UniSandbox, a framework using controlled synthetic datasets to reveal key insights into how understanding truly informs generation.
1
2
16
@FellMentKE
FELIX
39 minutes
Ant Group's LingGuang just hit 2M downloads and it's turning heads. This multimodal AI assistant is free, codes flash programs, and breaks down lengthy text-based answers into 3D models, animations, voice clips, and data charts in a minimalist style - Features that sound almost
13
2
20
@TheTuringPost
Ksenia_TuringPost
2 days
Must-read AI research of the week: ▪️ Downscaling Intelligence: Exploring Perception and Reasoning Bottlenecks in Small Multimodal Models ▪️ WorldGen ▪️ Mixture of States ▪️ What Does It Take to Be a Good AI Research Agent? ▪️ Souper-Model ▪️ Seer: Online Context Learning for
5
37
157