Hidir Yesiltepe @ NeurIPS’25
@d_yesiltepe
Followers
235
Following
338
Media
20
Statuses
75
PhD Student @virginia_tech | Research Intern @Adobe | Ex Intern @AmazonScience | @FAL Fellow
United States
Joined July 2024
🎬 2026 will be the year of autoregressive video models. As we wrap up 2025, we ask a critical question: How far can a diffusion model distilled with Self-Forcing on only 5-second, 16-FPS videos be pushed into long-form video generation without any supervision? ✨We introduce
3
10
62
So grateful for the incredible students and alumni of GEMLAB! Together, we published 4 main conference papers + 4 workshop papers at @NeurIPSConf. You’re all absolute gems 💎
1
4
32
Today we met the legendary @docmilanfar and in the back @yusuf_dalva looked terrified by his sudden appearance :)
2
2
45
We are live today with @d_yesiltepe ! Make sure to drop by to hear about LoRAShop, where we perform multi concept generation + editing without any training! 📍 Thu, Dec 4 · 4:30–7:30 PM PST 📷 Poster #4306 Project: https://t.co/YY7kHSfaXP Code:
github.com
Implementation for project LoRAShop (NeurIPS 2025 - Spotlight) - gemlab-vt/LoRAShop
🚀 Excited to present LoRAShop at NeurIPS (Spotlight)! A training-free framework for multi-concept image generation & editing using rectified flow transformers. Seamless identity-preservation, no retraining, no complex prompts: just clean compositional control. @d_yesiltepe
0
2
19
📸 Presenting Dynamic View Synthesis as an Inverse Problem! 📍Exhibit Hall C, #5306 ⏱️11:00 AM - 2:00 PM
✨I will be at #NeurIPS2025 in San Diego presenting 2 projects. 💬 Would be glad to chat on long-form video generation, video distillation, dynamic view synthesis and video super-resolution! • Dynamic View Synthesis as an Inverse Problem: https://t.co/8Ly2poDLoi • LoRAShop:
0
0
4
Thanks @_akhaliq for sharing our work. Self-Forcing unlocks ✨ real-time ✨ long-form action-controllable dynamic video generation when combined with Infinity-RoPE! 💫 We will be releasing the code next week, and I would be glad to chat about the project in person at NeurIPS!
Infinity-RoPE Action-Controllable Infinite Video Generation Emerges From Autoregressive Self-Rollout
0
0
3
✨I will be at #NeurIPS2025 in San Diego presenting 2 projects. 💬 Would be glad to chat on long-form video generation, video distillation, dynamic view synthesis and video super-resolution! • Dynamic View Synthesis as an Inverse Problem: https://t.co/8Ly2poDLoi • LoRAShop:
0
1
8
Introducing Canvas-to-Image (C2I): A new paradigm where you define all controls within a single RGB canvas. 🎨 We simplify complex generation into one intuitive interface. Place specific Identities, Poses, and Boxes to control exactly who appears, how they pose, and where they
16
41
283
@tunahansalih @akaan_akan @oktay_kaan @PINguAR @fal 🌐 Project Page: https://t.co/NCNSiZ7HIH 📝 Arxiv:
arxiv.org
Current autoregressive video diffusion models are constrained by three core bottlenecks: (i) the finite temporal horizon imposed by the base model's 3D Rotary Positional Embedding (3D-RoPE), (ii)...
0
1
6
🎊I sincerely thank all my excellent team members @tunahansalih, @akaan_akan, @oktay_kaan for their hard work and my supervisor @PINguAR for her valuable feedbacks throughout the project. Special thanks goes to @fal for providing the resources that made this research possible.
1
1
6
🔜To introduce dynamic scene cuts and add cinematic variation, we use RoPE Cut. It applies a controlled discontinuous jump in the temporal RoPE coordinates, re-anchoring the current block to a new temporal location. This allows background, environment, or time-of-day to change
1
1
4
✂ To achieve fast prompt responsiveness, we introduce KV Flush. It renews the KV cache by keeping only two anchors, the global sink frame and the last generated latent frame, allowing new prompts to take effect immediately while preserving smooth local motion.
1
0
4
💫When the KV cache size is beyond the base model’s natural horizon, Block-Relativistic RoPE transitions into the semanticization regime. Instead of assigning unique temporal indices to very old frames, their positions are collapsed to a shared minimum index, turning them into
1
0
5
♾️At the core of our method is Block-Relativistic RoPE, which redefines temporal encoding as a moving frame of reference. New blocks are rotated relative to the base model’s horizon, earlier blocks are rotated backward, and relative geometry is preserved.
2
1
6
Excited to share that LoRAShop has been accepted at #NeurIPS25 as a Spotlight! 🎉 Huge thanks to my collaborator @d_yesiltepe and my advisor @PINguAR 🙏 Code will be released soon, looking forward to connecting with the community in San Diego!
👨🎨 LoRAShop is now out! We introduce LoRAShop, that enables both generation and editing with multiple personalized concepts (no training), pushing the boundaries of the task of image editing! Kudos to my long time collaborator @d_yesiltepe and my advisor @PINguAR !
5
2
36
And accepted at #NeurIPS2025! See you in San Diego.
✨ We introduce Dynamic View Synthesis as an Inverse Problem, a training free framework for generating novel views from a single monocular video by operating entirely in the diffusion noise initialization phase, with no weight updates, no architecture changes.
1
0
27
Dinner at #CVPR turned into a GenAI think tank. 🍜 Veo3, personalization rants, and nonstop energy from amazing folks across Fal and Google. Couldn’t have asked for a better crew. @natanielruizg @gorkemyurt @d_yesiltepe @d_yesiltepe @yusuf_dalva
1
5
34
Dynamic View Synthesis as an Inverse Problem Contributions: • We identify and formalize the Zero-Terminal SNR Collapse Problem, showing that while zero terminal SNR schedules improve generation quality, they inherently break injectivity, preventing deterministic inversion and
1
6
44