Daniel Duckworth Profile
Daniel Duckworth

@duck

Followers
2,495
Following
143
Media
26
Statuses
330

Researcher/Engineer at Google DeepMind, Berlin.

Berlin, DE
Joined September 2010
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@duck
Daniel Duckworth
5 months
Introducing SMERF: a streamable, memory-efficient method for real-time exploration of large, multi-room scenes on everyday devices. Our method brings the realism of Zip-NeRF to your phone or laptop! Project page: ArXiv: (1/n)
19
191
877
@duck
Daniel Duckworth
4 years
Our paper, “NeRF in the Wild”, is out! NeRF-W is a method for reconstructing 3D scenes from internet photography. We apply it to the kinds of photos you might take on vacation: tourists, poor lighting, filters, and all. (1/n)
76
1K
6K
@duck
Daniel Duckworth
4 years
For lighting and image post-processing, we introduce a low-dimensional embedding space controlling NeRF’s radiance field. This not only gives NeRF-W the capacity to model photo-specific lighting, it enables us to “relight” a scene from new angles. (3/n)
5
33
227
@duck
Daniel Duckworth
4 years
We build on NeRF, a method for learning a volumetric radiance field from a posed photo collection. We introduce two extensions to soften NeRF’s “static world” assumption: one for lighting/post-processing, the other for transient objects. (2/n)
1
8
84
@duck
Daniel Duckworth
4 years
NeRF-W improves on the SOTA by >5dB in PSNR and reduces error on other metrics by 20-50%. Qualitatively, NeRF-W produces consistent, crisp 3D geometry without fog or checkerboard artifacts. Check out the project website for more videos and the paper. (5/n)
Tweet media one
2
8
82
@duck
Daniel Duckworth
4 years
For transient objects, we introduce a secondary volumetric radiance field combined with an uncertainty field. The former explicitly captures transient objects; the latter uncertainty about the color of a pixel passing through part of the 3D space. (4/n)
Tweet media one
3
5
75
@duck
Daniel Duckworth
5 months
SMERF has the best of both worlds: we produce renders nearly indistinguishable from Zip-NeRF while rendering at 60 fps or more on desktops, laptops, and even recent smartphones, all while scaling to scenes as big as a house! (3/n)
3
6
53
@duck
Daniel Duckworth
4 years
How does one trade-off sample quality and diversity in a language model? Which decoding method is best? We introduce a multi-objective framework maximizing human judgement score subject to a constraint on diversity (entropy). (1/7)
2
8
39
@duck
Daniel Duckworth
4 years
@Knusper2000 We used a few hundred to low-digit thousands. Based on the results of NeRF, if you capture your images in a controlled environment, you might be able to get away with as few as one hundred!
3
4
36
@duck
Daniel Duckworth
4 years
@NextWorldOfTech Infinite! We don't use polygons. We learn a "volumetric radiance field" scene representation.
4
0
34
@duck
Daniel Duckworth
5 months
How do we achieve this? We distill a teacher model into a family of MERF-like student submodels, each of which specializes to a different part of the scene. Each submodel captures the entire scene, so rendering stays fast and GPU memory consumption stays low. (4/n)
Tweet media one
1
0
23
@duck
Daniel Duckworth
4 years
@pradeepviswav @benedictevans @photosynth @MSFTResearch Indeed, we were heavily inspired by Photosynth! I remember being in awe when I first saw it back in high school.
2
0
20
@duck
Daniel Duckworth
5 months
Only a single submodel needs to be in memory at a time, and while the user explores the space, we swap out old submodels and stream in new ones. We train submodels to be mutually consistent, making transitions barely noticeable. (6/n)
Tweet media one
1
0
21
@duck
Daniel Duckworth
10 years
Proximal Gradient Descent? ADMM? They're more similar than you think! http://t.co/UT4Dakoisf
1
3
4
@duck
Daniel Duckworth
5 months
We also modify MERF to significantly improve visual fidelity on small-to-medium size scenes. Our submodels capture thin geometry, high-resolution textures, and specular highlights better than ever before. (5/n)
1
1
18
@duck
Daniel Duckworth
5 months
The result: a set of compact, streaming-ready submodels ready to run at up to 60 fps in your browser. The best part: you can try it out yourself: (7/n)
1
1
16
@duck
Daniel Duckworth
4 years
@zellyn @pradeepviswav @benedictevans @photosynth @MSFTResearch Like Photosynth, we require one to run a registration pipeline such as COLMAP to derive camera parameters (position, direction, focal length, etc). Once that's done, we learn the scene representation.
0
2
16
@duck
Daniel Duckworth
5 months
Existing approaches for view-synthesis are torn between two conflicting goals: high quality and fast rendering. Most methods only achieve one or the other. (2/n)
1
1
15
@duck
Daniel Duckworth
11 years
For all -1 of you reading this, I started a blog @ http://t.co/U9M3bZFAJx. It's almost entirely optimization proofs. LAGRANGIANS TO THE FACE
0
4
9
@duck
Daniel Duckworth
6 years
Proud to have my first arXiv publication with Sam, Jascha, and Quoc!
@jaschasd
Jascha Sohl-Dickstein
6 years
Stochastic natural gradient descent corresponds to Bayesian training of neural networks, with a modified prior. This equivalence holds *even away from local minima*. Very proud of this work with Sam Smith, Daniel Duckworth, and Quoc Le.
Tweet media one
2
36
125
0
0
6
@duck
Daniel Duckworth
2 years
I'm stoked to be a contributor on Object SRT, a new method for unsupervised, posed-images-to-3D-scene representation and segmentation! It's crazy fast and, while far from perfect, is leaps and bounds better than anything I've seen yet :)
@tkipf
Thomas Kipf
2 years
So excited to share Object Scene Representation Transformer (OSRT): OSRT learns about complex 3D scenes & decomposes them into objects w/o supervision, while rendering novel views up to 3000x faster than prior methods! 🖥️ 📜 1/7
5
119
647
0
0
7
@duck
Daniel Duckworth
4 years
I'm proud to announce the release of our new paper relating Whitening, Newton's Method, and Generalization! tl;dr whitening w/o regularization significantly reduces a model's ability to generalize. Work with @negative_result @sschoenholz @ethansdy @jaschasd
@jaschasd
Jascha Sohl-Dickstein
4 years
Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible: We examine what information is usable for training neural networks, and how second order methods destroy exactly that information.
Tweet media one
9
113
551
3
0
6
@duck
Daniel Duckworth
5 months
@RadianceFields Want to learn more? Check out my OG explainer thread!
@duck
Daniel Duckworth
5 months
Introducing SMERF: a streamable, memory-efficient method for real-time exploration of large, multi-room scenes on everyday devices. Our method brings the realism of Zip-NeRF to your phone or laptop! Project page: ArXiv: (1/n)
19
191
877
0
1
5
@duck
Daniel Duckworth
4 years
Super proud of my ACL publication with @daphneipp ! tl;dr we find that the decoding methods that produce the most "human-like" text are also the easiest for BERT-style classifiers to identify. We humans and our models don't see text the same way!
1
0
5
@duck
Daniel Duckworth
5 months
@Snosixtytwo Thank you, Bernhard :). It's an honor to hear it from you!
0
0
3
@duck
Daniel Duckworth
11 years
0
0
4
@duck
Daniel Duckworth
4 years
@GiorgioPatrini Obligatory "This Video Does Not Exists", haha
1
0
4
@duck
Daniel Duckworth
4 years
None of this would be possible without my amazing collaborators! @negative_result , @sschoenholz , @ethansdyer , and @jaschasd .
0
0
3
@duck
Daniel Duckworth
4 years
But all is not lost! We also find that *regularized* second-order optimization leads to better generalization than un-regularized second-order optimization or gradient descent.
Tweet media one
2
0
3
@duck
Daniel Duckworth
4 years
@w4nderlus7 I believe and this work aren't in conflict! Two points: (1) The Meena paper says, "given two models, the one with better perplexity produces better samples." This work says, "given two samples, the more likely isn't always better."
2
0
3
@duck
Daniel Duckworth
10 years
@Samu_tweetz I'm living in SF, but I'll be in Europe till October. Let's have a reunion when I get back!
0
0
2
@duck
Daniel Duckworth
5 months
@GKopanas Thanks for the kind words, Georgios! I look forward to the next generation of 3DGS work as well. It's just a matter of time till 3D capture & presentation is accessible as 2D is today.
0
0
3
@duck
Daniel Duckworth
4 years
@NextWorldOfTech It is, but the scene representation is a "volumetric radiance field". I really like the original presentation by the NeRF authors on the subject:
1
1
3
@duck
Daniel Duckworth
5 months
@Kyrannio Don't forget my amazing collaborators at Google Research, Google Inc, and Tübingen! This was very much a team effort. More info here:
@duck
Daniel Duckworth
5 months
Introducing SMERF: a streamable, memory-efficient method for real-time exploration of large, multi-room scenes on everyday devices. Our method brings the realism of Zip-NeRF to your phone or laptop! Project page: ArXiv: (1/n)
19
191
877
2
1
2
@duck
Daniel Duckworth
12 years
@tanay @DOOMTREE fuck yeah, saw them do a mini-set in Berkeley and a full show in San Francisco. Totally reminded me of you.
0
0
2
@duck
Daniel Duckworth
4 years
Have you wondered how effective social distancing is? Or quarantining? What happens if a few people ignore social distancing? How bad is it to go to the grocery store? In short, everything helps -- especially early testing and quarantine! We're all in this together.
@3blue1brown
Grant Sanderson
4 years
New video: Simulating an epidemic. What happens when people avoid each other for the most part but still go to a common central location like a store? What if you can track and isolate cases, but 20% slip through the cracks? 50%? And much more.
111
2K
6K
1
0
1
@duck
Daniel Duckworth
4 years
@w4nderlus7 ...we didn't know how to measure its entropy. If you know how, let me know :)
1
0
2
@duck
Daniel Duckworth
4 years
Key takeaways: (i) very high likelihood samples are bad, (ii) compare decoding methods fairly by controlling entropy, and (iii) there's more to decoding methods than favoring high-likelihood samples. (6/7)
1
0
2
@duck
Daniel Duckworth
11 years
A new day, a new proof. That's right kiddies, Accelerated Proximal Gradient Descent works. http://t.co/mzvyBVKA0f
0
0
2
@duck
Daniel Duckworth
5 years
Spot on article on the state of AI and the Mind. Definitely worth the read! "Despite the remarkable commercial success of current AI systems...we still have a long way to go in mimicking truly human like intelligence."
1
0
2
@duck
Daniel Duckworth
4 years
When using log likelihood as a proxy for human judgement ("quality"), we obtain "Global Temperature Sampling", a globally-normalized decoding method that optimally traverses the quality-diversity curve. (2/7)
1
0
2
@duck
Daniel Duckworth
5 months
@conroydave Thanks for the heads up. It looks like it's running to me, except for a few dropped images. Will fix ASAP.
2
0
2
@duck
Daniel Duckworth
11 years
@johnmyleswhite I'm happy to switch as long as I'm not fighting (too many) compiler bugs! I'll give it a shot, it can't hurt.
1
0
2
@duck
Daniel Duckworth
11 months
While this blog post may only have two authors, the project itself is the hard work of a number of amazing teammates. Take a peak at the "Acknowledgments" section -- you may spot a few familiar names :)
0
0
2
@duck
Daniel Duckworth
4 years
@w4nderlus7 On the decoding side, the Meena paper also advocates for a sample-and-rank method, N=20. I hypothesize that the method doesn't surface decodes on the "too likely" side of the Likelihood Trap. We didn't compare sample-and-rank as...
1
0
2
@duck
Daniel Duckworth
11 years
@weargustin I missed your Kickstarter! Any way to get an order in? If not, any friends you recommend in your place?
1
0
2
@duck
Daniel Duckworth
12 years
@bumptech Do you guys have a public API?
0
0
1
@duck
Daniel Duckworth
4 years
@josephreisinger Thanks Joe :). It warms my heart to hear your kind words!
0
0
2
@duck
Daniel Duckworth
10 years
My first blog post in over a year: ADMM revisited http://t.co/PCsPuKhgca
1
1
2
@duck
Daniel Duckworth
5 months
@RadianceFields Big thanks to @RadianceFields for the quick write-up! Fantastic work :)
0
0
2
@duck
Daniel Duckworth
5 months
@supremebeme @RadianceFields This and other large scenes are captured with a DSLR camera and a fisheye lens. Approximately ~1500 photos are used. Capture takes 30~60 min.
1
0
2
@duck
Daniel Duckworth
4 years
We perform the first large-scale human study (>38,000 ratings) comparing decoding method/hyperparameter combinations against each other. When controlling for entropy, we find nucleus > top-k > temperature sampling in low-entropy regimes. (4/7)
Tweet media one
1
0
2
@duck
Daniel Duckworth
5 months
@Royal790721 @WholeMarsBlog No GSplats were harmed in the making of SMERF :)
1
0
1
@duck
Daniel Duckworth
12 years
Just published a Kalman Filter library for Python. Go check it out! documentation: http://t.co/PTEXRe0w , and source:
2
0
2
@duck
Daniel Duckworth
13 years
@DOOMTREE I'm digging the Summer Tour, but is there no hope of a West Coast stop?
0
0
0
@duck
Daniel Duckworth
5 months
@GG3000 @alexcarliera Thanks! My collaborators and I are super proud of our work :)
0
0
1
@duck
Daniel Duckworth
5 months
@WholeMarsBlog Want to learn more? Check out my OG explainer thread:
@duck
Daniel Duckworth
5 months
Introducing SMERF: a streamable, memory-efficient method for real-time exploration of large, multi-room scenes on everyday devices. Our method brings the realism of Zip-NeRF to your phone or laptop! Project page: ArXiv: (1/n)
19
191
877
0
0
1
@duck
Daniel Duckworth
10 years
Thanks to @jdanbrown , Mirai is cleaner and safer than ever. Mirai: multithreading done right for Python! http://t.co/Nfq4tkMWcZ
0
0
1
@duck
Daniel Duckworth
11 years
@Alpa @echen Very nice! How do you select and manage your custom worker pool?
1
0
1
@duck
Daniel Duckworth
4 years
We further find that, when pairing samples from decoding methods with random samples from the model *with equal likelihood*, temperature sampling is preferred to nucleus and top-k sampling by human raters. (5/7)
Tweet media one
1
0
1
@duck
Daniel Duckworth
4 years
@AaronHertzmann @jon_barron Thanks for noticing! Fixed.
0
0
1
@duck
Daniel Duckworth
13 years
Lookout, Sims just dropped his new EP #wildlife for free! http://t.co/Lk8ia4y
0
0
1
@duck
Daniel Duckworth
5 months
@mmalex Thank you! We're all super happy with how this turned out :)
0
0
0
@duck
Daniel Duckworth
4 years
We fit a small CNN, MLPs, and a linear model w/ and w/o Natural Gradient Descent (a second order optimizer) and find all generalize more poorly than those trained w/ gradient descent.
Tweet media one
1
1
1
@duck
Daniel Duckworth
13 years
@noahlt I DON'T KNOW WHAT YOU'RE TALKING ABOUT
0
0
1
@duck
Daniel Duckworth
13 years
@kyledoherty list comprehensions are a higher form of happiness, directly followed by grilled cheese sandwiches.
0
0
1
@duck
Daniel Duckworth
12 years
There is a God, and he wrote XStream. Shit is magic I swear.
0
0
1
@duck
Daniel Duckworth
13 years
@rezcubed man, that's what I get further going to a school that starts early!
0
0
0
@duck
Daniel Duckworth
4 years
Surprisingly, this method is *worse* than token-by-token decoding methods according to human raters! We discover this is a consequence of the "Likelihood Trap", wherein samples with exceptionally high likelihood receive low human judgement scores. (3/7)
Tweet media one
1
0
1
@duck
Daniel Duckworth
4 years
This is a highly unintuitive result! In linear regression, training on a whitened dataset w/ fewer data points than dimensions results in a model that *cannot* of doing better than random chance on a validation set!
1
0
1
@duck
Daniel Duckworth
12 years
@rxchoi never too old for Legos. Never. http://t.co/rUS3T1pq
0
0
1
@duck
Daniel Duckworth
5 months
@AISavvyKushan Not just DeepMind! This wouldn't have been possible without my amazing colleagues in Google Research, Google Inc, and Tübingen.
1
0
1
@duck
Daniel Duckworth
4 years
@ali_thespaceguy We accomplished this using a couple thousand photos from the Image Matching Challenge 2020 dataset.
0
0
1
@duck
Daniel Duckworth
4 years
@bengeliscious If your observations can be expressed as integrals a la CT scans, I don't see why not :)
1
0
1
@duck
Daniel Duckworth
5 years
@johnmyleswhite *Fidgets excitedly in my seat* Oh Boyd Oh Boyd Oh Boyd Oh Boyd.
0
0
1
@duck
Daniel Duckworth
11 years
@DataKind I'd like to come to the DataDive, but will be traveling from MA. Can you hook me up with anyone else to split a place to stay?
1
1
1
@duck
Daniel Duckworth
13 years
Is this not the most beautiful thing you've ever seen? http://t.co/j7yV5s8
1
0
1
@duck
Daniel Duckworth
13 years
Happiness is Mochi + Strawberry http://t.co/QsnIlvr
1
0
1
@duck
Daniel Duckworth
11 months
From the moment NeRF was first published, the research community knew it would be something game-changing. I'm proud to be part of the team turning this amazing line of work into a real product experience!
@GoogleAI
Google AI
11 months
Immersive View gives users a virtual, close-up look at indoor spaces in 3D! Learn how it uses neural radiance fields to seamlessly fuse photos to produce realistic, multidimensional reconstructions of your favorite businesses and public spaces →
16
98
370
2
0
1
@duck
Daniel Duckworth
12 years
0
0
1
@duck
Daniel Duckworth
13 years
@Samu_tweetz I had no idea you were a metalhead, lol
1
0
1
@duck
Daniel Duckworth
13 years
@cuttlewig as measured in opportunities to visit the donutdonut shop?
1
0
1
@duck
Daniel Duckworth
13 years
@ritikm @cuttlewig free ice cream in Cory Courtyard!
1
0
1
@duck
Daniel Duckworth
4 years
Further, applying gradient descent on a whitened dataset is *exactly* equivalent to applying Newton's Method on the original dataset. This suggests that models trained w/ second order methods may generalize as well as those trained w/ SGD.
1
0
1
@duck
Daniel Duckworth
13 years
Goodbye, @rxbofficial . It's been a wonderful 5 years to know you, and I look forward to whatever comes next.
0
0
1
@duck
Daniel Duckworth
11 years
@syhw I've seen that picture many times and know what it means, but somehow the intuition doesn't pop out to me. Still, I should include it.
0
0
0
@duck
Daniel Duckworth
4 years
@wxswxs The original NeRF folks produced depth maps and meshes learned from those maps on their project website. You can see our depth map of Trevi Fountain in the overview video @ 2:25
1
0
1
@duck
Daniel Duckworth
4 years
Here’s a head-to-head comparison of nucleus, top-k, and temperature sampling, and our newly proposed decoding method. Sampling directly from the model is by far the worst and nucleus p=0.3 is best according to human judgement. (7/7)
Tweet media one
1
0
1