Stephan Hoyer Profile Banner
Stephan Hoyer Profile
Stephan Hoyer

@shoyer

Followers
7,836
Following
755
Media
63
Statuses
2,960

Senior staff engineer at Google Research building AI weather/climate models. Creator of @xarray_dev and scientific Python contributor. He/him.

San Francisco, CA
Joined August 2009
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@shoyer
Stephan Hoyer
3 months
New open source release from my team at Google: Dinosaur, a differentiable dynamical core for global atmospheric modeling, written in JAX: Dinosaur is a core component of NeuralGCM and we hope it is useful for the weather/climate research community.
7
66
439
@shoyer
Stephan Hoyer
2 years
Are you a PhD student interested in machine learning and numerical modeling for weather & climate? My team at Google Research is looking to hire a student researcher for this summer (and likely beyond) in either Mountain View, CA or Cambridge, MA.
18
252
866
@shoyer
Stephan Hoyer
3 years
I'm really excited to share this project that my team at @GoogleAI has been working on for the past year. We show that ML + TPUs can accelerate fluid simulations by up to two orders of magnitude without compromising accuracy or generalization.
@dkochkov1
Dmitrii Kochkov
3 years
1/2 Excited to share "Machine learning accelerated computational fluid dynamics" We use ML inside a CFD simulator to advance the accuracy/speed Pareto frontier with/ Jamie A. Smith Ayya Alieva Qing Wang Michael P. Brenner @shoyer
Tweet media one
12
123
532
16
143
822
@shoyer
Stephan Hoyer
3 years
Just hit quadruple digits for new unnamed Colab notebooks! 🎉
Tweet media one
10
18
818
@shoyer
Stephan Hoyer
2 years
Academics often ask me if we have "academic freedom" in industrial research. Of course we don't, but generally that isn't the real concern -- they want to know if I have control over my own destiny. I tell them this is my favorite part about industry. 🧵
15
56
434
@shoyer
Stephan Hoyer
4 months
I love when ML researchers get excited about science, but seriously the reviewing process for scientific applications at ML conferences (e.g., ICLR) is entirely broken. Papers with glaring errors are sailing through, without a single review from somebody with domain expertise.
20
26
353
@shoyer
Stephan Hoyer
2 years
please tell me I'm not the only one who thinks "consciousness" is not even a scientific concept
63
8
331
@shoyer
Stephan Hoyer
2 years
Gradient checkpointing (aka rematerialization) is an easy trick that can save massive amounts of memory for calculating gradients. If you differentiate through computation involving long iterative processes (like ODE solving), learn it and make it part of your toolkit! 👇🧵
5
42
329
@shoyer
Stephan Hoyer
6 years
We're developing a new protocol (__array_function__) that allows for alternate implementations of NumPy functions (e.g., for GPU, autodiff or units support): It's now implemented on the master branch of NumPy -- please give it a try and report back!
3
91
325
@shoyer
Stephan Hoyer
3 years
Can machine learning improve physics constrained optimization tasks, like those at the heart of numerical weather forecasting? Happy to share some our work on "Variational Data Assimilation with a Learned Inverse Observation Operator", tomorrow/today at ICML. 🧵
6
48
305
@shoyer
Stephan Hoyer
4 years
JAX now supports Google Cloud TPUs! I contributed this example, solving a 2D wave equation with a spatially partitioned grid. The code is remarkably simple and all in pure Python!
2
54
296
@shoyer
Stephan Hoyer
3 years
Does anyone know a good intuitive explanation for the central limit theorem? I realized the other day that even though I use it all the time I can't really justify *why* it's true.
55
25
258
@shoyer
Stephan Hoyer
3 months
My team at Google is looking to hire a PhD student intern for research on AI-based coupled Earth system modeling. This would be a full-time ~3 month position in summer or fall 2024 working in-person in Cambridge, MA with @dkochkov1 and @janniyuval .
6
58
254
@shoyer
Stephan Hoyer
7 months
Something that I think is under-appreciated in the current AI mania is that more compute does not always result in better models. Sometimes, even with perfect knowledge, you can hit a wall. A good example of this is weather prediction.
13
14
253
@shoyer
Stephan Hoyer
1 year
One lesson for writing NumPy/JAX code that took me a surprisingly long time to learn is to preserve array dimensionality whenever possible (i.e., always use keepdims=True).
3
14
237
@shoyer
Stephan Hoyer
3 years
The world's best weather forecast model is switching its numerics from double to single precision: They reinvested the 40% runtime savings in an increase in vertical resolution from 91 -> 137 levels 😍
@FPappenberger
Florian Pappenberger
3 years
New @ECMWF forecast system still on track for 11th May #IFS47r2 #newfcsystem @ECMWF Update your software to be ready! 👇
1
10
45
3
22
190
@shoyer
Stephan Hoyer
2 years
Honest question: why do ML researchers publish papers filled with giant tables? Yes, it's nice to have as a reference for comparing to prior results, but couldn't we put nice plots in the paper and the uninterpretable raw numbers in an appendix?
22
6
180
@shoyer
Stephan Hoyer
5 years
I'm happy to share a new paper, with @jaschasd and @samgreydanus : "Neural reparameterization improves structural optimization" We use neural nets to parameterize inputs of a finite elements method, and differentiable through the whole thing:
Tweet media one
2
38
170
@shoyer
Stephan Hoyer
3 years
We are hiring PhD interns for my team at Google Research: We use computational methods (especially ML) to advance research in a variety of scientific fields. For full consideration, please apply by January 15:
8
40
167
@shoyer
Stephan Hoyer
3 years
This paper "Optimal control of PDEs using physics informed neural networks" looks really nice: Finally, an assessment of PINNs that includes a runtime comparison to classical adjoint methods!
4
20
151
@shoyer
Stephan Hoyer
2 years
I'm happy to share a new project on using machine learning for computational fluid dynamics, led by @gideoknite : "Learning to correct spectral methods for simulating turbulent flows"
2
24
144
@shoyer
Stephan Hoyer
4 years
In this project led by @leeley18 , we show that end-to-end training with differentiable physics results in extremely effective hybrid physics/ML models for density functional theory.
@leeley18
Li Li
4 years
We discover that the prior knowledge embedded in the physics computation itself acts as an implicit regularization that greatly improves generalization of machine learning models for physics. Please check out our recent paper:
Tweet media one
7
114
392
2
30
145
@shoyer
Stephan Hoyer
6 years
I'm pleased to announce xarray v0.11: This release includes: - file-storage refactor for performance with @dask_dev - better support for calendars used in climate science - lots of other miscellaneous API clean-ups, bug-fixes and enhancements
3
47
131
@shoyer
Stephan Hoyer
4 years
@fchollet In 4 years of using TensorFlow for deep learning, I don't think I've ever used Keras once without cursing it
3
0
128
@shoyer
Stephan Hoyer
7 months
This year, Google's Research Scholar program for early-career professors is specifically solicitating proposals on large-language and multi-modal machine learning models for science: Applications will open next week and are due by the end of November.
1
36
130
@shoyer
Stephan Hoyer
2 years
My amazing colleagues on the Google AI for Weather & Climate team are hiring for a research/software role:
1
25
129
@shoyer
Stephan Hoyer
2 years
It is hard for me to imagine a better feeling than passing off a project to a total stranger from half way around the world, who builds on it and takes it to new heights you never imagined 🥰 This is the true magic of open source software.
5
2
130
@shoyer
Stephan Hoyer
3 years
I hope I never read another paper claiming faster simulation with ML that only compares to sims used for training data. Congrats, your method is faster and less accurate. So what? That costly reference simulation could almost certainly be made faster and less accurate, too.
12
13
127
@shoyer
Stephan Hoyer
6 months
I am so ready for LLMs to blow over so I can use data centers worth of TPUs for climate modeling 😂
@tunguz
Bojan Tunguz
6 months
We are in a midst of perhaps the greatest computing power buildup ever. Sure, the vast majority of it is going towards the AI products, but at some point, between two LLM trainings, someone will decide to use at least a fraction of it for other purposes. Hey, let’s simulate
55
63
808
4
6
124
@shoyer
Stephan Hoyer
2 years
In industry, you do not have complete freedom within a job, but you have freedom to pick an organization to work for that aligns with your values. Within a (good) job, nobody wants to tell you what to do. You succeed by showing that can deliver on the organization's mission.
5
7
116
@shoyer
Stephan Hoyer
4 years
Today's PSA: don't use continuous colormaps if quantitative comparisons are important. Our eyes aren't good at comparing colors! Example: left & right plots have same data & color scale, but colors on the right are binned. Despite strictly less info, the shape is easier to see.
Tweet media one
Tweet media two
4
9
107
@shoyer
Stephan Hoyer
2 years
New state of the art for global weather forecasting with machine learning!
@RyanKeisler
Ryan Keisler
2 years
📢 Time to share a project I’ve been working on: Forecasting Global Weather with Graph Neural Networks 🧵 (1/N)
44
309
2K
1
14
108
@shoyer
Stephan Hoyer
5 years
If only it were this easy! Everytime I've thought this was true, domain scientists have proved me wrong. Machine learning is no shortcut for domain knowledge.
@sirajraval
Siraj Raval
5 years
I'm seeing a lot of skepticism from Physicists around my "Learn Physics in 2 Months" curriculum. Machine Learning enables people to make scientific discoveries without needing as much domain knowledge. See my lecture at CERN last year for an example
21
46
318
9
23
101
@shoyer
Stephan Hoyer
4 years
Flax is worth taking a look at if you're interested in training neural nets in JAX. It's definitely a big step up from Stax!
@hardmaru
hardmaru
4 years
Flax: A neural network library for JAX designed for flexibility (pre-release)
Tweet media one
11
111
413
3
18
99
@shoyer
Stephan Hoyer
6 years
My first PR to CPython got merged. Yes, it's only a three line doc fix -- but it's exciting to have finally worked my way down to the bottom of the stack!
3
2
94
@shoyer
Stephan Hoyer
2 years
This makes it all worth it
@QuintonLawton
Quinton Lawton
2 years
What did people do before xarray? Not sure where my research would be without it
18
4
136
4
3
90
@shoyer
Stephan Hoyer
2 years
All that said -- if you're an academic considering industry, don't let "freedom" hold you back.
6
4
90
@shoyer
Stephan Hoyer
7 years
Our paper on xarray is finally out!
@_jhamman
Joe Hamman
7 years
This just out: xarray: N-D labeled Arrays and Datasets in Python:
1
25
43
2
34
89
@shoyer
Stephan Hoyer
3 years
The secret is a differentiable CFD simulator written in JAX (soon to be open sourced!), which lets us do end-to-end optimization with hard physics priors.
6
1
80
@shoyer
Stephan Hoyer
2 years
And if you do that, you will soon find yourself accumulating more freedom in your job than you know what to do with. It's not the same as the freedom of academia. It doesn't come all at once with a grant or tenure. But you earn it all the same, and it can be even more powerful.
1
3
75
@shoyer
Stephan Hoyer
7 years
xarray v0.10 has been released! Highlights include: - indexing with broadcasting over dimensions - easier wrapping of functions written for NumPy (+ auto-parallelization with dask)
1
53
76
@shoyer
Stephan Hoyer
2 years
The arguments about symbolic vs deep learning for AI reminds me a lot of arguments about numerical methods vs ML for solving physics problems. Seems like hybrid methods are the way to go in both cases.
@GaryMarcus
Gary Marcus
2 years
Deep Learning Is Hitting a Wall. What would it take for artificial intelligence to make real progress? #longread in ⁦ @NautilusMag ⁩ on one of the key technical questions in AI.
62
248
843
8
3
77
@shoyer
Stephan Hoyer
1 year
When did everyone forget how hard reinforcement learning is? I'll believe in AGI when we see chatbots behind the wheel of self-driving cars.
9
0
75
@shoyer
Stephan Hoyer
9 months
Our latest iteration on benchmarking global weather foreacasts, lead by @raspstephan
@GoogleAI
Google AI
9 months
The weather forecast is improving… literally! Introducing WeatherBench 2, a benchmark for the next generation of data-driven, global weather forecast models, providing data, tools, & an evaluation platform. Learn how to use it and check out the website →
18
127
520
0
6
75
@shoyer
Stephan Hoyer
2 years
The hardest part of research is distinguishing proofs of concept that will actually scale vs perpetual toy examples. Sadly every attempt I've seen at "intelligence" with neural nets seems to fall in the later category.
5
8
72
@shoyer
Stephan Hoyer
3 years
@DynamicWebPaige A slightly more finalized version is in the JAX docs now:
3
9
69
@shoyer
Stephan Hoyer
2 years
Finally, we're open sourced our 2D spectral Navier-Stokes solver with high-order time-stepping in JAX-CFD: It's differentiable and fast on GPU/TPU, so we hope it may be a useful reference for future ML for PDE efforts, e.g., like
@unsorsodicorda
andrea panizza
2 years
@shoyer 1. A set of standard benchmarks. Compare against high-accuracy baselines (DNS/LES, flow fields should be shared publicly). The spectral code used to compute the flow field should be shared, as well as the CPU/GPU time needed to create the baselines. We DO need the equivalent 2/n
1
0
1
2
18
71
@shoyer
Stephan Hoyer
6 years
New #xarray release v0.10.1 is out, with new IO (Iris and Zarr support) and plotting options, among many other option. Thanks to everyone who contributed!
1
26
70
@shoyer
Stephan Hoyer
2 years
This is perhaps my favorite Xarray release in years. @cherian_deepak 's Flox library speeds up many common groupby & resample operations by over 10x!
@xarray_dev
Xarray
2 years
We've just release v2022.06.0! This release brings a major internal refactor of the indexing functionality (thanks @benbovy , @cziscience ), the use of flox in groupby operations, and experimental support for the new Python Array API standard.
1
28
130
1
10
68
@shoyer
Stephan Hoyer
5 years
xarray 0.12.2 was released over the weekend: My favorite new features are the new N-D combine functions by @TomTomnicholas1 and the ability to append to existing #Zarr stores by a team of four (!) contributors, including @davidbrochart & @rabernat .
1
30
68
@shoyer
Stephan Hoyer
5 years
I wrote a demo of using JAX's auto-vectorization to implement NumPy style generalized ufuncs. It's pretty sweet!
1
19
63
@shoyer
Stephan Hoyer
4 years
I can't quite believe what showed up in my inbox today... I'm not sure whether it's more exploitive to ask students to do unpaid internships (instead, they're charged £500), or open source projects to mentor them without compensation.
Tweet media one
Tweet media two
4
13
63
@shoyer
Stephan Hoyer
2 months
This is a nice example from EarthMover of how to tune Zarr data pipelines for loading ML training data into @xarray_dev . Processing Zarr data on the fly is such a better paradigm than trying to anticipate access patterns with a preshuffled dataset on disk.
@EarthmoverHQ
Earthmover
2 months
New blog post from @_jhamman : cloud-native data loaders for machine learning using @zarr_dev and @xarray_dev 1/6
3
21
61
1
6
63
@shoyer
Stephan Hoyer
3 years
@rabernat In 30 years, will Google even still exist? Who knows. But will you still be able to compile your Fortran code from 1990? For sure!
3
7
62
@shoyer
Stephan Hoyer
6 years
Slides for my talk on xarray at the ECMWF Python workshop #Py4ESS : It was excited to hear from so many in the weather data community ( @ECMWF @bopensolutions @PyTrollOrg ) about how they're using Xarray & Dask.
0
14
62
@shoyer
Stephan Hoyer
6 years
The public version of Google's Python style guide has been updated for first time in ~5 years:
1
30
61
@shoyer
Stephan Hoyer
3 years
@docmilanfar I don't think you need any of these to be a researcher! Being a researcher means you work on problems that you don't know are solvable by anyone.
1
4
59
@shoyer
Stephan Hoyer
9 years
Just averaged 52GB of weather data out of core with xray + dask #pydata . My laptop has only 16GB of memory. http://t.co/3OE7ffseyt
Tweet media one
1
34
58
@shoyer
Stephan Hoyer
7 years
I'm really excited about plotnine. It looks like a full featured port of ggplot2 to Python! See also:
@danrothenberg
Daniel Rothenberg
7 years
plotnine - a functional Python port of ggplot2
0
4
8
2
33
55
@shoyer
Stephan Hoyer
3 years
Learning the laws of physics: 🥱😴 Learning how to solve the laws of physics: 🤔 Learning how to solve the laws of physics more efficiently than SOTA methods from scientific computing: 😍
1
2
55
@shoyer
Stephan Hoyer
3 months
For the record: Gemini's insistence on producing diverse images of people is a slightly glitchy feature, not a bug. If ahistorical images bother you much more than the myriad other generative AI issues, maybe this would be a good opportunity for self-reflection...
4
3
54
@shoyer
Stephan Hoyer
3 years
To elaborate: I spent most of my 20s in grad school and feeling really dumb, surrounded by people who were better at math & computers than I ever could be. Maybe this is the natural experience of doing a PhD but it sucked.
@shoyer
Stephan Hoyer
3 years
@alexeyguzey Way more. When I was 20, I just wanted to learn interesting things. It took real world experience to realize what I was capable of.
0
0
17
1
0
53
@shoyer
Stephan Hoyer
2 years
Industry is not a perfect system -- I've only experienced a few corporate cultures, and have undoubtedly benefited from a tremendous amount of privilege. And yes, there absolutely pathological cases where it fails entirely.
1
2
51
@shoyer
Stephan Hoyer
3 years
Slightly surreal experience with this article. I'm quoted liberally, but I never spoke with this journalist! AFAICT all the quotes are cribbed from a recorded talk that was posted on YouTube.
@bsaeta
Brennan Saeta
3 years
I'm regularly floored by all the amazing things people are doing with JAX, and I'm so glad we can do our little bit to help researchers push the boundaries of humanity's knowledge and capabilities. Nice work @shoyer et al.!
0
9
79
2
1
48
@shoyer
Stephan Hoyer
4 years
I’m excited to be part of this effort!
@bapadadada
Jason Hickey
4 years
In a Google Keyword blog post, I describe how we are using AI for advanced weather and climate prediction in partnership with @noaa and @ai2enviro .
0
9
45
1
1
49
@shoyer
Stephan Hoyer
3 years
@Thomas_ensc @OpenAI XLA (which JAX uses under the hood) is great, but it's solving a much higher level optimization problem to generate CUDA kernels from NumPy like code. Triton is lower level and thus offers more control to the programmer, for better or worse.
2
1
48
@shoyer
Stephan Hoyer
1 year
This is nice work, but improved RMSE on 10+ day deterministic forecasts doesn't mean your model is "substantially better." It means your model is blurrier. The right baseline is ECMWF's probabilistic ensemble.
@tungnd_13
Tung Nguyen
1 year
Introducing ClimaX, the first foundation model for weather and climate. A fast and accurate one-stop AI solution for a range of atmospheric science tasks. Paper: Blog: Thread🧵 #ML #Climate #Weather #FoundationModel
Tweet media one
35
179
850
3
1
49
@shoyer
Stephan Hoyer
3 years
It's nice to see a paper on efficient neural networks that measures speed on TPU/GPU, not FLOPs! 😍
@arankomatsuzaki
Aran Komatsuzaki
3 years
Revisiting ResNets: Improved Training and Scaling Strategies - The original ResNets w/ better training + scaling strategies + minor arch change achieve SotA perf w/ faster speed. - Training and scaling strategies matters more than architectural changes.
Tweet media one
2
32
196
0
6
47
@shoyer
Stephan Hoyer
4 years
Here's an interactive visualization of 728 GB of @ECMWF ERA5 temperature data curated by @pangeo_data stored in @zarr_dev using Neuroglancer. Be patient: each chunk is 64MB, so it's a little slow to load!
Tweet media one
@shoyer
Stephan Hoyer
4 years
@stardazed0 @zarr_dev We've been using it to read dataset created with @xarray_dev , and it works well! Neuroglancer even knows Xarray's convention for saving dimension names. Would love to get a @pangeo_data demo going on a big climate dataset
0
0
8
1
8
47
@shoyer
Stephan Hoyer
2 years
Dan writes fantastic and well-documented software. Check it out if you're interested in Gaussian processes!
@exoplaneteer
Dan F-M
2 years
I'm excited to announce the first beta of a new Python package that I've been working on: It's (yet another) Gaussian process library in Python, this time built on JAX. It's meant to be both performant & pedagogical with ~4x as many lines of docs as code.
8
51
452
0
3
46
@shoyer
Stephan Hoyer
5 years
@xarray_dev @ProjectJupyter @dask_dev Also: a special shout-out to @benbovy and @JSignell for xarray's new HTML repr for notebooks! It makes it really easy to explore complex datasets (you can click all over to expand/collapse sections). It even wraps @dask_dev 's HTML repr for showing off nested dask arrays!
Tweet media one
2
7
45
@shoyer
Stephan Hoyer
2 years
The best part is that you maybe be only one short function decorator away from solving your memory problems for good! If you use Python, take a look at jax.checkpoint, torch.utils.checkpoint or tf.recompute_grad. I'm sure it's easy in Julia, too -- please reply if you know how!
1
0
45
@shoyer
Stephan Hoyer
5 years
@jeremyphoward NumPy uses sorting for set operations, because it doesn't have any hashtable data structures. So np.isin(a, b) is OK for a large a and small b, but not the other way around. I agree that it would be good to have some performance warnings in the docs.
2
4
42
@shoyer
Stephan Hoyer
6 years
Now flying to Austin for #SciPy2018 . I'm looking forward to productive discussions about #NumPy , #pandas , #dask , #xarray and @pangeo_data !
2
1
45
@shoyer
Stephan Hoyer
5 years
I'm excited to share my work with @yohaibs on using differentiable programming to learn better discretizations for PDEs, now published in PNAS.
@GoogleAI
Google AI
5 years
New research shows how #machinelearning can improve high-performance computing for solving partial differential equations, with potential applications that range from modeling #climatechange to simulating fusion reactions. Learn all about it here ↓
5
186
501
4
5
45
@shoyer
Stephan Hoyer
6 months
I will be at AGU next week. If you want to talk about ML-based weather/climate modeling (NeuralGCM, WeatherBench2, etc) or @xarray_dev please reach out!
1
0
45
@shoyer
Stephan Hoyer
1 year
@lawdroid @jbrowder1 @donotpay He knows this already... this whole thing is disingenuous
2
1
43
@shoyer
Stephan Hoyer
4 years
I'll be presenting this work with @samgreydanus and @jaschasd tomorrow morning (Friday Dec 13) at 11:30am at the NeurIPS Deep Inverse Problems workshop ()
@shoyer
Stephan Hoyer
5 years
I'm happy to share a new paper, with @jaschasd and @samgreydanus : "Neural reparameterization improves structural optimization" We use neural nets to parameterize inputs of a finite elements method, and differentiable through the whole thing:
Tweet media one
2
38
170
2
9
42
@shoyer
Stephan Hoyer
2 years
To clarify, the likely focus is on fluid dynamics problems in climate & weather using hybrid physics/ML techniques:
0
2
43
@shoyer
Stephan Hoyer
2 years
This program is for PhD students planning to graduate in 2024 or later: Please apply on the linked page *and* let me know by email at shoyer @google .com
6
9
42
@shoyer
Stephan Hoyer
4 years
FYI, if you're excited about getting *paid* as a student to work on an open source project, check out @gsoc
2
6
40
@shoyer
Stephan Hoyer
4 years
@chrmanning There is undoubtedly loads of compute wasted on supercomputers, but simple AI shortcuts (like in this paper) are far from ready to replace physics based simulations. They don’t generalize in any meaningful way.
2
6
41
@shoyer
Stephan Hoyer
3 years
Interesting note: by far the most expensive part of training these sorts of models is *validation*. We have to run reference solvers on 32x higher resolution grids in space + time in order to rigorously measure the accuracy of our ML solver.
4
0
38
@shoyer
Stephan Hoyer
3 years
I just figured out what @pangeo_data is up to with pangeo-forge 😍🌎🌍🌏🚀
4
2
40
@shoyer
Stephan Hoyer
4 years
@raymondh I think this is ill-advised in most cases - it's better to pick a stable serialization format for persistent data, and there's no shame in explicitly writing to disk. That said, joblib does a reasonable implementation of this: .
1
0
40
@shoyer
Stephan Hoyer
7 months
But as I learned from Tim Palmer's delightful book "The primacy of doubt" the situation for weather forecasting is actually far more dire: as Lorenz showed in 1969, each doubling of resolution only gives half previous gains in accuracy.
1
1
40
@shoyer
Stephan Hoyer
1 year
Unpopular opinion: rainbow colormaps can be an excellent way to visualize scientific data! My favorite is "turbo:"
@trentag0n
Trent Thomas ⭔
1 year
Seen in the printing room
Tweet media one
24
125
1K
10
1
38
@shoyer
Stephan Hoyer
3 years
Congrats to the @pangeo_data team. This is the way!
@cisemag
Computing in Science & Engineering
3 years
In the current issue of #CiSEmag : “Cloud-Native Repositories for Big Scientific Data,” by @rabernat and colleagues—free to read!
Tweet media one
1
15
49
1
7
39
@shoyer
Stephan Hoyer
5 years
@jaschasd @samgreydanus I'm happy to share: (1) this paper was accepted for an oral presentation at the NeurIPS 2019 Deep Inverse workshop! (2) we're released the source code, so you can now run it yourself!
1
6
38
@shoyer
Stephan Hoyer
2 months
@bilalmahmood Bilal, I supported your campaign, but this is not a good look! You should never had led a campaign ad by calling yourself a "neuroscientist" based on a year or two of part time lab experience as part of your undergrad degree. It diminishes the credentials of real scientists.
3
2
38
@shoyer
Stephan Hoyer
3 years
Love this paper! A nice example of when gradients have questionable value in geoscience is optimizing parameters of a climate model.
@Luke_Metz
Luke Metz
3 years
New paper: when to use gradients DL researchers often compute derivatives though just about everything (physics simulators, optimization procedures, renderers). Sometimes these gradients are useful, other times they are not. We explore why. 1/7
13
158
797
1
1
38
@shoyer
Stephan Hoyer
2 years
Hitting refresh in TensorBoard is the ML equivalent of pulling the lever on a slot machine. It's every bit as addictive, but possibly even more expensive!
4
0
37
@shoyer
Stephan Hoyer
8 years
I'm sad to leave my colleagues at @ClimateCorp but stoked to be joining @GoogleResearch today!
Tweet media one
12
0
38
@shoyer
Stephan Hoyer
2 years
I designed the @xarray_dev logo in TikZ 🙈
@LongFormMath
Jay Cummings
2 years
interviewer: can you explain this gap in your CV me: yeah I was trying to make a complicated figure in TikZ and lost track of time and— interviewer: say no more
10
142
2K
2
1
38
@shoyer
Stephan Hoyer
2 years
One of my regrets with @xarray_dev is that I copied end-inclusive slicing rules from pandas. At this point there's basically no way to fix it without silently breaking lots of code.
@andrewwhite01
Andrew White 🐦‍⬛/acc
2 years
I should say that Pandas is a great resource for the Python community and enables my research. I just wish it didn't have such a strong "personality." Its own slicing rules (end of range is included), its own membership function (isin), etc.
3
1
25
2
1
37
@shoyer
Stephan Hoyer
1 year
If I was rewriting NumPy's API from scratch today, I would be tempted to make both array indexing and reductions over specific axes preserve rank, i.e., 1. `x[i]` would be equivalent to `x[i:i+1]` 2. `x.sum(axis=0)` would be equivalent to `x.sum(axis=0, keepdims=True)`
4
1
36
@shoyer
Stephan Hoyer
4 months
Maybe I'm the problem: I want my work to influence domain scientists (and get sensible reviews), so I rarely submit to ML conferences (and thus don't get asked to review). 🤷
2
1
36
@shoyer
Stephan Hoyer
3 years
Fun bonus fact: data assimilation was one of the first use-cases motivating the development of auto-diff software. E.g., see this 1993 paper So using modern libraries like JAX for data assimilation is really going back to the roots of the field :)
0
1
35