
theseriousadult
@gallabytes
Followers
7K
Following
214K
Media
185
Statuses
4K
father, ML enjoyer. @midjourney v2-7.
Oakland
Joined April 2014
everyone is building really big short-term memories and calling it long-term memory. they're different things. almost nobody is building long-term memory. I barely ever see papers on it. it's the last remaining puzzle piece imo. should be the thing people are trying to build.
29
14
324
I have seen people say this but this literally isn't a 3d model? I guess you could make multiview and hope it's consistent enough to run through a real multiview -> mesh pipeline?.
Nanobanana from Google just dropped today. The best use case I heard so far is taking objects out of pictures and creating 3D models from them for games!. Any THING from a movie can be put into a game.
7
0
22
and, similarly to how claude code doesn't assume you're a good typist, a good image generator assumes you're a good artist but not necessarily especially skilled with a pencil.
Claude Code works because it assume you know how to code- it lives in the terminal. I wish more AI tools would do the same. Image gen tools that assume you're a good artist, writing tools that assume you're good at writing. Not just prompt -> output.
0
0
6
RT @gallabytes: I feel like there's an interesting @patio11 style piece that could be written on the regulatory constraints of investing OP….
0
1
0
I feel like there's an interesting @patio11 style piece that could be written on the regulatory constraints of investing OPM and why the reader's oh-so-clever simple tweaks to common practice might be less adaptive than they think.
1
1
3
my hot take is that investors are underexcited about AI and overexcited about "AI" and this is basically downstream of the same regulatory barriers that create most of the other toxic vc dynamics.
be Sam Altman.openai gonna spend TRILLIONS on datacenters soon. We have better models and no, you can’t see them lmao . “Are we in an AI bubble?”.Altman: My opinion is YES.
2
1
21
this is a really cool approach and I'm frankly surprised it works!.
Bifrost-1: Bridging Multimodal LLMs and Diffusion Models with Patch-level CLIP Latents. "We present BIFROST-1, a unified framework that bridges pretrained multimodal LLMs (MLLMs) and diffusion models using patch-level CLIP image embeddings as latent variables, which are natively
0
0
7
long term memory posting will continue until the models can sleep.
the reason why gpt-5 can't do <insert thing humans can do> is that we've pretty much exhausted what's possible using a single-turn chat format. Yes, chatgpt can't discover new science, but neither can humans in ~20 minutes with only access to a search API and a python notebook.
3
2
56