MachineLearning Thread Highlights Flower, a Warp-Centric Neural PDE Solver
Original: [R] Neural PDE solvers built (almost) purely from learned warps View original →
Thread context
The r/MachineLearning post [R] Neural PDE solvers built (almost) purely from learned warps reached 79 points and 20 comments. The author explicitly labeled it as their own work and linked both a ResearchGate paper and a public GitHub repository, making the discussion unusually concrete for an early-stage research share.
Core architectural idea
According to the post, Flower treats learned spatial warps as the main interaction primitive. At each position, the model predicts displacements and samples features from shifted coordinates. While the implementation borrows transformer-era engineering choices such as multi-head paths, projections, skip connections, and U-Net scaffolding, the key claim is that in-scale spatial mixing comes primarily from warping rather than heavy convolution or attention blocks.
The author argues this can keep cost closer to linear in grid points, which is relevant for 3D PDE workloads where memory and compute scale rapidly.
Claimed benchmark outcomes
- On 16 mostly The Well datasets, Flower reportedly leads one-step prediction versus similarly sized FNO, convolutional U-Net, and attention baselines.
- For 20-step autoregressive rollouts, reported gains remain in most tasks, with one difficult regime where all models degrade.
- A larger 150M-parameter variant is claimed to beat a much larger pretrained model (Poseidon, 628M) on a compressible Euler setting.
Limitations and community questions
The post also lists caveats: advantages can shrink in long rollouts, and there are stability issues under some conditions. Commenters asked the right next-step questions, including transfer to harder operational domains (for example weather-like scenarios) and behavior around discontinuities or shocks where smooth warps may be stressed.
Because this is primarily author-reported and described as pre-arXiv at posting time, independent replication remains important. Even so, the thread is technically valuable: it surfaces a credible systems-and-architecture alternative in scientific ML where efficiency and physical structure both matter.
Sources: r/MachineLearning post, paper link, code link
Related Articles
r/MachineLearning pushed this paper up because it did not promise a miracle. It argued that deep learning theory is finally accumulating enough converging evidence to resemble a genuine scientific program, and commenters liked the paper's concrete framing more than another grand AI manifesto.
This turns Google DeepMind’s science stack into a named national partnership instead of another generic AI pledge. The plan starts with a Seoul AI Campus, work with SNU and KAIST, and an AlphaFold base already used by more than 85,000 researchers in Korea.
Hacker News latched onto this paper because it was not selling a new benchmark or model, but a bigger claim: deep learning may finally be mature enough for a real scientific theory. That mix of excitement and skepticism kept the thread moving.
Comments (0)
No comments yet. Be the first to comment!