May 25, 2025
Welcome to the second entry in my Holiday in Sicily series. Today, I reflect on the interplay between algorithms and theory—how practice often leads, theory follows, and how the two shape each other in a dynamic loop that drives much of applied math progress.
In many fields of applied mathematics and engineering, the relationship between theory and practice isn’t a one-way street. It’s a conversation. At its best, it’s a lively argument.
Often, the most exciting developments come from this back-and-forth—a dynamic feedback loop between algorithms and theory. Practice speaks first: a new method appears, driven by need or intuition. Then theory replies, asking tough questions. Does it work? When? Why?
Sometimes, theory pushes back, exposing flaws or failure modes. Practice listens, adjusts, and counters with a better design. And so the exchange continues—design, analysis, failure, improvement. A real dialogue. Each side learning, challenging, shaping the other.
This pattern can be broken down into five stages:
This cycle is more than just a historical curiosity. It reflects a healthy research ecosystem where algorithms and theory co-evolve, each pushing the other forward.
In the 1960s and 1970s, engineers began using the finite element method (FEM) for solving PDEs in structural mechanics. Early algorithms were developed based on physical intuition and computational feasibility, often with minimal theoretical backing.
As usage spread, mathematicians stepped in to analyze convergence, consistency, and error bounds. This theoretical work revealed the importance of mesh quality, element choice, and function spaces—leading to remedies such as adaptive mesh refinement and stabilized formulations. These in turn became subjects of rigorous mathematical analysis, culminating in the powerful FEM theory we now rely on.
The past decade has seen a similar cycle in machine learning. Deep neural networks first exploded in popularity due to empirical success in computer vision and natural language processing. The original architectures—like AlexNet—worked well, but few could explain why.
Soon after, theoretical work began to address generalization, expressivity, and training dynamics. Researchers identified failure modes: vanishing gradients, overfitting, sensitivity to initialization. Each insight led to remedies like residual connections, batch normalization, and dropout.
These new components became standard architecture elements, and new theoretical frameworks arose to understand them—such as neural tangent kernels, overparameterization theory, and mean-field limits.
Understanding this pattern helps us better appreciate the non-linear path of scientific progress. It reminds us that:
If you’re a theorist, don’t wait for the perfect question. Look at what practitioners are doing—it’s full of open problems.
If you’re a practitioner, don’t be afraid to work with unproven methods. The theory might catch up—and when it does, it’ll make your work even better.