I attended an AI school in Nepal and heard geometric deep learning from Michael Bronstein himself. Cool stuff.I attended an AI school in Nepal and heard geometric deep learning from Michael Bronstein himself. Cool stuff.

Deep learning through the lens of Felix Klein's Erlangen's

Geometric deep learning by starting with the XOR affair

I’ve just spent two days at an AI school (ANAIS ) in Kathmandu listening to lectures on geometric deep learning from bright folks from Oxford, TU Wien, and top institutions. Geometric deep learning from Michael Bronstein  (who also coined the term; needs no introduction really).

The program is super organised, covering symmetry groups, graph neural networks and fundamental principles underlying modern AI before it eventually catches up to modern trends. Papers. Maths. Code.

I had checked the list of confirmed speakers before I applied but wasn’t expecting them to start with the XOR problem from 1969. Marvin Minsky and Seymour Papert’s proof that simple perceptron couldn’t represent XOR.

It’s clever, it’s a good point for understanding why geometric deep learning matters.

The human element

Bronstein’s opening lecture traced an interesting path. He went back to the 1950s, to Frank Rosenblatt’s perceptron experiments, showed newspaper clippings (1958) calling it “First serious rival to the human brain..”, Minsky and Papert’s book proving perceptrons couldn’t even learn XOR.

I’ve attended a bunch of lectures over the years, and mostly it’s - “here’s a neural network, here’s what I need you to do, here’s the code”. Some treat the field like it emerged in 2012 with AlexNet.

It’s interesting that the book that killed AI funding for a decade also had the idea for Group Invariance theorem. Says - if a neural network is invariant to a group, it’s output can be expressed as functions of the orbits of that group. In English, that’s - if a function (nn) is designed to ignore how things are moved around, it still focuses on the essential part that doesn’t change. Say you pick a cat upside down. It might be hissing at you, but it’s still a cat. Rotating/ flipping the cat doesn’t change what the object is. Still the same object, albeit pretty mad at you.

Essentially, the idea behind modern geometric deep learning.

If you think about it, geometric DL is super intuitive

I understand convolutions but Bronstein’s connections are interesting:

Felix Klein’s Erlangen programme (1872): Geometry is the study of properties that are invariant under transformations.

CNNs (1980s): Image recognition is the study of features that are invariant under translations.

Dates aren’t important here (maybe they are), notice that it’s basically the same idea, just over a century apart. If you think about it, a CNN just tries to learn the “geometry” of images. Learned geometry means accurate detection.

Klein’s insight was that you can classify geometries by their symmetry groups. Euclidean geometry preserves distances and angles. Affine geometry preserves parallelism but not angles. Projective geometry preserves only intersections.

CNNs do the same thing for images. A teddy is still a teddy whether you’re holding it or hanging it upside down. Translation invariance isn’t a new trick to make neural networks better. It’s actually a geometric “prior” about how neural networks learn from images. Cool stuff. Progression makes sense.

Back to the lecture, the plan was roughly:

  • grids -> translation symmetry -> CNNs
  • graphs -> permutation symmetry -> graph neural networks
  • manifolds -> isometry/gauge symmetry -> geometric CNNs

Do you notice the pattern? Each architecture encodes a unique geometric prior about its input. Think about it. If it’s a cat image (a grid of pixels), you want CNNs to learn it’s a cat even if it’s upside down. In a graph of social networks, order of nodes doesn’t matter - it’s the overall shape of the graph that does. And, in a manifold (a 3D surface lets say), it shouldn’t matter whether it’s rotated, stretched or distorted as long as it’s the same object.

Fundamental yet very relevant

The ideas behind geometric networks are nothing new. It’s somehow also the current state of the art. There are obviously alterations (many; like SwiGLU instead of ReLU in some blocks), but Deepmind’s AlphaFold2 is basically a geometric model that uses equivariant graph networks. Rotational equivariance means it can learn 3D structures during protein folding. AF2 won a Nobel prize in Chemistry (2024). Quite relevant.

Erlangen programme, modernized

Bronstein’s framework is basically a modernized version of Klein’s Erlangen programme. So you’d ask what symmetries does my data have instead of asking what worked on ImageNet.

  • images -> translation symmetry -> CNNs
  • graphs-> permutation symmetry -> GNNs
  • 3D point clouds -> rotation + permutation symmetry -> Equivariant GNNs
  • sequences -> time warping -> LSTMs
  • sets -> permutation invariance -> DeepSets/transformers

Simply put, just design models that respect and exploit geometric priors. Without considering the latter, you’d need O(ε^(-d)) samples for accuracy ε in d dimensions.

Notice how we don’t look at the architecture first.

Closing thoughts

ML is function approximation - you’re fitting a function to some data. More data, better results, but also an art in many ways. If you find the right invariances, you get data with structures, then you can build architectures that respect and utilize those priors, then you’d have models learning faster and generalizing better. Quite intuitive.

The geometric lens is an interesting way to see how CNNs are more than a clever trick with shared weights. Also interesting that it took an AI professor and slides with images of 19th century mathematicians for me to get why group invariance theorem matters.

The program was quite technical. Teodora Reu’s  lecture on group theory assumed comfort with abstract algebra and Ismail Ceylan’s graph theory sessions went deep into the Weisfeiler-Lehman hierarchy. So far so good.


Footnotes:

  1. The slides have people. I’ve come to realise that I tend to ignore the bits about people in favor of the content. It has only recently dawned on me that researchers/scientists are real people. It’s also something Prof. Rajani Malla  would say often.
  2. Interesting read on the perceptron controversy
  3. Felix Klein’s Erlangen programme (1872) in “A Comparative Review of Recent Researches in Geometry” unifies different geometries by characterizing their transformation groups. Applies to ML too. Just took us 150 years to realize that.
  4. The Weisfeiler-Lehman test is named after Boris Weisfeiler , who disappeared in 1985 while hiking in Chile during Pinochet’s dictatorship. He never got to see it, but his contribution to graph theory lives in modern GNNs.
  5. Kunihiko Fukushima’s Neocognitron (1980)  is interesting; resembles a modern CNNs: deep architecture, local connectivity, weight sharing, pooling, ReLU. It’s a near-miss, just needed backprop.
  6. It’s not supposed to be a detailed account (unless they’re paying me, which is unlikely), but might add interesting bits from Alisia’s  sessions and more.
  7. Writing after some work, a 9-6 day of lectures + lab, debugging, and now this post (it’s 1 am). The kind of tired where you’re not sure if the ideas are good or if you’re just sleep-deprived. We’ll see. I’ve missed lots of details and sessions by amazing people.
  8. This post is also published on my personal site. Conflicts of interest: None. I’m not affiliated to NAAMII.

\

Market Opportunity
DeepBook Logo
DeepBook Price(DEEP)
$0.057965
$0.057965$0.057965
+4.19%
USD
DeepBook (DEEP) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment?

Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment?

The post Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment? appeared on BitcoinEthereumNews.com. Crypto News 17 September 2025 | 17:39 Is dogecoin really fading? As traders hunt the best crypto to buy now and weigh 2025 picks, Dogecoin (DOGE) still owns the meme coin spotlight, yet upside looks capped, today’s Dogecoin price prediction says as much. Attention is shifting to projects that blend culture with real on-chain tools. Buyers searching “best crypto to buy now” want shipped products, audits, and transparent tokenomics. That frames the true matchup: dogecoin vs. Pepeto. Enter Pepeto (PEPETO), an Ethereum-based memecoin with working rails: PepetoSwap, a zero-fee DEX, plus Pepeto Bridge for smooth cross-chain moves. By fusing story with tools people can use now, and speaking directly to crypto presale 2025 demand, Pepeto puts utility, clarity, and distribution in front. In a market where legacy meme coin leaders risk drifting on sentiment, Pepeto’s execution gives it a real seat in the “best crypto to buy now” debate. First, a quick look at why dogecoin may be losing altitude. Dogecoin Price Prediction: Is Doge Really Fading? Remember when dogecoin made crypto feel simple? In 2013, DOGE turned a meme into money and a loose forum into a movement. A decade on, the nonstop momentum has cooled; the backdrop is different, and the market is far more selective. With DOGE circling ~$0.268, the tape reads bearish-to-neutral for the next few weeks: hold the $0.26 shelf on daily closes and expect choppy range-trading toward $0.29–$0.30 where rallies keep stalling; lose $0.26 decisively and momentum often bleeds into $0.245 with risk of a deeper probe toward $0.22–$0.21; reclaim $0.30 on a clean daily close and the downside bias is likely neutralized, opening room for a squeeze into the low-$0.30s. Source: CoinMarketcap / TradingView Beyond the dogecoin price prediction, DOGE still centers on payments and lacks native smart contracts; ZK-proof verification is proposed,…
Share
BitcoinEthereumNews2025/09/18 00:14
Fed Decides On Interest Rates Today—Here’s What To Watch For

Fed Decides On Interest Rates Today—Here’s What To Watch For

The post Fed Decides On Interest Rates Today—Here’s What To Watch For appeared on BitcoinEthereumNews.com. Topline The Federal Reserve on Wednesday will conclude a two-day policymaking meeting and release a decision on whether to lower interest rates—following months of pressure and criticism from President Donald Trump—and potentially signal whether additional cuts are on the way. President Donald Trump has urged the central bank to “CUT INTEREST RATES, NOW, AND BIGGER” than they might plan to. Getty Images Key Facts The central bank is poised to cut interest rates by at least a quarter-point, down from the 4.25% to 4.5% range where they have been held since December to between 4% and 4.25%, as Wall Street has placed 100% odds of a rate cut, according to CME’s FedWatch, with higher odds (94%) on a quarter-point cut than a half-point (6%) reduction. Fed governors Christopher Waller and Michelle Bowman, both Trump appointees, voted in July for a quarter-point reduction to rates, and they may dissent again in favor of a large cut alongside Stephen Miran, Trump’s Council of Economic Advisers’ chair, who was sworn in at the meeting’s start on Tuesday. It’s unclear whether other policymakers, including Kansas City Fed President Jeffrey Schmid and St. Louis Fed President Alberto Musalem, will favor larger cuts or opt for no reduction. Fed Chair Jerome Powell said in his Jackson Hole, Wyoming, address last month the central bank would likely consider a looser monetary policy, noting the “shifting balance of risks” on the U.S. economy “may warrant adjusting our policy stance.” David Mericle, an economist for Goldman Sachs, wrote in a note the “key question” for the Fed’s meeting is whether policymakers signal “this is likely the first in a series of consecutive cuts” as the central bank is anticipated to “acknowledge the softening in the labor market,” though they may not “nod to an October cut.” Mericle said he…
Share
BitcoinEthereumNews2025/09/18 00:23
Stronger capital, bigger loans: Africa’s banking outlook for 2026

Stronger capital, bigger loans: Africa’s banking outlook for 2026

African banks spent 2025 consolidating, shoring up capital, tightening risk controls, and investing in digital infrastructure, following years of macroeconomic
Share
Techcabal2026/01/14 23:06