
Research
I am interested in questions at the intersection of Algebraic Geometry and Machine Learning. More concretely, I am interested in how to apply algebro-geometric tools to learn the underlying structure of real-world data through neural networks.
-
In Machine Learning, I am studying neural network expressivity through the lens of a neuromanifold, a finite-dimensional embedding of a fixed architecture into the ambient space. Understanding the geometry of the neuromanifold provides insights into how and why neural networks train and behave.
-
Another direction of my work concerns neural network quantization, approached via reduced weight precision and pruning.
-
-
In Algebraic Geometry, I am interested in studying simultaneous (and other) tensor decompositions, which correspond to various polynomial/rational neural network architectures.
-
Another line of my research focuses on using neural networks to identify the location and type of singularities of meromorphic functions from data.
-
Here is my CV.

Work in Progress
-
Simultaneous diagonalization of tensors, Elina Robeva, Matija Tomic, Carol Wu, Maksym Zubkov
-
Detection of Blow-up Singularities in Data, Vladimir Baranovsky, Maricela Best Mckay, Maksym Zubkov
Submitted
-
Algebraic geometry of rational neural networks, Alexandros Grosdos, Elina Robeva, Maksym Zubkov, submitted, available on arXiv
-
Expressivity of Shallow Neural Networks Over Finite Fields, Maksym Zubkov, Carol Wu, Shiwei Yang, Param Mody, Yifei Chen, submitted, available on openreview
-
Sign patterns of principal minors of real symmetric matrices, Tobias Boege, Jesse Selover, and Maksym Zubkov, submitted, available on arXiv
Publications
-
Likelihood Geometry of Determinantal Point Processes (with Hannah Friedman and Bernd Sturmfels) submitted
-
Chromatic Graph Homology for Brace Algebras (with Vladimir Baranovsky) published in New York J. Math. 23 (2017) 1307–1319​