Digitala Vetenskapliga Arkivet

System disruptions
We are currently experiencing disruptions on the search portals due to high traffic. We are working to resolve the issue, you may temporarily encounter an error message.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Structured models for scientific machine learning: From graphs to kernels
Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Systems and Control.ORCID iD: 0009-0007-5465-7170
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

This thesis investigates the reciprocal relationship between science and machine learning, showing how embedding scientific principles within machine learning models enhances accuracy and interpretability in complex scientific domains. Through five contributions, this work addresses challenges spanning molecular modeling, fluid dynamics, and graph-based learning, illustrating how scientific insights can guide model development and improve performance across diverse applications.

A central focus is developing models that directly incorporate physical symmetries and laws into their structure, creating novel, scientifically grounded approaches. For instance, the GeqShift model leverages E(3)-equivariant graph neural networks to predict NMR spectra with a significant accuracy boost by capturing three-dimensional molecular structures. Similarly, our SE(2)-equivariant graph neural network models rotational and translational symmetries to enhance data efficiency and performance in fluid dynamics simulations, demonstrating the strength of symmetry-aware models in complex physical domains.

This thesis advances graph-based machine learning by developing physics-informed models. For example, we introduce a subsampling model that incorporates principles from the classic Ising model of magnetism, introducing a novel approach to graph subsampling that enhances tasks like graph explanation and mesh sparsification. Building on these graph-based techniques, we also present an approach to nonnegative matrix factorization (NMF), leveraging graph structures to accelerate low-rank factorization.

In addition to symmetry-aware frameworks, we introduce the elliptical process, a flexible extension of the Gaussian process that adapts to non-Gaussian noise. This innovation allows the model to learn noise characteristics directly from data, producing robust predictions that address a broad spectrum of real-world challenges. 

These contributions underscore the dynamic exchange between scientific principles and machine learning, illustrating how physical knowledge enhances model performance and inspires new solutions. This thesis establishes a foundational framework for advancing scientific machine learning, paving the way for future breakthroughs in the field.

Place, publisher, year, edition, pages
Uppsala: Acta Universitatis Upsaliensis, 2025. , p. 94
Series
Digital Comprehensive Summaries of Uppsala Dissertations from the Faculty of Science and Technology, ISSN 1651-6214 ; 2492
National Category
Computer Sciences
Research subject
Computer Science; Machine learning
Identifiers
URN: urn:nbn:se:uu:diva-546022ISBN: 978-91-513-2350-3 (print)OAI: oai:DiVA.org:uu-546022DiVA, id: diva2:1924341
Public defence
2025-02-25, Lecture hall Sonja Lyttkens, Ångströmlaboratoriet, hus 10, Lägerhyddsvägen 1, Uppsala, 09:15 (English)
Opponent
Supervisors
Available from: 2025-02-05 Created: 2025-01-04 Last updated: 2025-02-13
List of papers
1. Ising on the Graph: Task-specific Graph Subsampling via the Ising Model
Open this publication in new window or tab >>Ising on the Graph: Task-specific Graph Subsampling via the Ising Model
2024 (English)In: Proceedings of theThird Learning on Graphs Conference (LoG 2024), Proceedings of Machine Learning Research , 2024Conference paper, Published paper (Refereed)
Abstract [en]

Reducing a graph while preserving its overall properties is an important problem with many applications. Typically, reduction approaches either remove edges (sparsification) or merge nodes (coarsening) in an unsupervised way with no specific downstream task in mind. In this paper, we present an approach for subsampling graph structures using an Ising model defined on either the nodes or edges and learning the external magnetic field of the Ising model using a graph neural network. Our approach is task-specific as it can learn how to reduce a graph for a specific downstream task in an end-to-end fashion without requiring a differentiable loss function for the task. We showcase the versatility of our approach on four distinct applications: image segmentation, explainability for graph classification, 3D shape sparsification, and sparse approximate matrix inverse determination.

Place, publisher, year, edition, pages
Proceedings of Machine Learning Research, 2024
Series
Proceedings of Machine Learning Research, ISSN 2640-3498 ; 269
National Category
Computer Sciences
Research subject
Machine learning
Identifiers
urn:nbn:se:uu:diva-545653 (URN)
Conference
The Third Learning on Graphs Conference (LoG 2024), Virtual Event, November 26–29, 2024
Available from: 2024-12-19 Created: 2024-12-19 Last updated: 2025-01-10Bibliographically approved
2. Carbohydrate NMR chemical shift prediction by GeqShift employing E(3) equivariant graph neural networks
Open this publication in new window or tab >>Carbohydrate NMR chemical shift prediction by GeqShift employing E(3) equivariant graph neural networks
2024 (English)In: RSC Advances, E-ISSN 2046-2069, Vol. 14, no 36, p. 26585-26595Article in journal (Refereed) Published
Abstract [en]

Carbohydrates, vital components of biological systems, are well-known for their structural diversity. Nuclear Magnetic Resonance (NMR) spectroscopy plays a crucial role in understanding their intricate molecular arrangements and is essential in assessing and verifying the molecular structure of organic molecules. An important part of this process is to predict the NMR chemical shift from the molecular structure. This work introduces a novel approach that leverages E(3) equivariant graph neural networks to predict carbohydrate NMR spectral data. Notably, our model achieves a substantial reduction in mean absolute error, up to threefold, compared to traditional models that rely solely on two-dimensional molecular structure. Even with limited data, the model excels, highlighting its robustness and generalization capabilities. The model is dubbed GeqShift (geometric equivariant shift) and uses equivariant graph self-attention layers to learn about NMR chemical shifts, in particular since stereochemical arrangements in carbohydrate molecules are characteristics of their structures.

Place, publisher, year, edition, pages
Royal Society of Chemistry, 2024
National Category
Organic Chemistry Biochemistry Molecular Biology
Identifiers
urn:nbn:se:uu:diva-537764 (URN)10.1039/d4ra03428g (DOI)001296088100001 ()39175672 (PubMedID)
Funder
Swedish Research Council, 2022-03014Knut and Alice Wallenberg Foundation
Available from: 2024-09-05 Created: 2024-09-05 Last updated: 2025-02-20Bibliographically approved
3. Flexible SE (2) graph neural networks with applications to PDE surrogates
Open this publication in new window or tab >>Flexible SE (2) graph neural networks with applications to PDE surrogates
(English)Manuscript (preprint) (Other academic)
Abstract [en]

This paper presents a novel approach for constructing graph neural networks equivariant to 2D rotations and translations and leveraging them as PDE surrogates on non-gridded domains. We show that aligning the representations with the principal axis allows us to sidestep many constraints while preserving SE(2) equivariance. By applying our model as a surrogate for fluid flow simulations and conducting thorough benchmarks against non-equivariant models, we demonstrate significant gains in terms of both data efficiency and accuracy.

National Category
Computer Sciences
Research subject
Scientific Computing; Computer Science; Computer Science
Identifiers
urn:nbn:se:uu:diva-546608 (URN)
Available from: 2025-01-10 Created: 2025-01-10 Last updated: 2025-01-15
4. Variational Elliptical Processes
Open this publication in new window or tab >>Variational Elliptical Processes
2023 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856Article in journal (Refereed) Published
Abstract [en]

We present elliptical processes—a family of non-parametric probabilistic models that subsumes Gaussian processes and Student's t processes. This generalization includes a range of new heavy-tailed behaviors while retaining computational tractability. Elliptical processes are based on a representation of elliptical distributions as a continuous mixture of Gaussian distributions. We parameterize this mixture distribution as a spline normalizing flow, which we train using variational inference. The proposed form of the variational posterior enables a sparse variational elliptical process applicable to large-scale problems. We highlight advantages compared to Gaussian processes through regression and classification experiments. Elliptical processes can supersede Gaussian processes in several settings, including cases where the likelihood is non-Gaussian or when accurate tail modeling is essential.

National Category
Other Computer and Information Science
Research subject
Machine learning; Artificial Intelligence
Identifiers
urn:nbn:se:uu:diva-516855 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Swedish Foundation for Strategic Research, SM19-0029Kjell and Marta Beijer FoundationKnut and Alice Wallenberg Foundation
Available from: 2023-11-30 Created: 2023-11-30 Last updated: 2025-01-04Bibliographically approved
5. Graph-based Neural Acceleration for Nonnegative Matrix Factorization
Open this publication in new window or tab >>Graph-based Neural Acceleration for Nonnegative Matrix Factorization
(English)Manuscript (preprint) (Other academic)
Abstract [en]

We describe a graph-based neural acceleration technique for nonnegative matrix factorization that builds upon a connection between matrices and bipartite graphs that is well-known in certain fields, e.g., sparse linear algebra, but has not yet been exploited to design graph neural networks for matrix computations. We first consider low-rank factorization more broadly and propose a graph representation of the problem suited for graph neural networks. Then, we focus on the task of nonnegative matrix factorization and propose a graph neural network that interleaves bipartite self-attention layers with updates based on the alternating direction method of multipliers. Our empirical evaluation of synthetic and two real-world datasets shows that we attain substantial acceleration, even though we only train in an unsupervised fashion on smaller synthetic instances. 

National Category
Computer Sciences
Identifiers
urn:nbn:se:uu:diva-546606 (URN)
Available from: 2025-01-10 Created: 2025-01-10 Last updated: 2025-01-15

Open Access in DiVA

UUThesis_M-Bånkestad-2025(2139 kB)117 downloads
File information
File name FULLTEXT01.pdfFile size 2139 kBChecksum SHA-512
7de876ae554a40dec0e3730f98119c54f8a5ef1351f63deddb279173d814b51acd08a1c45865d8ecf0b32b7d31cea5d2679f34634c006ebb654f7174aac5281c
Type fulltextMimetype application/pdf

Search in DiVA

By author/editor
Bånkestad, Maria
By organisation
Division of Systems and Control
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 117 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 1402 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf