This website is under active construction — check out the new Projects section!
Welcome to my portfolio
Tom Sander
Ph.D. Candidate · TU Dortmund University
Currently, I am working at the TU Dortmund University as a researcher in the Image Analysis group. I am working towards my Doctor of Engineering in the field of machine learning (AI) and planetary science. Feel free to reach out anytime via the contact form or through social media!
6 years · primary language
6 years · deep learning
3 years · signal processing
5 years · version control
2 years · scalable workflows
Tip: swipe horizontally to view the full year.
Note: We switched to Gitea in December 2025. Activity from the previous GitLab instance cannot be displayed in this heatmap. All of my paper projects and code projects are hosted on Gitea.
I specialize in engineering and scaling multimodal foundation models, with deep expertise in PyTorch and computer vision architectures. My recent work includes optimizing Transformer models processing over 1.34 billion tokens to push the boundaries of topological surface reconstruction. I have successfully developed architectures utilizing 29.7 million trainable parameters and designed models capable of predicting four distinct modalities simultaneously, securing a first-author publication in a Q1 journal.
Tom Sander1
·
Moritz Tenthoff1
·
Kay Wohlfarth1
·
Christian Wöhler1
1 TU Dortmund University, Image Analysis Group, Dortmund, Germany
Multimodal learning is an emerging research topic across multiple disciplines but has rarely been applied to planetary science. In this contribution, we identify that reflectance parameter estimation and image-based 3D reconstruction of lunar images can be formulated as a multimodal learning problem. We propose a single, unified transformer architecture trained to learn shared representations between multiple sources like grayscale images, digital elevation models, surface normals, and albedo maps. The architecture supports flexible translation from any input modality to any target modality. Predicting DEMs and albedo maps from grayscale images simultaneously solves the task of 3D reconstruction of planetary surfaces and disentangles photometric parameters and height information. Our results demonstrate that our foundation model learns physically plausible relations across these four modalities. Adding more input modalities in the future will enable tasks such as photometric normalization and co-registration.
Status
Published
ISPRS J. Photogramm. Remote Sens. — 2026
DOI
10.1016/j.isprsjprs.2026.04.008
Type
Journal Article
Field
Planetary Science & ML
"Unlocking the potential of remote sensing data to cultivate a deeper understanding of our ever-changing world."