SpecPT — Transformer Foundation Model for Spectroscopy
Designed a transformer autoencoder for self-supervised spectral representation learning, trained on 13 million galaxy spectra. The model denoises high-dimensional sequential inputs and predicts continuous targets with R² = 0.99 — compressing a months-long manual analysis pipeline to seconds per sample. Enables zero-shot transfer to new instruments.
Most remarkably: a University of Maryland team adopted SpecPT as the backbone for a mass spectrometry classifier targeting biosignature detection on planetary rovers — with minimal fine-tuning. The representations learned from galaxies transferred to biology.