ynot
Virtually all modern insights derived from astronomical spectroscopy first arrive to the world on 2D digital camera chips such as visible CCDs or near-infrared focal plane arrays. Data reduction pipelines translate these raw 2D images into 1D extracted spectra, the more nimble and natural data format for practitioners. These data pipelines work great, and for the vast majority of practitioners, pipelines are a mere afterthought.
But some science use cases---such as spatially extended objects, faint ultracool brown dwarfs, emission line objects, and Extreme Precision Radial Velocity (EPRV)---defy the assumptions built into these automated reduction pipelines. Heuristics may exist for some of these scenarios, with mixed performance, but in general, a turnkey solution is lacking for these and other non-standard use cases.
The $y_0$ framework aspires to address these problems through the creation of a _spectrograph digital twin_. In this context, a digital twin indicates an ambition to forward model as many physically realistic attributes of the spectrograph system as possible. Above all, we seek a pixel-perfect scene model analogous to spectroperfectionism^1, but with an improved ability to learn the physical state of the spectrograph by pooling information among observations. The physically interpretable and extensible underlying design should provide the stepping stone needed to simulate and ultimately fit EPRV data that may someday find Earth analogs.
A typical modern astronomical echellogram has at least $1000 \times 1000$ pixels, and often mosaics or tiles of these can reach tens of millions of pixels in a single instrument. Scene modeling all of those pixels to arbitrary precision represents an enormous computational challenge. We therefore designed $y_0$ with these three uncompromising computational considerations in order to make the problem tractable:
Our project is funded in part through NASA Astrophysics Data Analysis Program (ADAP) grant 80NSSC21K0650 to improve archival data from the Keck NIRSPEC spectrograph. We are encouraged to see that version 0.1 of the code already delivers improved performance on an example brown dwarf spectrum in our first pass scene model. Below is a tensorboard screencap that shows how the PyTorch machine learning training proceeded, giving lower training loss and better 2D spectral reconstruction. This training process took a few minutes on my NVIDIA RTX2070 GPU. We think we can improve both the precision and speed of the ynot
code in upcoming versions.
The baseline PyTorch code and project scaffolding is based off of Ian Czekala's MPoL project. The project started in the early months of the COVID pandemic as a proof-of-concept demo, and was dormant as we built blasé, which enables interpretable machine learning for 1D extracted spectra^2.
As of July 2023, we are returning to $y_0$ to make it a broadly applicable tool for a wide variety of scientific use cases. If that sounds good to you, join our open source project! We welcome contibutions from a wide range of experience levels.
To get involved, please introduce yourself on our discussions page, or open an Issue describing your interest, or just contact me directly.
Copyright M. Gully-Santiago and contributors 2020, 2021, 2022, 2023
Version : 0.1