Abstract
Over the past decade, hundreds of nights have been spent on the world’s largest telescopes to search for and directly detect new exoplanets using high-contrast imaging (HCI). Thereby, two ...scientific goals are of central interest: first, to study the characteristics of the underlying planet population and distinguish between different planet formation and evolution theories. Second, to find and characterize planets in our immediate solar neighborhood. Both goals heavily rely on the metric used to quantify planet detections and nondetections. Current standards often rely on several explicit or implicit assumptions about noise. For example, it is often assumed that the residual noise after data postprocessing is Gaussian. While being an inseparable part of the metric, these assumptions are rarely verified. This is problematic as any violation of these assumptions can lead to systematic biases. This makes it hard, if not impossible, to compare results across data sets or instruments with different noise characteristics. We revisit the fundamental question of how to quantify detection limits in HCI. We focus our analysis on the error budget resulting from violated assumptions. To this end, we propose a new metric based on bootstrapping that generalizes current standards to non-Gaussian noise. We apply our method to archival HCI data from the NACO instrument at the Very Large Telescope and derive detection limits for different types of noise. Our analysis shows that current standards tend to give detection limits that are about one magnitude too optimistic in the speckle-dominated regime. That is, HCI surveys may have excluded planets that can still exist.
Context
.
High-contrast imaging of exoplanets hinges on powerful post-processing methods to denoise the data and separate the signal of a companion from its host star, which is typically orders of ...magnitude brighter.
Aims
.
Existing post-processing algorithms do not use all prior domain knowledge that is available about the problem. We propose a new method that builds on our understanding of the systematic noise and the causal structure of the data-generating process.
Methods
.
Our algorithm is based on a modified version of half-sibling regression (HSR), a flexible denoising framework that combines ideas from the fields of machine learning and causality. We adapted the method to address the specific requirements of high-contrast exoplanet imaging data obtained in pupil tracking mode. The key idea is to estimate the systematic noise in a pixel by regressing the time series of this pixel onto a set of causally independent, signal-free predictor pixels. We use regularized linear models in this work; however, other (nonlinear) models are also possible. In a second step, we demonstrate how the HSR framework allows us to incorporate observing conditions such as wind speed or air temperature as additional predictors.
Results
.
When we applied our method to four data sets from the VLT/NACO instrument, our algorithm provided a better false-positive fraction than a popular baseline method in the field. Additionally, we found that the HSR-based method provides direct and accurate estimates for the contrast of the exoplanets without the need to insert artificial companions for calibration in the data sets. Finally, we present a first piece of evidence that using the observing conditions as additional predictors can improve the results.
Conclusions
.
Our HSR-based method provides an alternative, flexible, and promising approach to the challenge of modeling and subtracting the stellar PSF and systematic noise in exoplanet imaging data.
ABSTRACT
NASA is engaged in planning for a Habitable Worlds Observatory (HabWorlds ), a coronagraphic space mission to detect rocky planets in habitable zones and establish their habitability. ...Surface liquid water is central to the definition of planetary habitability. Photometric and polarimetric phase curves of starlight reflected by an exoplanet can reveal ocean glint, rainbows, and other phenomena caused by scattering by clouds or atmospheric gas. Direct imaging missions are optimized for planets near quadrature, but HabWorlds ’ coronagraph may obscure the phase angles where such optical features are strongest. The range of accessible phase angles for a given exoplanet will depend on the planet’s orbital inclination and/or the coronagraph’s inner working angle (IWA). We use a recently created catalog relevant to HabWorlds of 164 stars to estimate the number of exo-Earths that could be searched for ocean glint, rainbows, and polarization effects due to Rayleigh scattering. We find that the polarimetric Rayleigh scattering peak is accessible in most of the exo-Earth planetary systems. The rainbow due to water clouds at phase angles of ∼20○ − 60○ would be accessible with HabWorlds for a planet with an Earth equivalent instellation in ∼46 systems, while the ocean glint signature at phase angles of ∼130○ − 170○ would be accessible in ∼16 systems, assuming an IWA = 62 mas (3λ/D). Improving the IWA = 41 mas (2λ/D) increases accessibility to rainbows and glints by factors of approximately 2 and 3, respectively. By observing these scattering features, HabWorlds could detect a surface ocean and water cycle, key indicators of habitability.
The new generation of observatories and instruments (VLT/ERIS, JWST, ELT) motivate the development of robust methods to detect and characterise faint and close-in exoplanets. Molecular mapping and ...cross-correlation for spectroscopy use molecular templates to isolate a planet's spectrum from its host star. However, reliance on signal-to-noise ratio metrics can lead to missed discoveries, due to strong assumptions of Gaussian-independent and identically distributed noise. We introduce machine learning for cross-correlation spectroscopy (MLCCS). The aim of this method is to leverage weak assumptions on exoplanet characterisation, such as the presence of specific molecules in atmospheres, to improve detection sensitivity for exoplanets. The MLCCS methods, including a perceptron and unidimensional convolutional neural networks, operate in the cross-correlated spectral dimension, in which patterns from molecules can be identified. The methods flexibly detect a diversity of planets by taking an agnostic approach towards unknown atmospheric characteristics. The MLCCS approach is implemented to be adaptable for a variety of instruments and modes. We tested this approach on mock datasets of synthetic planets inserted into real noise from SINFONI at the K-band. The results from MLCCS show outstanding improvements. The outcome on a grid of faint synthetic gas giants shows that for a false discovery rate up to $5<!PCT!>$, a perceptron can detect about $26$ times the amount of planets compared to an S/N metric. This factor increases up to $77$ times with convolutional neural networks, with a statistical sensitivity (completeness) shift from $0.7<!PCT!>$ to $55.5<!PCT!>$. In addition, MLCCS methods show a drastic improvement in detection confidence and conspicuity on imaging spectroscopy. Once trained, MLCCS methods offer sensitive and rapid detection of exoplanets and their molecular species in the spectral dimension. They handle systematic noise and challenging seeing conditions, can adapt to many spectroscopic instruments and modes, and are versatile regarding planet characteristics, enabling the identification of various planets in archival and future data.
We present the statistical analysis of a subsample of 45 young stars surrounded by protoplanetary disks (PPDs). This is the largest imaging survey uniquely focused on PPDs to date. Our goal is to ...search for young forming companions embedded in the disk material and to constrain their occurrence rate in relation to the formation mechanism. We used principal component analysis based point spread function subtraction techniques to reveal young companions forming in the disks. We calculated detection limits for our datasets and adopted a black-body model to derive temperature upper limits of potential forming planets. We then used Monte Carlo simulations to constrain the population of forming gas giant companions and compare our results to different types of formation scenarios. Our data revealed a new binary system (HD38120) and a recently identified triple system with a brown dwarf companion orbiting a binary system (HD101412), in addition to 12 known companions. Furthermore, we detected signals from 17 disks, two of which (HD72106 and TCrA) were imaged for the first time. We reached median detection limits of L =15.4 mag at 2.0 arcsec, which were used to investigate the temperature of potentially embedded forming companions. We can constrain the occurrence of forming planets with semi-major axis a in 20 - 500 au and Teff in 600 - 3000 K, in line with the statistical results obtained for more evolved systems from other direct imaging surveys. The NaCo-ISPY data confirm that massive bright planets accreting at high rates are rare. More powerful instruments with better sensitivity in the near- to mid-infrared are likely required to unveil the wealth of forming planets sculpting the observed disk substructures.
The main challenge of exoplanet high-contrast imaging (HCI) is to separate the signal of exoplanets from their host stars, which are many orders of magnitude brighter. HCI for ground-based ...observations is further exacerbated by speckle noise originating from perturbations in the Earth's atmosphere and imperfections in the telescope optics. Various data post-processing techniques are used to remove this speckle noise and reveal the faint planet signal. Often, however, a significant part of the planet signal is accidentally subtracted together with the noise. In the present work, we use explainable machine learning to investigate the reason for the loss of the planet signal for one of the most used post-processing methods: Principal Component Analysis (PCA). We find that PCA learns the shape of the telescope point spread function for high numbers of PCA components. This representation of the noise captures not only the speckle noise, but also the characteristic shape of the planet signal. Building upon these insights, we develop a new post-processing method (4S) that constrains the noise model to minimize this signal loss. We apply our model to 11 archival HCI datasets from the VLT-NACO instrument in the L'-band and find that our model consistently outperforms PCA. The improvement is largest at close separations to the star (\(\leq 4 \lambda /D\)) providing up to 1.5 magnitudes deeper contrast. This enhancement enables us to detect the exoplanet AF Lep b in data from 2011, 11 years before its subsequent discovery. We present updated orbital parameters for this object.
High-contrast imaging of exoplanets hinges on powerful post-processing methods to denoise the data and separate the signal of a companion from its host star, which is typically orders of magnitude ...brighter. Existing post-processing algorithms do not use all prior domain knowledge that is available about the problem. We propose a new method that builds on our understanding of the systematic noise and the causal structure of the data-generating process. Our algorithm is based on a modified version of half-sibling regression (HSR), a flexible denoising framework that combines ideas from the fields of machine learning and causality. We adapt the method to address the specific requirements of high-contrast exoplanet imaging data obtained in pupil tracking mode. The key idea is to estimate the systematic noise in a pixel by regressing the time series of this pixel onto a set of causally independent, signal-free predictor pixels. We use regularized linear models in this work; however, other (non-linear) models are also possible. In a second step, we demonstrate how the HSR framework allows us to incorporate observing conditions such as wind speed or air temperature as additional predictors. When we apply our method to four data sets from the VLT/NACO instrument, our algorithm provides a better false-positive fraction than PCA-based PSF subtraction, a popular baseline method in the field. Additionally, we find that the HSR-based method provides direct and accurate estimates for the contrast of the exoplanets without the need to insert artificial companions for calibration in the data sets. Finally, we present first evidence that using the observing conditions as additional predictors can improve the results. Our HSR-based method provides an alternative, flexible and promising approach to the challenge of modeling and subtracting the stellar PSF and systematic noise in exoplanet imaging data.
Over the past decade, hundreds of nights have been spent on the worlds largest telescopes to search for and directly detect new exoplanets using high-contrast imaging (HCI). Thereby, two scientific ...goals are of central interest: First, to study the characteristics of the underlying planet population and distinguish between different planet formation and evolution theories. Second, to find and characterize planets in our immediate Solar neighborhood. Both goals heavily rely on the metric used to quantify planet detections and non-detections. Current standards often rely on several explicit or implicit assumptions about the noise. For example, it is often assumed that the residual noise after data post-processing is Gaussian. While being an inseparable part of the metric, these assumptions are rarely verified. This is problematic as any violation of these assumptions can lead to systematic biases. This makes it hard, if not impossible, to compare results across datasets or instruments with different noise characteristics. We revisit the fundamental question of how to quantify detection limits in HCI. We focus our analysis on the error budget resulting from violated assumptions. To this end, we propose a new metric based on bootstrapping that generalizes current standards to non-Gaussian noise. We apply our method to archival HCI data from the NACO-VLT instrument and derive detection limits for different types of noise. Our analysis shows that current standards tend to give detection limit that are about one magnitude too optimistic in the speckle-dominated regime. That is, HCI surveys may have excluded planets that can still exist.
The new generation of observatories and instruments (VLT/ERIS, JWST, ELT) motivate the development of robust methods to detect and characterise faint and close-in exoplanets. Molecular mapping and ...cross-correlation for spectroscopy use molecular templates to isolate a planet's spectrum from its host star. However, reliance on signal-to-noise ratio (S/N) metrics can lead to missed discoveries, due to strong assumptions of Gaussian independent and identically distributed noise. We introduce machine learning for cross-correlation spectroscopy (MLCCS); the method aims to leverage weak assumptions on exoplanet characterisation, such as the presence of specific molecules in atmospheres, to improve detection sensitivity for exoplanets. MLCCS methods, including a perceptron and unidimensional convolutional neural networks, operate in the cross-correlated spectral dimension, in which patterns from molecules can be identified. We test on mock datasets of synthetic planets inserted into real noise from SINFONI at K-band. The results from MLCCS show outstanding improvements. The outcome on a grid of faint synthetic gas giants shows that for a false discovery rate up to 5%, a perceptron can detect about 26 times the amount of planets compared to an S/N metric. This factor increases up to 77 times with convolutional neural networks, with a statistical sensitivity shift from 0.7% to 55.5%. In addition, MLCCS methods show a drastic improvement in detection confidence and conspicuity on imaging spectroscopy. Once trained, MLCCS methods offer sensitive and rapid detection of exoplanets and their molecular species in the spectral dimension. They handle systematic noise and challenging seeing conditions, can adapt to many spectroscopic instruments and modes, and are versatile regarding atmospheric characteristics, which can enable identification of various planets in archival and future data.