General classification of light curves using extreme boosting. (arXiv:1906.06628v1 [astro-ph.SR])
<a href="http://arxiv.org/find/astro-ph/1/au:+Kgoadi_R/0/1/0/all/0/1">Refilwe Kgoadi</a>, <a href="http://arxiv.org/find/astro-ph/1/au:+Engelbrecht_C/0/1/0/all/0/1">Chris Engelbrecht</a>, <a href="http://arxiv.org/find/astro-ph/1/au:+Whittingham_I/0/1/0/all/0/1">Ian Whittingham</a>, <a href="http://arxiv.org/find/astro-ph/1/au:+Tkachenko_A/0/1/0/all/0/1">Andrew Tkachenko</a>

A significant degree of misclassification of variable stars through the
application of machine learning methods to survey data motivates a search for
more reliable and accurate machine learning procedures, especially in light of
the very large data cubes that will be generated by future surveys and the need
for immediate production of accurate, formalised catalogues of variable
behaviour to enable science to proceed. In this study, the efficiency of an
ensemble machine learning procedure utilising extreme boosting was determined
by application to a large sample of data from the OGLE III and IV surveys and
from the textit{Kepler} mission. Through recursive training of classifiers,
the study developed a variable star classification workflow which produced an
average efficiency determined with the average precision of the model (0.81 for
textit{Kepler} and 0.91 for OGLE) and the $f-score$ of predictions on the test
sets. This suggests that extreme boosting can be presented as one of the
favourable shallow learning methods in developing a variable star classifier
for future large survey projects.

A significant degree of misclassification of variable stars through the
application of machine learning methods to survey data motivates a search for
more reliable and accurate machine learning procedures, especially in light of
the very large data cubes that will be generated by future surveys and the need
for immediate production of accurate, formalised catalogues of variable
behaviour to enable science to proceed. In this study, the efficiency of an
ensemble machine learning procedure utilising extreme boosting was determined
by application to a large sample of data from the OGLE III and IV surveys and
from the textit{Kepler} mission. Through recursive training of classifiers,
the study developed a variable star classification workflow which produced an
average efficiency determined with the average precision of the model (0.81 for
textit{Kepler} and 0.91 for OGLE) and the $f-score$ of predictions on the test
sets. This suggests that extreme boosting can be presented as one of the
favourable shallow learning methods in developing a variable star classifier
for future large survey projects.

http://arxiv.org/icons/sfx.gif