@misc{rajesh_feature_2021, address = {Super Computing}, type = {Poster}, title = {Feature {Reduction} of {Darshan} {Counters} {Using} {Evolutionary} {Algorithms}}, url = {https://sc21.supercomputing.org/proceedings/tech_poster/tech_poster_pages/rpost147.html}, abstract = {Feature reduction is an integral part of data preparation in machine learning. It helps denoise the data and makes it easier to fit the model. Predicting the performance of an application using Darshan counters can be tricky due to the large amount of data available, with not all of them being pertinent to predicting the I/O performance. There exist methods for feature reduction, the most common being recursive feature elimination (RFE). The RFE method aims to correlate the features to a specific data point. We aim to get a subset of features that are able to distinguish between the different applications, then compare the effectiveness of the subset by creating a model to predict I/O performance. We then aim to compare that with a similar model created with all the features and with a subset of features determined using RFE implemented on Scikit Learn.}, urldate = {2022-02-09}, author = {Rajesh, Neeraj and Koziol, Quincey and Byna, Suren and Tang, Houjun and Bez, Jean Luca and Kougkas, Anthony and Sun, Xian-He}, month = nov, year = {2021}, }