Publications

Some selected publications below. A curated full though probably outdated list can be found in my CV. The most recent list can be found in my Google Scholar profile.

Journal Articles


Identifying domains of applicability of machine learning models for materials science

Published in Nature Communications, 2020

Demonstrates how statistical rule learning enables the discovery of trustworthy input ranges of machine learning models for materials properties.

Recommended citation: C Sutton, M Boley, LM Ghringhelli, M Rupp, J Vreeken, M Scheffler. (2020). "Identifying domains of applicability of machine learning models for materials science." Nature Communications. 11(1),4428.
Download Paper

Conference Papers


Orthogonal Gradient Boosting for Simpler Additive Rule Ensembles

Published in International Conference on Artificial Intelligence and Statistics, 2024

Improving the accuracy / comprehensibility trade-off of rule ensembles and other additive models via proper adaption of boosting with weight correction.

Recommended citation: F Yang, P Le Bodic, M Kamp, M Boley. (2024). "Orthogonal Gradient Boosting for Simpler Additive Rule Ensembles." AISTATS.
Download Paper

Bayes beats cross validation: fast and accurate ridge regression via expectation maximization

Published in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS), 2023

Presents a novel method for tuning the regularization hyper-parameter, λ, of a ridge regression that is faster to compute than leave-one-out cross-validation (LOOCV) while yielding estimates of the regression parameters of equal, or particularly in the setting of sparse covariates, superior quality to those obtained by minimising the LOOCV risk

Recommended citation: SY Tew, M Boley, DF Schmidt. (2023). "Bayes beats cross validation: fast and accurate ridge regression via expectation maximization." NeurIPS. 36
Download Paper

Better Short than Greedy: Interpretable Models through Optimal Rule Boosting

Published in Proceedings of the 2021 SIAM International Conference on Data Mining (SDM), 2021

Improving the accuracy / comprehensibility trade-off of additive rule ensembles via exactly optimising the gradient boosting objective for conjunctive rules.

Recommended citation: M Boley, S Teshuva, P Le Bodic, G Webb. (2021). "Better Short than Greedy: Interpretable Models through Optimal Rule Boosting." SDM.
Download Paper