International Business Machines Corporation
Adding adversarial robustness to trained machine learning models
Last updated:
Abstract:
One or more hardened machine learning models are secured against adversarial attacks by adding adversarial protection to one or more previously trained machine learning models. To generate the hardened machine learning models, the previously trained machine learning models are retrained and extended using preprocessing layers or using additional network layers which test model performance on benign or adversarial samples. A rollback strategy is additionally implemented to retain intermediate model states during the retraining to provide recovery if a training collapse is detected.
Status:
Grant
Type:
Utility
Filling date:
14 Oct 2019
Issue date:
17 May 2022