AIME: Watermarking AI Models by Leveraging Errors

Dhwani Mehtaa, Nurun Mondolb, Farimah Farahmandic and Mark Tehranipoord
ECE, University of Florida Gainesville, United States
adhwanimehta@ufl.edu
bnmondol@ufl.edu
cfarimah@ece.ufl.edu
dtehranipoor@ece.ufl.edu

ABSTRACT


The recent evolution of deep neural networks (DNNs) has made running complex data analytics tasks, which range from natural language processing, object detection to autonomous cars, artificial intelligence (AI) warfare, cloud, healthcare, industrial robots, and edge devices feasible. The benefits of AI are indisputable. However, there are several concerns regarding the security of the deployed AI models, such as reverse engineering and Intellectual Property (IP) piracy. Accumulating a sufficiently large amount of data - building, training, improvement, and model deployment require immense human and computational power, making the process expensive. Therefore, it is of utmost importance to protect the model against IP infringement. We propose AIME, a novel watermarking framework that captures model inaccuracy during the training phase and converts it into the owner-specific unique signature. The watermark is embedded within the class mispredictions of the DNN model. Watermark extraction is performed when the model is queried by an ownerspecific sequence of key inputs, and the signature is decoded from the sequence of model predictions. AIME works with negligible watermark embedding runtime overhead while preserving the accurate functionality of the DNN. We have performed a comprehensive evaluation of AIME, which models on MNIST, Fashion- MNIST, and CIFAR-10 dataset and corroborated its effectiveness, robustness, and performance.

Keywords: Artificial Intelligence,Watermarking, Intellectual Property piracy, Deep Learning.



Full Text (PDF)