Emerging neural networks based machine learning techniques such as deep
learning and its variants have shown tremendous potential in many application
domains. However, they raise serious privacy concerns due to the risk of
leakage of highly privacy-sensitive data when data collected from users is used
to train neural network models to support predictive tasks. To tackle such
serious privacy concerns, several privacy-preserving approaches have been
proposed in the literature that use either secure multi-party computation (SMC)
or homomorphic encryption (HE) as the underlying mechanisms. However, neither
of these cryptographic approaches provides an efficient solution towards
constructing a privacy-preserving machine learning model, as well as supporting
both the training and inference phases. To tackle the above issue, we propose a
CryptoNN framework that supports training a neural network model over encrypted
data by using the emerging functional encryption scheme instead of SMC or HE.
We also construct a functional encryption scheme for basic arithmetic
computation to support the requirement of the proposed CryptoNN framework. We
present performance evaluation and security analysis of the underlying crypto
scheme and show through our experiments that CryptoNN achieves accuracy that is
similar to those of the baseline neural network models on the MNIST dataset.