Bernstein, J, Xiang Wang, Y, Azizzadenesheli, K, Anandkumar & A . 2018. Signsgd: Compressed optimization for non-convex problems. International Conference on Machine Learning 559–568.
Li, M, Andersen, D, Park, J W, Smola, A J, Ahmed, , Josifovski, A, Long, J, Shekita, I & Su, Yiing B . 2014. Scaling distributed machine learning with the parameter server. In: Proceedings of the 11th USENIX Conference on Operating Systems Design and Implementation, OSDI’14. (pp. 583-598)
Kakade, S M, Shwartz, S S & Tewari, A . 2012. Regularization techniques for learning with matrices. ournal of Machine Learning Research. 13:1865–1890
Chen, X W & Lin, X . 2014. Big data deep learning: Challenges and perspectives. Application of deep belief networks for opcode based 2:514–525.
Bonawitz, Keith, Eichner, Hubert, Grieskamp, Wolfgang, Huba, Dzmitry, Ingerman, Alex, Ivanov, Vladimir, Kiddon, Chloe, Konecny, Jakub, Mazzocchi, Stefano & H Brendan Mcmahan, . 2019. Towards federated learning at scale: System design.
Bijral, A S, Sarwate, Anand D & Srebro, N . 2016. On data dependence in distributed stochastic optimization.
Mcmahan, B & Ramage, R . 2017. Federated learning: Collaborative machine learning without centralized training data. Google Research Blog 3.
Ramage Brendan Mcmahan, Daniel . 2017. Federated Learning: Collaborative Machine Learning without Centralized Training Data. https://ai.googleblog.com/2017/04/federated-learning-collaborative.html
Zinkevich, M . 2003. Online convex programming and generalized infinitesimal gradient ascent. International Conference on Machine Learning 928–936.
Bekkerman, R, Bilenko, M & Langford, J . 2011. Scaling up machine learning: Parallel and distributed approaches. Cambridge University Press
Patil, A, Shah, A, Gaikwad, S, Mishra, A, Kohli, S S & Dhage, S . 2011. Fault Tolerance in Cluster Computing System. Int. Conf. P2P, Parallel, Grid 408–412.
Yuan, Y & Jia K . 2015. A distributed anomaly detection method of operation energy consumption using smart meter data. Proc. Int. Conf. Intell. Inf. Hiding Multimedia Signal Process. (IIH-MSP) 310–313.
Byzantine stochastic gradient descent. Advances in Neural Information Processing Systems 4613–4623.
Nielsen, M . 2018. 19. Dan Alistarh, Zeyuan Allen-Zhu & and Jerry Li , eds. Neural Networks and Deep Learning.
Bagdasaryan, Eugene, Veit, Andreas, Yiqinghua, Deborah, Estrin, Vitaly & Shmatikov, . 2018. How to backdoor federated learning.
Hatcher, W G & Yua, W . 2018. Survey of Deep Learning: Platforms, Applications, and Emerging Research Trends .
Zhu, D, Jin, H, Y, Wu, Y, D, Chen & W . 2017. DeepFlow: Deep learning-based malware detection by mining Android application for abnormal usage of sensitive data. Proc. IEEE Symp 438–443.
Shwartz, S S & David, S B . 2014. Understanding machine learning: From theory to algorithms. Cambridge University Press
Chaturapruek, S, John, C D, R´e, C & C . 2015. Asynchronous stochastic convex optimization: the noise is in the noise and sgd don’t care. Advances in Neural Information Processing Systems 1531–1539.
Bonawitz, Keith, Ivanov, Vladimir, Kreuter, Ben, Marcedone, Antonio, H Brendan Mcmahan, Sarvar, Patel, Daniel, Ramage, Aaron, Segal, Karn & Seth, . 2017. Practical secure aggregation for privacy-preserving machine learning. Proceedings of the (2017) ACM SIGSAC Conference on Computer and Communications Security 1175–1191.
Blanchard, P, Guerraoui, R & Stainer, J . 2017. Machine learning with adversaries: Byzantine tolerant gradient descent. Advances in Neural Information Processing Systems 119–129.
Engelmann, C, Vallée, G R, Naughton, T & Scott, S L . 2009. Proactive fault tolerance using preemptive migration. Proc. 17th Euromicro Int. Conf. Parallel, Distrib. Network-Based Process 252–257.
He, K, Zhang, X, Ren, S & Jian, S . 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition 770–778.
Alistarh, D, Allen-Zhu, Z & Li, J . 2018. Byzantine stochastic gradient descent. Advances in Neural Information Processing Systems 4613–4623.
Calheiros, R N, Ranjan, R, De Rose, C A F & Buyya, R . 2009. CloudSim: A Novel Framework for Model and Simulation of. Cloud Computing Infrastructures and Services 1–9.
Bernstein, Jeremy, Zhao, Jiawei, Azizzadenesheli, Kamyar & Kumar, Anima Anand . 2019. signSGD with Majority Vote is Communication Efficient and Fault Tolerant. In: 7th International Conference on Learning Representations, ICLR.
Malware detection. Proc. Int. Joint Conf. Neural Netw. (IJCNN) 3901–3908.
Kocher, D & Hilda, A K J . 2017. An approach for fault tolerance in cloud computing using a machine learning technique. Int. J. Pure Appl. Math 117(22):345–351.
Ujjwalkarn, . 2016. an Intuitive Explanation of Convolution Neural Networks.
Hazan, E . 2016. Introduction to online convex optimization. Foundations and Trends in Optimization. 2:157–325
Caldas, Sebastian, Wu, Peter, Li, Tian, Kone?ny, Jakub, Mcma- Han, Brendan, Smith, Virginia & Talwalkar, Ameet . 2018. Leaf: A benchmark for federated settings.