Abstract
Machine learning algorithms boast remarkable predictive capabilities and deep learning, a branch of machine learning, has already provided the much required breakthroughs for recognition and authentication. This has enabled the deployment of a face recognition based biometric identification system by the Department of Homeland Security at U.S. airports. Deep learning algorithms require huge amounts of data for training. However, this shall not be an issue in smart cities. Instead, the ability to use the same deep learning technology for ulterior motives raise some issues. Machine learning algorithms have already been developed which can generate fake images and videos and rendering humans incapable of differentiating between real and generated content. Such content can be used for spreading disinformation regarding individuals and may lead to legal issues. Any system, relying on face recognition may be mislead using such technology. Similarly, researchers were able to create master finger prints, i.e., a set of finger prints that may be used instead of an original finger print to defeat authentication biometric systems. Alongside data pollution, these image and video processing issues may present significant challenges to governance of smart cities that shall rely on automated processing of such data. This paper presents an introduction to interpretability, attempts made to improve understanding of ML algorithms and their results. Area where we believe the definition of interpretability is lacking are also highlighted and a few example scenarios are presented. Finally, some possible directions for addressing the raised issues are introduced.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
California Department of Corrections and Rehabilitation. https://www.cdcr.ca.gov/Rehabilitation/docs/FS_COMPAS_Final_4-15-09.pdf. Last accessed 25 Jan 2019
Blomberg, T., Bales, W., Mann, K., Meldrum, R., Nedelec, J.: Validation of the COMPAS risk assessment classification instrument. Florida State University, Tallahassee, FL, College of Criminology and Criminal Justice (2010)
Chouldechova, A.: Fair prediction with disparate impact: a study of bias in recidivism prediction instruments. Big Data 5(2), 153–163 (2017)
Krizhevsky, A., Sutskever, I., Hinton, G.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 1097–1105 (2012)
Lipton, Z.: The doctor just won’t accept that! arXiv:1711.08037v2 (2017)
Lipton, Z.: The Mythos of Model Interpretability. arXiv:1606.03490v3 (2017)
Bibal, A., Frenay, B.: Interpretability of machine learning models and representations: an introduction. In: European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning. Bruges, Belgium, 27–29 April 2016
Feng, C., Michie, D.: Machine learning of rules and trees. Machine Learning, Neural and Statistical Classification. Ellis Horwood, Hemel Hempstead (1994)
Weihs, C., Sondhauss, U.: Combining mental t and data t for classification rule selection. In: Exploratory Data Analysis in Empirical Research, pp. 188–203. Springer, Berlin (2003)
Zeiler, M., Fergus, R.: Visualizing and Understanding convolutional networks. In: European Conference on Computer Vision (ECCV), pp. 818–833 (2014)
Yosinski, J., Clune, J., Nguyen, A. M., Fuchs, T. J., Lipson, H.: Understanding Neural Networks through Deep Visualization. arxiv:1506.06579
Li, J., Chen, X., Hovy, E., Jurafsky, D.: Visualizing and Understanding Neural Models in NLP. arxiv:1506.01066
Zhang, Q., Wu, Y. N., Zhu, S-C.: Interpretable convolutional neural networks. In: IEEE Computer Vision and Pattern Recognition, pp. 8827–8836 (2018)
Deepfakes online application. https://deepfakesapp.online/. Last accessed 25 Jan 2019
Fitbit fitness tracker cracks Connecticut murder case. https://www.nydailynews.com/news/crime/police-solve-connecticut-murder-clues-fitbit-activity-article-1.3094802. Last accessed 25 Jan 2019
Detectives find new evidence in a high profile murder case. https://www.youtube.com/watch?v=JfBAg5t5Zoc. Last accessed 25 Jan 2019
Thies, J., Zollhofer, M., Stamminger, M., Theobalt, C., Niessner, M.: Face2Face: real-time face capture and reenactment of RGB videos. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2387–2395 (2016)
Karras, T., Laine, S., Aila, T.: A Style-Based Generator Architecture for Generative Adversarial Networks. arxiv:1812.04948
Virtual Holocaust Survivor. https://www.theguardian.com/technology/2016/jun/18/holocaust-survivor-hologram-pinchas-gutter-new-dimensions-history. Last accessed 25 Jan 2019
Roy, A., Memon, N., Ross, A.: MasterPrint: exploring the vulnerability of partial fingerprint-based authentication systems. IEEE Trans. Inf. Secur. 12(9), 2013–2025 (2017)
Hamilton, R.J.: User-generated evidence. Columbia J. Trans. Law 1(57) (2018) (American University, WCL Research Paper No. 2018-11)
ICC issues arrest warrant for Mahmoud Mustafa Busayf Al-Werfalli. https://www.icc-cpi.int/Pages/record.aspx?docNo=ICC-01/11-01/17-2. Last accessed 25 Jan 2019
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Khan, M.M., Ilyas, M.U., Saleem, S., Alowibdi, J.S., Alkatheiri, M.S. (2019). Emerging Computer Vision Based Machine Learning Issues for Smart Cities. In: Visvizi, A., Lytras, M. (eds) Research & Innovation Forum 2019. RIIFORUM 2019. Springer Proceedings in Complexity. Springer, Cham. https://doi.org/10.1007/978-3-030-30809-4_29
Download citation
DOI: https://doi.org/10.1007/978-3-030-30809-4_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-30808-7
Online ISBN: 978-3-030-30809-4
eBook Packages: EducationEducation (R0)