Abstract
Traffic data are obtained from various distributed sources such as infrastructure and vehicle sensors developed by various organisations, and often cannot be processed together because of data privacy regulations. Thus, distributed machine learning methods are required to process the data without sharing them. Federated learning allows the processing of data distributed by transmitting only the parameters without sharing the real data. The federated learning architecture is based mainly on deep learning, which is often more accurate than other machine learning approaches. However, deep-learning-based models are black-box models, and should be explained to increase trust in the system for both users and developers. Despite the fact that various explainability methods have been proposed, the solutions for explainable federated models are insufficient.
In this study, we used a federated deep learning model to predict a taxi trip duration within Brunswick region. We showed situations for which federated learning improves the prediction quality, allowing for an accuracy comparable to that obtained on the complete dataset. Moreover, we investigated how the amount of transmitted information in federated learning can be optimised while maintaining the same accuracy. Finally, we propose how the federated deep learning model can be interpreted using explainability methods without transmitting raw data and compare the results of various explainability approaches.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
References
Al-Abbasi, A.O., Ghosh, A., Aggarwal, V.: Deeppool: distributed model-free algorithm for ride-sharing using deep reinforcement learning. IEEE Trans. ITSs 20(12), 4714–4727 (2019). https://doi.org/10.1109/TITS.2019.2931830
Ancona, M., Ceolini, E., Öztireli, C., Gross, M.: Towards better understanding of gradient-based attribution methods for deep neural networks. In: Proceedings of 6th International Conference on Learning Representations, ICLR (2018). https://openreview.net/forum?id=Sy21R9JAW
Bonawitz, K., et al.: Towards federated learning at scale: system design. CoRR abs/1902.01046 (2019)
Castro, J., Gómez, D., Tejada, J.: Polynomial calculation of the Shapley value based on sampling. Comput. Oper. Res. 36(5), 1726–1730 (2009). https://doi.org/10.1016/j.cor.2008.04.004https://www.sciencedirect.com/science/article/pii/S0305054808000804, selected papers presented at the Tenth International Symposium on Locational Decisions (ISOLDE X)
Ciskowski, P., Drzewiński, G., Bazan, M., Janiczek, T.: Estimation of travel time in the city using neural networks trained with simulated urban traffic data. In: Zamojski, W., Mazurkiewicz, J., Sugier, J., Walkowiak, T., Kacprzyk, J. (eds.) DepCoS-RELCOMEX 2018. AISC, vol. 761, pp. 121–134. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-91446-6_13
Claes, R., Holvoet, T.: Ad hoc link traversal time prediction. In: IEEE Conference on ITS, pp. 1803–1808 (2011). https://doi.org/10.1109/ITSC.2011.6082970
Fiosina, J., Fiosins, M., Bonn, S.: Explainable deep learning for augmentation of sRNA expression profiles. J. Comput. Biol. (2019). (to appear)
Fiosina, J., Fiosins, M., Müller, J.: Big data processing and mining for next generation intelligent transportation systems. Jurnal Teknologi 63(3) (2013)
Fiosina, J., Fiosins, M., Müller, J.P.: Decentralised cooperative agent-based clustering in intelligent traffic clouds. In: Klusch, M., Thimm, M., Paprzycki, M. (eds.) MATES 2013. LNCS (LNAI), vol. 8076, pp. 59–72. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-40776-5_8
Fiosina, J.: Explainable federated learning for taxi travel time prediction. In: Proceedings of the 7th International Conference on Vehicle Technology and Intelligent Transport Systems: VEHITS2021, pp. 670–677. SciTePress (2021)
Fisher, A., Rudin, C., Dominici, F.: All models are wrong, but many are useful: learning a variable’s importance by studying an entire class of prediction models simultaneously. J. Mach. Learn. Res. 20(177), 1–81 (2019)
Holzinger, A.: Explainable AI (ex-AI). Informatik-Spektrum 41(2), 138–143 (2018). https://doi.org/10.1007/s00287-018-1102-5
Kokhlikyan, N., et al.: Captum: a unified and generic model interpretability library for pytorch (2020)
Konecný, J., McMahan, H.B., Ramage, D., Richtárik, P.: Federated optimization: distributed machine learning for on-device intelligence. CoRR abs/1610.02527 (2016). http://arxiv.org/abs/1610.02527
Kraus, S., et al.: AI for explaining decisions in multi-agent environment. In: AAAI-2020. AAAI Press (2020)
Lapuschkin, S., Binder, A., Montavon, G., Klauschen, F., Müller, K.R., Samek, W.: On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS One 10, e0130140 (2015). https://doi.org/10.1371/journal.pone.0130140
Lin, H., Zito, R., Taylor, M.: A review of travel-time prediction in transport and logistics. In: Proceedings of the Eastern Asia Society for Transportation Studies, vol. 5, pp. 1433–1448. Hamburg (2005)
Linardatos, P., Papastefanopoulos, V., Kotsiantis, S.: Explainable ai: a review of machine learning interpretability methods. Entropy 23(1) (2021). https://doi.org/10.3390/e23010018, https://www.mdpi.com/1099-4300/23/1/18
Lundberg, S., Lee, S.: A unified approach to interpreting model predictions. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems, vol. 30, pp. 4765–4774. Curran Associates, Inc. (2017). http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf
Meyes, R., Lu, M., de Puiseau, C.W., Meisen, T.: Ablation studies in artificial neural networks. CoRR abs/1901.08644 (2019). http://arxiv.org/abs/1901.08644
Molnar, C.: Interpretable machine learning. a guide for making black box models explainable. Lulu.com (2020). https://christophm.github.io/interpretable-ml-book/index.html
Ramanan, P., Nakayama, K., Sharma, R.: Baffle: blockchain based aggregator free federated learning. CoRR abs/1909.07452 (2020)
Ribeiro, M., Singh, S., Guestrin, C.: "why should i trust you?": explaining the predictions of any classifier. In: Proceedings of 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144. KDD 2016, ACM, New York, NY, USA (2016). https://doi.org/10.1145/2939672.2939778
Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences. In: Proceedings of ML Research, vol. 70, pp. 3145–3153. PMLR, Int. Convention Centre, Sydney, Australia, 06–11 August 2017
Shrikumar, A., Greenside, P., Shcherbina, A., Kundaje, A.: Not just a black box: learning important features through propagating activation differences. CoRR arXiv:1605.01713 (2017)
Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: visualising image classification models and saliency maps. In: 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, 14–16 April 2014. Workshop Track Proceedings (2014). http://arxiv.org/abs/1312.6034
Springenberg, J., Dosovitskiy, A., Brox, T., Riedmiller, M.: Striving for simplicity: the all convolutional net. In: ICLR (Workshop Track) (2015). http://lmb.informatik.uni-freiburg.de/Publications/2015/DB15a
Wang, G.: Interpret federated learning with Shapley values. CoRR abs/1905.04519 (2019)
Yang, Q., Liu, Y., Chen, T., Tong, Y.: Federated machine learning: concept and applications. ACM Trans. Intell. Syst. Technol. 10(2) (2019). https://doi.org/10.1145/3298981
Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 818–833. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10590-1_53
Acknowledgements
The research was funded by the Lower Saxony Ministry of Science and Culture under grant number ZN3493 within the Lower Saxony “Vorab” of the Volkswagen Foundation and supported by the Center for Digital Innovations (ZDIN). The Brunswick taxi FCD data were provided by SocialCars Research Training Group Project of the German Research Foundation.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Fiosina, J. (2022). Interpretable Privacy-Preserving Collaborative Deep Learning for Taxi Trip Duration Forecasting. In: Klein, C., Jarke, M., Helfert, M., Berns, K., Gusikhin, O. (eds) Smart Cities, Green Technologies, and Intelligent Transport Systems. VEHITS SMARTGREENS 2021 2021. Communications in Computer and Information Science, vol 1612. Springer, Cham. https://doi.org/10.1007/978-3-031-17098-0_20
Download citation
DOI: https://doi.org/10.1007/978-3-031-17098-0_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-17097-3
Online ISBN: 978-3-031-17098-0
eBook Packages: Computer ScienceComputer Science (R0)