Drift detection on feature attributions for monitoring visual reinforcement learning models in maritime port surveillance
Authors/Creators
- 1. Department of Computer Science and Artificial Intelligence, University of the Basque Country (UPV/EHU), Bilbao, Basque Country, 20018, Spain
- 2. Fundación Vicomtech, Basque Research and Technology Alliance (BRTA), Donostia/San Sebastián, Gipuzkoa, 20009, Spain
- 3. INDRA Sistemas S.A., Alcobendas, 28108, Spain
- 4. Ikerbasque, Bilbao, Basque Country, 48009, Spain
Description
This article shows how it is possible to monitor an AI model interpretation of the data it is being fed, rather than monitoring only the data itself, to improve our capability to understand how an AI model is behaving while it works. This allows us to detect changes in the way it is behaving, as well as understand these changes better, in order to detect more quickly when the model is not behaving correctly and fix it. In this article, we apply this approach to a maritime port surveillance environment, testing it under different simulated scenarios, such as overabundance of obstacles in the port or malicious attacks on the vessels, to see if the method meets its potential. We conclude that it does, with certain caveats such as the need of stronger hardware, resulting in a good balance between detection of irregular AI model behavior and interpretability of said behavior.
Files
openreseurope-6-23936.pdf
Files
(2.4 MB)
| Name | Size | Download all |
|---|---|---|
|
md5:4269a871483866c106dc967689bb6be6
|
2.4 MB | Preview Download |
Additional details
References
- (2023). Review of maritime transport 2023.
- (2022). European drug report 2022: trends and developments. doi:10.2810/75644
- (null). Coordination Of maritime assets for Persistent And Systematic Surveillance | COMPASS2020 | project | fact sheet | H2020. doi:10.3030/833650
- (null). Risk-aware Automated Port Inspection Drone(s) | RAPID | project | fact sheet | H2020. doi:10.3030/861211
- (null). Underwater security | UnderSec | project | fact sheet | HORIZON. doi:10.3030/101121288
- (null). Smart maritime and Underwater Guardian | SMAUG | project | fact sheet | HORIZON. doi:10.3030/101121129
- Wu W, Gao C, Chen J (2025). Reinforcement learning in vision: a survey. arXiv. doi:10.48550/arXiv.2508.08189
- Gikay AA, Lau PL, Sengul C (2023). High-risk Artificial Intelligence systems under the European Union's AI Act: systemic flaws and practical challenges. doi:10.2139/ssrn.4621605
- Patchipala SG (2023). Tackling data and model drift in AI: strategies for maintaining accuracy during ML model inference. Int J Sci Res Arch. doi:10.30574/ijsra.2023.10.2.0855
- Testi M, Ballabio M, Frontoni E (2022). MLOps: a taxonomy and a methodology. IEEE Access. doi:10.1109/ACCESS.2022.3181730
- Irpan A, Rao K, Bousmalis K (2019). Off-Policy evaluation via Off-Policy classification. doi:10.48550/arXiv.1906.01624
- Massey FJ (1951). The kolmogorov-smirnov test for goodness of fit. J Am Stat Assoc. doi:10.2307/2280095
- Gretton A, Borgwardt KM, Rasch MJ (2012). A kernel two-sample test. J Mach Learn Res.
- (2024). Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act) (Text with EEA relevance).
- Lundberg SM, Lee SI (2017). A unified approach to interpreting model predictions. doi:10.48550/arXiv.1705.07874
- Sundararajan M, Taly A, Yan Q (2017). Axiomatic attribution for deep networks. doi:10.48550/arXiv.1703.01365
- Lee Y, Lee Y, Lee E (2023). Explainable Artificial Intelligence-based model drift detection applicable to unsupervised environments. Comput Mater Contin. doi:10.32604/cmc.2023.040235
- Lee YE, Lee TJ (2023). A study on efficient ai model drift detection methods for MLOps. J Internet Comput Serv. doi:10.7472/jksii.2023.24.5.17
- Zimmermann B, Boussard M (2022). Improving drift detection by monitoring shapley loss values. doi:10.1007/978-3-031-09282-4_38
- Mnih V, Kavukcuoglu K, Silver D (2013). Playing atari with deep reinforcement learning. arXiv. doi:10.48550/arXiv.1312.5602
- Mnih V, Badia AP, Mirza M (2016). Asynchronous methods for deep reinforcement learning. arXiv. doi:10.48550/arXiv.1602.01783
- Schulman J, Wolski F, Dhariwal P (2017). Proximal policy optimization algorithms. arXiv. doi:10.48550/arXiv.1707.06347
- Gamage C, Dinalankara R, Samarabandu J (2023). A comprehensive survey on the applications of machine learning techniques on maritime surveillance to detect abnormal maritime vessel behaviors. WMU J Marit Aff. doi:10.1007/s13437-023-00312-7
- Yun WJ, Park S, Kim J (2022). Cooperative multiagent deep reinforcement learning for reliable surveillance via autonomous multi-UAV control. IEEE Trans Ind Inform. doi:10.1109/TII.2022.3143175
- Nandhini TJ, Thinakaran K (2023). Optimizing forensic investigation and security surveillance with deep reinforcement learning techniques. doi:10.1109/ICDSAAI59313.2023.10452551
- Zhao R, Li Y, Fan Y (2024). A survey on recent advancements in autonomous driving using deep reinforcement learning: applications, challenges, and solutions. IEEE Trans Intell Transp Syst. doi:10.1109/TITS.2024.3452480
- Vela D, Sharp A, Zhang R (2022). Temporal quality degradation in AI models. Sci Rep. doi:10.1038/s41598-022-15245-z
- Hassija V, Chamola V, Mahapatra A (2024). Interpreting black-box models: a review on explainable Artificial Intelligence. Cogn Comput. doi:10.1007/s12559-023-10179-8
- Sharief F, Ijaz H, Shojafar M (2025). Multi-class imbalanced data handling with concept drift in fog computing: a taxonomy, review, and future directions. ACM Comput Surv. doi:10.1145/3689627
- Anderson TW (1962). On the distribution of the two-sample Cramer-von mises criterion. Ann Math Stat. doi:10.1214/aoms/1177704477
- Agrahari S, Singh AK (2022). Concept drift detection in data stream mining : a literature review. J King Saud Univ Comput Inf Sci. doi:10.1016/j.jksuci.2021.11.006
- Hovakimyan G, Bravo JM (2024). Evolving strategies in machine learning: a systematic review of concept drift detection. Information. doi:10.3390/info15120786
- Bodor A, Hnida M, Daoudi N (2023). Machine learning models monitoring in MLOps context: metrics and tools. Int J Interact Mob Technol. doi:10.3991/ijim.v17i23.43479
- Böhle M, Singh N, Fritz M (2024). B-Cos alignment for inherently interpretable CNNs and vision transformers. IEEE Trans Pattern Anal Mach Intell. doi:10.1109/TPAMI.2024.3355155
- Rodriguez AM, Unzueta L, Geradts Z (2023). Multi-task explainable quality networks for large-scale forensic facial recognition. IEEE J Sel Top Signal Process. doi:10.1109/JSTSP.2023.3267263
- Selvaraju RR, Cogswell M, Das A (2017). Grad-CAM: visual explanations from deep networks via gradient-based localization. doi:10.1109/ICCV.2017.74
- Bora RP, Terhorst P, Veldhuis R (2024). SLICE: Stabilized LIME for Consistent Explanations for Image Classification. doi:10.1109/CVPR52733.2024.01045
- Ibrahim R, Shafiq MO (2023). Explainable convolutional neural networks: a taxonomy, review, and future directions. ACM Comput Surv. doi:10.1145/3563691
- Fantozzi P, Naldi M (2024). The explainability of transformers: current status and directions. Computers. doi:10.3390/computers13040092
- Arrieta AB, Díaz-Rodríguez N, Ser J (2020). Explainable Artificial Intelligence (Xai): concepts, taxonomies, opportunities and challenges toward responsible Ai. Inf Fusion. doi:10.1016/j.inffus.2019.12.012
- Chen H, Lundberg SM, Lee SI (2022). Explaining a series of models by propagating shapley values. Nat Commun. doi:10.1038/s41467-022-31384-3
- Chattopadhay A, Sarkar A, Howlader P (2018). Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks. doi:10.1109/WACV.2018.00097
- Muhammad MB, Yeasin M (2020). Eigen-CAM: class activation map using principal components. doi:10.1109/IJCNN48605.2020.9206626
- Molnar C (2025). Interpretable machine learning: a guide for making black box models explainable.
- Niu L, Li Z, Li S (2024). MMD fence GAN unsupervised anomaly detection model based on maximum mean discrepancy. Int J Cogn Inform Nat Intell. doi:10.4018/IJCINI.344813
- Rabanser S, Günnemann S, Lipton ZC (2019). Failing loudly: an empirical study of methods for detecting dataset shift. doi:10.48550/arXiv.1810.11953
- Iriarte FJ (2025). Zenodo.