The Influence of Non-learnable Activation Functions on the Positioning Performance of Deep Learning-Based Fingerprinting Models Trained with Small CSI Sample Sizes

Loading...
Thumbnail Image
Date
2022
Journal Title
Journal ISSN
Volume Title
Publisher
Springer Science and Business Media LLC
Abstract
Activation functions, being mathematical ‘gates’ in between the input feeding the current neuron and its output going to the next layer, is very crucial in the training of deep learning models. They play a big part in determining the output of a model, its accuracy, and computational efficiency. In some cases, activation functions have a major effect on the model’s ability to converge and the convergence speed. To be able to train deep learning based fingerprint positioning models using small CSI sample sizes and have satisfactory positioning results, the choice of appropriate activation functions is very important. In this paper we explore several non-learnable activation functions and conduct a comprehensive analysis to study the influence they have on the positioning performance of deep learning fingerprint-based positioning models using small CSI sample sizes. We then propose a better model training approach with a view of getting the best out of those activation functions.
Description
Full-text Article. Also available at https://doi.org/10.1007/s41403-022-00347-x
Keywords
Deep learning, Small sample size, Channel state information, Activation functions
Citation
Lutakamale, A. S., & Manyesela, Y. Z. (2022). The Influence of Non-learnable Activation Functions on the Positioning Performance of Deep Learning-Based Fingerprinting Models Trained with Small CSI Sample Sizes. Transactions of the Indian National Academy of Engineering, 7(3), 1059-1067.
Collections