Analysis of Sparse Trajectory Features Based on Mobile Device Location for User Group Classification Using Gaussian Mixture Model
<p>Data distribution for missing value ratio <math display="inline"><semantics> <mrow> <msub> <mi>θ</mi> <mi>k</mi> </msub> <mo>,</mo> <mo>∀</mo> <mi>k</mi> <mo>∈</mo> <mi>K</mi> </mrow> </semantics></math> for the PTDP data. These data were acquired from January to June 2021 in Narashino, Chiba, Japan. There are differences in the quantities of data among the missing value ratios. In particular, because the data range of 0.99–1.00 accounts for 60% of all user data, this figure shows that the user trajectories obtained from mobile devices in the area and in that period are very sparse.</p> "> Figure 2
<p>Imputation method based on the assumption as described in <a href="#sec3dot3-applsci-15-00982" class="html-sec">Section 3.3</a>. Data between two temporally continuous nodes are linearly imputed.</p> "> Figure 3
<p>Procedure for generating feature vectors <math display="inline"><semantics> <msubsup> <mrow> <mi mathvariant="bold-italic">f</mi> </mrow> <mi>k</mi> <mi>GMM</mi> </msubsup> </semantics></math> from <math display="inline"><semantics> <msub> <mi>L</mi> <mi>k</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>T</mi> <mi>k</mi> </msub> </semantics></math>.</p> "> Figure 4
<p>(<b>a</b>) SDA. (<b>b</b>) Fine-tuning of the DNN obtained via SDA.</p> "> Figure 5
<p>Data distribution by class and month. The raw location data are imbalanced. Undersampling is conducted based on the black line, i.e., <math display="inline"><semantics> <msup> <mi>S</mi> <mo>′</mo> </msup> </semantics></math>.</p> "> Figure 6
<p>Data distribution for the missing value rate <math display="inline"><semantics> <mrow> <msub> <mi>θ</mi> <mi>k</mi> </msub> <mo>,</mo> <mo>∀</mo> <mi>k</mi> <mo>∈</mo> <mi>K</mi> </mrow> </semantics></math> by month. Although there are differences in the data quantity, the distributions’ shapes for each missing value rate between months are almost the same.</p> "> Figure 7
<p>Diagram of fivefold GS+SV for tuning the hyperparameters of the ML models. This process increases the model performance and reduces overfitting of the model on the training data.</p> "> Figure 8
<p>Summary of a series of experimental procedures.</p> "> Figure 9
<p>Average and standard deviation (1SD) of 20 accuracies on test data by month for the <math display="inline"><semantics> <mrow> <msub> <mi>θ</mi> <mi>th</mi> </msub> <mo>=</mo> <mn>1.00</mn> </mrow> </semantics></math> and ML models. Each graph title indicates the maximum average and its standard deviation by components <span class="html-italic">n</span> in the form of average (1SD) and the <span class="html-italic">p</span>-value for the Shapiro–Wilk test for the accuracy distribution of 20 seeds obtained by components <span class="html-italic">n</span>. There is a tendency that the higher the number of GMD components is, the higher the accuracy across all months. Additionally, for all months, distribution normality for 20 accuracies is confirmed with <math display="inline"><semantics> <mrow> <mi>p</mi> <mo>></mo> <mn>0.050</mn> </mrow> </semantics></math>, and accuracies are higher than those of random classification (<math display="inline"><semantics> <mrow> <mi>accuracy</mi> <mo>≈</mo> <mn>0.170</mn> </mrow> </semantics></math>) in the range of 3SD on components <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>></mo> <mn>2</mn> </mrow> </semantics></math>, with any accuracy average by SVC and RF.</p> "> Figure 10
<p>Average and standard deviation (1SD) of 20 accuracies on training data by month in <math display="inline"><semantics> <mrow> <msub> <mi>θ</mi> <mi>th</mi> </msub> <mo>=</mo> <mn>1.00</mn> </mrow> </semantics></math> and ML models. This metric is used to evaluate the generalization performance of the ML models, and the lower the difference in the accuracy of the test data is, the higher the generalization performance. Compared with the test data, SVC and DNN exhibit a gradual increase in accuracy with respect to the number of components <span class="html-italic">n</span>, whereas RF shows an increase in accuracy from smaller values of <span class="html-italic">n</span> compared with the other two models.</p> "> Figure 11
<p>This figure illustrates the trends in classification accuracy for the missing value ratio <math display="inline"><semantics> <msub> <mi>θ</mi> <mi>th</mi> </msub> </semantics></math> based on the combinations of feature extraction methods and ML models for each month. Averages of 20 accuracies on test data. The dashed horizontal line indicates the best accuracy among the combinations. The title shows the month with the best average accuracy (standard deviation), the combinations of the feature extraction method and the ML model provide the best accuracy, and the results of the Shapiro–Wilk test for the accuracy distribution of 20 seeds, where n.s. is <math display="inline"><semantics> <mrow> <mi>p</mi> <mo>≥</mo> <mn>0.050</mn> </mrow> </semantics></math>. The best average accuracy values are higher than those of random classification (<math display="inline"><semantics> <mrow> <mi>accuracy</mi> <mo>≈</mo> <mn>0.170</mn> </mrow> </semantics></math>) in the range of 3SD in any month.</p> "> Figure 12
<p>Average of 20 accuracies on training data by <math display="inline"><semantics> <msub> <mi>θ</mi> <mi>th</mi> </msub> </semantics></math> for each month. The lower the difference between the accuracies of the training and test data is, the higher the generalization performance. The ML models using the IDNN show a wider range of accuracy differences between the training and test data than those using the GMM. This shows that features based on the GMM are more robust to overfitting than those based on the IDNN.</p> ">
Abstract
:1. Introduction
2. Literature Review
2.1. Trajectory Analysis Using Semantic Information
2.2. Trajectory Analysis Using Only Spatial–Temporal Information
2.2.1. Similarity-Based Methods
2.2.2. Trajectory Prediction and Recovery for Sparse Trajectory
2.3. Feature Extraction Applicable to Sparse Trajectories Without Semantic Information: A Comparative Baseline
3. Trajectory Data
3.1. PTDP Data
- Daily ID: User ID associated with the mobile device. This ID is reset on a daily basis.
- Timestamp: timestamp recorded in yyyy-mm-dd HH:MM for obtaining location information.
- Longitude: longitude of the location data.
- Latitude: latitude of the location data.
- Age group: age groups composed of those aged under 15, 70 and over and other age intervals divided into five-year increments, i.e., 0–14, 15–19, 20–24, …, 65–69, 70 and over.
3.2. Data Assumption
3.3. Imputation for Missing Location Data
4. Feature Extraction Method for Sparse Trajectories
4.1. Feature Extraction by GMM
4.1.1. Determining RPs via GMM
4.1.2. Aggregation of User Trajectories into RPs
4.2. Deep Feature Extraction: A Comparative Baseline
- Trajectory image generationThe target area is divided into rectangular regions of a specified size , and each region is represented as a pixel. This allows the target area to be treated as an image, where each element of the trajectory is converted into a pixel value for the region . The pixel value is incremented by one for every minute the user k spends in the region. Thus, represents the total time that user k spends in region , satisfying the following equation:Finally, the pixel values are normalized such that they fall within the range . This process generates a grayscale image associated with each ID k.
- Pretraining by SDAThe autoencoder (AE) is a type of NN designed to compress and reconstruct input vectors, and it is primarily used for feature extraction and dimensionality reduction. For example, when a flattened vector of an image is denoted as , an AE constructs an NN that minimizes the loss function between and the reconstructed vector . The denoising autoencoder (DAE) is an extension of the AE that enhances robustness by adding noise, sampled from a distribution , to the input vector before constructing the AE. SDA is a type of DNN built by stacking multiple DAEs as middle layers, where each layer is trained independently by its respective DAE [15]. Specifically, the l-th middle layer of an SDA is trained by minimizing the loss function between the features learned by the -th layer DAE and the output of the l-th layer. The number of middle layers in a DNN constructed by an SDA equals the number of stacked DAEs. In the l-th DAE, the input vector is first passed through the encoding function after noise is added according to and then reconstructed as a vector through the decoding function . Typically, employs masking noise, which randomly forces selected elements to zero. The conceptual diagram of an SDA is shown in Figure 4a.
- Fine tuning and feature extractionFine-tuning is performed by adding an output layer to the top of the pretrained DNN constructed by the SDA and minimizing the loss function between the output and the target labels (Figure 4b). In the fine-tuning process, the target labels are defined as binary vectors representing the classes. Finally, the fine-tuned IDNN takes the vector as input, and the feature vector is extracted from the -th intermediate layer of the IDNN. The dimensionality of the extracted feature vector is equal to the number of neurons in the -th middle layer.
5. Experimental Setup
5.1. Data Summary
5.2. Under Sampling for Imbalanced Data
5.3. ML Models for Classification
5.4. Metrics
6. Numerical Experiment
6.1. Methods
6.1.1. Tuning GMD Components n for Model Performance
6.1.2. Comparison of GMM and IDNN
- Image size:
- Noise distribution : masking noise (rate = 0.2)
- Activation functions:
- −
- , : ReLU
- −
- : Softmax
- Loss functions:
- −
- , : mean squared error
- Optimizer for each DEA and fine tuning of the DNN: Adam
6.1.3. Performance by the Missing Value Rate of the User Trajectory
6.2. Results and Discussion: Tuning GMD Components n for Model Performance
6.3. Results and Discussion: Comparison of the GMM and IDNN
6.4. Results and Discussions: Performance by the Missing Value Rate of the User Trajectory
7. Conclusions
- Although simple linear interpolation was adopted for trajectory imputation, its impact on classification performance was not verified.
- The proposed method was verified only for a single region. In addition, owing to constraints in the data acquisition period, the influence of behavioral changes from prepandemic norms during the COVID-19 era could not be taken into account.
- The features did not contain the connection relationships of RPs.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Van der Spek, S.; Van Schaick, J.; De Bois, P.; De Haan, R. Sensing Human Activity: GPS Tracking. Sensors 2009, 9, 3033–3055. [Google Scholar] [CrossRef] [PubMed]
- Giannotti, F.; Nanni, M.; Pinelli, F.; Pedreschi, D. Trajectory pattern mining. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Jose, CA, USA, 12–15 August 2007; pp. 330–339. [Google Scholar] [CrossRef]
- Lee, J.G.; Han, J.; Li, X.; Gonzalez, H. TraClass: Trajectory classification using hierarchical region-based and trajectory-based clustering. Proc. VLDB Endow. 2008, 1, 1081–1094. [Google Scholar] [CrossRef]
- Kapadais, K.; Varlamis, I.; Sardianos, C.; Tserpes, K. A Framework for the Detection of Search and Rescue Patterns Using Shapelet Classification. Future Internet 2019, 11, 192. [Google Scholar] [CrossRef]
- Kontopoulos, I.; Makris, A.; Tserpes, K. A Deep Learning Streaming Methodology for Trajectory Classification. ISPRS Int. J. Geo-Inf. 2021, 10, 250. [Google Scholar] [CrossRef]
- Li, Q.; Zheng, Y.; Xie, X.; Chen, Y.; Liu, W.; Ma, W.Y. Mining user similarity based on location history. In Proceedings of the ACM International Symposium on Advances in Geographic Information Systems, Irvine, CA, USA, 5–7 November 2008; pp. 298–307. [Google Scholar] [CrossRef]
- Calabrese, F.; Diao, M.; Lorenzo, G.D.; Ferreira, J.; Ratti, C. Understanding individual mobility patterns from urban sensing data: A mobile phone trace example. Transp. Res. Part C Emerg. Technol. 2013, 26, 301–313. [Google Scholar] [CrossRef]
- Zhong, E.; Tan, B.; Mo, K.; Yang, Q. User demographics prediction based on mobile data. Pervasive Mob. Comput. 2013, 9, 823–837. [Google Scholar] [CrossRef]
- Hu, T.; Luo, J.; Kautz, H.; Sadilek, A. Home Location Inference from Sparse and Noisy Data: Models and Applications. In Proceedings of the 15th IEEE International Conference on Data Mining Workshop (ICDMW 2015), Atlantic City, NJ, USA, 14–17 November 2015; pp. 1382–1387. [Google Scholar] [CrossRef]
- Sun, F.; Wang, P.; Zhao, J.; Xu, N.; Zeng, J.; Tao, J.; Song, K.; Deng, C.; Lui, J.C.; Guan, X. Mobile Data Traffic Prediction by Exploiting Time-Evolving User Mobility Patterns. IEEE Trans. Mob. Comput. 2022, 21, 4456–4470. [Google Scholar] [CrossRef]
- Sinclair, M.; Maadi, S.; Zhao, Q.; Hong, J.; Ghermandi, A.; Bailey, N. Assessing the socio-demographic representativeness of mobile phone application data. Appl. Geogr. 2023, 158, 102997. [Google Scholar] [CrossRef]
- Ferrero, C.A.; Petry, L.M.; Alvares, L.O.; da Silva, C.L.; Zalewski, W.; Bogorny, V. MasterMovelets: Discovering heterogeneous movelets for multiple aspect trajectory classification. Data Min. Knowl. Discov. 2020, 34, 652–680. [Google Scholar] [CrossRef]
- Han, P.; Wang, J.; Yao, D.; Shang, S.; Zhang, X. A Graph-based Approach for Trajectory Similarity Computation in Spatial Networks. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Virtually, 14–18 August 2021; pp. 556–564. [Google Scholar] [CrossRef]
- Zhou, S.; Han, P.; Yao, D.; Chen, L.; Zhang, X. Spatial-temporal fusion graph framework for trajectory similarity computation. World Wide Web 2023, 26, 1501–1523. [Google Scholar] [CrossRef]
- Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.A. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
- Endo, Y.; Toda, H.; Nishida, K.; Ikedo, J. Classifying spatial trajectories using representation learning. Int. J. Data Sci. Anal. 2016, 2, 107–117. [Google Scholar] [CrossRef]
- Kakimoto, Y.; Omae, Y. Evaluation of location-data based features using Gaussian mixture models for age group estimation. J. Phys. Conf. Ser. 2024, 2701, 012070. [Google Scholar] [CrossRef]
- Boser, B.E.; Guyon, I.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Agoop Corp. Dynamic Population Data. Available online: https://agoop.co.jp/ (accessed on 16 January 2025).
- Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 16 January 2025).
- Kwon, Y.H.; Lobo, N.d.V. Age Classification from Facial Images. Comput. Vis. Image Underst. 1999, 74, 1–21. [Google Scholar] [CrossRef]
- Angulu, R.; Tapamo, J.R.; Adewumi, A.O. Age estimation via face images: A survey. EURASIP J. Image Video Process. 2018, 2018, 42. [Google Scholar] [CrossRef]
- Mustapha, M.F.; Mohamad, N.M.; Osman, G.; Hamid, S.H.A. Age Group Classification using Convolutional Neural Network (CNN). J. Phys. Conf. Ser. 2021, 2084, 012028. [Google Scholar] [CrossRef]
- Saxena, A.K.; Chaurasiya, V.K. Fingerprint based human age group estimation. In Proceedings of the 2014 Annual IEEE India Conference (INDICON), Pune, India, 11–13 December 2014; pp. 1–4. [Google Scholar] [CrossRef]
- Nabila, M.; Mohammed, A.I.; Yousra, B.J. Gait-based human age classification using a silhouette model. IET Biom. 2018, 7, 116–124. [Google Scholar] [CrossRef]
- Guimaraes, R.G.; Rosa, R.L.; De Gaetano, D.; Rodríguez, D.Z.; Bressan, G. Age Groups Classification in Social Network Using Deep Learning. IEEE Access 2017, 5, 10805–10816. [Google Scholar] [CrossRef]
- Wang, P.; Sun, F.; Wang, D.; Tao, J.; Guan, X.; Bifet, A. Inferring demographics and social networks of mobile device users on campus from ap-trajectories. In Proceedings of the 26th International World Wide Web Conference 2017 (WWW 2017), Perth, Australia, 3–7 April 2017; pp. 139–147. [Google Scholar] [CrossRef]
- Montasser, O.; Kifer, D. Predicting Demographics of High-Resolution Geographies with Geotagged Tweets. In Proceedings of the 31st AAAI Conference on Artificial Intelligence (AAAI 2017), San Francisco, CA, USA, 4–9 February 2017; pp. 1460–1466. [Google Scholar] [CrossRef]
- Berndt, D.; Clifford, J. Using Dynamic Time Warping to Find Patterns in Time Series. In Proceedings of the 3rd International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 31 July–1 August 1994. [Google Scholar]
- Holt, G.T.; Reinders, M.; Hendriks, E. Multi-Dimensional Dynamic Time Warping for Gesture Recognition. In Proceedings of the Thirteenth Annual Conference of the Advanced School for Computing and Imaging, Heijen, The Nederland, 13–15 June 2007. [Google Scholar]
- Shokoohi-Yekta, M.; Hu, B.; Jin, H.; Wang, J.; Keogh, E. Generalizing DTW to the multi-dimensional case requires an adaptive approach. Data Min. Knowl. Discov. 2017, 31, 1–31. [Google Scholar] [CrossRef] [PubMed]
- Wang, H.; Zeng, S.; Li, Y.; Zhang, P.; Jin, D. Human Mobility Prediction Using Sparse Trajectory Data. IEEE Trans. Veh. Technol. 2020, 69, 10155–10166. [Google Scholar] [CrossRef]
- Xia, T.; Qi, Y.; Feng, J.; Xu, F.; Sun, F.; Guo, D.; Li, Y. AttnMove: History Enhanced Trajectory Recovery via Attentional Network. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 4494–4502. [Google Scholar] [CrossRef]
- Sun, H.; Yang, C.; Deng, L.; Zhou, F.; Huang, F.; Zheng, K. PeriodicMove: Shift-aware Human Mobility Recovery with Graph Neural Network. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (CIKM ’21), Gold Coast, QLD, Australia, 1–5 November 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 1734–1743. [Google Scholar] [CrossRef]
- Si, J.; Yang, J.; Xiang, Y.; Wang, H.; Li, L.; Zhang, R.; Tu, B.; Chen, X. TrajBERT: BERT-Based Trajectory Recovery With Spatial-Temporal Refinement for Implicit Sparse Trajectories. IEEE Trans. Mob. Comput. 2024, 23, 4849–4860. [Google Scholar] [CrossRef]
- Chen, W.; Liang, Y.; Zhu, Y.; Chang, Y.; Luo, K.; Wen, H.; Li, L.; Yu, Y.; Wen, Q.; Chen, C.; et al. Deep Learning for Trajectory Data Management and Mining: A Survey and Beyond. arXiv 2024, arXiv:2403.14151. [Google Scholar] [CrossRef]
- García-Palomares, J.C.; Gutiérrez, J.; Cardozo, O.D. Walking Accessibility to Public Transport: An Analysis Based on Microdata and GIS. Environ. Plan. Plan. Des. 2013, 40, 1087–1102. [Google Scholar] [CrossRef]
- Joo, S.; Kashiyama, T.; Sekimoto, Y.; Seto, T. An Analysis of Factors Influencing Disaster Mobility Using Location Data from Smartphones: Case Study of Western Japan Flooding. J. Disaster Res. 2019, 14, 903–911. [Google Scholar] [CrossRef]
- Mori, M.; Omae, Y.; Kakimoto, Y.; Sasaki, M.; Toyotani, J.; Mori, M.; Omae, Y.; Kakimoto, Y.; Sasaki, M.; Toyotani, J. Analyzing factors of daily travel distances in Japan during the COVID-19 pandemic. Math. Biosci. Eng. 2024, 21, 6936–6974. [Google Scholar] [CrossRef]
- Eom, S.; Kim, H.; Hasegawa, D.; Yamada, I. Pedestrian movement with large-scale GPS records and transit-oriented development attributes. Sustain. Cities Soc. 2024, 102, 105223. [Google Scholar] [CrossRef]
- Statistics Bureau of Japan. Statistics. Available online: https://www.stat.go.jp/english/data/index.html (accessed on 12 March 2024).
- Geospatil Information Authority of Japan. Planimetric Reports on the Land Area by Prefectures and Municipalities in Japan. Available online: https://www.gsi.go.jp/KOKUJYOHO/MENCHO-title.htm (accessed on 12 March 2024).
- Cabinet Agency for Infectious Disease Crisis Management. About The Cabinet Agency for Infectious Disease Crisis Management. Available online: https://www.caicm.go.jp/en/about/index.html (accessed on 12 March 2024).
- Guo, H.; Li, Y.; Shang, J.; Gu, M.; Huang, Y.; Bing, G. Learning from class-imbalanced data: Review of methods and applications. Expert Syst. Appl. 2017, 73, 220–239. [Google Scholar] [CrossRef]
- Hua, J.; Xiong, Z.; Lowey, J.; Suh, E.; Dougherty, E.R. Optimal number of features as a function of sample size for various classification rules. Bioinformatics 2005, 21, 1509–1515. [Google Scholar] [CrossRef]
- Rajput, D.; Wang, W.J.; Chen, C.C. Evaluation of a decided sample size in machine learning applications. BMC Bioinform. 2023, 24, 48. [Google Scholar] [CrossRef] [PubMed]
- Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
- Moon, T.K. The expectation-maximization algorithm. IEEE Signal Process. Mag. 1996, 13, 47–60. [Google Scholar] [CrossRef]
- Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the JMLR Workshop and Conference Proceedings, Taoyuan, Taiwan, 13–15 November 2011; pp. 315–323. [Google Scholar]
- Hui, L.; Belkin, M. Evaluation of Neural Architectures Trained with Square Loss vs Cross-Entropy in Classification Tasks. arXiv 2021, arXiv:2006.07322. [Google Scholar]
Data | Month | 0.95 | 0.96 | 0.97 | 0.98 | 0.99 | 1.00 |
---|---|---|---|---|---|---|---|
- | 96 | 198 | 300 | 510 | 1044 | 2286 | |
January | 42 | 84 | 126 | 216 | 444 | 972 | |
February | 102 | 180 | 330 | 612 | 1176 | 2538 | |
March | 834 | 1338 | 2004 | 3222 | 6342 | 16,752 | |
April | 1146 | 1686 | 2592 | 4194 | 7848 | 21,660 | |
May | 1044 | 1476 | 2556 | 4206 | 7680 | 20,952 | |
June | 1638 | 2280 | 3996 | 6810 | 12,888 | 38,136 |
Month | Data | GMM-SVC | GMM-RF | GMM-DNN | IDNN-SVC | IDNN-RF | IDNN-DNN |
---|---|---|---|---|---|---|---|
January | Train | 0.631 (0.008) | 0.597 (0.011) | 0.415 (0.027) | 0.907 (0.019) | 0.794 (0.017) | 0.350 (0.039) |
Test | 0.336 (0.013) | 0.331 (0.011) | 0.310 (0.012) | 0.323 (0.014) | 0.298 (0.012) | 0.268 (0.018) | |
Param. | (200) | (60) | (180) | (2, 200) | (2, 200) | (2, 200) | |
February | Train | 0.575 (0.033) | 0.599 (0.016) | 0.390 (0.030) | 0.907 (0.017) | 0.785 (0.024) | 0.313 (0.033) |
Test | 0.298 (0.010) | 0.296 (0.008) | 0.276 (0.009) | 0.296 (0.009) | 0.270 (0.009) | 0.236 (0.012) | |
Param. | (160) | (50) | (200) | (2, 200) | (2, 150) | (2, 200) | |
March | Train | 0.482 (0.065) | 0.635 (0.035) | 0.310 (0.020) | 0.905 (0.058) | 0.812 (0.029) | 0.285 (0.036) |
Test | 0.234 (0.006) | 0.243 (0.004) | 0.221 (0.007) | 0.226 (0.007) | 0.224 (0.006) | 0.205 (0.011) | |
Param. | (120) | (180) | (120) | (2, 200) | (3, 200) | (2, 200) | |
April | Train | 0.453 (0.080) | 0.590 (0.035) | 0.326 (0.028) | 0.712 (0.203) | 0.743 (0.059) | 0.290 (0.037) |
Test | 0.233 (0.005) | 0.244 (0.005) | 0.227 (0.007) | 0.226 (0.004) | 0.229 (0.006) | 0.216 (0.006) | |
Param. | (130) | (180) | (190) | (2, 200) | (2, 150) | (2, 200) | |
May | Train | 0.500 (0.095) | 0.599 (0.038) | 0.331 (0.025) | 0.715 (0.222) | 0.756 (0.054) | 0.294 (0.033) |
Test | 0.235 (0.004) | 0.250 (0.004) | 0.230 (0.004) | 0.228 (0.005) | 0.232 (0.004) | 0.220 (0.006) | |
Param. | (180) | (150) | (190) | (2, 200) | (2, 200) | (2, 200) | |
June | Train | 0.443 (0.102) | 0.586 (0.062) | 0.316 (0.027) | 0.578 (0.195) | 0.736 (0.107) | 0.285 (0.040) |
Test | 0.224 (0.004) | 0.238 (0.004) | 0.221 (0.006) | 0.220 (0.005) | 0.223 (0.004) | 0.211 (0.010) | |
Param. | (160) | (80) | (160) | (2, 200) | (2, 200) | (2, 200) |
Month | n | Shapiro–Wilk † | |||||
---|---|---|---|---|---|---|---|
January | 60 | 0.95 | 0.535 (0.087) | 0.562 (0.085) | 0.535 (0.087) | 0.534 (0.086) | n.s. |
0.96 | 0.545 (0.064) | 0.564 (0.062) | 0.545 (0.064) | 0.547 (0.063) | n.s. | ||
0.97 | 0.528 (0.046) | 0.543 (0.044) | 0.528 (0.046) | 0.529 (0.045) | n.s. | ||
0.98 | 0.485 (0.036) | 0.496 (0.031) | 0.485 (0.036) | 0.484 (0.034) | n.s. | ||
0.99 | 0.421 (0.021) | 0.426 (0.022) | 0.421 (0.021) | 0.418 (0.022) | n.s. | ||
1.00 | 0.331 (0.011) | 0.330 (0.013) | 0.331 (0.011) | 0.320 (0.009) | n.s. | ||
February | 50 | 0.95 | 0.443 (0.040) | 0.464 (0.041) | 0.443 (0.040) | 0.443 (0.040) | n.s. |
0.96 | 0.476 (0.038) | 0.483 (0.042) | 0.476 (0.038) | 0.473 (0.039) | n.s. | ||
0.97 | 0.428 (0.024) | 0.434 (0.027) | 0.428 (0.024) | 0.426 (0.025) | n.s. | ||
0.98 | 0.423 (0.018) | 0.426 (0.019) | 0.423 (0.018) | 0.421 (0.018) | n.s. | ||
0.99 | 0.371 (0.016) | 0.364 (0.018) | 0.371 (0.016) | 0.363 (0.017) | n.s. | ||
1.00 | 0.296 (0.008) | 0.297 (0.009) | 0.296 (0.008) | 0.292 (0.009) | n.s. | ||
March | 180 | 0.95 | 0.326 (0.024) | 0.341 (0.024) | 0.326 (0.024) | 0.327 (0.023) | n.s. |
0.96 | 0.342 (0.017) | 0.343 (0.020) | 0.342 (0.017) | 0.338 (0.019) | n.s. | ||
0.97 | 0.339 (0.017) | 0.339 (0.017) | 0.339 (0.017) | 0.334 (0.018) | n.s. | ||
0.98 | 0.329 (0.012) | 0.330 (0.012) | 0.329 (0.012) | 0.325 (0.012) | n.s. | ||
0.99 | 0.292 (0.009) | 0.289 (0.008) | 0.292 (0.009) | 0.287 (0.009) | n.s. | ||
1.00 | 0.243 (0.004) | 0.245 (0.006) | 0.243 (0.004) | 0.239 (0.005) | n.s. | ||
April | 180 | 0.95 | 0.319 (0.023) | 0.328 (0.026) | 0.319 (0.023) | 0.317 (0.024) | n.s. |
0.96 | 0.335 (0.023) | 0.335 (0.025) | 0.335 (0.023) | 0.332 (0.024) | n.s. | ||
0.97 | 0.325 (0.014) | 0.325 (0.017) | 0.325 (0.014) | 0.321 (0.017) | n.s. | ||
0.98 | 0.319 (0.012) | 0.318 (0.017) | 0.319 (0.012) | 0.315 (0.014) | n.s. | ||
0.99 | 0.292 (0.008) | 0.289 (0.011) | 0.292 (0.008) | 0.283 (0.010) | n.s. | ||
1.00 | 0.244 (0.005) | 0.241 (0.006) | 0.244 (0.005) | 0.232 (0.006) | n.s. | ||
May | 150 | 0.95 | 0.315 (0.013) | 0.317 (0.015) | 0.315 (0.013) | 0.312 (0.014) | ** |
0.96 | 0.331 (0.015) | 0.328 (0.016) | 0.331 (0.015) | 0.327 (0.015) | n.s. | ||
0.97 | 0.325 (0.012) | 0.323 (0.014) | 0.325 (0.012) | 0.321 (0.013) | n.s. | ||
0.98 | 0.325 (0.014) | 0.325 (0.017) | 0.325 (0.014) | 0.320 (0.016) | * | ||
0.99 | 0.298 (0.009) | 0.292 (0.010) | 0.298 (0.009) | 0.288 (0.011) | n.s. | ||
1.00 | 0.250 (0.004) | 0.245 (0.005) | 0.250 (0.004) | 0.235 (0.006) | n.s. | ||
June | 80 | 0.95 | 0.281 (0.021) | 0.288 (0.025) | 0.281 (0.021) | 0.280 (0.022) | ** |
0.96 | 0.310 (0.011) | 0.309 (0.013) | 0.310 (0.011) | 0.307 (0.012) | n.s. | ||
0.97 | 0.316 (0.011) | 0.311 (0.012) | 0.316 (0.011) | 0.311 (0.011) | n.s. | ||
0.98 | 0.308 (0.010) | 0.306 (0.010) | 0.308 (0.010) | 0.304 (0.010) | n.s. | ||
0.99 | 0.278 (0.007) | 0.274 (0.009) | 0.278 (0.007) | 0.270 (0.007) | n.s. | ||
1.00 | 0.238 (0.004) | 0.232 (0.005) | 0.238 (0.004) | 0.227 (0.005) | n.s. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kakimoto, Y.; Omae, Y.; Takahashi, H. Analysis of Sparse Trajectory Features Based on Mobile Device Location for User Group Classification Using Gaussian Mixture Model. Appl. Sci. 2025, 15, 982. https://doi.org/10.3390/app15020982
Kakimoto Y, Omae Y, Takahashi H. Analysis of Sparse Trajectory Features Based on Mobile Device Location for User Group Classification Using Gaussian Mixture Model. Applied Sciences. 2025; 15(2):982. https://doi.org/10.3390/app15020982
Chicago/Turabian StyleKakimoto, Yohei, Yuto Omae, and Hirotaka Takahashi. 2025. "Analysis of Sparse Trajectory Features Based on Mobile Device Location for User Group Classification Using Gaussian Mixture Model" Applied Sciences 15, no. 2: 982. https://doi.org/10.3390/app15020982
APA StyleKakimoto, Y., Omae, Y., & Takahashi, H. (2025). Analysis of Sparse Trajectory Features Based on Mobile Device Location for User Group Classification Using Gaussian Mixture Model. Applied Sciences, 15(2), 982. https://doi.org/10.3390/app15020982