Computers, Materials & Continua DOI:10.32604/cmc.2022.025112 | |
Article |
Self-Care Assessment for Daily Living Using Machine Learning Mechanism
1Department of Computer Science, Air University, Islamabad, 44000, Pakistan
2Department of Computer Science and Software Engineering, Al Ain University, Al Ain, 15551, UAE
3Department of Computer Science, College of Computer, Qassim University, Buraydah, 51452, Saudi Arabia
4Department of Humanities and Social Science, Al Ain University, Al Ain, 15551, UAE
5Department of Computer Engineering, Korea Polytechnic University, 237 Sangidaehak-ro Siheung-si, Gyeonggi-do, 15073, Korea
*Corresponding Author: Jeongmin Park. Email: jmpark@kpu.ac.kr
Received: 12 November 2021; Accepted: 11 January 2022
Abstract: Nowadays, activities of daily living (ADL) recognition system has been considered an important field of computer vision. Wearable and optical sensors are widely used to assess the daily living activities in healthy people and people with certain disorders. Although conventional ADL utilizes RGB optical sensors but an RGB-D camera with features of identifying depth (distance information) and visual cues has greatly enhanced the performance of activity recognition. In this paper, an RGB-D-based ADL recognition system has been presented. Initially, human silhouette has been extracted from the noisy background of RGB and depth images to track human movement in a scene. Based on these silhouettes, full body features and point based features have been extracted which are further optimized with probability based incremental learning (PBIL) algorithm. Finally, random forest classifier has been used to classify activities into different categories. The n-fold cross-validation scheme has been used to measure the viability of the proposed model on the RGBD-AC benchmark dataset and has achieved an accuracy of 92.71% over other state-of-the-art methodologies.
Keywords: Angular geometric features; decision tree classifier; human activity recognition; probability based incremental learning; ridge detection
In the world of artificial intelligence, Activities of Daily Living (ADL) has gained much of research interest for assisted living environments. The fundamental activities of daily life (ADL) are related to personal self-care and are defined as the basic activities of daily life. Doctor Deaver and Brown were the first to establish the concept of ADL in America [1]. The ADL is a medical instrument used to evaluate old persons, those with mental illnesses, and others. It is also gaining popularity in the measurement of human body motion [2]. Nowadays, videos with huge volumes have a lot of irrelevant content, therefore accurate recognition is required. Various sensors have been utilized in ADLs, namely, wearable sensors, optical sensors, and Radio Frequency Identification (RFID) but optical sensors-based systems have gained a lot of interest. Most of the previous work on human life-logging activity has been based on Red Green Blue (RGB) video and RFID sensors [3]. The RGB video systems achieved a very low accuracy of 78.5% even in the absence of clutter. On the other hand, the RFID method is generally too intrusive as it needs RFID tags on the people [4]. The accurate and effective recognition of ADLs plays an important role in determining a person's condition in realistic environment.
Different techniques have been developed for ADLs in recent years [5]. The space-time feature in-combination with bag of features is the state-of-the-art method for video representation [6]. The feature-based methods have usually been used for object recognition, which is further successfully tested for activity recognition [7]. The sparse-based action recognition system reduces the overall computational cost but results in performance degradation [8].
In this article, a unique machine learning framework is adopted to recognize the ADL system using the sequence of RGB and depth images. First, we extracted human RGB silhouette by applying background subtraction over the unwanted areas of images, and the silhouette has been extracted from the depth images by using the morphological operation model. Later on, system is designed that consists of ridge and angular-geometric features for feature extraction. Finally, combination of probability-based incremental learning (PBIL) and random forest classifier has been used for feature optimization and classification. The RGBD-AC publicly available benchmark dataset has been used for the validation of our results against state-of-the-art models. The proposed technique is based on five different aspects, silhouette extraction from the images, ridge detection, angular-geometric features modeling of a human silhouette, optimization, and classification. In addition, the proposed model has been applied over RGBD-AC dataset and achieved substantial results over other state-of-the-art methodologies.
The rest of this paper has been structured as: Section 2 comprehends related work, Section 3 covers the architectural representation of an anticipated model. Section 4 depicts the performance evaluation of the proposed work, Section 5 contains related discussions. Section 6 provides the conclusion and future directions on the proposed methodology.
For ages, RGB and depth images have been separately anticipated for ADL recognition systems. Lee et al. [9] has used RGB images for the classification of human actions. The OpenPose and 3D-baseline open source libraries has been used to extract skeleton joint features of skeleton. The CNN classifier has been finally applied on the extracted features. Their model has achieved an accuracy of 70% on NTU-RGBD dataset. In Mahmood et al. [10] WHITE STAG model is proposed for the RGB dataset. They have applied space-time and angular-geometric features over full-body silhouette and skeleton joint features. The performance of the model has been evaluated with a leave-one-out cross-validation scheme and achieved an overall accuracy of 87%. Varshney et al. [11] have utilized a skeleton tracking algorithm for the recognition of discrete skeleton positions on a real time RGB-D data streams. The proposed methodology has tested on 14 different activities that has been performed by two person in home environment and lecture hall. The main limitation in this model is that the proposed model has only tested on limited number of subject and has not been tested on any benchmark dataset. Si et al. [12] have extracted spatial and temporal features of RGB skeleton sequences images and CNN has applied to classify the discriminative features for human activity recognition. This model has tested over 3 benchmark datasets and has achieved an accuracy of 85% over SYSU-3D datasets. Kim et al. [13] have targeted human pose estimation problem using the improved architecture of hourglass network. The conventional hourglass network requires high computational power which is reduced by several skip connections. Hence the architecture of the network is improved with minimal modification. The proposed model has achieved an accuracy of 90.8%. Madhuranga et al. [14] employed CNN in-depth images that were previously used for RGB video for silhouette segmentation. They have developed the model for real-time recognition, which is capable of recognizing ADL on commodity computers. Khalid et al. [15] proposed K-ary Tree Hashing classifier to categorize NTU RGB-D sequential images. Buys et al. [16] extracted human skeleton from complex scenes using RGB-D sensors. The model was designed for color and depth image sequence structure in an environment. The results are further fed to RDF (Random Decision Forest) for cluster-based learning. Fu et al. [17] presented a Bayesian conditional probability problem over RGB-D based dataset. They have tested their model on twelve different activities performed by four different people and achieved good recognition accuracy.
The main drawback in RGB methodology is that depth information has been lost during 3D to the 2D project process. Moreover, the depth information depends only on the distance, which leads to difficulty in finding invariant features of objects. The RGB-D sensors can capture color and depth information simultaneously. The color and appearance information has been contained by RGB image whereas variations in color and illumination has been depicted from depth image. Thus, RGB-D images have gained popularity in the ADL recognition mechanism. In this paper, we proposed a novel mechanism for the RGB-D image-based ADL recognition system. In this work, RGB and depth image sequences has been adopted to bring robustness into our ADL model. For accurate pre-processing of data, separate methods have been taken into account for RGB and depth images. While, full-body features and angle-based geometric features bring more strength to the model. Moreover, PBIL and random forest classifier enhanced the classification of activities. The major contribution in this work are: (1) pre-processing of RGB and depth images effectively, (2) integration of full-body feature and angle based geometric feature for the feature extraction process, (3) using PBIL and decision classifier together for accurate classification of action recognition, and (4) validation of proposed model over RGBD-AC dataset.
The proposed methodology has taken RGB and depth images together as input to the system. In the pre-processing step, the human silhouette has been extracted from the raw images (RGB and depth) by applying background subtraction on RGB images, and morphological operation on depth images. Further, full-body features and angular-geometric features have been extracted from the silhouette. Ridge has been used to detect full-body features and the angular-geometric feature has been applied on the skeleton model to extract point-based features of human silhouette. Moreover, the probability-based incremental learning algorithm has been applied in order to symbolize the full-body features and angular-geometric features. Lastly, a decision tree classifier has been applied to the resultant optimized features. The overview of the model has been depicted in Fig. 1.
3.1 Preprocessing of Raw Images
First, all the raw images (RGB and depth) have been processed through the image normalization technique in order to enhance the quality of RGB-D images. Then, image contrast has been adjusted and a histogram equation has been applied to the image to uniformly distribute the intensity values through the entire image. The median filter has been applied to efficiently remove noise from the image [18]. It works by replacing the pixels with neighbouring pixels [19]. Finally, the region of interest (ROI) has been extracted from the image that have been segmented from their background, which has been described as below.
3.1.1 Background Subtraction for RGB Images
The silhouette of RGB image has been extracted through the background subtraction procedure [20]. The frame difference has been applied to the images from which current frame has been subtracted from the previous frames in each iteration [21]. The pixels of current frame
where
The threshold of the images has been selected by using Otsu's threshold technique [22]. In Otsu's method, the subtracted human silhouette frame has been first converted to a grayscale image, and then Otsu automatically selects the best value for threshold
3.1.2 Morphological Operations for Depth Images
The human silhouette from the depth images have been obtained through a morphological operation [24]. Firstly, binary images have been obtained from the depth images by applying threshold-based segmentation. The morphological operation has been applied to the segmented images by using the binary dilation process. The binary dilation process works by adding pixels to human edges and binary erosion has been applied on the resultant images that removes the boundary pixels [25]. The description for binary dilation and erosion has been portrayed in Eqs. (3)–(4) as follows:
where p is the location of pixels for element Z. The reflection of element Z is
The ridge detection of human silhouette comprises of two steps that are binary edge extraction and ridge data generation [26]. In the binary edge extraction step, the binary edges have been extracted from the RGB and depth silhouette obtained in the described pre-processing stage. The distance maps have been produced on the edges by using distance transform. While, in ridge data generation step, the local maximal has been obtained from the pre-computed maps, which produced ridge data within the binary edges [27]. A further description of binary edge detection and generation of ridge data is described below:
3.2.1 Edge Detection of Binary Image
The edge of the binary image has been calculated by computing the binary edge information of human silhouette. The binary edge information of RGB and depth silhouette is obtained using a window searching mechanism that computes statistical values of intensities from their nearest neighbor's pixel values [28]. As a result, correct edge connectivity of bounded body structure has been obtained using Eq. (5).
where
In this step, the distance maps of the binary edges have been calculated that have been utilized later to calculate the local maximal called ridge data [29]. The binary edges of the ridge data have been calculated using Eq. (6) as:
where
In skeleton point modeling, firstly, the contours of a human silhouette have been calculated that detect the outer pixels of human shape [30]. Secondly, torso has been traced by approximating the center point of calculated human contours [31], which is depicted in Eq. (7).
where
where
Finally, thirteen key points of e body has been detected (see Algorithm 1) to get the optimal location of body points.
The skeletonization process detects thirteen body key points in three main body segments including lower body (LB), upper body (UB), and mid body (MB). UB connects the shoulders
The angular-geometric feature is a body point feature that extracts shape-based entity features of human skeleton point modeling [32]. The extreme points identified on the human silhouette has been extracted including head, feet, arms, and shoulders. Then, extreme points have been joined together to form three geometric shapes of triangle, quadrilateral, and pentagon. Moreover, changes in the angles have been measured at each extreme point within sequential frames [33]. The geometric shapes have been prepared made by joining the extreme points on the body including inter-silhouette triangle, quadrangular, and pentagon, as shown in Tab. 1.
The extreme body points have been joined together to form inter-silhouette geometric shapes. Later on, cosine angle within the shapes have been measured as shown in Eq. (13):
where u and v are the vectors that have been used to measure the angle. Next, the area of a triangle is measured using Eq. (14):
where a, b, and c are the three sides of a triangle i.e., extreme points, and S depicts semi-perimeter of the triangle.
The silhouette movement may result in the increase or decrease of each geometric shape. Hence, geometric features and angles also variate between sequential frames. The differentiation in angle is higher in a few moments namely, open, pull, pick as they involved rapid movement, than others movements, namely drink, plug, and switch because include less pronounced movement.
Population Based Incremental Learning (PBIL) is a stochastic search optimization technique that converges to the optimal solution based on the learning of current excellent individuals. PBIL algorithm terminates automatically when three of its parameters including learning rate, population size, and search rate converge on a single optimal solution [34]. First, the feature vectors obtained in the Sections 3.2 and 3.4 have been coded with a binary chromosome of population length to form a single chromosome.
The prototype vector P has been generated to bias the generation of chromosomes population [35]. The P with a length equal to the population array has been initialized to 0.5 for each corresponding location to bias the generation of bits. For each chromosome, the bits have been randomly set in the range [0, 1]. The chromosome has been set to either one in case the corresponding random number in the vector P is less than a certain threshold or zero. Next, chromosomes have been evaluated with Eqs. (15)–(16) to incorporate the best chromosome.
where CBshows best chromosomes, l represents learning rate, and s represents search rate. The algorithm runs iteratively until all chromosomes converge to optimal solution with a convergence level of s or 1-s, see Eq. (17).
Fig. 6 shows the optimization result of probability based incremental learning (PBIL).
3.6 Classification Using Random Forest Classifier
The results of the optimized features have been then fed to the random forest classifier in order to classify features into different activities. Random forest is an ensemble learning algorithm for data classification [36].
Initially, a random forest algorithm has been formed ample number of impartial decision trees to form a forest. The random forest has been then supervised learning impact by assigning the score to each feature. The decision trees have been then selected for classification that contained abundant number of votes among other decision trees in the whole forest [37]. Eq. (18) illustrates mathematical expression for random forest classifier.
where x represents predictions of unseen samples. To outfit high dimensional data, the bootstrapping aggregation methodology has been used in the random forest classifier. Hence, the best possible classification tree has been obtained by combining multiple trees together and then fit the training data into regression trees. Finally, all the samples have been employed with replacement to achieve consistency and maximum possible accuracy. Fig. 7 shows visualization clusters of a random forest classifier.
4 Experimental Setting and Results
In the experimental section, the n-fold cross-validation scheme has been used to evaluate the performance of proposed model over RGBD-AC benchmark datasets. The description of the dataset is also given in this section. Furthermore, the proposed model has been assessed with recognition accuracy, number of precision, observations, recall, computation time, F1 score, and number of states. Finally, the proposed model has been compared against other state-of-the-art methods.
4.1 RGBD-AC Datasets Description
The RGBD-AC dataset [38] embodies RGB and depth images data, taken from optical measuring devices like Microsoft Kinetic v2. This RGBD-AC dataset is divided into 414 sequences of complete and incomplete six different actions. The RGB folders contain RGB sequences of images related to six different actions i.e., RGB-drink, RGB-open, RGB-pick, RGB-plug, RGB-pull, and RGB-switch. Similarly, each depth folders contain depth sequences of images related to six different actions including depth-drink, depth-open, depth-pick, depth-plug, depth-pull, and depth-switch. Each action has been cataloged with 383 images. In this dataset, action has been completed if the goal was achieved i.e., the plug action is completed when a person plugs the switch. In the proposed model, the main focus is on the accuracy for action detection rather than on the action completeness.
4.2 Performance Parameters and Evaluations
The proposed ADL methodology has been validated with the numerous parameters i.e., recognition accuracy, number of precision, observations, recall, computation time, F1-score, and number of states. Details of parameters for individual experiments have been deliberated as:
In the first experiment, the n-fold cross-validation method has been used to evaluate the benchmark dataset for the average accuracy of the proposed model. Tab. 2 shows the confusion matrix for dataset and the mean accuracy of dataset is 92.71%.
By means of the proposed effective model, the pick activity has been achieved an accuracy of 95.9%. While some vectors of the plug and switch have been interleaved with each other. Some percentage of drink, open, and pull are also interleaved. It is difficult for the system at some points to differentiate between drink/open and plug/pull due to the repetition of similar movements involved in the activities. However, the overall accuracy of different actions proved the robustness of the proposed model by achieving higher mean accuracy of 92.71%.
The proposed system has been evaluated with an n-fold cross-validation scheme and dividing the dataset into training and testing sets. The results of the PBIL optimizer and random forest classifier have been validated using precision, recall, and F1-score to identify ADL activities. Moreover, the evaluation of RGBD-AC over three parameters have been shown in Tab. 3.
It is observed from Tab. 3 that the drink and switch activities have the least precision rate due to the maximum percentage of false positive. This is due to the body movement of silhouette performed during drink and switch activities involving many actions quite similar to interactions with each other.
In the third experiment, various experiments has been performed for the dataset. In this experiment, different sequences of observations has been fused with different states to estimate the accuracy of proposed system based on recognition accuracy and computational time. The performance of the model has effected with the number of states as the current state influenced the performance of previous states. Moreover, the greater the number of observations and states, the greater will be accuracy rate as shown in Tab. 4.
From Tab. 4, starting with four states and by varying the observations from 10 to 30 has gradually improved the accuracy rate. This accuracy has been further enhanced in states 4 and 5. Finally, in state 6 the computational time will remained static with no major change in accuracy.
In fourth experiment, the proposed model has been compared with the other proposed scheme. The Tabs. 5 and 6 shows that the proposed model has achieved a significance performance of 92.7% than other proposed methods.
Practical Discussion:
The experiments have been performed on a Desktop PC equipped with Intel(R) Process(R) CPU G3260, 8 GB RAM, 3.30 GHz processing power, 64-bit operating system, Windows 10, and using Anaconda Spyder Python IDE. Our proposed framework has been based on RGB and depth images which help to attain the information of color, appearance, and illumination information simultaneously. The fused information of RGB and depth has helped to gain the high accuracy and F1-Score. Initially, different preprocessing methodologies has been used to do filtration on RGB-D images to eliminate the noise from raw images of RGB and depth images. Furthermore, full body features and point based features has helped the proposed system in obtaining better performance. Next, attained features have been optimized through Probability based incremental learning. Finally, random forest classifier has been applied to evaluate the performance of the proposed system. The leave-one-subject-out cross-validation (LOSO) method has been used to evaluate the performance of our proposed model on RGB-D benchmark dataset. Moreover, the proposed AAL has been authorized over precision, recall, and F-measure parameters to assess the validity of our model. Finally, a comparison with other state-of-the-art methods has been done to prove that our model out-performed in its best accuracy rate. The results have been showed in Fig. 8.
The following are the limitations of our systems.
1. The proposed methodology has been evaluated only on one benchmark dataset which is based on indoor activities. The different datasets could be considered based on diverse patterns of activities and with both indoor and outdoor scenarios for the better performance of our model.
2. The proposed methodology is comprised of full-body and point based features in combination with optimization and classification required high computational time to process the whole results but has achieved the high accuracy of 92.71%.
3. The main problem in PBIL algorithm is the early convergence to the local optimal solution which made it difficult to obtained the excellent results in drink and switch activity.
Theoretical Discussion:
From the obtained results in Tab. 2, the drink activity has obtained an accuracy of 90.8, while rest of the results are confused with pick, plug, pull, and switch activities. Similarly, open, pick, plug, pull, and switch activities have obtained an accuracy of 92.06%, 95.9%, 93.5%, 92.3%, 91.7%, and rest of their results are confused with other activities. The drink activity has obtained best accuracy of 90.8% and pick has obtained the high accuracy of 95.9%. The open and pull activity has obtained almost same accuracy of 92%. While, switch activity has obtained the average accuracy of 91.7%. It can be concluded from the experimental results that combination of features along with optimization and classification delivers better results for the proposed ADL system.
In this work, we have proposed a novel RGB-D based ADL recognition system. The main accomplishments of this research effort include identification of human silhouette, integration of full-body feature with a geometric feature, PBIL optimization, and recognition of activities with random forest classifier. The full body features and point based features analyses the optimal pattern, temporal patterns, repetitive pattern, and invariability in motion and thus help to improve the performance of proposed ADL recognition system. This paper has also done comparison against RGB and depth based research work. During the experiment, RGBD-AC benchmark dataset has been used to evaluate the performance of our proposed model. The proposed system achieved an accuracy rate of 92.7% over the other state-of-the-art methods. The proposed system can be applied to various real-life activities including health care, video indexing, and security monitoring.
In the future, we plan to use point cloud features extraction methodology for the ADL recognition model on the RGBD-AC dataset. We will also use static scenes of RGB-D (2.5D) and vision-based 3D visual data of indoor scenes for the enhancement of ADL activities recognition. We will also test our current system on more ADL-based challenging datasets.
Funding Statement: This research was supported by a grant (2021R1F1A1063634) of the Basic Science Research Program through the National Research Foundation (NRF) funded by the Ministry of Education, Republic of Korea.
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.
1. L. M. Powell, “Aging and performance of home tasks,” Human Factors, vol. 32, no. 5, pp. 527–536, 1990. [Google Scholar]
2. D. J. Magermans, E. K. J. Chadwick, H. E. J. Veeger, “Requirements for upper extremity motions during activities of daily living,” Clinical Biomechanics, vol. 20, no. 6, pp. 591–599, 2005. [Google Scholar]
3. A. Prati, C. Shan and K. I. -K. Wang, “Sensors, vision and networks: From video surveillance to activity recognition and health monitoring,” Journal of Ambient Intelligence and Smart Environments, vol. 11, no. 1, pp. 5–22, 2019. [Google Scholar]
4. A. Jalal, S. Kamal and D. Kim, “Facial expression recognition using 1D transform features and Hidden Markov Model,” Journal of Electrical Engineering & Technology, vol. 12, no. 4, pp. 1657–1662, 2017. [Google Scholar]
5. A. Jalal, S. Kamal, and D. Kim, “A depth video-based human detection and activity recognition using multi-features and embedded hidden Markov models for health care monitoring systems,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 4, no. 4, pp. 54–62, 2017. [Google Scholar]
6. F. Farooq, A. Jalal, and L. Zheng, “Facial expression recognition using hybrid features and self-organizing maps,” in Proc. IEEE Int. Conf. on Multimedia and Expo, Hong Kong, China, pp. 409–414, 2017. [Google Scholar]
7. A. Jalal, S. Kamal and D. S. Kim, “Detecting complex 3D human motions with body model low-rank representation for real-time smart activity monitoring system,” KSII Transactions on Internet and Information Systems, vol. 12, no. 3, pp. 1189–1204, 2018. [Google Scholar]
8. M. Mahmood, A. Jalal and H. A. Evans, “Facial expression recognition in image sequences using 1D transform and Gabor wavelet transform,” in IEEE Conf. on Int. Conf. on Applied and Engineering Mathematics, Taxila, Pakistan, pp. 1–6, 2018. [Google Scholar]
9. J. Lee and B. Ahn, “Real-time human action recognition with a low-cost RGB camera and mobile robot platform,” Sensors, vol. 20, no. 10, pp. 1–12, 2020. [Google Scholar]
10. M. Mahmood, A. Jalal and K. Kim, “WHITE STAG model: Wise human interaction tracking and estimation (WHITE) using spatio-temporal and angular-geometric (STAG) descriptors,” Multimed. Tools Appl., vol. 79, pp. 6919–6950, 2020. [Google Scholar]
11. N. Varshney, B. Bakariya and A. K. S. Kushwaha et al., “Rule-based multi-view human activity recognition system in real time using skeleton data from RGB-D senso,” Soft Comput., vol. 210, pp. 1-–17, 2021. [Google Scholar]
12. C. Si, Y. Jing, W. Wang, L. Wang, T. Tan, “Skeleton-based action recognition with hierarchical spatial reasoning and temporal stack learning network,” Pattern Recognit., vol. 107, pp. 107511--107532, 2020. [Google Scholar]
13. S. -T. Kim, H. J. Lee, “Lightweight stacked hourglass network for human pose estimation,” Appl. Sci., vol. 10, no. 6497, pp. 1--15, 2020. [Google Scholar]
14. D. Madhuranga, R. Madushan and C. Siriwardane, “Real-time multimodal ADL recognition using convolution neural networks,” Vis Comput., vol. 37, pp. 1263–1276, 2021. [Google Scholar]
15. N. Khalid, Y. Y. Ghadi, M. Gochoo, A. Jalal and K. Kim, “Semantic recognition of human-object interactions via Gaussian-based elliptical modeling and pixel-level labeling,” IEEE Access, vol. 9, pp. 111249–111266, 2021. [Google Scholar]
16. K. Buys, C. Cagniart, A. Baksheev, T. D. Laet, J. D. Schutter et al., “An adaptable system for RGB-D based human body detection and pose estimation,” Journal of Visual Communication and Image Representation, vol. 25, no. 1, pp. 39–52, 2014. [Google Scholar]
17. J. Fu, C. Liu, Y. Hsu and L. Fu, “Recognizing context-aware activities of daily living using RGBD sensor,” in 2013 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, Tokyo, Japan, pp. 2222–2227, 2013. [Google Scholar]
18. A. Jalal and S. Kamal, “Improved behavior monitoring and classification using cues parameters extraction from camera array images,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 5, no. 2, pp. 71–78, 2018. [Google Scholar]
19. K. Alzahrani and M. Alnfiai, “Evaluation of NFC-guidable system to manage polypharmacy in elderly patients,” Computer Systems Science and Engineering, vol. 41, no. 2, pp. 445–460, 2022. [Google Scholar]
20. S. Abbas, Y. Alhwaiti, A. Fatima, M. A. Khan, T. M. Ghazal et al., “Convolutional neural network based intelligent handwritten document recognition,” Computers, Materials & Continua, vol. 70, no. 3, pp. 4563–4581, 2022. [Google Scholar]
21. Y. Guo, Z. Cui, X. Li, J. Peng, J. Hu et al., “MRI image segmentation of nasopharyngeal carcinoma using multi-scale cascaded fully convolutional network,” Intelligent Automation & Soft Computing, vol. 31, no. 3, pp. 1771–1782, 2022. [Google Scholar]
22. A. Jalal, M. A. K. Quaid and M. A. Sidduqi, “A triaxial acceleration-based human motion detection for ambient smart home system,” in IEEE Int. Conf. on Applied Sciences and Technology, Islamabad, Pakistan, pp. 353–358, 2019. [Google Scholar]
23. A. Jalal, M. Mahmood and A. S. Hasan, “Multi-features descriptors for human activity tracking and recognition in indoor-outdoor environments,” in IEEE Int. Conf. on Applied Sciences and Technology, Islamabad, Pakistan, pp. 371–376, 2019. [Google Scholar]
24. Q. Ye, H. Zhong, C. Qu and Y. Zhang, “Human interaction recognition based on whole-individual detection,” Sensors, vol. 20, no. 8, pp. 1–18, 2020. [Google Scholar]
25. O. Ouyed and M. A. Said, “Group-of-features relevance in multinomial kernel logistic regression and application to human interaction recognition,” Expert Systems with Applications, vol. 148, pp. 1–22, 2020. [Google Scholar]
26. A. A. Rafique, A. Jalal and A. Ahmed, “Scene understanding and recognition: Statistical segmented model using geometrical features and Gaussian naïve Bayes,” in IEEE Conf. on Int. Conf. on Applied and Engineering Mathematics, pp. 225–230, 2019. [Google Scholar]
27. M. Batool, A. Jalal and K. Kim, “Sensors technologies for human activity analysis based on SVM optimized by PSO algorithm,” IEEE ICAEM Conf., Taxila, Pakistan, pp. 145–150, 2019. [Google Scholar]
28. A. Shehzad, A. Jalal and K. Kim, “Multi-person tracking in smart surveillance system for crowd counting and normal/abnormal events detection,” in IEEE Conf. on Int. Conf. on Applied and Engineering Mathematics, Taxila, Pakistan, pp. 163–168, 2019. [Google Scholar]
29. A. Ahmed, A. Jalal and A. A. Rafique, “Salient segmentation based object detection and recognition using hybrid genetic transform,” IEEE ICAEM Conf., Taxila, Pakistan, pp. 203–208, 2019. [Google Scholar]
30. S. Bibi, N. Anjum and M. Sher, “Automated multi-feature human interaction recognition in complex environment,” Computers in Industry Elsevier, vol. 99, pp. 282–293, 2018. [Google Scholar]
31. Y. Ji, H. Cheng, Y. Zheng and H. Li, “Learning contrastive feature distribution model for interaction recognition,” Journal of Visual Communication and Image Representation, vol. 33, pp. 340–349, 2015. [Google Scholar]
32. J. Peng, C. Xia, Y. Xu, X. Li, X. Wu et al., “A multi-task network for cardiac magnetic resonance image segmentation and classification,” Intelligent Automation & Soft Computing, vol. 30, no. 1, pp. 259–272, 2021. [Google Scholar]
33. M. u. Rehman, S. H. Khan, S. M. Danish Rizvi, Z. Abbas, A. Zafar et al., “Classification of skin lesion by interference of segmentation and convolotion neural network,” in 2nd Int. Conf. on Engineering Innovation (ICEI), Bangkok, Thailand, pp. 81–85, 2018. [Google Scholar]
34. A. Ahmed, A. Jalal and K. Kim, “RGB-D images for object segmentation, localization and recognition in indoor scenes using feature descriptor and Hough voting,” in IEEE Conf. on Applied Sciences and Technology, Islamabad, Pakistan, pp. 290–295, 2020. [Google Scholar]
35. M. Mahmood, A. Jalal and K. Kim, “WHITE STAG model: Wise human interaction tracking and estimation (WHITE) using spatio-temporal and angular-geometric (STAG) descriptors,” Multimedia Tools and Applications, vol. 79, pp. 6919–6950, 2020. [Google Scholar]
36. M. A. K. Quaid and A. Jalal, “Wearable sensors based human behavioral pattern recognition using statistical features and reweighted genetic algorithm,” Multimedia Tools and Applications, vol. 79, pp. 6061–6083, 2019. [Google Scholar]
37. A. Nadeem, A. Jalal and K. Kim, “Human actions tracking and recognition based on body parts detection via artificial neural network,” in IEEE Int. Conf. on Advancements in Computational Sciences, Lahore, Pakistan, pp. 1–6, 2020. [Google Scholar]
38. F. Heidarivincheh, M. Mirmehdi and D. Damen, “Beyond action recognition: Action completion in RGB-D data,” in 27th British Machine Vision Conf., York, United Kingdom, 2016. [Google Scholar]
39. A. Miron and C. Grosan, “Classifying action correctness in physical rehabilitation exercises,” in IEEE Conf. on Int. Conf. on Applied and Engineering Mathematics, Stockholm, Sweden, 2021. [Google Scholar]
40. F. Heidarivincheh, M. Mirmehdi and D. Damen, “Detecting the moment of completion: Temporal models for localising action completion,” in 28th British Machine Vision Conf., London, UK, 2017. [Google Scholar]
41. F. Heidarivincheh, M. Mirmehdi and D. Damen, “Weakly-supervised completion moment detection using temporal attention,” in 30th British Machine Vision Conf., Seoul, South Korea, 2019. [Google Scholar]
42. F. Heidarivincheh, M. Mirmehdi and D. Damen, “Action completion: A temporal model for moment detection,” in 29th British Machine Vision Conf.- Northumbria University, Newcastle upon Tyne, United Kingdom, 2018. [Google Scholar]
This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |