1
|
Fatima R, Khan MH, Nisar MA, Doniec R, Farid MS, Grzegorzek M. A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data. Sensors (Basel) 2023; 24:75. [PMID: 38202937 PMCID: PMC10780594 DOI: 10.3390/s24010075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/15/2023] [Revised: 12/09/2023] [Accepted: 12/19/2023] [Indexed: 01/12/2024]
Abstract
This paper addresses the problem of feature encoding for gait analysis using multimodal time series sensory data. In recent years, the dramatic increase in the use of numerous sensors, e.g., inertial measurement unit (IMU), in our daily wearable devices has gained the interest of the research community to collect kinematic and kinetic data to analyze the gait. The most crucial step for gait analysis is to find the set of appropriate features from continuous time series data to accurately represent human locomotion. This paper presents a systematic assessment of numerous feature extraction techniques. In particular, three different feature encoding techniques are presented to encode multimodal time series sensory data. In the first technique, we utilized eighteen different handcrafted features which are extracted directly from the raw sensory data. The second technique follows the Bag-of-Visual-Words model; the raw sensory data are encoded using a pre-computed codebook and a locality-constrained linear encoding (LLC)-based feature encoding technique. We evaluated two different machine learning algorithms to assess the effectiveness of the proposed features in the encoding of raw sensory data. In the third feature encoding technique, we proposed two end-to-end deep learning models to automatically extract the features from raw sensory data. A thorough experimental evaluation is conducted on four large sensory datasets and their outcomes are compared. A comparison of the recognition results with current state-of-the-art methods demonstrates the computational efficiency and high efficacy of the proposed feature encoding method. The robustness of the proposed feature encoding technique is also evaluated to recognize human daily activities. Additionally, this paper also presents a new dataset consisting of the gait patterns of 42 individuals, gathered using IMU sensors.
Collapse
Affiliation(s)
- Rimsha Fatima
- Department of Computer Science, University of the Punjab, Lahore 54590, Pakistan (M.S.F.)
| | - Muhammad Hassan Khan
- Department of Computer Science, University of the Punjab, Lahore 54590, Pakistan (M.S.F.)
| | - Muhammad Adeel Nisar
- Department of Information Technology, University of the Punjab, Lahore 54000, Pakistan;
| | - Rafał Doniec
- Faculty of Biomedical Engineering, The Silesian University of Technology, 44-100 Gliwice, Poland;
| | - Muhammad Shahid Farid
- Department of Computer Science, University of the Punjab, Lahore 54590, Pakistan (M.S.F.)
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, 23562 Lübeck, Germany
| |
Collapse
|
2
|
Amjad F, Khan MH, Nisar MA, Farid MS, Grzegorzek M. A Comparative Study of Feature Selection Approaches for Human Activity Recognition Using Multimodal Sensory Data. Sensors (Basel) 2021; 21:s21072368. [PMID: 33805368 PMCID: PMC8036571 DOI: 10.3390/s21072368] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/04/2021] [Revised: 03/16/2021] [Accepted: 03/26/2021] [Indexed: 12/16/2022]
Abstract
Human activity recognition (HAR) aims to recognize the actions of the human body through a series of observations and environmental conditions. The analysis of human activities has drawn the attention of the research community in the last two decades due to its widespread applications, diverse nature of activities, and recording infrastructure. Lately, one of the most challenging applications in this framework is to recognize the human body actions using unobtrusive wearable motion sensors. Since the human activities of daily life (e.g., cooking, eating) comprises several repetitive and circumstantial short sequences of actions (e.g., moving arm), it is quite difficult to directly use the sensory data for recognition because the multiple sequences of the same activity data may have large diversity. However, a similarity can be observed in the temporal occurrence of the atomic actions. Therefore, this paper presents a two-level hierarchical method to recognize human activities using a set of wearable sensors. In the first step, the atomic activities are detected from the original sensory data, and their recognition scores are obtained. Secondly, the composite activities are recognized using the scores of atomic actions. We propose two different methods of feature extraction from atomic scores to recognize the composite activities, and they include handcrafted features and the features obtained using the subspace pooling technique. The proposed method is evaluated on the large publicly available CogAge dataset, which contains the instances of both atomic and composite activities. The data is recorded using three unobtrusive wearable devices: smartphone, smartwatch, and smart glasses. We also investigated the performance evaluation of different classification algorithms to recognize the composite activities. The proposed method achieved 79% and 62.8% average recognition accuracies using the handcrafted features and the features obtained using subspace pooling technique, respectively. The recognition results of the proposed technique and their comparison with the existing state-of-the-art techniques confirm its effectiveness.
Collapse
Affiliation(s)
- Fatima Amjad
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (F.A.); (M.A.N.); (M.S.F.)
| | - Muhammad Hassan Khan
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (F.A.); (M.A.N.); (M.S.F.)
- Correspondence:
| | - Muhammad Adeel Nisar
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (F.A.); (M.A.N.); (M.S.F.)
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23538 Lübeck, Germany;
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (F.A.); (M.A.N.); (M.S.F.)
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23538 Lübeck, Germany;
| |
Collapse
|
3
|
Ilyas A, Farid MS, Khan MH, Grzegorzek M. Exploiting Superpixels for Multi-Focus Image Fusion. Entropy (Basel) 2021; 23:247. [PMID: 33670018 PMCID: PMC7926613 DOI: 10.3390/e23020247] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/29/2021] [Revised: 02/15/2021] [Accepted: 02/17/2021] [Indexed: 12/03/2022]
Abstract
Multi-focus image fusion is the process of combining focused regions of two or more images to obtain a single all-in-focus image. It is an important research area because a fused image is of high quality and contains more details than the source images. This makes it useful for numerous applications in image enhancement, remote sensing, object recognition, medical imaging, etc. This paper presents a novel multi-focus image fusion algorithm that proposes to group the local connected pixels with similar colors and patterns, usually referred to as superpixels, and use them to separate the focused and de-focused regions of an image. We note that these superpixels are more expressive than individual pixels, and they carry more distinctive statistical properties when compared with other superpixels. The statistical properties of superpixels are analyzed to categorize the pixels as focused or de-focused and to estimate a focus map. A spatial consistency constraint is ensured on the initial focus map to obtain a refined map, which is used in the fusion rule to obtain a single all-in-focus image. Qualitative and quantitative evaluations are performed to assess the performance of the proposed method on a benchmark multi-focus image fusion dataset. The results show that our method produces better quality fused images than existing image fusion techniques.
Collapse
Affiliation(s)
- Areeba Ilyas
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Muhammad Hassan Khan
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23538 Lübeck, Germany;
| |
Collapse
|
4
|
Zafar R, Farid MS, Khan MH. Multi-Focus Image Fusion: Algorithms, Evaluation, and a Library. J Imaging 2020; 6:60. [PMID: 34460653 PMCID: PMC8321074 DOI: 10.3390/jimaging6070060] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2020] [Revised: 06/21/2020] [Accepted: 06/24/2020] [Indexed: 11/17/2022] Open
Abstract
Image fusion is a process that integrates similar types of images collected from heterogeneous sources into one image in which the information is more definite and certain. Hence, the resultant image is anticipated as more explanatory and enlightening both for human and machine perception. Different image combination methods have been presented to consolidate significant data from a collection of images into one image. As a result of its applications and advantages in variety of fields such as remote sensing, surveillance, and medical imaging, it is significant to comprehend image fusion algorithms and have a comparative study on them. This paper presents a review of the present state-of-the-art and well-known image fusion techniques. The performance of each algorithm is assessed qualitatively and quantitatively on two benchmark multi-focus image datasets. We also produce a multi-focus image fusion dataset by collecting the widely used test images in different studies. The quantitative evaluation of fusion results is performed using a set of image fusion quality assessment metrics. The performance is also evaluated using different statistical measures. Another contribution of this paper is the proposal of a multi-focus image fusion library, to the best of our knowledge, no such library exists so far. The library provides implementation of numerous state-of-the-art image fusion algorithms and is made available publicly at project website.
Collapse
Affiliation(s)
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (R.Z.); (M.H.K.)
| | | |
Collapse
|
5
|
Khan MH, Zöller M, Farid MS, Grzegorzek M. Marker-Based Movement Analysis of Human Body Parts in Therapeutic Procedure. Sensors (Basel) 2020; 20:E3312. [PMID: 32532113 PMCID: PMC7313697 DOI: 10.3390/s20113312] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/11/2020] [Revised: 06/05/2020] [Accepted: 06/08/2020] [Indexed: 12/20/2022]
Abstract
Movement analysis of human body parts is momentous in several applications including clinical diagnosis and rehabilitation programs. The objective of this research is to present a low-cost 3D visual tracking system to analyze the movement of various body parts during therapeutic procedures. Specifically, a marker based motion tracking system is proposed in this paper to capture the movement information in home-based rehabilitation. Different color markers are attached to the desired joints' locations and they are detected and tracked in the video to encode their motion information. The availability of this motion information of different body parts during the therapy can be exploited to achieve more accurate results with better clinical insight, which in turn can help improve the therapeutic decision making. The proposed framework is an automated and inexpensive motion tracking system with execution speed close to real time. The performance of the proposed method is evaluated on a dataset of 10 patients using two challenging matrices that measure the average accuracy by estimating the joints' locations and rotations. The experimental evaluation and its comparison with the existing state-of-the-art techniques reveals the efficiency of the proposed method.
Collapse
Affiliation(s)
- Muhammad Hassan Khan
- Research Group for Pattern Recognition, University of Siegen, Hölderlinstr 3, 57076 Siegen, Germany;
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan;
| | - Martin Zöller
- Research Group for Pattern Recognition, University of Siegen, Hölderlinstr 3, 57076 Siegen, Germany;
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan;
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23538 Lübeck, Germany;
| |
Collapse
|
6
|
Abstract
Malaria is caused by Plasmodium parasite. It is transmitted by female Anopheles bite. Thick and thin blood smears of the patient are manually examined by an expert pathologist with the help of a microscope to diagnose the disease. Such expert pathologists may not be available in many parts of the world due to poor health facilities. Moreover, manual inspection requires full concentration of the pathologist and it is a tedious and time consuming way to detect the malaria. Therefore, development of automated systems is momentous for a quick and reliable detection of malaria. It can reduce the false negative rate and it can help in detecting the disease at early stages where it can be cured effectively. In this paper, we present a computer aided design to automatically detect malarial parasite from microscopic blood images. The proposed method uses bilateral filtering to remove the noise and enhance the image quality. Adaptive thresholding and morphological image processing algorithms are used to detect the malaria parasites inside individual cell. To measure the efficiency of the proposed algorithm, we have tested our method on a NIH Malaria dataset and also compared the results with existing similar methods. Our method achieved the detection accuracy of more than 91% outperforming the competing methods. The results show that the proposed algorithm is reliable and can be of great assistance to the pathologists and hematologists for accurate malaria parasite detection.
Collapse
Affiliation(s)
- Tehreem Fatima
- Punjab University College of Information Technology, University of the Punjab, Lahore, Pakistan
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore, Pakistan
| |
Collapse
|
7
|
Khehrah N, Farid MS, Bilal S, Khan MH. Lung Nodule Detection in CT Images Using Statistical and Shape-Based Features. J Imaging 2020; 6:jimaging6020006. [PMID: 34460555 PMCID: PMC8321000 DOI: 10.3390/jimaging6020006] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2020] [Revised: 02/20/2020] [Accepted: 02/21/2020] [Indexed: 12/20/2022] Open
Abstract
The lung tumor is among the most detrimental kinds of malignancy. It has a high occurrence rate and a high death rate, as it is frequently diagnosed at the later stages. Computed Tomography (CT) scans are broadly used to distinguish the disease; computer aided systems are being created to analyze the ailment at prior stages productively. In this paper, we present a fully automatic framework for nodule detection from CT images of lungs. A histogram of the grayscale CT image is computed to automatically isolate the lung locale from the foundation. The results are refined using morphological operators. The internal structures are then extracted from the parenchyma. A threshold-based technique is proposed to separate the candidate nodules from other structures, e.g., bronchioles and blood vessels. Different statistical and shape-based features are extracted for these nodule candidates to form nodule feature vectors which are classified using support vector machines. The proposed method is evaluated on a large lungs CT dataset collected from the Lung Image Database Consortium (LIDC). The proposed method achieved excellent results compared to similar existing methods; it achieves a sensitivity rate of 93.75%, which demonstrates its effectiveness.
Collapse
Affiliation(s)
- Noor Khehrah
- Punjab University College of Information Technology, University of the Punjab, Lahore-54000, Pakistan; (N.K.)
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore-54000, Pakistan; (N.K.)
- Correspondence:
| | - Saira Bilal
- Department of Radiology, General Hospital, Lahore-54000, Pakistan
| | - Muhammad Hassan Khan
- Punjab University College of Information Technology, University of the Punjab, Lahore-54000, Pakistan; (N.K.)
| |
Collapse
|
8
|
Khan MH, Schneider M, Farid MS, Grzegorzek M. Detection of Infantile Movement Disorders in Video Data Using Deformable Part-Based Model. Sensors (Basel) 2018; 18:E3202. [PMID: 30248968 PMCID: PMC6210538 DOI: 10.3390/s18103202] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/03/2018] [Revised: 09/17/2018] [Accepted: 09/20/2018] [Indexed: 12/20/2022]
Abstract
Movement analysis of infants' body parts is momentous for the early detection of various movement disorders such as cerebral palsy. Most existing techniques are either marker-based or use wearable sensors to analyze the movement disorders. Such techniques work well for adults, however they are not effective for infants as wearing such sensors or markers may cause discomfort to them, affecting their natural movements. This paper presents a method to help the clinicians for the early detection of movement disorders in infants. The proposed method is marker-less and does not use any wearable sensors which makes it ideal for the analysis of body parts movement in infants. The algorithm is based on the deformable part-based model to detect the body parts and track them in the subsequent frames of the video to encode the motion information. The proposed algorithm learns a model using a set of part filters and spatial relations between the body parts. In particular, it forms a mixture of part-filters for each body part to determine its orientation which is used to detect the parts and analyze their movements by tracking them in the temporal direction. The model is represented using a tree-structured graph and the learning process is carried out using the structured support vector machine. The proposed framework will assist the clinicians and the general practitioners in the early detection of infantile movement disorders. The performance evaluation of the proposed method is carried out on a large dataset and the results compared with the existing techniques demonstrate its effectiveness.
Collapse
Affiliation(s)
- Muhammad Hassan Khan
- Research Group for Pattern Recognition, University of Siegen, 57076 Siegen, Germany.
| | - Manuel Schneider
- Research Group for Pattern Recognition, University of Siegen, 57076 Siegen, Germany.
| | - Muhammad Shahid Farid
- College of Information Technology, University of the Punjab, 54000 Lahore, Pakistan.
| | - Marcin Grzegorzek
- Research Group for Pattern Recognition, University of Siegen, 57076 Siegen, Germany.
| |
Collapse
|
9
|
Khan MH, Helsper J, Farid MS, Grzegorzek M. A computer vision-based system for monitoring Vojta therapy. Int J Med Inform 2018; 113:85-95. [PMID: 29602437 DOI: 10.1016/j.ijmedinf.2018.02.010] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2017] [Revised: 01/23/2018] [Accepted: 02/15/2018] [Indexed: 11/17/2022]
Abstract
A neurological illness is t he disorder in human nervous system that can result in various diseases including the motor disabilities. Neurological disorders may affect the motor neurons, which are associated with skeletal muscles and control the body movement. Consequently, they introduce some diseases in the human e.g. cerebral palsy, spinal scoliosis, peripheral paralysis of arms/legs, hip joint dysplasia and various myopathies. Vojta therapy is considered a useful technique to treat the motor disabilities. In Vojta therapy, a specific stimulation is given to the patient's body to perform certain reflexive pattern movements which the patient is unable to perform in a normal manner. The repetition of stimulation ultimately brings forth the previously blocked connections between the spinal cord and the brain. After few therapy sessions, the patient can perform these movements without external stimulation. In this paper, we propose a computer vision-based system to monitor the correct movements of the patient during the therapy treatment using the RGBD data. The proposed framework works in three steps. In the first step, patient's body is automatically detected and segmented and two novel techniques are proposed for this purpose. In the second step, a multi-dimensional feature vector is computed to define various movements of patient's body during the therapy. In the final step, a multi-class support vector machine is used to classify these movements. The experimental evaluation carried out on the large captured dataset shows that the proposed system is highly useful in monitoring the patient's body movements during Vojta therapy.
Collapse
Affiliation(s)
- Muhammad Hassan Khan
- Research Group of Pattern Recognition, University of Siegen, Germany; College of Information Technology, University of the Punjab, Pakistan.
| | - Julien Helsper
- Research Group of Pattern Recognition, University of Siegen, Germany
| | | | - Marcin Grzegorzek
- Research Group of Pattern Recognition, University of Siegen, Germany; Faculty of Informatics and Communication, University of Economics in Katowice, Poland
| |
Collapse
|
10
|
Farid MS, Lucenteforte M, Grangetto M. Panorama view with spatiotemporal occlusion compensation for 3D video coding. IEEE Trans Image Process 2015; 24:205-219. [PMID: 25438310 DOI: 10.1109/tip.2014.2374533] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
The future of novel 3D display technologies largely depends on the design of efficient techniques for 3D video representation and coding. Recently, multiple view plus depth video formats have attracted many research efforts since they enable intermediate view estimation and permit to efficiently represent and compress 3D video sequences. In this paper, we present spatiotemporal occlusion compensation with panorama view (STOP), a novel 3D video coding technique based on the creation of a panorama view and occlusion coding in terms of spatiotemporal offsets. The panorama picture represents the most of the visual information acquired from multiple views using a single virtual view, characterized by a larger field of view. Encoding the panorama video with state-of-the-art HECV and representing occlusions with simple spatiotemporal ancillary information STOP achieves high-compression ratio and good visual quality with competitive results with respect to competing techniques. Moreover, STOP enables free viewpoint 3D TV applications whilst allowing legacy display to get a bidimensional service using a standard video codec and simple cropping operations.
Collapse
|