Next Article in Journal
Practical Applications of a Set-Based Camera Deployment Methodology
Previous Article in Journal
Measuring the Canopy Architecture of Young Vegetation Using the Fastrak Polhemus 3D Digitizer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SOCA-PRNet: Spatially Oriented Attention-Infused Structured-Feature-Enabled PoseResNet for 2D Human Pose Estimation

1
Department of Informatics, Graduate School of Informatics and Engineering, The University of Electro-Communications, Tokyo 182-8585, Japan
2
Artificial Intelligence Exploration Research Center/Meta-Networking Research Center, The University of Electro-Communications, Tokyo 182-8585, Japan
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(1), 110; https://doi.org/10.3390/s24010110
Submission received: 5 November 2023 / Revised: 3 December 2023 / Accepted: 20 December 2023 / Published: 25 December 2023
(This article belongs to the Section Wearables)

Abstract

:
In the recent era, 2D human pose estimation (HPE) has become an integral part of advanced computer vision (CV) applications, particularly in understanding human behaviors. Despite challenges such as occlusion, unfavorable lighting, and motion blur, advancements in deep learning have significantly enhanced the performance of 2D HPE by enabling automatic feature learning from data and improving model generalization. Given the crucial role of 2D HPE in accurately identifying and classifying human body joints, optimization is imperative. In response, we introduce the Spatially Oriented Attention-Infused Structured-Feature-enabled PoseResNet (SOCA-PRNet) for enhanced 2D HPE. This model incorporates a novel element, Spatially Oriented Attention (SOCA), designed to enhance accuracy without significantly increasing the parameter count. Leveraging the strength of ResNet34 and integrating Global Context Blocks (GCBs), SOCA-PRNet precisely captures detailed human poses. Empirical evaluations demonstrate that our model outperforms existing state-of-the-art approaches, achieving a Percentage of Correct Keypoints at 0.5 ([email protected]) of 90.877 at a 50% threshold and a Mean Precision ([email protected]) score of 41.137. These results underscore the potential of SOCA-PRNet in real-world applications such as robotics, gaming, and human–computer interaction, where precise and efficient 2D HPE is paramount.

1. Introduction

The widespread use and improvement in computer vision (CV) technology in various everyday settings, such as smartphones, digital cameras, and surveillance systems, generate a constant stream of image and video data. Extracting information about human activities from this data is of great importance. Central to these interaction mechanisms is HPE. HPE focuses on identifying and categorizing various joints in the human body. It captures each joint’s coordinates, such as arms, head, and torso—often termed keypoints, to delineate a person’s posture. Over recent decades, the automated interpretation of HPE has become a significant research interest within the field of CV. It forms the foundation for numerous complex CV tasks. It provides a base for predicting 3D HPE, identifying human actions and motion prediction, parsing human body components, and retargeting human movements. Additionally, 2D HPE offers substantial support across applications, from understanding human dynamics, monitoring crowd anomalies or riots, spotting instances of violence, detecting unusual behaviors, and enhancing human–computer interaction (HCI) to aiding autonomous vehicle advancements [1]. The complexity of 2D HPE stems from various factors, occluded keypoints, challenging lighting and background conditions, motion blur, and the intimidating task of implementing the model in real-time due to its vast number of parameters [2].
In the initial phases of research in 2D HPE, the field predominantly relied on traditional methods such as probabilistic graphical models [3,4]. These approaches were characterized by a considerable dependence on manually designed features incorporated into models. While effective to an extent, this reliance on handcrafted features often posed significant limitations, restricting the models’ capacity for broader generalization and optimal performance. The intricate nature of human poses, varying across diverse contexts and environments, posed challenges that these traditional methods struggled to consistently address.
As the field evolved, a paradigm shift occurred with the advent of deep learning techniques. This marked a substantial transformation in the approach to 2D HPE. Deep learning, diverging from the constraints of manual feature engineering, brought the capability of automatically extracting relevant features and learning from data. This shift was particularly catalyzed by the advancements in convolutional neural networks (CNNs). CNNs’ ability to process complex image data effectively and their versatility in learning feature hierarchies propelled 2D HPE into a new era. The success of CNNs and their applications in pose estimation underscored the potential of deep learning, paving the way for the development and incorporation of various sophisticated deep learning strategies that built on the foundational achievements of CNNs [5].
With this backdrop, the primary objective of our paper is to further enhance the prediction accuracy of 2D HPE methods while optimizing efficiency through a reduced parameter set. We recognize the challenges posed by large deep learning models, particularly when deployed in real-time or resource-constrained settings. Such models, while powerful, can be computationally demanding, memory-intensive, and may require specialized hardware. Additionally, their complexity often risks overfitting, where performance on training data does not translate to unseen data. Addressing these concerns, our research aims to strike a balance between accuracy and efficiency, creating versatile and cost-effective models suitable for a range of applications, from edge computing devices to large-scale cloud infrastructures. This endeavor leads us to propose SOCA-PRNet, a framework that epitomizes this balance by integrating advanced features within a streamlined architecture.
Our research led us to the simple baseline network [6], which has demonstrated superior performance compared to other top-down methodologies. Its streamlined and efficient architecture positions it as a prime foundation for further advancement in 2D HPE. Building on this foundation, we introduce SOCA-PRNet. This framework is characterized by integrating a Spatially Oriented Attention-Infused Structured Features module, with a modified version of ResNet serving as its primary feature extractor [7,8]. Within this ResNet adaptation, we have omitted the average pooling and the last fully connected layers, emphasizing convolutional layers. To further simplify the model and decrease its complexity, we have employed ResNet34 over the more elaborate variants such as ResNet50, 101, or 152, all of which possess a larger parameter count. We have added two deconvolution layers designed to enhance visual processing capabilities and mitigate quantization distortions from large output stride sizes. While it is understood that a smaller network size might impact the model’s accuracy due to the trade-off between precision and parameter quantity, we have addressed this by another significant inclusion is the integration of Global Context Blocks (GCBs) [9], which aims to expand the performance of both the downsampling and upsampling modules. Furthermore, our innovative SOCA module merges and amplifies feature representations through spatial attention, channeling these refined features to the upsampler layers, thereby bypassing traditional skip connections [10,11]. This methodology fosters hierarchical representations with enhanced spatial awareness, adeptly capturing complex details. These modifications and attributes are designed to offer a detailed, context-rich representation of data, ensuring the model’s stability.
The threefold contribution of the SOCA-PRNet model can be summarized as follows:
  • We introduced the SOCA-PRNet, deliberately choosing ResNet34 over more intricate models to streamline its structure. This decision promotes efficiency without sacrificing capability. Further enhancements include adding two deconvolution layers, bolstering the model’s visual processing, and addressing the quantization distortion from large output stride sizes. We integrated GCBs into the downsampler and upsampler modules to endow the model with robust global context features.
  • Central to our design is the SOCA module. It merges features derived from various downsampler layers. These collective features undergo refinement via a spatial attention mechanism and are subsequently channeled to the appropriate upsampler layers. The outcome is a generation of hierarchical representations with enhanced spatial awareness, adeptly capturing intricate pose details.
  • To evaluate the merits of our proposed model, we subjected it to rigorous testing on the MPII dataset. Both quantitative and qualitative assessments revealed that SOCA-PRNet outperforms existing 2D human pose estimation techniques in terms of accuracy while maintaining a more favorable computational cost.
This article follows a structured approach, with several sections. Section 2 presents an overview of prior research conducted in the same field. Section 3 elaborates on the comprehensive methodology of our proposed SOCA-PRNet. Section 4 covers pertinent information regarding the experimental setup and implementation details. An analysis of both qualitative and quantitative results is exhibited in Section 5. Section 6 offers an in-depth analysis of our results.The final section, Section 7, draws conclusions and lays out plans for future exploration.

2. Related Works

Deep learning approaches are utilized in designing network architectures for 2D HPE to extract robust features that span from low to high levels. These approaches are typically categorized into two frameworks: the top-down and bottom-up frameworks. The method of the top-down paradigm involves a sequential process where the initial step is to identify the human bounding boxes in an image, followed by executing the single HPE for every identified box. This type of approach is not a suitable method for managing large crowds as the computational time for the second step increases in association with the number of individuals present [1,8]. A. Toshev et al. [12] have made a pioneering contribution to the field of HPE by introducing a CNN for the first time.
They leveraged the CNN’s robust fitting capability to regress the coordinates of human joints implements a cascading structure to refine the outcomes continuously, though the model tends to overfit because the weights of the fully connected layer depend on the distribution of the training dataset. The convolutional pose machine (CPM) [13] and stacked hourglass networks [10] solved this issue by predicting heatmaps of 2D joint locations. Two main object detection techniques exist in 2D HPE: the RCNN [14] series and the SSD series [15]. The RCNN series employs a complicated network structure that achieves high accuracy and introduces the Mask-RCNN approach, which builds upon the faster RCNN architecture [14] by incorporating keypoint prediction. As a result, this method achieves excellent results in HPE, demonstrating strong competitiveness in this domain. Conversely, the SSD series offers an average compromise between precision and Y. Chen et al. [16] present the concept of a cascaded pyramid network (CPN) that uses GlobalNet to identify simple keypoints and Refine-Net to handle more challenging keypoints. To be more precise, Refine-Net includes multiple standard convolutional layers that merge feature representations from all levels of GlobalNet.
The process of bottom-up methods starts with detecting keypoints for every human instance present in an image. Subsequently, the keypoints of the same individual are joined to form skeletons of multiple instances. This grouping optimization problem is crucial in determining the outcome of the bottom-up approach. Some representative methods utilize this approach, and they are [5,17]. Open-Pose, as described in [5], utilized two branches—one of which employed a CNN to predict all keypoints based on heatmaps, and the other used a CNN to acquire part affinity fields. The part affinity fields represent 2D direction vectors, and they serve as a confidence metric to determine if the keypoints are associated with the same person. Ultimately, both branches are merged to generate the concluding prediction. The approach known as associative embedding [11], derived from hourglass networks [10], is end-to-end trainable. The source detected and accumulated keypoints in one step without requiring two separate processes.
Implementing bottom-up approaches can be challenging due to the difficulty of combining information from multiple scales and grouping features together. Even with the introduction of effective grouping procedures, these methods still struggle to contest top-down strategies for pose estimation. In recent times, the majority of cutting-edge outcomes have been achieved through top-down methodologies. Our research traced the top-down approach and developed a successful 2D HPE model. This addresses the issue of top-down approaches by modifying a baseline network with Spatially Oriented Attention-Infused Structured Features. We utilized a simpler ResNet34 model and removed specific layers to reduce complexity. We then added deconvolution layers and GCB to improve visual processing and global context features. The proposed SOCA module combines and enhances feature representations from various layers, enabling better capture of finer details through hierarchical representations with spatial awareness.

3. Proposed SOCA-PRNet

We introduce SOCA-PRNet, a novel framework in the field of 2D HPE, distinguished by its integration of a SOCA module with a modified ResNet architecture. This framework is designed to address the intricate requirements of pose estimation by enhancing feature representation and spatial awareness. Our approach begins with the primary objective of 2D HPE; given an RGB image or a video frame labeled as I, the goal is to identify the posture. The pose P of any individual is represented in this visual content. This posture, expressed as P , is characterized by a set of N specific keypoints. Each keypoint is denoted by a two-dimensional coordinate ( x n , y n ) . The number of keypoints, N, can vary based on the dataset used for training a model. Thus, our objective is to pinpoint the pose P = { P i } i = 1 N for every k individual within the input. Algorithm 1, while general, represents the fundamental process of pose estimation in the field of 2D HPE. It serves as a baseline framework from which the innovations of SOCA-PRNet are developed. The algorithm outlines the standard procedure of initializing the posture set, detecting individuals in the image, identifying keypoints, and compiling these into a posture representation.
Algorithm 1: Foundational process of 2D human pose estimation (HPE)
Sensors 24 00110 i001
In developing SOCA-PRNet, we adapted the ResNet architecture, emphasizing convolutional operations and reducing complexity. Specifically, we chose ResNet34 for its balance of efficiency and performance and added two deconvolution layers to enhance visual processing. We also integrated GCBs to improve both downsampling and upsampling modules and introduced the SOCA module, a key innovation that merges and amplifies feature representations through spatial attention. This module directs refined features to the upsampler layers, effectively bypassing traditional skip connections. These modifications aim to provide detailed, context-rich data representations, ensuring both stability and accuracy in pose estimation.
Building on the foundational process outlined in Algorithm 1, the SOCA-PRNet introduces specific enhancements. These include the integration of the SOCA module and the ResNet architecture’s adaptation, which collectively enhance the accuracy and efficiency of pose estimation. This advanced approach, leveraging novel techniques for feature representation and spatial awareness, marks a significant evolution from the general framework of 2D HPE.
Figure 1 presents the detailed structure of SOCA-PRNet, clearly explaining it from the simple baseline network as shown in Figure 2 [6], upon which our research builds. The figure is designed to distinctly show the architectural changes and the inclusion of novel components unique to SOCA-PRNet. Key differences are highlighted, such as the replacement of certain ResNet layers with GCBs and the addition of the SOCA module. These differences are visually contrasted against the architecture of existing networks, emphasizing the enhancements and optimizations we have incorporated.
In the following subsections, we explore the structure of SOCA-PRNet as depicted in Figure 1. We will provide a comprehensive explanation of each component, from the modified ResNet base through the integration of deconvolution layers and GCBs to the final integration of the SOCA module. This detailed breakdown will clarify the functionality of each element in our model and explain how these components collaboratively contribute to the model’s overall performance, highlighting the advancements our network brings to 2D HPE.

3.1. Enhancing Backbone Model with Modified ResNet and Deconvolution Module

The structure of the residual network is commonly utilized for dense labeling tasks. To achieve this, we employed such kind of network structure that slowly decreases the resolution of embeddings to capture extended-range details, which subsequently increases feature maps while recovering spatial resolution. Hourglass and simple baseline networks create smaller output feature maps than their input feature maps, which are then resized using a simple transformation technique that can cause quantization errors. When data processing is biased, prediction errors can occur due to horizontal flipping and how the model processes the output resolution [18]. We incorporated two deconvolution modules into our approach to tackling the above mentioned challenges. These modules were designed to generate a complete output feature map and were integrated within the architecture of the simple baseline network. We opted to use ResNet34, which has fewer parameters than more complex ResNet models like 50, 101, or 152. We modified ResNet [7] by removing the average pooling segment and fully connected part and replacing them with four ResNet blocks after a convolutional and pooling layer. The modifications are visually depicted in Figure 3. The first set of layers in the network, which includes a convolutional layer and a pooling layer, reduces the size of the feature maps by half. As the input passes through each block of the network, additional convolutional layers are used to decrease the feature maps by two strides while simultaneously increasing the number of filters by a factor of two. We added five deconvolutional modules with batch normalization and HardSwish activation, each doubling the feature resolution map until the output matches the input. The fourth and fifth deconvolutional layers have channel sizes of 64 and 32, respectively.

3.2. Amplifying Model Performance with GCB

In computer vision, a Global Context Block is a module designed to capture the overall spatial information of an input feature map, aiming to improve object recognition in an image. In convolutional layers, the association among pixels is only considered within a local neighborhood and baseline network. We opted to use ResNet34, which have fewer parameters compared to more complex ResNet; capturing long-range dependencies requires multiple convolution layers. To address this limitation, researchers proposed a non-local operation [19], which employed a self-attention mechanism from [20] to model long range dependencies. Using a global network creates an attention map tailored to each query position, enabling the collection of contextual features that can then be integrated into the features of the corresponding position. GCNet is presented as a highly well organized and operative method for global context modeling [9]. This method employs a query agnostic attention map to generate a contextual representation that can be globally shared and then incorporates it into the features of each query location in the network.
Our proposed method uses GCBs [9] to enhance the spatial information of input feature maps. Specifically, as illustrated by the sky-blue blocks in Figure 1, GCBs are incorporated into each ResNet block as well as the first four blocks of the deconvolution modules. We generate a spatially aware attention heatmap using a 1 × 1 convolution and SoftMax to produce attention weights, which are then used in attention pooling to extract a global context feature. Channel-wise dependencies are obtained using the bottleneck transform technique. Afterward, the resulting global context features are combined with the features of each position in the network, as shown in the following Equation (1).
f g = i = 1 h j = 1 w w i j f i j
where in Equation (1), f g represents the global context feature, h and w are the height and width of the input feature map, w i j is the attention weights at position ( i , j ) , and f i j is the feature vector at the position ( i , j ) .

3.3. SOCA Module

The Spatially Oriented Attention-Infused Structured Features (SOCA) module overcomes the limitations of earlier frameworks, such as the simple baseline framework [6], which did not integrate skip connections [10,21]. These connections have proven effective in U-Net and hourglass networks for retaining spatial information at each feature map, allowing for an efficient transfer of spatial information across the network, and leading to improved localization.
In contrast to these earlier approaches, our SOCA module, as depicted in Figure 4, represents a significant advancement. Unlike traditional skip connections that typically rely on direct concatenation or summation of feature maps, SOCA employs a novel approach of combining hierarchical features from various layers. It utilizes spatial attention to selectively enhance features that are critical for pose estimation. This process involves the elementwise multiplication of feature maps from the first four Global Context Blocks, ResNet blocks, and spatially oriented attention feature maps. As a result, SOCA provides a more targeted enhancement of features, emphasizing areas crucial for accurate pose estimation. The design of the SOCA module is specifically tailored to generate more relevant details by focusing on key locations for pose estimation while effectively suppressing less relevant background information. This leads to a significant improvement in feature specificity, which is crucial for pose estimation tasks. The spatially aware attention mechanism of SOCA ensures that the enhanced features are optimally tuned to the demands of pose estimation, contributing to robust and accurate model performance, especially in complex scenes.
Our analysis further highlights the advantages of the SOCA module over traditional skip connections. The method of feature integration used by SOCA, through spatial attention and elementwise multiplication, aligns well with tasks that require high accuracy in localization. This approach offers a more refined and context-aware integration of features compared to the simpler methods used in traditional skip connections. By examining the existing literature on skip connections and spatial attention mechanisms, we underline the improvements that SOCA brings in terms of feature representation and model performance. The enhancements in performance with the integration of the SOCA module are evident in the experimental results and analysis section.attest to its effectiveness. Analyzing these data in various pose estimation scenarios reveals the practical benefits of SOCA over conventional methods.
Our observation indicates that the SOCA module is a more effective feature combination mechanism for 2D HPE models compared to traditional skip connections. SOCA’s focus on spatially oriented feature enhancement is expected to lead to improved accuracy in pose estimation, particularly in complex and varied scenarios.

3.4. Heatmap Joint Prediction

Our model employs a sophisticated approach to estimate joint positions by transforming pixel-level predictions into a spatial probability distribution, represented as heatmaps. This transformation is facilitated by a 2D Gaussian function centered on each joint’s true location within the confines of a bounding box. The intensity at each pixel location ( x , y ) on the heatmap H k ( x , y ) is computed in Equation (2):
H k ( x , y ) = exp x y k 2 + y y k 2 2 σ 2
In Equation (2), H k represents the heatmap for k th joint where k { 1 , 2 , , K } , and ( x , y ) show the position of the specified pixel in the heatmap. The k th joints coordinated are denoted by ( x k , y k ) . After several experimental iterations, we found that setting σ to 6 offers an optimal balance, capturing the joint’s essence without excessive spreading.

4. Experimental Setup

4.1. Dataset

Our experimentation to evaluate the effectiveness of our proposed model utilized the widely recognized MPII (Max Planck Institute for Informatics) Dataset [22]. This expansive dataset encompasses over 25,000 annotated images, representing more than 40,000 individuals, each annotated with 16 unique keypoints. We strategically divided this dataset into two subsets: one for training and the other for testing. For the training phase, we used 28,000 images to develop and fine-tune our model. Subsequently, we reserved a separate set of 11,000 images to evaluate the model’s performance, offering an objective measure of its robustness and accuracy.

4.2. Implementation Details

We employed a range of data augmentation techniques to enhance our model’s robustness against scale and spatial rotation variations. We introduced a random horizontal flip to the dataset, diversifying its content and helping to mitigate overfitting. Additionally, we made rotation adjustments, enabling the model to process images tilted between −40 and +40 degrees, which improves its adaptability to varied orientations. To ensure the model effectively recognizes objects of different sizes, we adjusted scaling within a range of 0.7 to 1.3. The model was built using the PyTorch framework. During training, we set a learning rate of 1 × 10 5 , used a batch size of 16, and deployed six workers for parallel data processing. The training process extended over 150 epochs.

4.3. Loss Calculation Function

Using the correct loss function is crucial for precise model training. In our methodology, we opted to utilize the Mean Square Error (MSE) loss function, which has been utilized in previous works such as [6,8,23], to evaluate the model’s error. The formula for MSE is presented in Equation (3).
L = k = 1 K H k H ^ k K
In Equation (3), H ^ k represent the estimated heatmap for the k th joint, whereas H k is the heatmap for the k th joint k { 1 , 2 , , K } . In our implementation, we have provided the option to weight the errors for different joints differently using target-weight. If this option is enabled, the MSE for each joint is computed after multiplying the heatmaps by their respective weights. This allows us to prioritize or deprioritize certain joints based on their importance or reliability in the dataset.

4.4. Optimization of Model

The optimization of models can be enhanced with the implementation of a new variant of the Adam optimizer named AdamW. The key difference between Adam and AdamW lies in their approach to weight decay. While Adam intertwines weight decay with its adaptive learning rate—sometimes resulting in suboptimal training dynamics—AdamW distinctly decouples weight decay from learning rate updates. This separation minimizes the interference between the adaptive learning rate and weight decay, fostering more stable and consistent optimization. Consequently, models optimized with AdamW tend to exhibit less overfitting than those using the original Adam, making AdamW a favored choice among many deep learning practitioners. We conducted a series of experiments, the comprehensive details of which are presented in Section 6 and summarized in Table 3.

4.5. Activation Functions

In deep learning, activation functions are pivotal in determining a neural network’s output. For our research, we adopted the HardSwish activation function, which has showcased notable advantages over the conventional ReLU function. Mathematically, ReLU is expressed as in Equation (4).
f ( x ) = max ( 0 , x )
In contrast, HardSwish is described in Equation (5).
f ( x ) = x × min ( max ( 0 , x + 3 ) , 6 ) 6
This distinction is visually represented in Figure 5, where the blue curve depicts the ReLU function that nullifies all negative values, and the green curve illustrates the HardSwish function, which offers a smoother gradient and does not nullify negative values as abruptly. Furthermore, a common challenge with ReLU is the ’dying neuron’ problem, where specific neurons become inactive and stop learning. With HardSwish, we improved the overall performance of our network model and achieved better results in our experiments. Our results, detailed in Section 6 and summarized in Table 3, indicate that the integration of the HardSwish activation function significantly enhances the performance of our model. This improvement suggests that HardSwish may be advantageous in a wide range of deep learning applications.

4.6. Evaluation Metrics

In HPE tasks, evaluation metrics are crucial for measuring model performance. Among the commonly used metrics, we adopted PCK (Percentage of Correct Keypoints) and [email protected] for our evaluation. PCKh is a specific variant of PCK. Instead of using absolute distances, PCKh leverages the head bone link length as a reference. A prediction is deemed correct if the distance between the predicted and actual keypoints is within 50% of this reference length, denoted as [email protected]. Meanwhile, [email protected] quantifies the average discrepancy between predicted and actual keypoints, but the head bone link length normalizes it. This normalization makes it scale-invariant, ensuring consistent evaluation across different image resolutions and subject sizes.

5. Experimental Results and Discussion

Our comparative analysis involved an array of models, evaluated across distinct input resolutions of 256 × 256 , 384 × 288 , and 384 × 384 , as illustrated in Table 1. The baseline models, which included SimpleBaseLine [6], PRTR [24], HRNet-W32 [25], and macro–micro [26] configurations, were examined at 256 × 256 and 384 × 384 for a subset of configurations.
When considering the baseline models at 256 × 256 and 384 × 384 resolutions, the SOCA-PRNet34 model demonstrated a remarkable [email protected] score of 89.846 at 256 × 256 , which further increased to 90.877 at 384 × 384 . These scores were significantly higher than those achieved by SimpleBaseline [6] and HRNet-W32 [25], as evidenced by the [email protected] scores of 36.417 and 41.137, respectively, underlining the effectiveness of SOCA-PRNet’s approach. The model’s robustness to resolution scaling was particularly notable when compared to these benchmarks. The exceptional efficiency of SOCA-PRNet34 is further emphasized by its requirement of only 30 million parameters for training, which is considerably lower than the baseline models. This parameter efficiency, juxtaposed with its superior performance metrics, highlights the unique advantages of SOCA-PRNet over SimpleBaseline [6] and HRNet-W32 [25].
The SOCA-PRNet18 model was also evaluated across the same input size range. In line with the SOCA-PRNet34, the SOCA-PRNet18 surpassed baseline models in terms of [email protected] and [email protected] metrics while operating with fewer parameters, reinforcing the efficacy of our approach. Furthermore, the HRNet-W32 [25] model demonstrated commendable performance with a [email protected] of 90.300 at the 256 × 256 resolution. However, even this strong competitor was marginally outperformed by our SOCA-PRNet models at higher resolutions, as reflected in the [email protected] scores.
Figure 6 visually contrasts the accuracy and parameter counts of various 2D HPE models, including our SOCA-PRNet18 and SOCA-PRNet34, as well as the Pose-Resnet series of SimpleBaseline [6] and Pose-hrnet32 [25], which are listed in Table 2. This comparison highlights the efficiency and performance balance achieved by different models. Notably, SOCA-PRNet34 excels with a high [email protected] score of 90.875, using only 30 million parameters, demonstrating an optimal balance between accuracy and model economy. This is particularly impressive when compared to models like Pose-Resnet152, which has over double the parameters but similar accuracy levels. SOCA-PRNet18 also performs competitively, achieving close accuracy to more complex models with just 21 million parameters, illustrating the effectiveness of our approach in resource-limited scenarios. This analysis demonstrates the strength of SOCA-PRNet models in providing high accuracy with a reduced parameter count, affirming the success of our architectural optimizations in 2D HPE.
The data we present in graphical form offer a more intuitive understanding of our research outcomes. Specifically, Figure 7a displays the [email protected] scores for individual joints, contrasting the performance of our proposed SOCA-PRNet models against the SimpleBaseline [6] at a resolution of 256 × 256 . Notably, SOCA-PRNet34 demonstrates significant improvements in challenging keypoints like the Elbow, Wrist, Hip, Knee, and Ankle over the PoseResNet baseline models. SOCA-PRNet18, despite being slightly less performant than its SOCA-PRNet34 counterpart, also shows a competitive edge, particularly in accurately estimating the Hip and Knee joints. These outcomes are significant as they highlight the efficacy of the SOCA-PRNet models, which are specifically designed to balance reduced model complexity with high accuracy. The superior performance of SOCA-PRNet34 in keypoints like the Hip and Knee, even surpassing the more complex PoseResNet152 model, underscores the success of integrating the SOCA module and our streamlined architecture. This balance is crucial for applications where model efficiency is as important as accuracy, particularly in real-world scenarios with limited computational resources.
In Figure 7b, the Mean and [email protected] scores across all joints offer a clear comparison of our SOCA-PRNet models against the PoseResNet baseline models. The SOCA-PRNet34 stands out with the highest Mean accuracy of 90.877% and [email protected] score of 41.137%, indicating its superior overall accuracy and precision in joint localization. This performance, especially in the [email protected] metric, highlights its capability in accurately detecting joints in challenging conditions. The SOCA-PRNet18 also demonstrates notable performance, outperforming the PoseResNet50 and PoseResNet101 in Mean accuracy, which reinforces the effectiveness of our model design. Although slightly behind SOCA-PRNet34, it maintains high accuracy with fewer parameters. Comparatively, the PoseResNet152, despite its competitiveness, does not match the performance of SOCA-PRNet34, emphasizing the advancements our models bring in balancing efficiency and accuracy in 2D HPE.
The SOCA-PRNet18 model was also evaluated across the same input size range. In line with the SOCA-PRNet34, the SOCA-PRNet18 surpassed baseline models in terms of [email protected] and [email protected] metrics while operating with fewer parameters, reinforcing the efficacy of our approach. Furthermore, the HRNet-W32 model demonstrated commendable performance with a [email protected] of 90.300 at the 256 × 256 resolution. However, even this strong competitor was marginally outperformed by our SOCA-PRNet models at higher resolutions, as reflected in the [email protected] scores. This trend is visualized in Figure 6, which showes our models’ competitive edge in accuracy and model economy.
The data we present in graphical form offer a more intuitive understanding of our research outcomes. Specifically, Figure 7a displays the [email protected] scores for individual joints, contrasting the performance of our proposed SOCA-PRNet models against the baseline models at a resolution of 256 × 256 . This visual comparison highlights the relative proficiency of each model in joint estimation accuracy. Complementing this, Figure 7b aggregates the performance metrics, presenting a concise overview of the Mean and [email protected] scores across all joints at the same resolution. These collective metrics serve to encapsulate the models’ precision in joint localization in a single, comparative glance.
To contextualize our findings within practical applications, Figure 8 illustrates the practical efficacy of the SOCA-PRNet34 model by showing its pose estimation results on images from the MPII dataset. This visual representation demonstrates the model’s real-world applicability and solidifies its potential for accurate human pose estimation in varied and complex scenarios.

6. Ablation Study

Our ablation study aimed to optimize performance in 2D HPE with a focus on the individual contributions of different components that are deconvolution layers, the SOCA module, activation functions, and optimizers.

6.1. Initial Model Development

We initiated with the ResNet34 architecture, leading to our base model, PRNet34-3xDeconvolution. Equipped with three deconvolution layers, this model achieved a Mean accuracy of 86.607 and a [email protected] value of 24.421, as presented in Table 3.

6.2. Assessing Deconvolution Layers

To assess the impact of additional deconvolution layers, we subsequently developed the PPNet34-5xDeconvolution model with five deconvolution layers. This variant showed an increase in Mean accuracy to 89.113 and [email protected] to 37.359, in Table 3, indicating the effectiveness of added deconvolution layers in visual processing.

6.3. Evaluating the SOCA Module

Our exploration, however, was not limited to these results. Recognizing the potential of integrating global context into our model, we introduced GCN. This led to the emergence of the SOCA-PRNet34 model. The novel SOCA module, embedded within, integrates feature representations from varied layers. The resultant architecture thus becomes adept at capturing intricate details through layered representations while maintaining spatial awareness. The outcomes of this integration were obvious: SOCA-PRNet34 outperformed its predecessors, achieving an impressive Mean accuracy of 90.877 and a [email protected] of 41.137, as presented in Table 3.

6.4. Comparative Analysis of Activation Functions and Optimizers

Further, we conducted a comparative analysis of activation functions and optimizers across these model variants. For PPNet34-5xDeconvolution, switching from ReLU and Adam to HardSwish and AdamW improved Mean accuracy from 88.925 to 89.113. In the case of SOCA-PRNet34, the ReLU-Adam configuration achieved a Mean accuracy of 88.259, while the AdamW optimizer, particularly when paired with HardSwish, elevated performance to a Mean accuracy of 90.877, as detailed in Table 4. This demonstrates the notable influence of activation functions and optimizers on model performance.
Our systematic ablation study reveals that each component—additional deconvolution layers, the SOCA module, and the choice of activation function and optimizer—independently contributes to enhancing the performance of 2D HPE models. These insights highlight the potential of component-specific optimizations in advancing the field.

7. Conclusions and Future Work

In this study, we introduced the SOCA-PRNet for 2D HPE, a novel approach that binds the efficient ResNet34 architecture to find a balance between computational simplicity and visual processing capability. The model’s design is further encouraged by including GCBs in the downsampler and upsampler modules, ensuring the assimilation of comprehensive global context features. Our proposed SOCA module plays a crucial role in merging and directing features with heightened spatial attention, allowing the model to generate detailed hierarchical representations. When compared to standard models on the MPII dataset, SOCA-PRNet’s enhanced performance becomes evident, driven by its refined feature processing, optimal activation function, and advanced optimizer. As we look to the future, SOCA-PRNet’s adaptability presents it as a promising option for applications beyond 2D HPE, such as 3D human pose estimation, object recognition, and hand pose estimation. Given its versatility, the model is anticipated to contribute significantly to enhancing interactive experiences in the rapidly expanding fields of HCI, robotics, and gaming.

Author Contributions

Conceptualization A.Z. and S.A.S.; writing—original draft preparation A.Z.; writing—review and editing, A.Z., S.A.S. and H.T.; supervision, H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SOCA-PRNetSpatially Oriented Attention-Infused Structured-Feature-enabled PoseResNet
CVComputer Vision
HPEHuman Pose Estimation
GCBGlobal Context Blocks
HCIHuman–Computer Interaction
CPMConvolutional Pose Machine
CNNConvolutional Neural Network
CPNCascaded Pyramid Network
PCKhPercentage of Correct Keypoints
MPIIMax Planck Institute for Informatics
MSEMean Square Error

References

  1. Bertasius, G.; Feichtenhofer, C.; Tran, D.; Shi, J.; Torresani, L. Learning temporal pose estimation from sparsely-labeled videos. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  2. Chen, H.; Feng, R.; Wu, S.; Xu, H.; Zhou, F.; Liu, Z. 2D Human pose estimation: A survey. Multimed. Syst. 2023, 29, 3115–3138. [Google Scholar] [CrossRef]
  3. Sapp, B.; Toshev, A.; Taskar, B. Cascaded models for articulated pose estimation. In Proceedings of the Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 406–420. [Google Scholar]
  4. Wang, F.; Li, Y. Beyond physical connections: Tree models in human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 596–603. [Google Scholar]
  5. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  6. Xiao, B.; Wu, H.; Wei, Y. Simple baselines for human pose estimation and tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 466–481. [Google Scholar]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  8. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  9. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In Proceedings of the IEEE/CVF international Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  10. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 483–499. [Google Scholar]
  11. Newell, A.; Huang, Z.; Deng, J. Associative embedding: End-to-end learning for joint detection and grouping. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  12. Toshev, A.; Szegedy, C. Deeppose: Human pose estimation via deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1653–1660. [Google Scholar]
  13. Wei, S.E.; Ramakrishna, V.; Kanade, T.; Sheikh, Y. Convolutional pose machines. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4724–4732. [Google Scholar]
  14. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  16. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded pyramid network for multi-person pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7103–7112. [Google Scholar]
  17. Salman, S.A.; Zakir, A.; Takahashi, H. Cascaded deep graphical convolutional neural network for 2D hand pose estimation. In Proceedings of the International Workshop on Advanced Imaging Technology (IWAIT), Langkawi, Malaysia, 7–8 January 2023; SPIE: Bellingham, WA, USA, 2023; Volume 12592, pp. 227–232. [Google Scholar]
  18. Huang, J.; Zhu, Z.; Guo, F.; Huang, G. The devil is in the details: Delving into unbiased data processing for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5700–5709. [Google Scholar]
  19. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  20. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  22. Andriluka, M.; Pishchulin, L.; Gehler, P.; Schiele, B. 2d human pose estimation: New benchmark and state of the art analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3686–3693. [Google Scholar]
  23. Hua, D.; Zhang, Q.; Liao, W.; Wang, B.; Yan, T. CAttNet: A Compound Attention Network for Depth Estimation of Light Field Images. J. Inf. Process. Syst. 2023, 19, 483–497. [Google Scholar] [CrossRef]
  24. Li, K.; Wang, S.; Zhang, X.; Xu, Y.; Xu, W.; Tu, Z. Pose recognition with cascade transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1944–1953. [Google Scholar]
  25. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  26. Zhou, L.; Chen, Y.; Cao, C.; Chu, Y.; Wang, J.; Lu, H. Macro-micro mutual learning inside compositional model for human pose estimation. Neurocomputing 2021, 449, 176–188. [Google Scholar] [CrossRef]
Figure 1. Detailed architecture of our proposed SOCA-PRNet model for 2D HPE.
Figure 1. Detailed architecture of our proposed SOCA-PRNet model for 2D HPE.
Sensors 24 00110 g001
Figure 2. Detailed architecture of the simple base line model for 2D HPE [6].
Figure 2. Detailed architecture of the simple base line model for 2D HPE [6].
Sensors 24 00110 g002
Figure 3. Modified ResNet with deconvolution module.
Figure 3. Modified ResNet with deconvolution module.
Sensors 24 00110 g003
Figure 4. Visualization of the SOCA module’s feature integration and weights W generation and distribution mechanisms.
Figure 4. Visualization of the SOCA module’s feature integration and weights W generation and distribution mechanisms.
Sensors 24 00110 g004
Figure 5. Comparative visualization of HardSwish and ReLU activation functions.
Figure 5. Comparative visualization of HardSwish and ReLU activation functions.
Sensors 24 00110 g005
Figure 6. Visual analysis of 2D HPE models in terms of accuracy and parameter count.
Figure 6. Visual analysis of 2D HPE models in terms of accuracy and parameter count.
Sensors 24 00110 g006
Figure 7. Graphical Illustration of the proposed model and simple baseline models. (a) Illustration of [email protected] results: proposed model and simple baseline models. (b) Graphical analysis of Mean and [email protected]: proposed models and simple baseline models.
Figure 7. Graphical Illustration of the proposed model and simple baseline models. (a) Illustration of [email protected] results: proposed model and simple baseline models. (b) Graphical analysis of Mean and [email protected]: proposed models and simple baseline models.
Sensors 24 00110 g007
Figure 8. Qualitative results on MPII pose estimation result, containing viewpoint change, occlusion, and self-occlusion.
Figure 8. Qualitative results on MPII pose estimation result, containing viewpoint change, occlusion, and self-occlusion.
Sensors 24 00110 g008
Table 1. Performance comparisons of our SOCA-PRNet with other model results on MPII dataset.
Table 1. Performance comparisons of our SOCA-PRNet with other model results on MPII dataset.
ModelBackboneInputHeadShElbowWristHipKneeAnkleMean[email protected]
SimpleBaseLine [6]Pose-Resnet50 256 × 256 96.35195.32988.98983.17688.42083.96079.59488.53233.911
384 × 384 96.65895.75489.79084.61488.52384.66679.28789.06638.046
Pose-Resnet101 256 × 256 96.86295.87389.51884.37688.43784.48680.70389.13134.020
384 × 384 96.96595.90790.26885.78089.59785.93582.09890.00338.860
Pose-Resnet152 256 × 256 97.03395.94190.04684.97689.16485.31181.27189.62035.025
384 × 384 96.79495.61890.08086.22589.70086.86282.85390.20039.433
PRTR [24]Pose-Resnet101 384 × 384 96.30095.00088.30082.40088.10083.60077.40087.900-
Pose-Resnet152 96.40094.90088.40082.60088.60084.10078.40088.200-
HRNet-W32 [25]Pose-hrnet_w32 256 × 256 97.10095.90090.30086.40089.10087.10083.30090.30037.700
macro–micro [26]Pose-Resnet50 256 × 256 96.65095.49089.22083.65088.29084.44080.91088.890-
SOCA-PRNetPose-Resnet18 256 × 256 96.96595.68889.39884.05190.25485.02980.72889.42534.483
384 × 288 97.16995.78890.13184.46290.34185.33181.69689.76636.435
384 × 384 97.20396.26490.47285.48990.98186.37981.89090.29739.670
Pose-Resnet34 256 × 256 97.23795.80590.01284.89190.06485.97681.50789.84636.417
384 × 288 97.27196.24790.60885.64291.01686.98482.71290.53638.158
384 × 384 96.93096.29891.18886.07291.53587.66883.13790.87741.137
Table 2. Comparison of model complexity: parameter counts in various models.
Table 2. Comparison of model complexity: parameter counts in various models.
ModelNo. Parameters (M)
Pose-Resnet50 [6]34.0
Pose-Resnet101 [6]53.0
Pose-Resnet152 [6]68.6
Pose-hrnet32 [25]28.5
SOCA-PRNet 1821.0
SOCA-PRNet 3430.0
Table 3. An ablation analysis: evaluating network performance across varied additional modules.
Table 3. An ablation analysis: evaluating network performance across varied additional modules.
ModelInputHeadShoulderElbowWristHipKneeAnkleMean[email protected]
PRNet34-3 × Decon384 × 38495.80594.46386.97880.40087.46780.94075.20086.60724.421
PPNet34-5 × Decon96.35195.31289.67184.61688.52384.80780.56389.11337.359
SOCA-PRNet3496.93096.29891.18886.07291.53587.66883.13790.87741.137
Table 4. Comprehensive Ablation analysis: assessing network performance across distinct activation functions and optimization techniques.
Table 4. Comprehensive Ablation analysis: assessing network performance across distinct activation functions and optimization techniques.
ModelActivation FunctionOptimizationMean[email protected]
PPNet34-5 × DeconReLUAdam88.92536.784
HardSwishAdamW89.11337.359
SOCA-PRNet34ReLUAdam88.25938.457
ReLUAdamW89.41539.493
HardSwishAdamW90.87741.137
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zakir, A.; Salman, S.A.; Takahashi, H. SOCA-PRNet: Spatially Oriented Attention-Infused Structured-Feature-Enabled PoseResNet for 2D Human Pose Estimation. Sensors 2024, 24, 110. https://doi.org/10.3390/s24010110

AMA Style

Zakir A, Salman SA, Takahashi H. SOCA-PRNet: Spatially Oriented Attention-Infused Structured-Feature-Enabled PoseResNet for 2D Human Pose Estimation. Sensors. 2024; 24(1):110. https://doi.org/10.3390/s24010110

Chicago/Turabian Style

Zakir, Ali, Sartaj Ahmed Salman, and Hiroki Takahashi. 2024. "SOCA-PRNet: Spatially Oriented Attention-Infused Structured-Feature-Enabled PoseResNet for 2D Human Pose Estimation" Sensors 24, no. 1: 110. https://doi.org/10.3390/s24010110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop