Next Article in Journal
Transmission Line Vibration Damper Detection Using Multi-Granularity Conditional Generative Adversarial Nets Based on UAV Inspection Images
Previous Article in Journal
A New Electric Field Mill Network to Estimate Temporal Variation of Simplified Charge Model in an Isolated Thundercloud
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Linear Regression vs. Deep Learning: A Simple Yet Effective Baseline for Human Body Measurement

by
Kristijan Bartol
1,*,
David Bojanić
1,
Tomislav Petković
1,
Stanislav Peharec
2 and
Tomislav Pribanić
1
1
Faculty of Electrical Engineering and Computing, University of Zagreb, Unska 3, 10000 Zagreb, Croatia
2
Peharec Polyclinic for Physical Medicine and Rehabilitation, 52100 Pula, Croatia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 1885; https://doi.org/10.3390/s22051885
Submission received: 14 January 2022 / Revised: 22 February 2022 / Accepted: 24 February 2022 / Published: 28 February 2022
(This article belongs to the Special Issue Deep Learning and Sensing Technologies for Anthropometry)

Abstract

:
We propose a linear regression model for the estimation of human body measurements. The input to the model only consists of the information that a person can self-estimate, such as height and weight. We evaluate our model against the state-of-the-art approaches for body measurement from point clouds and images, demonstrate the comparable performance with the best methods, and even outperform several deep learning models on public datasets. The simplicity of the proposed regression model makes it perfectly suitable as a baseline in addition to the convenience for applications such as the virtual try-on. To improve the repeatability of the results of our baseline and the competing methods, we provide guidelines toward standardized body measurement estimation.

1. Introduction

Body measurement or anthropometry is a study of the numerical description of human body segments and overall shape. Anthropometry is important for health applications [1], virtual try-on and the fashion industry [2], fitness [3], and ergonomics [4]. Advances in human body measurement and shape estimation have been significantly driven by statistical models and deep learning. The statistical models such as SMPL [5] describe the population of human bodies using pose and shape parameters. The parameters correspond to specifically posed and shaped template meshes. The template meshes are particularly useful for body measurement because each mesh contains a fixed number of vertices, and their semantics are common for the entire population of meshes. Therefore, the procedure for body measurement from the template meshes for a given statistical model, e.g., SMPL, can be standardized. We describe the procedure for extracting 15 body measurements from SMPL templates in Section 3 and use these measurements to compare the linear regression model to the state-of-the-art.
Statistical models enable human body mesh regression, which can be described as fitting point cloud (3D) or image (2D) features to the pose and shape parameters of a given template mesh of the statistical model to obtain the best correspondence between the estimated mesh and the input human body. The 3D point cloud-based approaches [6,7,8] are generally more accurate than their image-based counterparts [9], but they require 3D scans as input, making them impractical for most anthropometric applications. Image-based methods greatly simplify the data acquisition stage. A very popular research topic is pose and shape estimation using in-the-wild images of clothed people in various difficult poses [10,11,12,13,14,15,16,17,18,19,20]. Even though these models recover a person’s pose remarkably well, less attention is dedicated to the accurate estimation of shape parameters. Finally, there are several image-based methods that estimate the shape parameters based on the extracted silhouette(s) of a person [21,22,23,24,25,26,27,28,29]. Silhouette-based methods achieve state-of-the-art results in shape estimation and body measurement, but their applications are limited to a configuration with minimal clothing and fixed posture.
We propose a linear regression model that requires only the information that any person can self-estimate, such as height and weight. We demonstrate strong performance of the proposed model against the state-of-the-art for body measurement estimation on two public datasets, BODY-fit [28] and ANSUR [30]. The linear model provides a simple and straightforward way to obtain an estimate of body measurements, which makes it convenient for virtual try-on and augmented reality. In this work, we focus on measurement estimation of the bodies from the statistical human body population. Statistical models such as SMPL are extensively used today, especially in computer vision, so we propose the linear model as a baseline for future evaluations and provide guidelines for body measurement from template meshes.
The main contributions of our work are the following:
  • We propose the linear regression model, that uses self-estimated height and weight, to be used as a baseline for human body measurement estimation;
  • We demonstrate that the baseline performs strongly against the state-of-the-art methods for body measurement estimation and analyze its performance in detail;
  • We publish the source code, the demo for obtaining the body measurements given height and weight as input, and the protocol for extracting the standard body measurements from the SMPL mesh.

2. Related Work

Body measurements can be obtained using traditional anthropometric techniques and tools, 3D scanning data such as point clouds or meshes (non-parametric approaches), or features extracted from 3D or 2D data and then used to estimate the parameters of the statistical model (parametric approaches). The accuracy of body measurements obtained using traditional tools and 3D scanning are usually more accurate compared to feature-based, parametric approaches, but the former are more time consuming and relatively expensive.
Traditional Anthropometry. Traditional body measurement involves the use of tools such as calipers and tape measures [2]. Measurements taken by an expert are considered the gold standard and are generally used as the ground truth [31]. The public anthropometric databases such as ANSUR are collected by the expert measurers. However, even measurements taken by the experts are not perfectly accurate [32]. Therefore, the values for the allowable or “expert” error [30] for each measurement are defined in the standard ISO:7250 [33,34]. Based on the public databases, several works propose linear regression models for the estimation of measurements [35,36,37] and other body characteristics such as skeletal muscle mass [38]. We extend these analyses by focusing specifically on using self-estimated height and weight as input as well as on using the statistical body models.
Non-Parametric Approaches. With the advances in 3D scanning technology, more automatic approaches to body measurement have been proposed [7,39,40]. Most 3D-based body measurement methods use landmarks to determine distances and calculate measurements such as arm and leg length, shoulder-to-crotch, etc. Circumferences can be obtained by slicing the point cloud with a vertical plane and summing up distances between the nearest points [39]. The 3D-based methods are generally the most accurate among the (semi-)automatic body measurement methods. However, their main drawback is that they require 3D scanning, which is cumbersome and relatively expensive.
There are a number of image-based (2D) non-parametric models [20,41,42,43] that freely deform meshes to best fit the input features. To improve convergence, they start with the template human mesh. However, the final deformed mesh does not necessarily retain the original vertex semantics. This property makes current image-based non-parametric models less suitable for body measurements. On the other hand, compared to their parametric counterparts, non-parametric models might have the advantage of better fitting out-of-distribution samples.
Parametric Approaches. Apart from direct body measurements from point clouds, 3D scans are usually used to create the statistical models [5,44,45,46]. The initial template mesh at rest (or T-position) is deformed (registered) to each 3D scan of the dataset [47,48,49] using an optimization algorithm, such as L-BFGS [50]. To parameterize body poses, a kinematic model of the human body is applied. To describe the shape variation in the dataset, principal component analysis (PCA) is applied to the vertices of the deformed meshes [5]. The first ten principal components (shape parameters) are usually used for the mesh regression. Tsoli et al. [6] first registers template meshes to the 3D scans and additionally learns features for body measurement prediction. BUFF [51] addresses the problem of 3D body shape estimation under clothing by allowing the mesh to deviate from the template but regularizes the optimization to satisfy the anthropometric constraints. Similar to non-parametric approaches, 3D parametric approaches are generally more accurate than image-based approaches, but they also require 3D scans as input.
Image-based approaches (2D) can be divided into shape-aware pose estimation methods, which typically regress pose and shape parameters in-the-wild either from 2D keypoints or directly from images [10,11,13,14,15,16,17,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66], and shape estimation methods, which regress shape from silhouettes, usually in fixed pose and minimal clothing [21,22,23,24,25,26,27,28,31,67]. We compare the proposed baseline against the state-of-the-art 3D- and 2D-based approaches for human body measurement estimation and achieve comparable performance to the best methods, while outperforming several deep learning models (see Section 4).

3. Method

In this section, we describe the proposed linear regression model and the extraction of body measurements from a template SMPL mesh. We fit separate regression models for males and females. The purpose of the proposed method is to demonstrate that body measurements can be determined using only the information that each person typically knows about themselves, i.e., height and weight, without making any further measurement, which makes it suitable as a baseline.

3.1. Linear Regression Model

The method is shown in Figure 1. Each human body sample i is defined by 10 shape parameters, θ S i , and 63 pose parameters, θ P i . The height and the 15 body measurements of the model are extracted from the template mesh. The input to the model consists of height (h) and weight (w). Weight should be available in the dataset; otherwise, it can be estimated from the calculated mesh volume. The output consists of the 15 other body measurements (A–O). The measurement names are listed in Table 1, and their extraction is described in Section 3.2. Then, the linear model for the j-th measurement is described as:
y j = x T a j + b j , j { 1 , 2 , , 15 }
where x T R 2 is a row vector consisting of the height and weight of the samples (an independent variable), a j R 2 is a column vector of the linear coefficients (the slope), b j R is an intercept of the regression model, and y j R is an output measurement (a dependent variable). These equations can be written more compactly in matrix form as:
Y j = X A j , j { 1 , 2 , , 15 }
where X R N × 3 are the heights and weights from the N samples from the training dataset and Y j R N × 1 are their output measurements. Additionally, a column of ones is added to X to account for b j , which is now included in A j R 3 × 1 , representing all the model parameters.
Therefore, the least-squares closed-form solution is:
A j = X T X 1 X T Y j , j { 1 , 2 , , 15 }
The linear regression has four assumptions: linearity, homoscedasticity, independence, and normality. The linearity assumes that the relationship between X and the mean Y is linear. The homoscedasticity assumes that the variance of residual is the same for any value of X. Independence assumes that the observations are independent of each other. Finally, the normality assumes that for any fixed value of X, Y is normally distributed. The latter three assumptions are related to residuals, and we verify those in Section 5.1. For further details on linear regression, we refer readers to the relevant literature [68].
Note that we also experiment with the interaction terms (h is height, w is weight): w h 2 (BMI), w h , w 2 , h 2 , and add them to the input vector x T . The model with no interaction terms is called the baseline, and the models with added interaction terms are called the augmented baselines. The augmented baselines are marked in the remainder of the paper as Baseline (I = N), where N is the number of interaction terms. The interaction terms I = 2 correspond to w h 2 and w h , and I = 4 corresponds to w h 2 , w h , w 2 , and h 2 .
In case that the weight measurement is not available in the dataset, we estimate it using the extracted volume. The volume is extracted using a standard algorithm [69]. Then, the body weight (w) is estimated based on the human body density, which is approximately ρ = 1 ± 0.005 kg/L [31,70], and the extracted volume (V). To account for the variation in body density w.r.t. weight, we model the volume as a normal stochastic variable:
V = V e x t r a c t e d + N ( μ V = 0 , σ V = 5 ) [ L ] .
Note that the standard deviation of 5 L applied to the extracted volume propagates to the standard deviation of 5 kg applied to the estimated weight. Additionally, to account for the variation in self-estimation of height and weight, we model self-estimation using another two stochastic variables, h = h e x t r a c t e d + N ( μ h = 0 , σ h = 1 ) cm, and w = V · ρ + N ( μ w = 0 , σ w = 1.5 ) kg.

3.2. Extraction of Body Measurements

We use a total of 18 body measurements, 15 of which are a standard set of measurements used in previous works [21,22,23,24,25,26,27,28,31,67], and 3 of which are used specifically to compare with Virtual Caliper [31] (see Table 1). The measurements are either lengths or circumferences and are calcuated using their corresponding landmarks. The complete list of landmarks with their corresponding SMPL vertex index is shown in Table 2. To extract the lengths, such as the shoulder width and arm length, we calculate the Euclidean distances between the two respective landmarks. To extract the circumferences such as the waist or thigh circumference, we slice the mesh with a horizontal or vertical plane at the specified landmark location and sum up the resulting line segments [8,28].

4. Evaluation

We evaluate the linear baseline on BODY-fit [28] and ANSUR [30] datasets. More specifically, the BODY-fit dataset is extended by weight estimations obtained from the extracted mesh volumes. We call the extended dataset BODY-fit+W. We compare the baseline on BODY-fit+W with the following state-of-the-art image-based approaches: SMPLify [10], ExPose [14], and Yan et al. [28]. Unfortunately, most of the other methods have not published their source code or data, so we use their reported results on other datasets instead, such as CAESAR [49] and NOMO3D [8] datasets. In total, six datasets are referenced in this work, as listed in Table 3.

4.1. Datasets

The BODY-fit dataset contains 1474 male and 2675 female SMPL meshes. The template meshes are obtained by fitting them to the original 3D scans of people, which are not publicly available. In addition, the BODY-fit+W dataset contains weights that are estimated from corresponding mesh volumes. The distributions of male body measurements on BODY-fit+W, obtained by measuring the template meshes as described in Section 3.2, are shown in Figure 2. The range of the height is between 145 and 196 cm for male and between 135 and 190 cm for female, while weight is between 40 and 130 kg for male, and between 30 and 130 kg for female. Regarding other measurements, they also vary proportionally to their absolute values. For example, ankle circumferences are generally smaller than waist circumferences, etc. The body measurements on BODY-fit+W are similarly diverse as the body measurements from the ANSUR dataset, as shown in Figure 3. The ANSUR dataset is significant, as it represents the realistic human body population, while the BODY-fit+W represents the population of SMPL meshes fitted to 3D scans. Note that not all of the body measurements used in the BODY-fit+W dataset exist in the ANSUR dataset. Table 4 specifies the ANSUR attributes and expressions used to obtain the corresponding measurements from BODY-fit+W.

4.2. Quantitative Evaluation

The methods are compared quantitatively against the 15 standard body measurements and the three additional measurements to compare the baseline with the Virtual Caliper [31]. The metrics used for comparison are as follows:
  • Mean absolute error (MAE), E j , M A E = 1 N i N y e s t , j ( i ) y g t , j ( i ) , where i is the sample index, j represents the measurement, and N is the number of samples;
  • Mean relative error (MRE), E j , M R E = 1 N i N y e s t , j ( i ) y g t , j ( i ) y g t , j ( i ) , where i is the sample index, j represents measurement, and N is the number of samples;
  • Expert ratio (%<Expert), % < E x p e r t j = # < E x p e r t j N , where j is the measurement and N is the number of samples. This metric shows the ratio of samples that are within the expert errors [30,33,34]. The expert errors are shown in Table 5 and Table 6 (Expert error rows).
We compare our linear models against the competing methods in several configurations:
  • Against the methods that use ground-truth features as input, such as ground truth silhouettes [21,22,23,25,26,27]. In this case, we evaluate the baseline using the ground truth volume from the original BODY-fit data, i.e., the volume, height, and weight are modeled as deterministic variables.
  • Against the state-of-the-art methods that use estimated or extracted features, including both 3D-based [6,8,72] and 2D-based [10,11,14,16,21,22,23,25,26,27,28] methods. The volume, height, and weight are modeled as stochastic variables (see Section 3).
  • Against other methods such as the Virtual Caliper [31] that estimates body lengths using a VR headset.
  • More detailed comparison with the representative 2D-based [10] and 3D-based methods [8]. On top of MAE, we also report the mean relative error (MRE) and the percentage of the samples within the expert errors (%<Expert).
Ground Truth Methods. We compare the baseline against the methods that use ground truth information as input. Several previous silhouette-based body measurement estimation methods [21,22,23,25,26,27] report their results using the ground truth silhouettes. To evaluate the baseline, we use the BODY-fit dataset and the volume as input. This way, we also exploit the ground truth information instead of estimating the weight. The results are shown in Table 5. Note that several methods [21,22,23] perform within the expert errors. All of these methods, including the baselines, perform comparably and achieve significantly low body measurement errors below 1 cm for all body measurements. However, using ground-truth silhouettes or body volume is unrealistic for real-world anthropometric applications. We provide these analyses for the sake of completeness, following the previous works.
State-of-the-Art Methods. Table 6 shows the performance of state-of-the-art body measurement estimation methods, compared to our baselines fitted on the BODY-fit+W and ANSUR datasets. The baseline fitted on the BODY-fit+W dataset models the volume, height, and weight as stochastic variables. For comparison with SMPLify [10,11] and ExPose [14], we scale their meshes to match ground truth height; otherwise, their mesh estimates would be significantly degraded by height estimation errors. Our baseline demonstrates comparable performance with the competing methods, even outperforming several popular deep learning approaches such as HMR [16], SMPLify [10,11], and ExPose [14]. Note that the baseline achieves MAEs within the expert errors for neck circumference (B), shoulder-to-crotch (C), and forearm circumference (I). The baseline evaluated on the ANSUR dataset is less accurate on average, but it still shows competitive performance. A more detailed comparison between BODY-fit+W and ANSUR is given in Section 5.3. The discussion on the performance of deep mesh regression approaches is given in Section 5.4.
Other Methods. We also compare to the Virtual Caliper [31] that proposes using a VR headset for body measurement. Compared to our work, the Virtual Caliper evaluates its performance on real subjects and compares it to measurements obtained by their expert, which is more realistic. They also use self-reported weight as input, but they estimate height. For fair comparison, we predict three additional body measurements such as arm span, inseam height, and hip width, as shown in the additional part of Table 1. The results are shown in Table 7. The baseline outperforms the measurements obtained using the Virtual Caliper. The significant advantage of the proposed baseline is that it does not require a VR headset.
More Detailed Comparison. To compare the representative 2D- and 3D-based methods to the baseline, we calculate MAE, MRE, and %Expert metrics. The representative 2D-based method is SMPLify [10], even though it does not achieve the best performance among the 2D-based methods (see Table 6). However, the best performing method, by Smith et al. [21], does not provide the source code to evaluate on BODY-fit+W. As shown in Table 8, the 3D-based method by Yan et al. [8], evaluated on the NOMO3D dataset, achieves the best overall performance, with the exception of wrist (G) and ankle circumference (N). The baseline fitted to ANSUR achieves MREs below 5% for all body measurements, which is reasonable and useful for the anthropometric applications. Interestingly, the ratio of samples that are within the expert errors of above 50% are shoulder-to-crotch distance (C) and wrist circumference (G). In general, most ratios are above 25% for most of the body measurements.

5. Discussion

The presented baseline demonstrates strong performance on the public datasets. In this section, we analyze the residual hypotheses, p-values, and R 2 scores (Section 5.1), comment on using height and weight for body measurement estimation in more detail (Section 5.2), compare the BODY-fit+W and ANSUR datasets in more detail (Section 5.3), discuss previous image-based mesh regression approaches (Section 5.4), and provide the guidelines for future body measurement estimation methods (Section 5.5).

5.1. Residuals, p-Values, and  R 2 Scores

In this section, we verify the three assumptions of linear regression regarding residuals (homoscedasticity, independence, and normality) and check the p-values and R 2 scores for the regression models of each body measurement. Figure 4 shows the residuals for the BODY-fit+W and ANSUR models on train and test splits. The variance of the residuals is generally constant for all the values of both models, which satisfies the homoscedasticity assumption. The values are relatively randomly spread, which satisfies the independence assumption. Finally, as shown in the right sides of the two figures, the means of the residuals are zero and they are normally distributed, which satisfies the normality.
Table 9 and Table 10 show the p-values, R 2 scores, MAEs, and RMSEs for male and female models, on BODY-fit+W and ANSUR, respectively, with two interaction terms (I = 2). We can observe that the vast majority of p-values are within the <0.05 threshold. For the simplicity of the analyses, we keep all the input variables that might lead to increased variance in the predictions and hence larger RMSEs [73]. Ideally, the R 2 scores should be as high as possible. Most of the scores for BODY-fit+W datasets are above or close to 0.8 , except for head circumference (A) and shoulder breadth (O). It is reasonable that the head circumference is more difficult to estimate based only on height and weight and their derivative terms. Note that based on p B M I of the shoulder breadth, it would make sense to fit the model without the BMI input term, which may improve the R 2 score. The ANSUR model has somewhat lower R 2 scores, particularly for shoulder-to-crotch (C), wrist circumference (G), and ankle circumference (N). Note that the shoulder-to-crotch measure is derived from three manual measurements, as specified in Table 4, which might the part of the reason for the lower score. Intuitively, wrist and ankle circumferences only somewhat correspond to a person’s height and weight, which makes those more difficult to estimate. Finally, even though the linear models could be further improved, the current performance is competitive w.r.t the state-of-the-art 2D- and 3D-based methods, which is the most important observation of these experiments.

5.2. Height and Weight for Body Measurement Estimation

It is not completely surprising that height and weight strongly correlate with the body measurements for the statistical model population. The previous works, such as [5,74], analyze the principal components of the SMPL and SCAPE [44] statistical body models, respectively. They add several standard deviations separately to each principal component of the mean shape. The resulting explained variance for the first 10 components is shown in Figure 5, and the variation in shapes are shown in Figure 6. It can be observed that the first two principal components explain most of the variance in body shapes (Figure 5), which in turn define the extreme shape variations visible in Figure 6, particularly in terms of height and weight. Minor variations are visible for the third and fourth components. The remaining components do not significantly influence body shape and measurements. Whether or not these linear relationships hold for the general population, it is currently not easy to verify due to lack of public data. Still, the statistical models are expected to be made from a diverse set of human bodies; therefore, we consider this linear relationship relevant.

5.3. Comparing BODY-Fit and ANSUR Models

The two datasets used for the evaluation of the proposed baseline represent the synthetic, statistical population (BODY-fit+W) and the realistic population (ANSUR). The linear models fitted on the BODY-fit+W and ANSUR datasets perform differently for certain body measurements, such as neck circumference (B), shoulder-to-crotch (C), arm length (J), inside-leg length (K), and ankle circumference (N) (see Table 6).
The differences can be explained by the fact that these body measurements are less dispersed in the height–weight space, as shown in Figure 7, even though the volume, height, and weight are modeled as stochastic variables. This might suggest that certain body measurements of the bodies from the synthetic population retain a linear relationship with height and weight, even though Gaussian noise has been added to input. Another reason might be that the human body populations of the two datasets are significantly different w.r.t. to these measurements. ANSUR represents the population of military personnel, while the original population of BODY-fit subjects are probably different.
In order to verify these hypotheses, the actual weight information from BODY-fit subjects is required, at least. Unfortunately, the weight information is not available. Finally, note that we simplify the analysis of ANSUR by not accounting for the diversity in self-estimation of height and weight, because a certain amount of error already exists in the manual measurements, which is difficult to quantify.

5.4. On Image-Based Mesh Regression

Most of the previous image-based methods train deep learning models to regress pose and shape parameters that correspond to the template 3D mesh [10,14,16,21,22,23,28], which was made possible with the creation of the statistical models. This approach, while simple and powerful, has the problem of estimating the size of the person in an image. Due to 2D-to-3D scale ambiguity, it is impossible to determine absolute size without additional information [75]. Mesh regression approaches [10,14,16] typically ignore this problem by training convolutional networks on images and hoping that the network will fit proper height by exploiting the image context. As a result, they often estimate an incorrect absolute height when not explicitly trained for scale estimation. To allow a fair comparison with these methods, we scale their output meshes to match the ground truth height by multiplying with the ratio between the mesh height and the ground truth height. Nevertheless, errors are propagated to the body measurements, which is probably because the body measurements are distributed differently for different body heights. On the other hand, silhouette-based methods [21,22,23,28] typically render human meshes at a fixed distance from the camera and then train their model to estimate the height. The drawback of this approach is that the model only estimates the height correctly for that particular camera–distance configuration.

5.5. Limitations, Assumptions, and Future Guidelines

Limitations. The major limitation of body measurement estimation in general is a lack of large, public, and realistic benchmarks. The public benchmark used in this work, based on BODY-fit [28], contains 3D template meshes fitted to 3D scans. Even though the fitted meshes represent the scanning dataset, the SMPL fits are still an approximation of the original scans. Moreover, the 3D scanning process is also not perfect [9], so the scans themselves do not necessarily represent the original physical human bodies.
General Assumptions. Therefore, the first assumption is that the manual body measurements are comparable to the body measurements obtained from the 3D scan, i.e., from the SMPL template mesh. In addition, all ANSUR body measurements do not necessarily physically correspond to the measurements from the SMPL model, such as shoulder-to-crotch or waist circumference, as specified in Section 4.1, but we assume they are highly linearly correlated and thus comparable. The height is measured as the difference between the top head point and the heel point on the y (height) axis. However, most of the datasets’ subjects are expected to take approximately the A-pose, which is not fully erect. This might result in height being incorrectly estimated in some cases. We assume that the posture is not significantly affected by breathing, that the subjects wear minimal or tight clothes, and that the body measurements such as height and head circumference are not affected by hair artifacts. The problem with hair artifacts is particularly important for the female subjects. Finally, we model all stochastic variables by adding normal distributions that are uncorrelated with other variables, which is not realistic for all groups inside the population. For example, underweight people tend to overestimate their weight and vice versa [76].
Future Guidelines. To be able to evaluate our model, as well as other body measurement estimation models, even more reliably, future work should focus on creating more realistic, diverse, and public benchmarks. To avoid estimating body weight from volume, the benchmark and the training dataset should contain measured weights. For the evaluation of future methods, we propose the 15 body measurements as a basic set of measurements for mesh comparison. The body measurements are intuitive and simple to obtain from template mesh, such as SMPL. The evaluation of the 15 measurements can be used in addition to the previously used per-vertex error (PVE) metric, which is used to compare between the mesh regression methods [12,61,77]. For more standardized body measurement comparison, we specify the exact vertices used to obtain each body measurement from the SMPL mesh.
Measurement Standards. When creating anthropometric datasets, special attention must be paid to body measurement standards [33] and body postures [34]. Even though manual expert measurements are considered to be the golden standard, they will never be perfect, because the human body is not rigid [78]. There are many proposed methods for body measurement error analysis [79,80,81,82], but the conclusion is that no statistical procedure is optimal for manual anthropometry in general. [83]. The most common, simple to calculate, and easily interpretable approach, proposed by Gordon et al. [30], includes the allowable or “expert” errors [30], which we mentioned earlier and compare against in Table 5 and Table 6. The authors took into account many factors that potentially affect measurement accuracy, such as posture, time of the day (morning, evening), measurement technique and instrument, etc. The allowable errors are the inter-observer values obtained by the expert measurers based on the mean of absolute differences. Those values have been incorporated into the international standards [33,34].
Synthetic Data. Finally, the advantage of using synthetic data and 3D template meshes is that the body measurements can be extracted in a standardized way, which guarantees a standardized measurement. Another advantage is that synthetic data avoid privacy issues and approvals. Therefore, the promising future direction is the creation of large, more realistic, and more diverse synthetic anthropometric benchmarks. The statistical models such as SMPL are currently the best approximation of the overall population and are likely to be used as a tool for generating the body measurement benchmarks in the coming years.

5.6. Implementation Details

We implement the model and the experiments in Python 3.8. For linear regression and other models in Section 5, we use the scikit-learn package [84]. For creating mesh objects, calculating volume, slicing with planes, and visualization, we use the trimesh package [85]. For processing 3D meshes as SMPL template meshes and calculating body measurements, we use part of the public SMPL-X source code [11]. For other efficient computations, we use NumPy [86] and SciPy [87]. The seed for generating random numbers is set to ( seed = 2021 ) . All experiments are done on a single desktop computer, with an Intel Core i7-9700 CPU (3GHz, 8-core), 16GB RAM, and NVidia GeForce RTX 2080 Super, under Ubuntu 20.04 LTS. The source code, demos, and instructions are attached as Supplementary Files.

6. Conclusions

The presented regression method is a simple but significantly accurate tool for the convenient and automatic estimation of body measurements, without the need to remove clothes or capture images. Our work demonstrates that the self-reported height and body weight predict body dimensions as good as the state-of-the-art deep learning methods, or even better. The linear regression based on self-reported height and weight should be used as the baseline for future methods; i.e., any body measurement estimation method should never perform worse. The results reported in this paper mostly serve to show the approximate relative performances between the baseline and the competing methods on the currently available public benchmarks, but for the future reference and datasets, we recommend fitting the model to more diverse and realistic data, if applicable. The baseline is perfectly suitable for but not limited to the applications such as virtual try-on, VR, or ergonomics. The next step is to create more realistic, statistical model-based, public benchmarks to further evaluate the relationship between height and weight as well as other anthropometric measurements.

Supplementary Materials

The supplementary materials are available online at https://www.mdpi.com/article/10.3390/s22051885/s1.

Author Contributions

Conceptualization, K.B. and T.P. (Tomislav Pribanić); Data curation, T.P. (Tomislav Petković), S.P.; Formal analysis, D.B.; Methodology, K.B., T.P. (Tomislav Petković), T.P. (Tomislav Pribanić); Project administration, S.P. and T.P. (Tomislav Pribanić); Software, K.B. and D.B.; Supervision, T.P. (Tomislav Pribanić); Writing—original draft, K.B.; Writing—review and editing, D.B., T.P. (Tomislav Petković), S.P. and T.P. (Tomislav Pribanić) All authors have read and agreed to the published version of the manuscript.

Funding

This work has been funded by Croatian Science Foundation under the grant number IP-2018-01-8118 (STEAM) and by European Regional Development Fund under the grant number KK.01.1.1.01.0009 (DATACROSS).

Data Availability Statement

The benchmark data is available as a part of Supplementary Materials.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Heymsfield, S.; Bourgeois, B.; Ng, B.; Sommer, M.; Li, X.; Shepherd, J. Digital anthropometry: A critical review. Eur. J. Clin. Nutr. 2018, 72, 680–687. [Google Scholar] [CrossRef]
  2. Zakaria, N.; Gupta, D. Anthropometry, Apparel Sizing and Design; The Textile Institute Book Series; Elsevier Science: Amsterdam, The Netherlands, 2019. [Google Scholar]
  3. Casadei, K.; Kiel, J. Anthropometric Measurement; StatPearls Publishing: Treasure Island, FL, USA, 2020. [Google Scholar]
  4. Dianat, I.; Molenbroek, J.; Castellucci, I. A review of the methodology and applications of anthropometry in ergonomics and product design. Ergonomics 2018, 61, 1696–1720. [Google Scholar] [CrossRef]
  5. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; Black, M.J. SMPL: A Skinned Multi-Person Linear Model. ACM Trans. Graph. (Proc. SIGGRAPH Asia) 2015, 34, 248:1–248:16. [Google Scholar] [CrossRef]
  6. Tsoli, A.; Loper, M.; Black, M. Model-based Anthropometry: Predicting Measurements from 3D Human Scans in Multiple Poses. In Proceedings of the 2014 IEEE Winter Conference on Applications of Computer Vision (WACV 2014), Steamboat Springs, CO, USA, 24–26 March 2014. [Google Scholar] [CrossRef] [Green Version]
  7. Kaashki, N.N.; Hu, P.; Munteanu, A. Deep Learning-Based Automated Extraction of Anthropometric Measurements From a Single 3-D Scan. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
  8. Yan, S.; Wirta, J.; Kämäräinen, J.K. Anthropometric clothing measurements from 3D body scans. Mach. Vis. Appl. 2020, 31, 7. [Google Scholar] [CrossRef] [Green Version]
  9. Bartol, K.; Bojanić, D.; Petković, T.; Pribanić, T. A Review of Body Measurement Using 3D Scanning. IEEE Access 2021, 9, 67281–67301. [Google Scholar] [CrossRef]
  10. Bogo, F.; Kanazawa, A.; Lassner, C.; Gehler, P.; Romero, J.; Black, M.J. Keep it SMPL: Automatic Estimation of 3D Human Pose and Shape from a Single Image. In Proceedings of the 14th European Conference on Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 8–16 October 2016; Springer International Publishing: Amsterdam, The Netherlands, 2016. [Google Scholar]
  11. Pavlakos, G.; Choutas, V.; Ghorbani, N.; Bolkart, T.; Osman, A.A.A.; Tzionas, D.; Black, M.J. Expressive Body Capture: 3D Hands, Face, and Body from a Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  12. Kocabas, M.; Athanasiou, N.; Black, M.J. VIBE: Video Inference for Human Body Pose and Shape Estimation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5252–5262. [Google Scholar]
  13. Kocabas, M.; Huang, C.H.P.; Hilliges, O.; Black, M.J. PARE: Part Attention Regressor for 3D Human Body Estimation. arXiv 2021, arXiv:2104.08527. [Google Scholar]
  14. Choutas, V.; Pavlakos, G.; Bolkart, T.; Tzionas, D.; Black, M.J. Monocular Expressive Body Regression through Body-Driven Attention. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 20–40. [Google Scholar]
  15. Kolotouros, N.; Pavlakos, G.; Black, M.J.; Daniilidis, K. Learning to Reconstruct 3D Human Pose and Shape via Model-Fitting in the Loop. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019; pp. 2252–2261. [Google Scholar]
  16. Kanazawa, A.; Black, M.J.; Jacobs, D.W.; Malik, J. End-to-end Recovery of Human Shape and Pose. In Proceedings of the Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  17. Kolotouros, N.; Pavlakos, G.; Jayaraman, D.; Daniilidis, K. Probabilistic Modeling for Human Mesh Recovery. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  18. Dwivedi, S.K.; Athanasiou, N.; Kocabas, M.; Black, M.J. Learning To Regress Bodies From Images Using Differentiable Semantic Rendering. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 11250–11259. [Google Scholar]
  19. Moon, G.; Lee, K.M. I2L-MeshNet: Image-to-Lixel Prediction Network for Accurate 3D Human Pose and Mesh Estimation from a Single RGB Image. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  20. Kolotouros, N.; Pavlakos, G.; Daniilidis, K. Convolutional Mesh Regression for Single-Image Human Shape Reconstruction. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4496–4505. [Google Scholar]
  21. Smith, B.M.; Chari, V.; Agrawal, A.; Rehg, J.M.; Sever, R. Towards Accurate 3D Human Body Reconstruction from Silhouettes. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Quebec City, QC, Canada, 16–19 September 2019; pp. 279–288. [Google Scholar] [CrossRef]
  22. Dibra, E.; Jain, H.; Öztireli, C.; Ziegler, R.; Gross, M. HS-Nets: Estimating Human Body Shape from Silhouettes with Convolutional Neural Networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 108–117. [Google Scholar] [CrossRef]
  23. Dibra, E.; Jain, H.; Öztireli, C.; Ziegler, R.; Gross, M. Human Shape from Silhouettes Using Generative HKS Descriptors and Cross-Modal Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5504–5514. [Google Scholar] [CrossRef] [Green Version]
  24. Dibra, E.; Öztireli, C.; Ziegler, R.; Gross, M. Shape from Selfies: Human Body Shape Estimation Using CCA Regression Forests. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 8–16 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 88–104. [Google Scholar]
  25. Boisvert, J.; Shu, C.; Wuhrer, S.; Xi, P. Three-dimensional human shape inference from silhouettes: Reconstruction and validation. Mach. Vis. Appl. 2011, 24, 145–157. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, Y.; Kim, T.K.; Cipolla, R. Inferring 3D Shapes and Deformations from Single Views. In Proceedings of the Computer Vision—ECCV 2010, Crete, Greece, 5–11 September 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 300–313. [Google Scholar]
  27. Xi, P.; Lee, W.S.; Shu, C. A Data-driven Approach to Human-body Cloning Using a Segmented Body Database. In Proceedings of the 15th Pacific Conference on Computer Graphics and Applications (PG’07), Maui, HI, USA, 29 October–2 November 2007; pp. 139–147. [Google Scholar] [CrossRef] [Green Version]
  28. Yan, S.; Wirta, J.; Kämäräinen, J.K. Silhouette Body Measurement Benchmarks. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 7804–7809. [Google Scholar] [CrossRef]
  29. Wan, Z.; Li, Z.; Tian, M.; Liu, J.; Yi, S.; Li, H. Encoder-decoder with Multi-level Attention for 3D Human Shape and Pose Estimation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  30. Gordon, C.C.; Blackwell, C.L.; Bradtmiller, B.; Parham, J.L.; Barrientos, P.; Paquette, S.P.; Corner, B.D.; Carson, J.M.; Venezia, J.C.; Rockwell, B.M.; et al. 2012 Anthropometric Survey of U.S. Army Personnel: Methods and Summary Statistics; U.S. Army Natick Soldier Research, Development and Engineering Center: Natick, MA, USA, 2015. [Google Scholar]
  31. Pujades, S.; Mohler, B.J.; Thaler, A.; Tesch, J.; Mahmood, N.; Hesse, N.; Bülthoff, H.H.; Black, M.J. The Virtual Caliper: Rapid Creation of Metrically Accurate Avatars from 3D Measurements. IEEE Trans. Vis. Comput. Graph. 2019, 25, 1887–1897. [Google Scholar] [CrossRef] [Green Version]
  32. Kouchi, M.; Mochimaru, M. Errors in landmarking and the evaluation of the accuracy of traditional and 3D anthropometry. Appl. Ergon. 2011, 42, 518–527. [Google Scholar] [CrossRef]
  33. ISO 7250-1; Basic Human Body Measurements for Technological Design—Part 1: Body Measurement Definitions and Landmarks. International Organization for Standardization: Seoul, Korea, 2017.
  34. ISO 20685-1; 3-D Scanning Methodologies for Internationally Compatible Anthropometric Databases—Part 1: Evaluation Protocol for Body Dimensions Extracted from 3-D Body Scans. International Organization for Standardization: London, UK, 2018.
  35. You, H.; Ryu, T. Development of a hierarchical estimation method for anthropometric variables. Int. J. Ind. Ergon. 2005, 35, 331–343. [Google Scholar] [CrossRef]
  36. Brolin, E.; Högberg, D.; Hanson, L.; Örtengren, R. Adaptive regression model for synthesizing anthropometric population data. Int. J. Ind. Ergon. 2017, 59, 46–53. [Google Scholar] [CrossRef]
  37. PHEASANT, S.T. A technique for estimating anthropometric data from the parameters of the distribution of stature. Ergonomics 1982, 25, 981–992. [Google Scholar] [CrossRef]
  38. Kawakami, R.; Miyachi, M.; Tanisawa, K.; Ito, T.; Usui, C.; Midorikawa, T.; Torii, S.; Ishii, K.; Suzuki, K.; Sakamoto, S.; et al. Development and validation of a simple anthropometric equation to predict appendicular skeletal muscle mass. Clin. Nutr. 2021, 40, 5523–5530. [Google Scholar] [CrossRef]
  39. Lu, J.M.; Wang, M.J.J. Automated anthropometric data collection using 3D whole body scanners. Expert Syst. Appl. 2008, 35, 407–414. [Google Scholar] [CrossRef]
  40. Wang, M.J.; Wu, W.Y.; Lin, K.C.; Yang, S.N.; Lu, J.M. Automated anthropometric data collection from three-dimensional digital human models. Int. J. Adv. Manuf. Technol. 2007, 32, 109–115. [Google Scholar] [CrossRef]
  41. Lin, K.; Wang, L.; Liu, Z. End-to-End Human Pose and Mesh Reconstruction with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Conference, 19–25 June 2021. [Google Scholar]
  42. Lin, K.; Wang, L.; Liu, Z. Mesh Graphormer. In Proceedings of the International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  43. Choi, H.; Moon, G.; Lee, K.M. Pose2Mesh: Graph Convolutional Network for 3D Human Pose and Mesh Recovery from a 2D Human Pose. arXiv 2020, arXiv:2008.09047. [Google Scholar]
  44. Anguelov, D.; Srinivasan, P.; Koller, D.; Thrun, S.; Rodgers, J.; Davis, J. SCAPE: Shape Completion and Animation of People. ACM Trans. Graph. 2005, 24, 408–416. [Google Scholar] [CrossRef]
  45. Hirshberg, D.; Loper, M.; Rachlin, E.; Black, M. Coregistration: Simultaneous alignment and modeling of articulated 3D shape. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; Springer: Firenze, Italy, 2012; pp. 242–255. [Google Scholar]
  46. Osman, A.A.A.; Bolkart, T.; Black, M.J. STAR: A Sparse Trained Articulated Human Body Regressor. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 598–613. [Google Scholar]
  47. UK National Sizing Survey. Available online: http://www.size.org/ (accessed on 20 November 2020).
  48. SizeUSA Dataset. Available online: https://www.tc2.com/size-usa.html (accessed on 17 March 2021).
  49. Caesar I, The Most Comprehensive Source for Body Measurement Data. Available online: http://store.sae.org/caesar/ (accessed on 20 November 2020).
  50. Liu, D.C.; Nocedal, J. On the Limited Memory BFGS Method for Large Scale Optimization. Math. Program. 1989, 45, 503–528. [Google Scholar] [CrossRef] [Green Version]
  51. Zhang, C.; Pujades, S.; Black, M.; Pons-Moll, G. Detailed, accurate, human shape estimation from clothed 3D scan sequences. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 5484–5493. [Google Scholar] [CrossRef] [Green Version]
  52. Balan, A.O.; Sigal, L.; Black, M.J.; Davis, J.E.; Haussecker, H.W. Detailed Human Shape and Pose from Images. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1–8. [Google Scholar] [CrossRef] [Green Version]
  53. Pavlakos, G.; Zhu, L.; Zhou, X.; Daniilidis, K. Learning to Estimate 3D Human Pose and Shape from a Single Color Image. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 459–468. [Google Scholar]
  54. Iqbal, U.; Xie, K.; Guo, Y.; Kautz, J.; Molchanov, P. KAMA: 3D Keypoint Aware Body Mesh Articulation. arXiv 2021, arXiv:2104.13502. [Google Scholar]
  55. Li, J.; Xu, C.; Chen, Z.; Bian, S.; Yang, L.; Lu, C. HybrIK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition, Virtual Conference, 19–25 June 2021. [Google Scholar]
  56. Choi, H.; Moon, G.; Lee, K.M. Beyond Static Features for Temporally Consistent 3D Human Pose and Shape from a Video. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Conference, 19–25 June 2021. [Google Scholar]
  57. Joo, H.; Neverova, N.; Vedaldi, A. Exemplar Fine-Tuning for 3D Human Pose Fitting Towards In-the-Wild 3D Human Pose Estimation. arXiv 2020, arXiv:2004.03686. [Google Scholar]
  58. Doersch, C.; Zisserman, A. Sim2real transfer learning for 3D pose estimation: Motion to the rescue. arXiv 2019, arXiv:1907.02499. [Google Scholar]
  59. Guan, S.; Xu, J.; Wang, Y.; Ni, B.; Yang, X. Bilevel Online Adaptation for Out-of-Domain Human Mesh Reconstruction. arXiv 2021, arXiv:2103.16449. [Google Scholar]
  60. Rempe, D.; Birdal, T.; Hertzmann, A.; Yang, J.; Sridhar, S.; Guibas, L.J. HuMoR: 3D Human Motion Model for Robust Pose Estimation. arXiv 2021, arXiv:2105.04668. [Google Scholar]
  61. Chen, Y.C.; Piccirilli, M.; Piramuthu, R.; Yang, M.H. Self-attentive 3D human pose and shape estimation from videos. In Computer Vision and Image Understanding; Elsevier Science: Amsterdam, The Netherlands, 2021; p. 103305. [Google Scholar] [CrossRef]
  62. Moon, G.; Lee, K.M. Pose2Pose: 3D Positional Pose-Guided 3D Rotational Pose Prediction for Expressive 3D Human Pose and Mesh Estimation. arXiv 2020, arXiv:2011.11534. [Google Scholar]
  63. Yi, X.; Zhou, Y.; Xu, F. TransPose: Real-time 3D Human Translation and Pose Estimation with Six Inertial Sensors. ACM Trans. Graph. 2021, 40, 86:1–86:13. [Google Scholar] [CrossRef]
  64. Zhou, Y.; Habermann, M.; Habibie, I.; Tewari, A.; Theobalt, C.; Xu, F. Monocular Real-time Full Body Capture with Inter-part Correlations. In Proceedings of the CVPR, Virtual Conference, 19–25 June 2021. [Google Scholar]
  65. Madadi, M.; Bertiche, H.; Escalera, S. SMPLR: Deep learning based SMPL reverse for 3D human pose and shape recovery. Pattern Recognit. 2020, 106, 107472. [Google Scholar] [CrossRef]
  66. Georgakis, G.; Li, R.; Karanam, S.; Chen, T.; Košecká, J.; Wu, Z. Hierarchical Kinematic Human Mesh Recovery. In Proceedings of the European Conference on Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 768–784. [Google Scholar]
  67. Yan, S.; Kämäräinen, J.K. Learning Anthropometry from Rendered Humans. arXiv 2021, arXiv:2101.02515. [Google Scholar]
  68. Walpole, R.E.; Myers, R.H.; Myers, S.L.; Ye, K. Probability & Statistics for Engineers and Scientists, 8th ed.; Pearson Education: Upper Saddle River, NJ, USA, 2007. [Google Scholar]
  69. Lien, S.l.; Kajiya, J.T. A symbolic method for calculating the integral properties of arbitrary nonconvex polyhedra. IEEE Comput. Graph. Appl. 1984, 4, 35–42. [Google Scholar] [CrossRef]
  70. Krzywicki, H.J.; Chinn, K.S. Human body density and fat of an adult male population as measured by water displacement. Am. J. Clin. Nutr. 1967, 20, 305–310. [Google Scholar] [CrossRef]
  71. Hasler, N.; Stoll, C.; Sunkel, M.; Rosenhahn, B.; Seidel, H.P. A Statistical Model of Human Pose and Body Shape. Comput. Graph. Forum 2009, 28, 337–346. [Google Scholar] [CrossRef]
  72. Human Solutions GmbH. Anthroscan. Available online: https://www.human-solutions.com/ (accessed on 20 November 2020).
  73. Heinze, G.; Wallisch, C.; Dunkler, D. Variable selection—A review and recommendations for the practicing statistician. Biom. J. 2018, 60, 431–449. [Google Scholar] [CrossRef] [Green Version]
  74. Pishchulin, L.; Wuhrer, S.; Helten, T.; Theobalt, C.; Schiele, B. Building Statistical Shape Spaces for 3D Human Modeling. Pattern Recognit. 2017, 67, 276–286. [Google Scholar] [CrossRef] [Green Version]
  75. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: New York, NY, USA, 2003. [Google Scholar]
  76. Cawley, J.; Maclean, J.; Hammer, M.; Wintfeld, N. Reporting error in weight and its implications for bias in economic models. Econ. Hum. Biol. 2015, 19, 27–44. [Google Scholar] [CrossRef]
  77. Sengupta, A.; Budvytis, I.; Cipolla, R. Synthetic Training for Accurate 3D Human Pose and Shape Estimation in the Wild. In Proceedings of the British Machine Vision Conference (BMVC), Virtual Conference, 7–10 September 2020. [Google Scholar]
  78. Kouchi, M.; Mochimaru, M.; Tsuzuki, K.; Yokoi, T. Random Errors in Anthropometry. J. Hum. Ergol. 1996, 25, 155–166. [Google Scholar] [CrossRef]
  79. Jamison, P.L.; Zegura, S.L. A univariate and multivariate examination of measurement error in anthropometry. Am. J. Phys. Anthropol. 1974, 40, 197–203. [Google Scholar] [CrossRef]
  80. Kouchi, M.; Mochimaru, M.; Tsuzuki, K.; Yokoi, T. Interobserver errors in anthropometry. J. Hum. Ergol. 1999, 28, 15–24. [Google Scholar]
  81. Ulijaszek, S.J.; Kerr, D.A. Anthropometric measurement error and the assessment of nutritional status. Br. J. Nutr. 1999, 82, 165–177. [Google Scholar] [CrossRef] [Green Version]
  82. Branson, R.S.; Branson, R.S.; Vaucher, Y.E.; Harrison, G.G.; Harrison, G.G.; Vargas, M.; Thies, C. Inter- and Intra-Observer Reliability of Skinfold Thickness Measurements in Newborn Infants. Hum. Biol. 1982, 54, 137–143. [Google Scholar]
  83. Utermohle, C.J.; Zegura, S.L.; Heathcote, G.M. Multiple observers, humidity, and choice of precision statistics: Factors influencing craniometric data quality. Am. J. Phys. Anthropol. 1983, 61, 85–95. [Google Scholar] [CrossRef]
  84. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  85. Trimesh. Available online: https://trimsh.org/ (accessed on 22 February 2022).
  86. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  87. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [Green Version]
Figure 1. An overview of the linear regression model for a single sample from the statistical model. The sample mesh is defined using the shape and pose parameters, M ( θ S , θ P ) i . Both input and output are extracted from the sample template mesh, M i . The input consists of height and weight. Weight can be either available in the dataset or estimated using the calculated mesh volume. The output consists of the 15 body measurements (A–O), which are listed in Table 1. To account for the errors in self-reporting height and weight, we model height and weight as stochastic variables by adding Gaussian noise to the input. A linear regression model is fitted to each of the 15 body measurements.
Figure 1. An overview of the linear regression model for a single sample from the statistical model. The sample mesh is defined using the shape and pose parameters, M ( θ S , θ P ) i . Both input and output are extracted from the sample template mesh, M i . The input consists of height and weight. Weight can be either available in the dataset or estimated using the calculated mesh volume. The output consists of the 15 body measurements (A–O), which are listed in Table 1. To account for the errors in self-reporting height and weight, we model height and weight as stochastic variables by adding Gaussian noise to the input. A linear regression model is fitted to each of the 15 body measurements.
Sensors 22 01885 g001
Figure 2. The distribution of body measurements for male subjects in the BODY-fit+W dataset.
Figure 2. The distribution of body measurements for male subjects in the BODY-fit+W dataset.
Sensors 22 01885 g002
Figure 3. The distribution of body measurements for male subjects in the ANSUR dataset.
Figure 3. The distribution of body measurements for male subjects in the ANSUR dataset.
Sensors 22 01885 g003
Figure 4. An analysis of the residuals for hip circumference (F), for BODY-fit+W and ANSUR, on train and test splits.
Figure 4. An analysis of the residuals for hip circumference (F), for BODY-fit+W and ANSUR, on train and test splits.
Sensors 22 01885 g004
Figure 5. Explained variance for the first 10 principal components of the dataset. The graph is generated on CAESAR-fits data [74] by applying PCA to the given vertices. As expected, the first two components, which most significantly correlate to height and weight, explain most of the variance in the data.
Figure 5. Explained variance for the first 10 principal components of the dataset. The graph is generated on CAESAR-fits data [74] by applying PCA to the given vertices. As expected, the first two components, which most significantly correlate to height and weight, explain most of the variance in the data.
Sensors 22 01885 g005
Figure 6. Explained variance for the first 10 principal components of the dataset. As expected, the first two components, which highly correlate to height and weight, explain most of the variance in the data. The image is inspired by [74], originally made using SCAPE model [44]. We have drawn the above image using SMPL model [5].
Figure 6. Explained variance for the first 10 principal components of the dataset. As expected, the first two components, which highly correlate to height and weight, explain most of the variance in the data. The image is inspired by [74], originally made using SCAPE model [44]. We have drawn the above image using SMPL model [5].
Sensors 22 01885 g006
Figure 7. Out of all measurements, the above images show the extreme three cases of body measurements with different dispersions on the ANSUR (first row) and BODY-fit+W (second row) datasets: shoulder-to-crotch (C), arm length (J), and ankle circumference (N). The fitted planes correspond to the linear model with two interaction terms (I = 2).
Figure 7. Out of all measurements, the above images show the extreme three cases of body measurements with different dispersions on the ANSUR (first row) and BODY-fit+W (second row) datasets: shoulder-to-crotch (C), arm length (J), and ankle circumference (N). The fitted planes correspond to the linear model with two interaction terms (I = 2).
Sensors 22 01885 g007
Table 1. The list of 15 (+3) body measurements. The 15 measurements (A–O) are used to compare to the state-of-the-art. The arm length (J) and the three additional measurements are specifically used to compare with [31].
Table 1. The list of 15 (+3) body measurements. The 15 measurements (A–O) are used to compare to the state-of-the-art. The arm length (J) and the three additional measurements are specifically used to compare with [31].
Measurement Set MeasurementLandmark Index
StandardAHead circumference14
BNeck circumference10
CShoulder to crotch1, 10
DChest circumference4
EWaist circumference13
FHip circumference19
GWrist circumference9
HBicep circumference20
IForearm circumference15
JArm length2, 9
KInside leg length11, 12
LThigh circumference16
MCalf circumference17
NAnkle circumference18
OShoulder breadth2, 3
Additional [31]-Arm span7, 8
-Inseam height2, 19
-Hip width5, 6
Table 2. The list of 20 landmarks and their corresponding SMPL vertex indices.
Table 2. The list of 20 landmarks and their corresponding SMPL vertex indices.
Landmark IndexLandmark NameVertex IndexLandmark IndexLandmarkVertex Index
1Inseam point314911Low left hip3134
2Left shoulder301112Left ankle3334
3Right shoulder647013Lower belly point1769
4Left chest142314Forehead point335
5Left hip122915Right forearm point5084
6Right hip494916Right thigh point4971
7Left mid finger244517Right calf point4589
8Right mid finger590618Right ankle point6723
9Left wrist224119Mid hip point3145
10Shoulder top306820Right bicep point6281
Table 3. The list of datasets referenced in this work. Note that the baseline is evaluated on the first three datasets, above the dashed line.
Table 3. The list of datasets referenced in this work. Note that the baseline is evaluated on the first three datasets, above the dashed line.
DatasetSamplesData TypeAvailabilityApproachReported by
BODY-fit4149SMPL meshPublic2D-based[28], Our
BODY-fit+W4149SMPL meshPublic2D-based[10,14,28], Our
ANSUR6068TabularPublicRegressionISO [30], Our
CAESAR3800Point cloudProprietary3D-based[21,23,25,26,27]
[6,16,22,71,72]
NOMO3D375Point cloudPublic3D-based[8]
Virtual Caliper20Point cloudPrivateRegression[31]
Table 4. The specification of ANSUR attributes and expressions corresponding to the body measurements extracted from the SMPL meshes.
Table 4. The specification of ANSUR attributes and expressions corresponding to the body measurements extracted from the SMPL meshes.
SMPL Mesh (BODY-fit+W)ANSUR Attribute/Expression
AHead circumferenceheadcircumference
BNeck circumferenceneckcircumference
CShoulder to crotchsittingheight-(stature-acromialheight)
DChest circumferencechestcircumference
EWaist circumferencewaistcircumference
FHip circumferencebuttockcircumference
GWrist circumferencewristcircumference
HBicep circumferencebicepcircumferenceflexed
IForearm circumferenceforearmcircumferenceflexed
JArm lengthacromialheight-wristheight
KInside leg lengthcrotchheight-lateralmalleolusheight
LThigh circumferencethighcircumference
MCalf circumferencecalfcircumference
NAnkle circumferenceanklecircumference
OShoulder breadthbiacromialbreadth
Table 5. Quantitative comparison to image-based body measurement approaches that use ground truth silhouettes (MAEs, in mm). In this case, we also show the performance of the linear baseline using the volume, height, and weight as deterministic variables (unrealistic). Note that we demonstrate only the performance of the BODY-fit model, as ANSUR does not contain volume measurements. The baseline (I = 4) additionally uses four interaction terms, as described in Section 3. Methods marked with † are evaluated on different, non-public data, and the results are reported in [21]. The best results are shown in bold.
Table 5. Quantitative comparison to image-based body measurement approaches that use ground truth silhouettes (MAEs, in mm). In this case, we also show the performance of the linear baseline using the volume, height, and weight as deterministic variables (unrealistic). Note that we demonstrate only the performance of the BODY-fit model, as ANSUR does not contain volume measurements. The baseline (I = 4) additionally uses four interaction terms, as described in Section 3. Methods marked with † are evaluated on different, non-public data, and the results are reported in [21]. The best results are shown in bold.
MeasurementDatasetABCDEFGHIJKLMNOMean
†Xi ’07 [27]CAESAR50.059.011936.055.023.056.014618210919.035.033.061.024.067.1
†Chen ’10 [26]CAESAR23.027.052.018.037.015.024.059.076.053.09.019.016.028.012.031.2
†Boisvert ’13 [25]CAESAR10.011.04.010.022.011.09.017.016.015.06.09.06.014.06.011.1
Expert error [30]ANSUR5.06.015.012.012.0---6.0-4.0---8.08.5
†Dibra ’17 [23]CAESAR3.21.94.25.67.16.91.62.62.22.34.35.12.71.42.13.6
†Dibra ’16 [22]CAESAR2.02.03.02.07.04.02.02.01.03.09.06.03.02.02.03.3
†Smith ’19 [21]CAESAR5.13.01.54.74.83.02.52.71.91.71.52.42.32.11.92.7
Baseline (I = 4)BODY-fit7.91.25.610.19.23.60.61.41.35.38.08.42.61.46.64.9
Table 6. Quantitative comparison to image-based body measurement and shape estimation approaches (MAEs in mm). In this case, we show the performance of the linear baseline using self-reported height and weight as stochastic variables (more realistic). The BODY-fit+W and ANSUR datasets are used to present the baseline performance. The baseline used I = 2 interaction terms, as described in Section 3. The best results are shown in bold.
Table 6. Quantitative comparison to image-based body measurement and shape estimation approaches (MAEs in mm). In this case, we show the performance of the linear baseline using self-reported height and weight as stochastic variables (more realistic). The BODY-fit+W and ANSUR datasets are used to present the baseline performance. The baseline used I = 2 interaction terms, as described in Section 3. The best results are shown in bold.
MethodDatasetABCDEFGHIJKLMNOMean
HMR [16]CAESAR16.735.733.892.811868.712.229.320.629.944.338.525.814.026.539.8
ExPose [14]BODY-fit17.413.131.496.0116.754.87.733.315.312.329.537.318.28.923.034.3
SMPLify [10]BODY-fit15.37.78.757.574.739.75.121.09.55.711.427.212.36.510.421.6
Hasler ’09 [71]CAESAR7.517.07.513.019.016.2---10.4---6.6-12.2
Anthroscan [72]CAESAR7.421.17.512.417.07.5---11.7---7.6-11.5
Tsoli ’14 [6]CAESAR5.915.85.512.718.612.4---10.1---6.2-10.9
Yan ’20 [28]BODY-fit12.013.68.922.216.914.24.810.08.06.87.513.89.15.98.210.8
Dibra ’16 [22]CAESAR9.310.06.622.824.020.09.912.07.96.48.915.513.27.66.010.7
Expert Error [30]ANSUR5.06.015.012.012.0---6.0-4.0---8.08.5
Yan ’20 [8]NOMO3D-3.7-13.212.48.94.55.53.013.2-7.93.010.612.48.2
Smith ’19 [21]CAESAR6.78.05.112.515.89.39.38.15.75.16.88.87.25.04.57.9
Baseline (I = 2)BODY-fit+W9.14.26.630.339.528.02.710.04.95.79.516.07.33.39.012.4
Baseline (I = 2)ANSUR11.910.717.429.137.921.64.413.29.317.619.617.012.88.711.416.2
Table 7. Comparison to the Virtual Caliper [31] (MAEs in mm) w.r.t. four body measurements—arm length (J) and the three additional measurements (arm span, inseam height, and hip width). We present the baseline evaluated on the same data as in Table 6 (BODY-fit+W). Better results are shown in bold.
Table 7. Comparison to the Virtual Caliper [31] (MAEs in mm) w.r.t. four body measurements—arm length (J) and the three additional measurements (arm span, inseam height, and hip width). We present the baseline evaluated on the same data as in Table 6 (BODY-fit+W). Better results are shown in bold.
MeasurementDatasetArm SpanArm LengthInseam HeightHip WidthMean
Virtual Caliper [31]Virtual Caliper17.27.624.66.514.0
Baseline (I = 2)BODY-fit+W13.15.78.86.78.6
Table 8. Detailed comparison between 2D-based methods (SMPLify [10]), 3D-based methods (Yan et al. [8]), and the two linear baselines (with I = 2 interaction terms), one fitted to the BODY-fit+W dataset, and one fitted to the ANSUR dataset [30]. Note that to fairly compare with Yan et al., the expert error values are extended according to [8]. The best results in each row for MAEs and %<Experts are shown in bold.
Table 8. Detailed comparison between 2D-based methods (SMPLify [10]), 3D-based methods (Yan et al. [8]), and the two linear baselines (with I = 2 interaction terms), one fitted to the BODY-fit+W dataset, and one fitted to the ANSUR dataset [30]. Note that to fairly compare with Yan et al., the expert error values are extended according to [8]. The best results in each row for MAEs and %<Experts are shown in bold.
2D-Based3D-BasedBaseline (I = 2)Baseline (I = 2)
SMPLify [10] (BODY-fit+W)Yan et al. [8] (NOMO3D)BODY-fit+WANSUR
MAE [mm]MRE (%)%<Expert ↑MAEMRE%<ExpertMAEMRE%<ExpertMAEMRE%<Expert
A15.32.325.1---9.11.543.111.92.127.9
B7.74.450.73.7-87.64.21.174.910.72.934.9
C8.71.485.2---6.60.994.217.43.050.1
D57.55.513.613.2-67.630.31.423.829.12.927.3
E74.77.09.312.4-58.739.51.619.837.94.220.6
F39.75.912.38.9-72.428.01.123.421.62.135.8
G5.13.359.54.5-66.52.70.787.04.42.763.2
H21.07.517.25.5-65.810.01.433.813.24.028.5
I9.53.840.03.0-74.24.90.963.99.33.240.1
J5.71.6-13.2--5.71.2-17.62.3-
K11.41.721.0---9.51.426.819.62.613.9
L27.24.514.57.9-47.516.01.723.417.02.725.1
M12.33.427.53.0-82.57.31.040.712.83.425.5
N6.52.941.410.6-26.73.30.860.58.78.728.0
O10.43.249.212.4--9.01.856.211.42.943.1
Table 9. The linear regression statistics for the BODY-fit+W dataset, for males and females (I = 2). The p-values correspond to the intercept (b), height (h), weight (w), BMI ( w h 2 ), and w h , respectively. In addition, we report adjusted R 2 scores, MAEs (mm), and RMSEs (mm).
Table 9. The linear regression statistics for the BODY-fit+W dataset, for males and females (I = 2). The p-values correspond to the intercept (b), height (h), weight (w), BMI ( w h 2 ), and w h , respectively. In addition, we report adjusted R 2 scores, MAEs (mm), and RMSEs (mm).
MALEFEMALE
p b p h p w p BMI p wh Adj. R 2 MAERMSE p b p h p w p BMI p wh Adj. R 2 MAERMSE
A0.0040.0000.0000.0000.0000.5088.9012.360.0000.0000.0000.0000.0000.4428.3112.67
B0.0000.2550.0000.0330.0000.7964.005.040.0000.0000.0000.0000.0000.7954.125.22
C0.3510.0000.0000.0000.0000.8926.958.770.0010.0000.0000.0000.0000.9036.167.73
D0.4020.0000.0000.0670.0000.80526.6833.810.0530.0000.0000.0000.0000.80831.8140.29
E0.0940.0000.0000.0000.0000.81138.5549.180.0000.0000.0000.0000.0000.81938.2748.95
F0.9040.0000.0000.0010.0000.82922.2528.500.0000.0000.0000.0000.0000.83330.7739.45
G0.2640.0000.0000.0010.0000.8452.743.460.520.0000.0000.0000.0000.852.443.14
H0.7180.0050.0000.0550.0000.8118.4610.760.0000.0000.0000.0000.0000.82510.4713.40
I0.0190.0000.0000.3540.0000.8414.445.710.0000.0000.0000.0000.0000.8484.996.44
J0.4890.0000.4210.2810.3590.9305.817.810.0000.0000.0550.6480.0550.9236.078.19
K0.6000.0000.0000.1280.0010.90310.1013.260.0000.0000.0070.0090.1110.9208.9111.51
L0.5800.0000.0000.0210.0020.74214.0118.700.8460.0000.0000.0000.0000.7908.9121.57
M0.0110.0000.0000.1130.0000.8107.109.290.0120.0000.0000.0000.0000.8356.698.70
N0.2570.0000.0000.0000.0000.8562.763.490.9250.0000.0000.0000.0000.8483.174.11
O0.0000.0340.0000.9800.0150.6798.5410.780.0000.0000.0000.0000.0000.6898.4510.81
Table 10. The linear regression statistics for the ANSUR dataset, for males and females (I = 2). The p-values correspond to the intercept (b), height (h), weight (w), BMI ( w h 2 ), and w h , respectively. In addition, we report adjusted R 2 scores, MAEs (mm), and RMSEs (mm).
Table 10. The linear regression statistics for the ANSUR dataset, for males and females (I = 2). The p-values correspond to the intercept (b), height (h), weight (w), BMI ( w h 2 ), and w h , respectively. In addition, we report adjusted R 2 scores, MAEs (mm), and RMSEs (mm).
MALEFEMALE
p b p h p w p BMI p wh Adj. R 2 MAERMSE p b p h p w p BMI p wh Adj. R 2 MAERMSE
A0.0000.0110.1160.0140.3490.26610.5013.250.0000.3700.6540.0020.6420.15213.1917.02
B0.0000.8630.0000.0300.0010.67112.0115.090.0020.3740.0000.9020.0280.6069.4512.14
C0.2620.0000.1850.3120.4020.48416.9021.700.6350.0000.2340.7790.2630.40517.9021.96
D0.0000.1370.0000.0000.0000.86624.4631.340.0490.6320.0000.4000.0010.74033.9643.56
E0.0000.1230.0000.0800.0000.84836.5045.310.3440.8040.0000.4400.0000.77739.3749.78
F0.0000.5740.0000.0000.0010.88720.7926.680.0000.8920.0000.0000.0260.85522.3428.38
G0.1400.0000.0000.5480.0000.5474.906.150.0010.0200.3000.0150.8260.5433.894.90
H0.0000.0000.1220.0000.3700.71215.2619.720.0190.6250.0000.2490.0010.80011.0914.11
I0.0000.4350.0140.0000.6960.66210.6513.440.0000.7790.0040.0050.1820.6788.0310.15
J0.5920.0000.0540.2090.0910.64915.5919.390.3230.0000.8750.8620.8300.64614.2918.05
K0.0170.0000.2650.4400.4780.71419.6025.000.1210.0000.6050.3410.6440.69219.6324.11
L0.0000.0000.0760.0000.0260.85917.1222.150.0000.0000.1090.0000.4020.84016.7121.59
M0.0000.2720.0000.0000.2280.71112.2815.620.0000.0630.5870.0000.3810.64813.3516.62
N0.0000.0290.0040.0000.1540.5338.0710.280.0000.4940.4700.0000.2100.3859.3111.61
O0.0720.0000.0000.8080.0040.42011.2614.090.0540.0000.8450.0160.8830.37311.4914.50
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bartol, K.; Bojanić, D.; Petković, T.; Peharec, S.; Pribanić, T. Linear Regression vs. Deep Learning: A Simple Yet Effective Baseline for Human Body Measurement. Sensors 2022, 22, 1885. https://doi.org/10.3390/s22051885

AMA Style

Bartol K, Bojanić D, Petković T, Peharec S, Pribanić T. Linear Regression vs. Deep Learning: A Simple Yet Effective Baseline for Human Body Measurement. Sensors. 2022; 22(5):1885. https://doi.org/10.3390/s22051885

Chicago/Turabian Style

Bartol, Kristijan, David Bojanić, Tomislav Petković, Stanislav Peharec, and Tomislav Pribanić. 2022. "Linear Regression vs. Deep Learning: A Simple Yet Effective Baseline for Human Body Measurement" Sensors 22, no. 5: 1885. https://doi.org/10.3390/s22051885

APA Style

Bartol, K., Bojanić, D., Petković, T., Peharec, S., & Pribanić, T. (2022). Linear Regression vs. Deep Learning: A Simple Yet Effective Baseline for Human Body Measurement. Sensors, 22(5), 1885. https://doi.org/10.3390/s22051885

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop