Next Article in Journal
Energy-Efficient Harmonic Transponder Based on On-Off Keying Modulation for Both Identification and Sensing
Next Article in Special Issue
Imitating Emergencies: Generating Thermal Surveillance Fall Data Using Low-Cost Human-like Dolls
Previous Article in Journal
A Smart Colorimetric Platform for Detection of Methanol, Ethanol and Formic Acid
Previous Article in Special Issue
Towards the Automation of Infrared Thermography Inspections for Industrial Maintenance Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Clothing Insulation Rate and Metabolic Rate Estimation for Individual Thermal Comfort Assessment in Real Life

1
Visual Analysis and Perception Laboratory, CREATE, Aalborg University, 9000 Aalborg, Denmark
2
The Royal Danish Academy—Architecture, Design, Conservation, 1435 Copenhagen, Denmark
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(2), 619; https://doi.org/10.3390/s22020619
Submission received: 20 December 2021 / Revised: 7 January 2022 / Accepted: 11 January 2022 / Published: 14 January 2022
(This article belongs to the Special Issue Thermal Imaging Sensors and Their Applications)

Abstract

:
Satisfactory indoor thermal environments can improve working efficiencies of office staff. To build such satisfactory indoor microclimates, individual thermal comfort assessment is important, for which personal clothing insulation rate ( I c l ) and metabolic rate (M) need to be estimated dynamically. Therefore, this paper proposes a vision-based method. Specifically, a human tracking-by-detection framework is implemented to acquire each person’s clothing status (short-sleeved, long-sleeved), key posture (sitting, standing), and bounding box information simultaneously. The clothing status together with a key body points detector locate the person’s skin region and clothes region, allowing the measurement of skin temperature ( T s ) and clothes temperature ( T c ), and realizing the calculation of I c l from T s and T c . The key posture and the bounding box change across time can category the person’s activity intensity into a corresponding level, from which the M value is estimated. Moreover, we have collected a multi-person thermal dataset to evaluate the method. The tracking-by-detection framework achieves a mAP50 (Mean Average Precision) rate of 89.1% and a MOTA (Multiple Object Tracking Accuracy) rate of 99.5%. The I c l estimation module gets an accuracy of 96.2% in locating skin and clothes. The M estimation module obtains a classification rate of 95.6% in categorizing activity level. All of these prove the usefulness of the proposed method in a multi-person scenario of real-life applications.

1. Introduction

In the world today, more people have to rely on computers to tackle various tasks. This results in indoor office work being much more popular than ever before. From the commercial buildings energy consumption survey in 2012 [1], offices consume much more energy for heating and cooling than other types of buildings. If energy can be used according to office workers’ thermal needs, energy waste resulting from overheating or overcooling will be greatly reduced, and also staff will have better working efficiencies as they feel comfortable with the environment they work in.
To make each office staff feel thermal comfort and at the same time reduce energy waste, two main kinds of methods have been researched. One is directly relying on the worn clothes to control a person’s micro-environment between the body skin and the indoor atmosphere, which avoids controlling the entire indoor microclimate via heaters, ventilation, and air conditioners (HVAC) that consume lots of energy. This kind of method takes advantage of different thermal properties (thermal resistance, thermal conductivity, thermal radiation, thermal convection, water evaporation, etc.) of different clothes in materials, thicknesses, and layers to maintain the body temperature in a comfortable range [2,3,4,5]. The other kind of method still focuses on the entire indoor environment but in a way that adjusts the microclimate according to each occupant’s thermal need, which is the topic of this paper.
However, each person’s thermal need is unique and dynamic, which cannot be met well by existing microclimate-controlling systems like HVAC that all rely on static assumptions to serve the occupants in a room. For example, a standard air conditioner’s temperature is set to 25 to 27 degrees for cooling in summer, and 18 to 20 degrees for heating in winter, no matter whether this is what the office workers need.
To improve this situation, individual thermal comfort feeling has to be assessed, like in scales (cold, cool, slightly cool, neutral, slightly warm, warm, and hot) [6,7,8]. These scales depend on both environmental factors and personal factors. The environmental factors are air temperature ( t a ), mean radiation temperature ( t ¯ r ), relative humidity ( R H ), and air velocity ( V a ), which can be measured by sensors. The personal factors include clothing insulation rate ( I c l ) and metabolic rate (M); I c l describes the ability of the clothes to insulate the heat exchange between the skin and the environment outside the clothes, and M describes the amount of energy, in-unit time, consumed by a person. Both the personal factors are difficult to acquire for their complexity and dynamics.
Accordingly, international standards [8,9,10,11,12] have defined reference values of I c l and M in certain situations (see Table 1 and Table 2). Such values are empirical and fixed, and thus cannot describe a person’s dynamic property for that the situation in real life is much more complex than these noted ones. This hinders the development of systems and applications for adjusting indoor microclimates according to occupants’ thermal needs. Therefore, the solution dynamically estimating a person’s I c l and M is to be explored. To this end, we propose a method to do this, and the concrete contributions are:
  • The method inventively adapts state-of-the-art computer vision solutions to the thermal comfort domain, achieving a contactless approach that can be employed in multi-person real-life applications.
  • The method can detect and track each person, at the same time recognizing his or her clothing status (long-sleeved, short-sleeved) and key posture (sitting, standing).
  • The method can further output a person’s skin temperature and clothes temperature, based on which his or her I c l is estimated.
  • The method proposes three useful features from a person’s bounding box tracked across time. These features can category the person’s activity into a certain intensity level which indicates the M.
The rest contents are organized as follows. Section 2 introduces the related work. Section 3 describes our methodology. Section 4 tells the experiments. Section 5 concludes the paper and proposes future work.

2. Related Work

This paper applies computer vision solutions to the thermal comfort domain. Therefore, the related researches of both I c l and M estimation and computer vision methods are studied.

2.1. I c l and M Estimation

Several works have been published to calculate the two personal factors, I c l and M, for assessing the human thermal sensation. However, most works only focus on one of them, leaving the other one unsolved.
Some works take advantage of the relationship between clothing choice and environment temperature [13,14,15,16] to predict clothing insulation ability. This type of method is simple but neglects the inherent property of clothes themselves. To resolve this drawback, work [17] uses the weight of the clothes to predict I c l , which is unrealistic in real applications; studies estimate I c l from the temperature difference between the body skin and the clothes surface with infrared sensors [18,19], however, this is also inconvenient due to the attached sensors on the human body. To decouple such interference with personal life, researches [20,21,22] all adopt contactless infrared cameras to monitor persons. Unfortunately, refs. [20,21] do not mention the method of acquiring temperatures of interested body locations, limiting their applications in the real world; ref. [22] only considers five types of garments that cannot represent various clothing choices in daily life.
For metabolic rate estimation, almost all works have to use attached equipment. Correspondingly, a person’s M is estimated by measuring his or her oxygen consumption and carbon dioxide generation [23,24,25], heart rates [26,27,28,29], or blood pressure [30]. Though [31,32,33] adopt cameras for such a task, they still partly rely on sophisticated equipment mentioned above. These devices have to be worn by subjects, making them unrealistically used in daily life.
When estimating both I c l and M, refs. [34,35] use a CNN (Convolutional Neural Network)-based classifier to recognize a person’s clothes type and activity type, and then refer ISO (International Organization for Standardization) standard tables to get the I c l and M values from the recognized types. These works prove the importance of clothing status (short sleeves, long sleeves) and posture (sitting, standing) in estimating I c l and M. However, refs. [34,35] are only valid in a simple and controlled single-person environment. Expanding and enriching this kind of solution is in great need. Therefore, this paper closes this gap and is the first work targeted at a multi-person scenario in the real world.

2.2. Detection and Tracking

The ability to do individual processing from multiple persons is the crucial point of the proposed method, which mainly comes from our implemented human tracking-by-detection framework. To this end, widely used object detectors are studied, like Faster R-CNN (Region-based Convolutional Neural Network) [36], YOLO (You Only Look Once) series [37,38,39,40,41], and FPN (Feature Pyramid Network) [42] which all consist of a backbone network (to extract deep features) and headers (to predict bounding box locations and categories). All these methods perform well on RGB (Red Green Blue) benchmark datasets [43,44].
When it comes to the tracking part (referring in particular to online multi-object tracking in this paper), SORT (Simple Online and Realtime Tracking) [45] initially replaces the conventional object detector with a CNN-based detector and thus improves the tracking result by up to 18.9%,revealing the importance of accurate detections for tracking. The following DeepSort (Simple Online and Realtime Tracking with a Deep Association Metric) [46] and CDA_DDAL (Confidence-based Data Association and Discriminative Deep Appearance Learning) [47] incorporate appearance information into the data association phase and solve the ID (Identity)-switch problem. Other works focus on improving the correlation filter to estimate better positions of targets in the next frame [48], fusing multi-modality data in data association [49], and linking detection and tracking to let them benefit each other [50].
In general, though existing methods on human detection and tracking are quite mature in RGB datasets, studies applying them in thermal datasets like [51,52,53] are few and far between. This situation makes our research with the thermal camera more essential.

3. Methodology

In this section, we describe our approach, the overview of which is illustrated in Figure 1 including three key parts:
  • The thermal input goes through a tracking-by-detection framework (see the red dashed box) to track each individual (see the ID 1 and ID 2) and at the same time categorize each person to get his or her clothing status and key posture (see the red and green solid boxes around persons which indicate different categories).
  • With ID information, for each person, the clothing status classified by the tracking-by-detection part helps differentiate the skin region from the clothing-covered region. Then the detected key body points from these two regions can represent the skin temperature and the clothes temperature, based on which I c l is estimated.
  • With ID information, for each person, the optical flow within each person’s bounding box region, together with the bounding box (center location and box size) changes across time are calculated. These three features are good representations of the person’s activity intensity, which are used to estimate M.
Details of the three parts are described below.

3.1. Tracking-by-Detection

This part has two main components, one is an object detector, YOLOv5 [41], for human detection, the other is a tracker, DeepSort [46].
The video collected from a thermal camera is the input to the detector YOLOv5 for frame-by-frame human detection. To integrate clothing status and key posture recognition into this detection procedure, we classify persons into six categories (see Table 3). Here the clothing status is represented by the sleeve status (long, short) for three reasons: (i) these two are the most common clothing situations in an office environment while the lower part of the body is often totally occluded by the desk; (ii) according to to [10,34,35], sleeve status is significantly important in estimating I c l ; (iii) the change between a long-sleeved status to a short-sleeved status by rolling up sleeves or taking off outer jackets is a sign of feeling hot and vice versa, indicating a person’s thermal sensation directly; (iv) the sleeves status helps to locate skin region and clothes region separately for further skin and clothes temperatures acquisition. For example, the elbows of a person wearing short-sleeved clothes are skin regions, while the elbows of a person wearing long-sleeved clothes are clothes regions. This localization makes it possible to use such key body points to calculate a person’s skin temperature and clothes temperature, because key body points on arms are widely used sensitive heat receptors in thermal comfort assessment [35,54,55,56]. Besides the two statuses of long sleeves and short sleeves, another status called difficult to predict clothes type due to occlusion is also usual in daily life. For clear illustration, such cases are in Figure 2. The right persons in Figure 2a,b are partly occluded by the computer monitor; the right person in Figure 2c moves the arms out of the scene; the left person in Figure 2d occludes his lower arms by hiding them behind the torso. These occlusions make it unrealistic to know whether the sleeves are long or short. One thing to be noted is that even though a person is occluded in a few frames, his or her clothing status can be recognized in other frames. Therefore, voting of a classified category over a few seconds is important. When it comes to the key posture recognition, from ISO standards [8,9,11,12], a person’s metabolic rate M is closely related to the behaving posture (sitting, standing, lying down, etc.). And in a typical office environment the most common ones are sitting and standing, therefore, these two are considered in our study.
The ultimate goal of this research is to acquire every occupant’s personal factors and thus facilitate individual thermal comfort assessment. This means that each person must be tracked across time. To this end, we adopt DeepSort. This tracker receives the image information and YOLOv5-predicted detections, and then decides which tracking ID a detection should be associated to. Like Figure 1 shows, DeepSort can use the detected bounding box information in the ( t 1 ) th frame ( x i , t 1 , y i , t 1 , w i , t 1 , h i , t 1 indicating the ith box’s top-left coordinates, width, height, respectively) to infer the location of the same object in the tth frame in the form of x i , t , y i , t , w i , t , h i , t by Kalman filter. At the same time, DeepSort extracts and saves the deep features of the object as its appearance information. In this way, two similarity metrics (location and appearance) can be calculated, based on which each detected person can be linked to a specific identity thus making the same person be tracked with a consistent ID over time.
The reason why this DeepSort-by-YOLOv5 paradigm is chosen and applied to such a specific research field is explained further below. The data we use is in a thermal mode having significantly fewer details compared with its RGB counterpart. This makes the reuse of such limited details/features extremely important. Compared with other detectors, YOLOv5 introduces PANet (Path Aggregation Network) [57] as its neck, making the deeper layers access to the lower-layer features much more efficiently, so the thermal features are well reused. When it comes to the tracking part, the Maximum Age strategy in DeepSort that deletes a track only when it is not associated to any detection more than A m a x frames can guarantee a consistent ID with the existence of a few false negatives (FN) from YOLOv5. The Tentative Track strategy in DeepSort which confirms a track only after it is associated with detection in three continuous frames also guarantees that occasional false positives (FP) from YOLOv5 have no severe influence on the output. That is to say, this tracking-by-detection framework smooths the direct output from a detector by filtering the undesired consequences of FN and FP, making both the detector and the tracker benefit each other. Additionally, the low complexity and real-time performance of DeepSort fit well the relatively simple scene in our case compared with other cases like pedestrians/vehicles tracking in autonomous driving assistance systems.
Overall, this design not only locates and tracks each individual with a consistent ID in the scene, but also predicts the person’s clothing and posture status simultaneously that directly influence I c l and M estimation.

3.2. I c l Estimation

I c l estimation relying on lookup tables in ISO standards [8,9,10,12] and updated clothes databases [58,59] can be a fast solution for laboratory studies, but it is unfeasible to use such a scheme in real applications due to reasons: (i) looking up the I c l value for a person needs extra manual work which is tedious and expensive; (ii) if this look-up task is expected to be done automatically, the solution must have the ability to recognize hundreds of different garment combinations that vary in materials and number of layers as the latest research has revealed the significant importance of them in thermal comfort [2], which is far beyond the capability of existing algorithms.
Therefore, to realize automated estimation, we go another way—using the difference between the skin temperature T s and the clothes temperature T c to calculate I c l . This method is intuitive since the difference between T s and T c explicitly reveals the heat insulation of clothes to isolate the bare skin from the environmental air. The larger the temperature difference, the higher the clothing insulation rate.
To get T s and T c for each individual, the person’s skin region R s and clothing-covered region R c need to be differentiated from each other. Empirically, R s includes face, hands, and neck; R c includes shoulders, torso, and upper arms. However, in daily life, accessories (hat, glasses, scarf, watch, etc.), spontaneous behaviors (lower one’s head, turn one’s face away, hide one’s arm behind the torso, etc.), and inevitable occlusions by things in front make many body parts be detected unreliably and even totally invisible. After considering such situations, this research counts the lower arms (the middle point of the elbow and wrist) for short-sleeved clothes and the nose area as R s , and the elbows for long-sleeved clothes and the shoulders as R c . These regions are also widely used heat receptors in thermal comfort research [35,54,55,56]. Figure 3 illustrates R s in green crosses and R c in red crosses on four images.
To locate these body parts, we employ OpenPose [60]—a 2D pose estimation tool. OpenPose has a robust ability against occlusions to detect key body points. The level of the ability against occlusions is determined by a parameter called confidence threshold which means that only the detected key point whose confidence score is higher than the threshold will be counted as the output. The higher threshold, the lower the level of ability against occlusions but the higher accuracy of detection; the lower threshold, the higher-level ability against occlusions but more false positives. This can be shown in Figure 4 which draws the detected key body points by OpenPose with different confidence thresholds of 0.1, 0.3, 0.5, and 0.7.
Since the detected key points are representations of R s and R c and thus directly related to T s and T c , a higher accuracy instead of the ability against occlusions is much more important. Like in Figure 4a,b, the detected elbows of the left person are in fact in the computer monitor region; the result in Figure 4c is more accurate, but the detected wrists of the right person are in the laptop region which will influence the lower arm localization in R s . These preliminary trials inspire us to set the confidence threshold as high as possible, but a too high threshold produces more missing detections. Therefore, our work uses 0.6 as the threshold in the entire research which has been proved as an effective parameter in the experimental part Section 4.3. To further decrease the influence of miss detections, an accumulation strategy of all the detected key points within a duration like five minutes is introduced since a person’s clothes status is not changed very frequently, which at the same time filters out potential noises.
Another thing worth mentioning is that although OpenPose detects key body points for each person, it has no function of multi-person tracking, and hence our tracking-by-detection framework is still necessary.
In mathematics, based on the recognized sleeves status and OpenPose-predicted key body points, the skin region R s and the clothing-covered region R c are determined, both of which are a set of pixel coordinates ( x , y ) in the image plane like Equation (1) and (2).
R s = ( x t 1 s , y t 1 s ) , ( x t 2 s , y t 2 s ) , , ( x t + 1 m s , y t + 1 m s ) , , ( x t + i t v 1 n s , y t + i t v 1 n s )
R c = ( x t 1 c , y t 1 c ) , ( x t 2 c , y t 2 c ) , , ( x t + 1 m c , y t + 1 m c ) , , ( x t + i t v 1 n c , y t + i t v 1 n c )
In the equations, the subscript (t, t + 1 , t + i t v 1 ) refers to the index of each frame within a time period of i t v frames; the superscript ( 1 s , 2 s , m s , n s , 1 c , 2 c , m c , n c ) refers to the index of each detected key point. So in the consecutive i t v frames there are n s and n c key points detected in R s and R c , respectively.
The thermal camera we use is Xenics Gobi-384-GigE that can visualize a thermography of the scene it captures and measure the temperature of each pixel within the image with an accurate resolution of 0.08 °C. Therefore, temperatures of the detected key points ( T 1 s , T 2 s ,..., T n s ) in R s and ( T 1 c , T 2 c ,…, T n c ) in R c are easily read from the camera. Then an average calculation of the temperature values ( T 1 s , T 2 s ,…, T n s ) and ( T 1 c , T 2 c ,…, T n c ) gets T s and T c , respectively.
As long as T s and T c of each individual are calculated, the person’s I c l can be estimated by:
I c l = 1 0.155 · h · T s T c T c T o
where h equals to 8.6 referring to human’s heat transfer coefficient; T o is the operative temperature considering both the air temperature and the mean radiation temperature, so here it is calculated by the average temperature of the background region in each frame. This calculation comes from [35] according to [10,61], and all the temperatures T s , T c , and T o are in degrees Celsius. We claim that our emphasis is the OpenPose strategy for localizing R s and R c to get T s and T c , based on which any I c l calculation method can be applied.

3.3. M Estimation

In this part, we first in Section 3.3.1 propose three vision-based features to represent each person’s activity intensity, based on which M is estimated in Section 3.3.2.

3.3.1. Three Vision-Based Features

Though M can be estimated by a person’s key posture or activity type listed in ISO standards [8,9,11,12] and updated databases [62,63], this is a rough estimation in many cases, since we have observed that different people tend to have different activity intensities for the same posture. For example, some people will do a bit of stretching when standing up while others may just stand still. Therefore, a more accurate and dynamic M estimation is expected. This is done by computing three vision-based features—a person’s bounding box changes in two aspects (location and scale) and the optical flow intensity within the bounding box, over a few seconds like 10 s (210 frames) in our case. Here, the choice of 10 s comes from an observation that it takes similar durations for a smart bracelet to monitor a user’s heartbeats and blood oxygen content—two human physiological signals indicating the M value. This three-feature idea is motivated by that: the bounding box location change captures the general body movement; the bounding box scale change captures the motion of limbs; the optical flow intensity within the box captures the subtle movement that the box changes may ignore.
To realize this, for the location change of a certain person’s bounding boxes during 10 s (210 frames), the center coordinates ( c x , c y ) of the person’s bounding box in each frame is drawn as a point in a 2D plane, and totally the 210 2D points form a cluster-shaped pattern. The more spread out the points are, the larger the general body movement is. The degree of spread can be approximated by fitting an ellipse to the cluster and then calculating the area of this ellipse. In mathematics, first, the covariance matrix of the vector V c x (composed of the horizontal coordinates of the 210 points) and the vector V c y (composed of the vertical coordinates of the 210 points) is computed, and then the two eigenvalues of the covariance matrix are computed, at last, the multiplication of these two eigenvalues represents the area of the ellipse.
For the scale change of a certain person’s bounding boxes, after translating the 210 bounding boxes from 210 frames, they will have the same center at the origin, and then the upper-right coordinates ( u x , u y ) of each bounding box represents its scale. Similarly, the 210 upper right points form a cluster in a 2D plane, and the area of the ellipse fitting to the cluster will represent the scale change across time. The larger the area, the larger movement of limbs.
When it comes to the optical flow intensity in a person’s bounding box from the tth to ( t + i t v 1 ) th frame ( i t v equals to 210 here), for each frame two optical flows in horizontal and vertical directions are extracted by the TV-L1 algorithm [64] realized in a tool called MMAction [65]. Each optical flow is saved as an 8-bit image in which pixels with a grayscale value of 127 represent no movement while these pixels with grayscale values farther away from 127 represent larger movements. Therefore, within a duration of i t v frames, a person’s optical flow intensity I x y is calculated by:
I x y = τ = t τ = t + i t v 1 I x y τ i t v
I x y τ = ( I x τ ) 2 + ( I y τ ) 2
I x τ = ( x , y ) b o x τ f h r z τ ( x , y ) 127 ( x , y ) b o x τ 1
I y τ = ( x , y ) b o x τ f v t c τ ( x , y ) 127 ( x , y ) b o x τ 1
where τ indicates the frame index; I x y τ is the person’s optical flow intensity in the τ th frame; I x τ and I y τ are the person’s optical flow intensity in the horizontal and vertical directions in the τ th frame, respectively; ( x , y ) is any pixel in the optical flow; b o x τ is the bounding box region of the person in the τ th frame; f h r z and f v t c mean the two optical flows in the horizontal and vertical directions, respectively. In Equations (6) and (7), the number of pixels in the bounding box is acted as the denominator to normalize the influence of the size of the box.
In this way, the three features (bounding box location change, bounding box scale change, optical flow intensity) representing an individual’s activity intensity are acquired. A visualization showing the bounding box location change by a cluster of 210 2D points/circles, the bounding box scale change also by a cluster of 210 2D points/circles, and the optical flow intensity within the bounding box in each frame from a duration of 210 frames are in Figure 5, in which ID 1 person is standing with very limited movements while ID 2 person is standing and stretching with large movements. This figure intuitively illustrates that the larger body movements of ID 2, the more spread out the points/circles in Figure 5d,f, and the larger optical flow intensity in Figure 5h.

3.3.2. M Estimation from the Three Features

In real life, persons may have various activities which are unrealistic to be analyzed accurately. However, for an office environment, staff usually have scheduled routines and thus relatively fixed behaviors. Generally, the sitting staff are typing the keyboard, reading, taking notes, sorting through files, chatting with colleagues, online meetings, etc. And the standing staff are also occupied by the same tasks but may be involved with some walking or body stretching. This prior knowledge is such important that it gives a metabolic rate range from which each individual’s M varies.
Therefore, with the above prior knowledge of standard office behaviors, by referring Table A.1 and Table A.2 in ISO 8996 [11], the CBE (Center for the Built Environment) thermal comfort tool [66], and the 2011 compendium of physical activities tables [63,67], the usual metabolic rate range of a sitting office staff is quite narrow from 58 W/m2 (1.0 MET) to 87 W/m2 (1.5 MET), while a standing staff’s metabolic rate usually varies from 75 W/m2 (1.3 MET) to 174 W/m2 (3.0 MET). According to the CBE thermal comfort tool, the slight M change of a sitting person within the range [58 W/m2, 87 W/m2] has a mild influence on his or her thermal sensation, while the M change within the much larger range of a standing person significantly influences the thermal feeling. This result inspires us to use a middle value of 72.5 W/m2 to represent a sitting office staff’s M for simplicity and generalization which also relieves the three-feature extraction for him or her, but we need to specifically define a standing person’s M from his or her dynamic activity intensity situation represented by the three vision-based features.
To map such features to a value of M, a classification idea is introduced. Similar to Table A.2 in ISO 8896 where metabolic rates from 55 W/m2 to more than 260 W/m2 are categorized into resting, low, moderate, high, and very high levels, we decide to categorize the metabolic rate of a standing office staff into low, moderate, and high levels. Specifically, a low level means standing with very limited movements or transient spontaneous movements (standing quietly in a line, reading, using a cellphone, normally chatting, etc.); a moderate level means standing with spontaneous but lasting movements (natural and small paces, limbs movements, head movements, discussing with gestures, etc.); a high level means standing with significant movements usually indicating intentional actions like sustained location changes by walking, constant trunk movements to stretch/relax the body, etc.
It is extremely important that the three levels do not mean there are only three options for the M value. Instead, for a person’s activity intensity, there are three classification probabilities P l , P m , and P h indicating the possibilities of being viewed as low, moderate, and high level, respectively. Based on P l , P m , and P h , the person’s final M is estimated by:
M = P l · M l + P m · M m + P h · M h
where M l , M m , and M h are the lower boundary, the middle value, and the upper boundary of a standing person’s M, that are, 75 W/m2, 125 W/m2, and 174 W/m2, respectively.
To realize this solution, the classification probabilities P l , P m , and P h are in need. With only three features describing a person’s activity intensity within a few seconds as the input, a simple and flexible classification model instead of a CNN can be used. So, in this study, several lightweight models are employed and the random forest model works best. The training and testing details are in Section 4.4.
In summary, the proposed M estimation method has several advantages: (i) the three explicitly-extracted features can guide the metabolic rate estimation efficiently, considering that the features automatically extracted by a learning method are relatively difficult to anticipate and thus may potentially fail for a specific task; (ii) the three features are really low dimensional, making it possible to use lightweight machine learning classifiers which are flexible to be integrated into the whole system; (iii) the probability-weighted summation (Equation (8)) makes the estimated M continuously change in a range, which not only fits the real-life scenario than limited and discrete choices in existing methods but also avoids the very difficult annotation if a regression model is adopted.

4. Experiments

In this part, we first introduce the information of the dataset we collected from a multi-person environment, and then the proposed tracking-by-detection module, I c l estimation module, and M estimation module are evaluated.

4.1. Dataset Information

There is no available public dataset for visual analysis of I c l and M in a multi-person environment. We, therefore, collected such a dataset in December 2020 in Denmark. During the collection, two persons were sitting or standing with different types of clothes in a typical office environment where the indoor temperature and humidity were 22 °C and 32%, and they were encouraged to behave naturally. That means, typing the keyboard, texting with cellphones, chatting with each other, reading, stretching the body to relax, and others were captured in the collected videos. The horizontal distance between the camera and persons is around 3.5 meters, and the vertical distance between the camera and the ground is around 2.7 meters. In this way, ten subjects contributed to 114 videos with each video’s length about 2000 frames by using a thermal camera (Xenics Gobi-384-GigE whose sensor size is 384 × 288 ).

4.2. Evaluation of the Tracking-by-Detection Module

The tracking-by-detection (DeepSort-by-YOLOv5) module needs a well-trained human detector to detect persons in six categories mentioned before in Table 3. To train YOLOv5, from the dataset we sampled one frame every 50 frames for annotation and thus 5263 frames are selected in which each person’s bounding box and category are labeled. These 5263 images are then divided into a training set (4467), validation set (362), and testing set (434) to guarantee that subjects in the testing set never exist in the training set and validation set for a fair evaluation. Additionally, we selected and labeled 832 images from a single-person thermal dataset from [34] to increase the amount and diversity of the training set. The detailed information of the data to train and evaluate YOLOv5 is listed in Table 4. Accordingly, the 15 videos from which the 434 testing images are sampled are used to evaluate the whole DeepSort-by-YOLOv5 framework.
With a desktop equipped with Windows 10, CUDA (Compute Unified Device Architecture) 10.2, Pytorch 1.7.1, and one NVIDIA 2080Ti GPU (Graphics Processing Unit) card, the YOLOv5m version [41] is finetuned with the learning rate 0.0075 and stops at the 200th epoch at which the training loss is not decreasing any more. Other settings remain the same with the released YOLOv5m. The best model on the validation set is performed on the testing set and then achieves a mAP50 (Mean Average Precision) of 89.1% over six categories. Specifically, the AP50 rates of LongSit, ShortSit, OclSit, LongStand, ShortStand, and OclStand are 98.8%, 90.0%, 95.5%, 98.5%, 99.5%, and 52.5%, respectively. The AP50 drop in OclStand is due to the data imbalance problem. There are less than 300 images having OclStand persons in the training set, and there are only two images having Oclstand persons in the testing set (see Figure 6). In Figure 6, persons with bounding box SSD (ShortStand), SS(ShortSit), and OSD (OclStand) are categorized correctly, while the one with box LSD (LongStand) is categorized wrongly since the person’s sleeve status is unknown and thus should have been recognized as OclStand (OSD).
With the same hardware and software platforms, DeepSort-by-YOLOv5 runs on the 15 testing videos without further fine-tuning of the tracker itself. There are a total of 44,077 ground truth persons, 206 false negatives, 16 false positives, and 0 ID-switch in the 15 videos, which achieves an average MOTA (Multiple Object Tracking Accuracy) of 99.5% and the lowest MOTA of an individual video is 93.7%. Figure 7 shows four sampled tracking results. The eight persons from left to right in Figure 7 are in category ShortSit, LongStand, ShortSit, LongStand, ShortSit, OclSit, ShortStand, and ShortStand, respectively. Figure 7a,b are near frames from a video, and both persons are well tracked though the person with ID 2 is moving intensely. The false negative in Figure 7c is because there is no similar situation in the training set that a person is occluded so severely. The mug with hot coffee in Figure 7d has a similar temperature distribution as humans, which leads to the false positive.
In summary, the proposed DeepSort-by-YOLOv5 module achieves a mAP50 rate of 89.1% and a MOTA rate of 99.5% on the testing data. As this is the first work on multi-person analysis in terms of clothing and activity status recognition for thermal comfort, a direct comparison with other works is not possible. Instead, we refer to the latest performance of human detection/tracking on other thermal databases as an indirect comparison. Work [51] shows that the mAP50 values are from 62.0% to 96.0% on benchmark databases with different difficulties like OSU, KAIST, VOT-TIR2015, etc. Work [68] shows that the MOTA values are from 54.3% to 64.9% with different trackers on SCUT-FIR pedestrian dataset. These reference results indicate that our results are good enough and thus the proposed method can be included in a real application.

4.3. Evaluation of the I c l Estimation Module

The I c l estimation closely depends on the skin temperature T s and clothes temperature T c acquisition, which is bridged by the localization of skin region R s and clothing-covered region R c via OpenPose. Therefore, this evaluation first looks at the efficacy of applying OpenPose to our dataset.
4901 images are used to examine OpenPose’s performance. These 4901 images come from the 5263 annotated images for YOLOv5 but do not include the images where persons are wearing masks due to coronavirus restrictions. Such an evaluation set is evenly sampled from the 114 collected videos, guaranteeing comprehensiveness and fairness. The evaluation protocols are: (i) the OpenPose tool is not finetuned with our thermal dataset, and the confidence threshold is set as 0.6 as mentioned in Section 3.2; (ii) only these key points that influence R s and R c localization are checked, i.e., nose, shoulders, elbows, and wrists; (iii) any frame with even only one wrongly detected key body point is counted as one error frame, to make the evaluation strict and conservative.
After a frame-by-frame check, there are 187 error frames out of the whole 4901 frames, indicating an accuracy of 96.2%. We found that there are two types of representative errors—nose detected in the hair region due to a lowering head (Figure 8a) and nose detected in the background region due to a turned side face (Figure 8b). The good point is that with the average computation within a few minutes to get T s and T c , the influence of these errors can be eliminated effectively, and of course, a higher confidence threshold can further reduce such errors if needed.
Therefore, the efficacy of applying OpenPose to our multi-person thermal scenario to locate R s and R c is verified. The performance surpasses that of applying OpenPose to a controlled single-person thermal environment [35] and applying OpenPose to RGB MPII dataset [69], further proving the feasibility of our strategy relying on OpenPose.
Based on the above acquired R s and R c , here we calculate the T s and T c , and then estimate the I c l value. Since an individual I c l estimation also involves the human tracking part, we use the testing videos for the tracking module to evaluate this I c l estimation module too. From the testing videos, a female wearing a lightweight T-shirt is acting as the subject to be researched, because there is an available reference for her clothes type in the ISO tables so that we can make a comparison. And thus, two videos including various situations where the female is sitting, standing, reading, writing, typing the keyboard, chatting, and drinking coffee (some frames are shown in Figure 9) go through our methodology pipeline to get her I c l . In one video consisting of 1477 frames (70 seconds), 3326 skin points and 2849 clothes points are detected for the female, from which the T s and T c are calculated as 34.67 °C and 33.32 °C, respectively. Together with the T o as 24.96 °C, the female’s I c l is estimated as 0.1220 clo. In the other video of 1536 frames (73 seconds), 2496 skin points and 2502 clothes points are detected for the female; the resultant T s is 34.73 °C and T c is 33.48 °C; together with the T o as 25.58 °C, the female’s I c l is estimated as 0.1182 clo.
From above calculation, we find that: (i) within a time period like more than 1 minute, the accumulated detected points in R s and R c are way enough for an accurate T s and T c calculation as the potential noises can be filtered out efficiently; (ii) the estimated I c l values of 0.1220 clo and 0.1182 clo are quite similar, revealing the stability and robustness of the method; (iii) the reference value of the female’s I c l is 0.09 clo to 0.15 clo from Table B.1 in ISO 9920 [10], showing the consistency of our method with the international standards, and proving the feasibility of the proposed method.

4.4. Evaluation of the M Estimation Module

This subsection evaluates the effectiveness of the M estimation based on the three extracted vision features, specifically for a standing person. As this estimation is a probability-weighted summation, measuring the accuracy of the classifier is the key.
Therefore, by dividing the 114 collected videos into small clips of 10 seconds and then extracting the three vision features for each standing person in these clips, 315 sets of the three features are used as the training data to help the classifier learn the ability to category each person’s activity intensity into low, moderate, or high level, and another 68 sets are used as the testing data to evaluate the classifier’s performance.
During the phase of preparing the training and testing data—annotating a standing person’s activity intensity level, we met another dilemma that frequently happens in the real world—there are always the situations where a person’s movement is mixed with transient, lasting, mild, or intensive movements within a short period which makes it very difficult to label the intensity level. Therefore, these difficult cases are not included in the training/testing sets to not confuse the classifier. From the positive side, this situation further indicates the strength of our probability-weighted summation strategy that makes the estimated M a continuous value.
To avoid being one-sided, three widely-used classifiers—KNN (K-NearestNeighbor), SVM (Support Vector Machine), and RF (Random Forest) are used. The parameters and performances of the three classifiers are listed in Table 5, in which each parameter is tuned by grid searching using the training data and the meaning of each parameter is explained in the scikit-learn library [70]. These accuracy values in Table 5 prove that the three features are good representations of a person’s activity intensity, and thus the M estimation from them by a classifier’s probability-weighted summation is also reasonable. And then we decide to use RF as the classifier for M estimation due to its best performance on the testing data.
Based on RF’s classification probabilities P l , P m , and P h , by Equation (8), the M values of a same standing person with two totally different activity intensities are estimated. The person is shown in Figure 10, in which Figure 10a is a frame from a clip where the standing person is normally chatting with many gestures, and Figure 10b is a frame from another clip where the standing person is stretching his body like doing Pilates. For them, our method outputs the estimated M values of 99 W/m2 and 170 W/m2, respectively, which are very similar to the reference values of 104 W/m2 (CODE 09050 in [67]) and 174 W/m2 (CODE 02105 in [67]), further proving the feasibility and usability of the proposed M estimation module.

4.5. Application in Thermal Comfort Assessment

From all the above evaluations, the proposed method indeed has the ability to estimate individual I c l and M across time for each person in a room. With these two dynamic personal factors and the other four environmental factors easily measured from sensors, a thermal comfort model like Fanger’s model [6,7] can calculate individual thermal comfort sensation to see if the person feels hot, cold, or satisfied with the indoor environment. Although occupants may have different thermal feelings at the same time, by regulating the indoor microclimate in separate local regions, it is possible to achieve varied thermal conditions that respond to the different subjective thermal states. Moreover, the used thermal camera instead of an RGB camera, the computation in a local device, and the erasing function of captured image information as long as I c l and M are estimated will make the whole processing pipeline privacy-friendly.

5. Conclusions and Future Work

This paper proposes a contactless method to estimate each person’s clothing insulation rate I c l and metabolic rate M dynamically by use of a thermal camera, in an uncontrolled multi-person indoor environment.
Specifically, the method composes of a tracking-by-detection (DeepSort-by-YOLOv5) module to track each person and recognize his or her clothing status and key posture simultaneous, a key body points detection module to measure the skin temperature and clothes temperature for I c l estimation, and a random forest classifier module to categorize each individual’s activity intensity into different levels for M estimation. All three modules are evaluated with a new multi-person thermal dataset, verifying that the methodology is robust to be applied in real-life applications for individual thermal comfort assessment.
The future work will be to include this research into such an application to facilitate thermal comfort control systems for lower energy waste and higher working comfort in an office building.

Author Contributions

Conceptualization, J.L., I.W.F. and T.B.M.; methodology, J.L. and T.B.M.; software, J.L.; validation, J.L.; formal analysis, J.L.; investigation, J.L.; resources, J.L.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, I.W.F. and T.B.M.; visualization, J.L.; supervision, T.B.M.; project administration, I.W.F. and T.B.M.; funding acquisition, I.W.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Realdania Foundation and The Obel Family Foundation as part of the project Thermal Adaptive Architecture.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

To protect the facial information of some participators, only a partial dataset is available on request from the corresponding author.

Acknowledgments

We would like to thank all the participators for the data collection from Department of Architecture, Design, and Media Technology, Aalborg University, Denmark.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
mAPMean Average Precision
MOTAMultiple Object Tracking Accuracy
HVACHeaters, ventilation, and air conditioners
CNNConvolutional Neural Network
ISOInternational Organization for Standardization
R-CNNRegion-based Convolutional Neural Network
YOLOYou Only Look Once
FPNFeature Pyramid Network
RGBRed Green Blue
SORTSimple Online and Realtime Tracking
DeepSortSimple Online and Realtime Tracking with a Deep Association Metric
CDA_DDALConfidence-based Data Association and Discriminative Deep Appearance Learning
IDIdentity
PANetPath Aggregation Network
FNFalse Negative
FPFalse Positive
TV-L1Total Variation L1 Norm
CBECenter for the Built Environment
CUDACompute Unified Device Architecture
GPUGraphics Processing Unit
KNNK-NearestNeighbor
SVMSupport Vector Machine
RFRandom Forest

References

  1. EIA. 2012 Commercial Buildings Energy Consumption Survey Data. Available online: https://www.eia.gov/consumption/commercial/data/2012/ (accessed on 15 August 2021).
  2. Atalie, D.; Tesinova, P.; Tadesse, M.G.; Ferede, E.; Dulgheriu, I.; Loghin, E. Thermo-Physiological Comfort Properties of Sportswear with Different Combination of Inner and Outer Layers. Materials 2021, 14, 6863. [Google Scholar] [CrossRef]
  3. Ullah, H.M.K.; Lejeune, J.; Cayla, A.; Monceaux, M.; Campagne, C.; Devaux, É. A review of noteworthy/major innovations in wearable clothing for thermal and moisture management from material to fabric structure. Text. Res. J. 2021, 00405175211027799. [Google Scholar] [CrossRef]
  4. Peng, Y.; Sun, F.; Xiao, C.; Iqbal, M.I.; Sun, Z.; Guo, M.; Gao, W.; Hu, X. Hierarchically Structured and Scalable Artificial Muscles for Smart Textiles. ACS Appl. Mater. Interfaces 2021, 13, 54386–54395. [Google Scholar] [CrossRef]
  5. Zhu, F.; Feng, Q. Recent advances in textile materials for personal radiative thermal management in indoor and outdoor environments. Int. J. Therm. Sci. 2021, 165, 106899. [Google Scholar] [CrossRef]
  6. Fanger, P.O. Thermal comfort. Analysis and applications in environmental engineering. In Thermal Comfort. Analysis and Applications in Environmental Engineering; Danish Technical Press: Copenhagen, Denmark, 1970. [Google Scholar]
  7. Fanger, P.O. Assessment of man’s thermal comfort in practice. Occup. Environ. Med. 1973, 30, 313–324. [Google Scholar] [CrossRef] [Green Version]
  8. The International Organization for Standardization. Ergonomics of the Thermal Environment—Analytical Determination and Interpretation of Thermal Comfort Using Calculation of the PMV and PPD Indices and Local Thermal Comfort Criteria. Available online: https://www.iso.org/standard/39155.html/ (accessed on 1 September 2021).
  9. The International Organization for Standardization. Ergonomics of the Thermal Environment—Analytical Determination and Interpretation of Heat Stress Using Calculation of the Predicted Heat Strain. Available online: https://www.iso.org/standard/37600.html/ (accessed on 1 September 2021).
  10. The International Organization for Standardization. Ergonomics of the Thermal Environment—Estimation of Thermal Insulation and Water Vapour Resistance of a Clothing Ensemble. Available online: https://www.iso.org/standard/39257.html/ (accessed on 1 September 2021).
  11. The International Organization for Standardization. Ergonomics of the Thermal Environment—Determination of Metabolic Rate. Available online: https://www.iso.org/standard/34251.html/ (accessed on 1 September 2021).
  12. ASHRAE. Thermal Environmental Conditions for Human Occupancy. Available online: https://www.ashrae.org/technical-resources/bookstore/standard-55-thermal-environmental-conditions-for-human-occupancy (accessed on 1 September 2021).
  13. De Carli, M.; Olesen, B.W.; Zarrella, A.; Zecchin, R. People’s clothing behaviour according to external weather and indoor environment. Build. Environ. 2007, 42, 3965–3973. [Google Scholar] [CrossRef]
  14. Ngarambe, J.; Yun, G.Y.; Kim, G. Prediction of indoor clothing insulation levels: A deep learning approach. Energy Build. 2019, 202, 109402. [Google Scholar] [CrossRef]
  15. de Carvalho, P.M.; da Silva, M.G.; Ramos, J.E. Influence of weather and indoor climate on clothing of occupants in naturally ventilated school buildings. Build. Environ. 2013, 59, 38–46. [Google Scholar] [CrossRef]
  16. Liu, W.; Yang, D.; Shen, X.; Yang, P. Indoor clothing insulation and thermal history: A clothing model based on logistic function and running mean outdoor temperature. Build. Environ. 2018, 135, 142–152. [Google Scholar] [CrossRef]
  17. Matsumoto, H.; Iwai, Y.; Ishiguro, H. Estimation of Thermal Comfort by Measuring Clo Value without Contact; MVA; Citeseer: Princeton, NJ, USA, 2011; pp. 491–494. [Google Scholar]
  18. Konarska, M.; Soltynski, K.; Sudol-Szopinska, I.; Chojnacka, A. Comparative evaluation of clothing thermal insulation measured on a thermal manikin and on volunteers. Fibres Text. East. Eur. 2007, 15, 73. [Google Scholar]
  19. Lu, S.; Cochran Hameen, E. Integrated IR Vision Sensor for Online Clothing Insulation Measurement. In Proceedings of the 23rd Annual Conference of the Association for Computer-Aided Architectural Design Research in Asia, Beijing, China, 17–19 May 2018. [Google Scholar]
  20. Lee, J.H.; Kim, Y.K.; Kim, K.S.; Kim, S. Estimating clothing thermal insulation using an infrared camera. Sensors 2016, 16, 341. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Lee, K.; Choi, H.; Kim, H.; Kim, D.D.; Kim, T. Assessment of a real-time prediction method for high clothing thermal insulation using a thermoregulation model and an infrared camera. Atmosphere 2020, 11, 106. [Google Scholar] [CrossRef] [Green Version]
  22. Choi, H.; Na, H.; Kim, T.; Kim, T. Vision-based estimation of clothing insulation for building control: A case study of residential buildings. Build. Environ. 2021, 202, 108036. [Google Scholar] [CrossRef]
  23. Luo, M.; Zhou, X.; Zhu, Y.; Sundell, J. Revisiting an overlooked parameter in thermal comfort studies, the metabolic rate. Energy Build. 2016, 118, 152–159. [Google Scholar] [CrossRef]
  24. Zhai, Y.; Li, M.; Gao, S.; Yang, L.; Zhang, H.; Arens, E.; Gao, Y. Indirect calorimetry on the metabolic rate of sitting, standing and walking office activities. Build. Environ. 2018, 145, 77–84. [Google Scholar] [CrossRef]
  25. Ji, W.; Luo, M.; Cao, B.; Zhu, Y.; Geng, Y.; Lin, B. A new method to study human metabolic rate changes and thermal comfort in physical exercise by CO2 measurement in an airtight chamber. Energy Build. 2018, 177, 402–412. [Google Scholar] [CrossRef]
  26. Calvaresi, A.; Arnesano, M.; Pietroni, F.; Revel, G.M. Measuring metabolic rate to improve comfort management in buildings. Environ. Eng. Manag. J. (EEMJ) 2018, 17, 2287–2296. [Google Scholar]
  27. Hasan, M.H.; Alsaleem, F.; Rafaie, M. Sensitivity study for the PMV thermal comfort model and the use of wearable devices biometric data for metabolic rate estimation. Build. Environ. 2016, 110, 173–183. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Zhou, X.; Zheng, Z.; Oladokun, M.O.; Fang, Z. Experimental investigation into the effects of different metabolic rates of body movement on thermal comfort. Build. Environ. 2020, 168, 106489. [Google Scholar] [CrossRef]
  29. Lee, J.; Ham, Y. Physiological sensing-driven personal thermal comfort modelling in consideration of human activity variations. Build. Res. Inf. 2021, 49, 512–524. [Google Scholar]
  30. Gilani, S.I.u.H.; Khan, M.H.; Ali, M. Revisiting Fanger’s thermal comfort model using mean blood pressure as a bio-marker: An experimental investigation. Appl. Therm. Eng. 2016, 109, 35–43. [Google Scholar] [CrossRef]
  31. Jensen, M.M.; Poulsen, M.K.; Alldieck, T.; Larsen, R.G.; Gade, R.; Moeslund, T.B.; Franch, J. Estimation of energy expenditure during treadmill exercise via thermal imaging. Med. Sci. Sport. Exerc. 2016, 48, 2571–2579. [Google Scholar] [CrossRef] [Green Version]
  32. Gade, R.; Godsk Larsen, R.; Moeslund, T.B. Measuring energy expenditure in sports by thermal video analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 131–138. [Google Scholar]
  33. Na, H.; Choi, H.; Kim, T. Metabolic rate estimation method using image deep learning. In Building Simulation; Springer: Berlin/Heidelberg, Germany, 2020; Volume 13, pp. 1077–1093. [Google Scholar]
  34. Liu, J.; Foged, I.W.; Moeslund, T.B. Vision-Based Individual Factors Acquisition for Thermal Comfort Assessment in a Built Environment. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 18–22 May 2020; pp. 662–666. [Google Scholar]
  35. Liu, J.; Foged, I.W.; Moeslund, T.B. Automatic estimation of clothing insulation rate and metabolic rate for dynamic thermal comfort assessment. Pattern Anal. Appl. 2021, 1–16. [Google Scholar] [CrossRef]
  36. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  38. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  39. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  40. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  41. Ultralytics. YOLOv5. Available online: https://github.com/ultralytics/yolov5/ (accessed on 5 March 2021).
  42. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  43. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  44. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  45. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  46. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  47. Bae, S.H.; Yoon, K.J. Confidence-based data association and discriminative deep appearance learning for robust online multi-object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 595–610. [Google Scholar] [CrossRef]
  48. Zhao, D.; Fu, H.; Xiao, L.; Wu, T.; Dai, B. Multi-object tracking with correlation filter for autonomous vehicle. Sensors 2018, 18, 2004. [Google Scholar] [CrossRef] [Green Version]
  49. Zhang, W.; Zhou, H.; Sun, S.; Wang, Z.; Shi, J.; Loy, C.C. Robust multi-modality multi-object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 2365–2374. [Google Scholar]
  50. Lu, Z.; Rathod, V.; Votel, R.; Huang, J. Retinatrack: Online single stage joint detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 14668–14678. [Google Scholar]
  51. Krišto, M.; Ivasic-Kos, M.; Pobar, M. Thermal Object Detection in Difficult Weather Conditions Using YOLO. IEEE Access 2020, 8, 125459–125476. [Google Scholar] [CrossRef]
  52. Huda, N.U.; Hansen, B.D.; Gade, R.; Moeslund, T.B. The Effect of a Diverse Dataset for Transfer Learning in Thermal Person Detection. Sensors 2020, 20, 1982. [Google Scholar] [CrossRef] [Green Version]
  53. Liu., J.; Philipsen., M.; Moeslund., T. Supervised versus Self-supervised Assistant for Surveillance of Harbor Fronts. In Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications–Volume 5: VISAPP, INSTICC; SciTePress: Setúbal, Portugal, 2021; pp. 610–617. [Google Scholar]
  54. Yao, Y.; Lian, Z.; Liu, W.; Jiang, C. Measurement methods of mean skin temperatures for the PMV model. HVAC&R Res. 2008, 14, 161–174. [Google Scholar]
  55. Dai, C.; Zhang, H.; Arens, E.; Lian, Z. Machine learning approaches to predict thermal demands using skin temperatures: Steady-state conditions. Build. Environ. 2017, 114, 1–10. [Google Scholar] [CrossRef] [Green Version]
  56. Cosma, A.C.; Simha, R. Thermal comfort modeling in transient conditions using real-time local body temperature extraction with a thermographic camera. Build. Environ. 2018, 143, 36–47. [Google Scholar] [CrossRef]
  57. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  58. Petersson, J.; Halder, A. Updated Database of Clothing Thermal Insulation and Vapor Permeability Values of Western Ensembles for Use in ASHRAE Standard 55, ISO 7730, and ISO 9920. ASHRAE Trans. 2021, 127, 773–799. [Google Scholar]
  59. Tang, Y.; Su, Z.; Yu, H.; Zhang, K.; Li, C.; Ye, H. A database of clothing overall and local insulation and prediction models for estimating ensembles’ insulation. Build. Environ. 2022, 207, 108418. [Google Scholar] [CrossRef]
  60. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.E.; Sheikh, Y. OpenPose: Realtime multi-person 2D pose estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 172–186. [Google Scholar] [CrossRef] [Green Version]
  61. Miura, J.; Demura, M.; Nishi, K.; Oishi, S. Thermal comfort measurement using thermal-depth images for robotic monitoring. Pattern Recognit. Lett. 2020, 137, 108–113. [Google Scholar] [CrossRef]
  62. Ainsworth, B.E.; Haskell, W.L.; Whitt, M.C.; Irwin, M.L.; Swartz, A.M.; Strath, S.J.; O Brien, W.L.; Bassett, D.R.; Schmitz, K.H.; Emplaincourt, P.O.; et al. Compendium of physical activities: An update of activity codes and MET intensities. Med. Sci. Sport. Exerc. 2000, 32, S498–S504. [Google Scholar] [CrossRef] [Green Version]
  63. Ainsworth, B.E.; Haskell, W.L.; Herrmann, S.D.; Meckes, N.; Bassett, D.R.; Tudor-Locke, C.; Greer, J.L.; Vezina, J.; Whitt-Glover, M.C.; Leon, A.S. 2011 Compendium of Physical Activities: A second update of codes and MET values. Med Sci Sport. Exerc 2011, 43, 1575–1581. [Google Scholar] [CrossRef] [Green Version]
  64. Zach, C.; Pock, T.; Bischof, H. A duality based approach for realtime tv-l 1 optical flow. In Joint Pattern Recognition Symposium; Springer: Berlin/Heidelberg, Germany, 2007; pp. 214–223. [Google Scholar]
  65. Zhao, Y.; Xiong, Y.; Lin, D. MMAction. Available online: https://github.com/open-mmlab/mmaction (accessed on 5 March 2021).
  66. Tartarini, F.; Schiavon, S.; Cheung, T.; Hoyt, T. CBE Thermal Comfort Tool: Online tool for thermal comfort calculations and visualizations. SoftwareX 2020, 12, 100563. [Google Scholar] [CrossRef]
  67. Ainsworth, B.E.; Haskell, W.L.; Herrmann, S.D.; Meckes, N.; Bassett, D.R.; Tudor-Locke, C.; Greer, J.L.; Vezina, J.; Whitt-Glover, M.C.; Leon, A.S. 2011 Compendium of Physical Activities Tables. Available online: https://sites.google.com/site/compendiumofphysicalactivities/compendia (accessed on 10 September 2021).
  68. Chen, H.; Cai, W.; Wu, F.; Liu, Q. Vehicle-mounted far-infrared pedestrian detection using multi-object tracking. Infrared Phys. Technol. 2021, 115, 103697. [Google Scholar] [CrossRef]
  69. Andriluka, M.; Pishchulin, L.; Gehler, P.; Schiele, B. 2d human pose estimation: New benchmark and state of the art analysis. In Proceedings of the IEEE Conference on computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3686–3693. [Google Scholar]
  70. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Figure 1. Overview of the proposed method. The numbers 1 and 2 are the corresponding tracking ID numbers of the two persons.
Figure 1. Overview of the proposed method. The numbers 1 and 2 are the corresponding tracking ID numbers of the two persons.
Sensors 22 00619 g001
Figure 2. Persons difficult to predict clothing type due to occlusion. They are pointed by the green arrows. (a) The right person is partly occluded by the monitor. (b) The right person is partly occluded by the monitor. (c) The right person moves the arms out of the scene. (d) The left person hides the arms behind the torso.
Figure 2. Persons difficult to predict clothing type due to occlusion. They are pointed by the green arrows. (a) The right person is partly occluded by the monitor. (b) The right person is partly occluded by the monitor. (c) The right person moves the arms out of the scene. (d) The left person hides the arms behind the torso.
Sensors 22 00619 g002
Figure 3. Skin region R s and clothing-covered region R c . R s in green crosses and R c in red crosses. (ad) illustrate persons doing different tasks in different poses.
Figure 3. Skin region R s and clothing-covered region R c . R s in green crosses and R c in red crosses. (ad) illustrate persons doing different tasks in different poses.
Sensors 22 00619 g003
Figure 4. Detected key body points by OpenPose with different confidence thresholds, (a) threshold of 0.1, (b) threshold of 0.3, (c) threshold of 0.5, (d) threshold of 0.7. For better visualization, each key point is the end of the colorful line segment.
Figure 4. Detected key body points by OpenPose with different confidence thresholds, (a) threshold of 0.1, (b) threshold of 0.3, (c) threshold of 0.5, (d) threshold of 0.7. For better visualization, each key point is the end of the colorful line segment.
Sensors 22 00619 g004
Figure 5. Using bounding box changes in location and scale, and the optical flow in the bounding box to represent an individual’s activity intensity. (a,b) ID 1 person is with small movements and ID 2 person is with large movements. (c,d) Bounding box location change of ID 1 person and ID 2 person, respectively. (e,f) Bounding box scale change of ID 1 person and ID 2 person, respectively. (g,h) Optical flow intensity change of ID 1 person and ID 2 person, respectively.
Figure 5. Using bounding box changes in location and scale, and the optical flow in the bounding box to represent an individual’s activity intensity. (a,b) ID 1 person is with small movements and ID 2 person is with large movements. (c,d) Bounding box location change of ID 1 person and ID 2 person, respectively. (e,f) Bounding box scale change of ID 1 person and ID 2 person, respectively. (g,h) Optical flow intensity change of ID 1 person and ID 2 person, respectively.
Sensors 22 00619 g005
Figure 6. Detection results on two test images with OclStand (OSD) persons in them. (a) The right person is wrongly categorized as LongStand (LSD). (b) Both persons are detected and categorized correctly.
Figure 6. Detection results on two test images with OclStand (OSD) persons in them. (a) The right person is wrongly categorized as LongStand (LSD). (b) Both persons are detected and categorized correctly.
Sensors 22 00619 g006
Figure 7. Sampled tracking results on the testing set, no false positive and false negative in (a,b), one false negativeFN in (c), one false positive in (d). The numbers indicate tracked ID numbers.
Figure 7. Sampled tracking results on the testing set, no false positive and false negative in (a,b), one false negativeFN in (c), one false positive in (d). The numbers indicate tracked ID numbers.
Sensors 22 00619 g007
Figure 8. Two representative error frames with red arrows pointing to the wrongly detected noses. (a) Nose is detected in the hair region. (b) Nose is detected in the background region.
Figure 8. Two representative error frames with red arrows pointing to the wrongly detected noses. (a) Nose is detected in the hair region. (b) Nose is detected in the background region.
Sensors 22 00619 g008
Figure 9. The female to be researched is pointed by the green arrow, (a) chatting, (b) reading, (c) chatting with gestures, (d) drinking.
Figure 9. The female to be researched is pointed by the green arrow, (a) chatting, (b) reading, (c) chatting with gestures, (d) drinking.
Sensors 22 00619 g009
Figure 10. The standing person to be researched for M estimation is pointed by the green arrow. (a) normally chatting but with many gestures, (b) stretching body like doing Pilates.
Figure 10. The standing person to be researched for M estimation is pointed by the green arrow. (a) normally chatting but with many gestures, (b) stretching body like doing Pilates.
Sensors 22 00619 g010
Table 1. Insulation values of various typical garments [10].
Table 1. Insulation values of various typical garments [10].
Garment I cl (clo)
UnderwearSinglet0.04
T-shirt0.09
Shirts with long sleeves0.12
Shirts, blousesShort sleeves0.15
Lightweight, long sleeves0.2
Normal, long sleeves0.25
Table 2. Metabolic rates of typical activities [8].
Table 2. Metabolic rates of typical activities [8].
ActivityM (W/m2)
Reclining46
Seated, relaxed58
Sedentary activity70
Standing, light activity93
Standing, medium activity116
Walking on level ground:
  2 km/h110
  3 km/h140
  4 km/h165
  5 km/h200
Table 3. Persons in six categories.
Table 3. Persons in six categories.
CategoryMeaning
LongSitLong-sleeved clothes, sitting
ShortSitShort-sleeved clothes, sitting
OclSitDifficult to predict clothes type due to occlusion, sitting
LongStandLong-sleeved clothes, standing
ShortStandShort-sleeved clothes, standing
OclStandDifficult to predict clothes type due to occlusion, standing
Table 4. Detailed information of the data to train and evaluate YOLOv5.
Table 4. Detailed information of the data to train and evaluate YOLOv5.
Number of ImagesNumber of Persons
LongSitShortSitOclSitLongStandShortStandOclStand
Training52992099161582822802735254
Validation362172292741401009
Testing43422157921494432
Table 5. The parameters and performances of the used three classifiers.
Table 5. The parameters and performances of the used three classifiers.
ParametersTraining AccuracyTesting Accuracy
KNNmetric=‘manhattan’, weights= ‘distance’, n_neighbors=13100%92.7%
SVMC=50, kernel=‘rbf’, gamma=‘scale’83.5%88.2%
RFmax_depth=2, random_state=095.6%95.6%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, J.; Foged, I.W.; Moeslund, T.B. Clothing Insulation Rate and Metabolic Rate Estimation for Individual Thermal Comfort Assessment in Real Life. Sensors 2022, 22, 619. https://doi.org/10.3390/s22020619

AMA Style

Liu J, Foged IW, Moeslund TB. Clothing Insulation Rate and Metabolic Rate Estimation for Individual Thermal Comfort Assessment in Real Life. Sensors. 2022; 22(2):619. https://doi.org/10.3390/s22020619

Chicago/Turabian Style

Liu, Jinsong, Isak Worre Foged, and Thomas B. Moeslund. 2022. "Clothing Insulation Rate and Metabolic Rate Estimation for Individual Thermal Comfort Assessment in Real Life" Sensors 22, no. 2: 619. https://doi.org/10.3390/s22020619

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop