Literature DB >> 35671292

Improving work detection by segmentation heuristics pre-training on factory operations video.

Shotaro Kataoka1, Tetsuro Ito2, Genki Iwaka2, Masashi Oba2, Hirofumi Nonaka2.   

Abstract

The measurement of work time for individual tasks by using video has made a significant contribution to a framework for productivity improvement such as value stream mapping (VSM). In the past, the work time has been often measured manually, but this process is quite costly and labor-intensive. For these reasons, automation of work analysis at the worksite is needed. There are two main methods for computing spatio-temporal information: by 3D-CNN, and by temporal computation using LSTM after feature extraction in the spatial domain by 2D-CNN. These methods has high computational cost but high model representational power, and the latter has low computational cost but relatively low model representational power. In the manufacturing industry, the use of local computers to make inferences is often required for practicality and confidentiality reasons, necessitating a low computational cost, and so the latter, a lightweight model, needs to have improved performance. Therefore, in this paper, we propose a method that pre-trains the image encoder module of a work detection model using an image segmentation model. This is based on the CNN-LSTM structure, which separates spatial and temporal computation and enables us to include heuristics such as workers' body parts and work tools in the CNN module. Experimental results demonstrate that our pre-training method reduces over-fitting and provides a greater improvement in detection performance than pre-training on ImageNet.

Entities:  

Mesh:

Year:  2022        PMID: 35671292      PMCID: PMC9173629          DOI: 10.1371/journal.pone.0267457

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.752


1 Introduction

Recognition of employee activities in the production process is gaining attention [1-12]. Bell developed a conceptual framework about artificial intelligence data-driven Internet of Things systems [1]. Elvira et al carried out to evaluate and analyze artificial intelligence-supported workplace decision method such as big data algorithmic analytics, sensory and tracking technologies [2]. Ren et al carried out a comprehensive review of big data analytics throughout product lifecycle to support sustainable smart manufacturing [3]. Smith analyzes the outcomes of cyber-physical manufacturing systems such as real-time sensor networks, and Internet of Things smart devices [4]. Clarke draw on a substantial body of theoretical and empirical research on big data-driven manufacturing, and to explore this, I inspected, used, and replicated survey data from Accenture, Capgemini, PwC, Software AG, and we.CONECT, performing analyses and making estimates regarding sensing, smart, and sustainable technologies [5]. Nica et al performed analyses and made estimates regarding the link between smart connected sensors, industrial big data, and real-time process monitoring to examine cyber-physical system-based manufacturing [6]. Leng et al presents a digital twin-driven manufacturing cyber-physical system (MCPS) for parallel controlling of smart workshop under mass individualization paradigm [7]. Hyers et al. analyzed big data-driven decision-making processes, Industry 4.0 wireless networks, and digitized mass production to evaluate utility of cyber-physical system of smart factories [8]. Keane et al performed structural equation modeling to analyse cognitive automation, big data-driven manufacturing, and sustainable industrial value creation using data collected from BCG, CompTIA, Deloitte, IW Custom Research, Kronos, McKinsey, PAC, PwC, and Software AG [9]. Mircica et al. analyzed and estimated regarding the main areas of interest for manufacturers within Industry 4.0 and the effects of Industry 4.0 on the workforce [10]. Graessley et al performed analyses and making estimates regarding how Industry 4.0 is delivering revenue, cost and efficiency gains and top technologies being considered in-line with organizations’ strategic plan [11]. Meyers collected data from Bright & Company, Corporate Research Forum, Deloitte, Management Events, McKinsey, and Top Employers Institute and estimated data-driven approaches in industry [12]. Among them, employee behavior recognition in the manufacturing industry is attracting increasing attention from both industry and academia [13]. For example, to improve management and effectiveness of employees’ learning processes, Patalas-Maliszewska et al. [14] used CNN-SVM to extract features from video material concerning each work activity, and comparison with the features of the instruction picture. The importance of real-time process monitoring and analysis of employee behavior based on various sensors and other devices has been increasing from both industry and academia [15-18]. In such a situation, the measurement of work time for individual tasks by using video and other methods has made a significant contribution to a framework for productivity improvement such as value stream mapping (VSM) [19] which is a method for reviewing and improving productivity by dividing tasks into two categories—value-adding tasks and non-value-adding tasks—and by analyzing the time spent on the tasks [20-24]. Monteiro et al performed case study analysis of lean tool. In the case of metalworking company project, flowcharts and VSM (Value Stream Mapping) approaches were conducted to identify and map key processes and improved setup times. As a result, setup times were reduced in 40% on the vertical milling machine of the company, and in 57% on the horizontal milling machine [20]. UlfKTeichgraber et al. have empirically demonstrated that VSM can be applied to identify NVAs and improve productivity in the manufacturing process of endovascular stents. Specifically, they found that 5 out of 15 processes were unnecessary and could be eliminated by using VSM [21]. Mareike Heinzen et al. clarified the effect of co-location in the new drug development process through VSM. The results suggest that communication increases especially in child-located teams [22]. Heravi et al. used value stream mapping (VSM) as a lean method in the manufacturing process of prefabricated steel frames (PSF) and showed that 34% of the lead time could be reduced [23]. Wang et al. developed a training method combining VR and VSM, and showed that it can contribute to the efficiency of the construction process [24]. It has made a significant contribution to improve work efficiency and has been successfully introduced to a variety of industries [25-28]. SMED provides the most complete and detailed procedures available anywhere to transform the manufacturing environment in a way that speeds production and makes small lot inventory feasible [25]. Zhang et al. found that lean manufacturing contributes to knowledge creation through a questionnaire survey [26]. Sousa et al. applied VSM to improve production efficiency in the cork industry, and found that production time could be reduced by 43% [27]. Adanna et al. showed that VSM can reduce set-up time by 45% when applied to automotive equipment manufacturing [28]. An important part of implementing such a VSM is to measure work times in order to visualize them. For this purpose, video analysis is often used. For example, [29] used video to extract the cycle time and other factors that are important in analyzing the value added by a task. However, this type of work time analysis requires experts who are well versed in the work process and also requires considerable labor and involves economic costs because of the need to check the video footage during the work. Thus, the implementation hurdles are high, especially in small and medium-sized companies. For these reasons, automation of work analysis at the worksite is needed. In the mainstream research on this topic, sensors attached to the operator have traditionally been used for collecting measurement data. However, this method has certain problems, such as the burden of attaching a sensor to the worker, the resulting limitation on the worker’s flexibility of movement, and the high cost of introduction. Another approach is to detect action from videos; among such methods, convolutional neural network (CNN)–based methods are the most popular. There are two main methods for computing spatio-temporal information: by 3D-CNN, and by temporal computation using LSTM after feature extraction in the spatial domain by 2D-CNN. The former has high computational cost but high model representational power, and the latter has low computational cost but relatively low model representational power. In the manufacturing industry, the use of local computers to make inferences is often required for practicality and confidentiality reasons, necessitating a low computational cost, and so the latter, a lightweight model, needs to have improved performance. In addition, both methods tend to overlearn when the training dataset is small, and it is difficult to prepare a large dataset with a large number of work annotations for actual applications in the manufacturing industry. Therefore, the automation of work time analysis is a highly important topic of research. However, conventional video analysis and sensor-based methods are difficult to apply to the analysis of work hours in the manufacturing industry in environments where the granularity of work is not uniform and workers can move around freely. This has posed a major problem, especially in applying the system to manufacturing sites with high-mix low-volume production. In this paper, therefore, we propose a method using machine learning to train a model to extract the features of a worker’s position and posture from a set of video clips taken in a factory and to then use the model to detect the content of the worker’s work. Our contributions can be summarized as follows: We propose the printing factory video dataset, which is a dataset of videos taken from a third person’s point of view using a fixed camera and containing second-by-second work content annotations. The dataset is available at https://www.kaggle.com/datasets/shotarokataoka/factory-oparation-video-dataset. We propose a method for considering heuristics in a model that uses an image segmentation task as a pre-training method. To our best knowledge, there is no research on the application of such a method to work videos in factories. In particular, the video proposed in this study has completely new characteristics, as described in Table 2, and it would be of great academic value to confirm the effectiveness of heuristic pre-training on this video. We present our findings that with our dataset, the proposed pre-training method significantly reduced over-fitting and improved the F1 score on the test set by 0.2906 points over the ImageNet pre-training method. The code and raw result are available at https://github.com/ShotaroKataoka/WorkDetectionFactoryPretrain. The structure of this paper is as follows. In section 3, we review related works. Then, in Section 4 we define the task. In section 4, we explain about proposal method. In section 5, we explain about experiment. In section 6, we showed the experiment result. Finally, in Section 5. we dicussed abou the result. The paper finishes with some conclusions and points out some future work.

2 Related work

In order to automate the work process analysis, it is necessary to estimate the work content and the time it took to perform it from some data. In this paper, we refer to this sequence of steps as “work detection.” There are two approaches to work detection: using a sensor (e.g., an accelerometer) to collect data on the physical movement of a worker and using a camera to record videos of the work. In this section, we describe previous studies on sensor-based and video-based task estimation, video datasets, and methods of action recognition.

2.1 Action recognition from sensors vs. videos

With the development of sensing devices, research on motion recognition using sensor information has been attracting attention [30-50]. Peterek compared different algorithms for human physical activity recognition from accelerometric and gyroscopic data which are recorded by a smartphone [30]. Chang et al. proposed a hierarchical hand motion recognition method based on one inertial measurement unit (IMU) sensor and two surface electromyogram (sEMG) sensors. An SVM classifier was used for the EMG signal and a decision tree classifier was used for the IMU signal [31]. Ronao developped a two-stage continuous hidden Markov model (CHMM) approach for the task of activity recognition using accelerometer and gyroscope sensory data gathered from a smartphone [32]. Uddin et al. propose a model that extracts feature vectors from multi-sensor data such as ECG based on Gaussian kernel-based principal component analysis and Z-score normalization, and then trains them with CNN [33]. Wang et al. also use CNN to analyze human activity [34]. Lee et al. have developed a method to analyze time-series acceleration signals using 3D acceleration sensors of smart phones using a hierarchical hidden Markov model [35]. Ravi et al. found that using Plurality Voting showed good motion recognition performance even with a single accelerometer [36]. Kwapisz also used a single accelerometer to recognize human activities [37]. Casanova et al. have shown that recognition of motion features from acceleration data can be applied to biometric authentication [38]. Albinali proposed a motion recognition algorithm for individuals with the Autism Spectrum Disorder (ASD) by combining the Fourier transform of acceleration time series with decision trees [39]. Khan developed a model that combines an artificial neural network (ANN) and an autoregressive model (AR) to recognize motion from acceleration data [40]. Kaghyan et al. proposed an algorithm that uses accelerometer data and the K-NN algorithm to identify common activities performed by the user [41]. Brezmes also used K-NN to recognize human activities [42]. Mitchell et al. have developed a method for sports motion recognition from accelerometers using Discrete Wavelet Transform (DWT) and SVM [43]. Subasi et al. showed that an ensemble classifier based on the Adaboost algorithm using acceleration data can significantly improve the performance of automatic human activity recognition (HAR) [44]. Wang et al. proposed a real-time, hierarchical model to recognize both simple gestures and complex activities using a wireless body sensor network [45]. Garcia-Ceja used hidden Markov models and conditional probability fields to recognize actions based on time series of sensor information [46]. Hossain et al. proposed a method that combines accelerometers with LoRaWAN sensors to recognize basic actions such as walking, staying, and running by KNN and LDA [47]. Ryu et al. showed that accelerometer information can be used to classify a simple routine task consisting of four subtasks using a multi-class SVM [48]. Kim et al. analyzed IMU data by dynamic time warping (DTW) to classify the movements of construction machines [49]. These studies are of general action recognition. Sensors have been used in a number of studies, such as in an analysis of work motions in bicycle maintenance using ultrasonic and IMU sensors [51], an analysis of line motions using only IMU sensors [52], the estimation of hammering and other motions on an assembly line using wrist-mounted IMU sensors [53], and the estimation of work types using wristwatch IMU sensors [54, 55]. Modern sensor-based work recognition methods that use deep learning include those proposed by [56, 57], which use a convolutional neural network (CNN) to analyze data from IMU sensors. Such methods use IMU sensors and other wearable sensors primarily to sense the movement of a specific body part such as the hand and are intended for the analysis of assembly work in which the worker is stationed at a workbench and is not required to walk around. However, it is difficult to apply a sensor-based method to an environment in which workers can move freely because 1) parts of the body that are not equipped with sensors can move, 2) the direction of movement ranges widely, and 3) the body position must be accurately estimated. To capture workers’ movements in such high-flexibility environments, expensive devices such as multi-purpose, infrared, and line-of-sight sensors may be required. Under such conditions, it is difficult to sense the work as performed in a natural state because of the burden of mounting the sensor. Therefore, the application of sensor-based work recognition methods has been limited to the analysis of production lines in which only specific tasks with fixed actions are performed. Unlike sensors, video is a rich source of information and can extract the features of the work even if the degree of flexibility in the worker’s movement is high; in addition, locations can be accurately identified. Furthermore, the use of a small camera has the advantage of facilitating the construction of an inexpensive system. Table 1 summarizes the characteristics and capabilities of sensors and videos for action recognition.
Table 1

Comparison of sensors and video.

DeviceMovement flexibilityCostRepresentation
Sensors×××
Videoooo
As described in Section 3 below, in a printing factory such as the one used for the work analysis in this study, the worker has a high degree of movement flexibility; therefore, gathering data by using a sensor was not considered suitable from the standpoint of the burden on the worker. Therefore, in this study, we adopted a video-based work detection method.

2.2 Video datasets used for action recognition

Recently, there have been many studies on video-based general action recognition, focusing primarily on large public video datasets created through crowdsourcing. The characteristics of datasets commonly used in action recognition/detection research are listed in Table 2. UCF101 [63], Kinetics [59, 60], Something-Something [62] HMDB51 [79], and Sports-1M [65] are particularly well-known. However, they are used for action recognition tasks rather than for action detection because unlike our proposed dataset, each of their videos contains only one action. “AVA” is a movie action dataset, whichconsists of multi-label annotations of basic actions such as “talk to” and “watch” [66]. “NTU RGB+D” contains 60 classes of human activitivity and 56,880 video samples [67]. The labels in these two datasets are in three major categories: daily actions, mutual actions, and medical conditions. These labels consist of basic actions or states such as “drink water” or “staggering” [67]. “EgoGesture” is a multi-modal large scale dataset for egocentric hand gesture recognition. This dataset has 83 classes of static or dynamic gestures such as “OK” [68]. “THUMOS Challenge 2014 Temporal Action Detection” dataset defines specific motions in sports, such as dunking and pitching [69]. “Activity Net” Temporal Action Localization is a dataset for action recognition consisted of 200 different daily activities such as: “walking the dog”, “long jump” [70]. Ibrahim et al. collected a new dataset using publicly available YouTube volleyball videos. They annotated 4830 frames that were handpicked from 55 videos with 9 player’s basic action labels such as “spiking” [71]. “Win-Fail Action Recognition” is a task for judging the success or failure of an action in the following domain: “General Stunts,” “Internet Wins-Fails,” “Trick Shots,” & “Party Games.” [72]. “HAA500” contains fine-grained atomic actions where only consistent actions are classified under the same label (e.g., “baseball pitching” and “basketball free throws” to minimize ambiguity in action classification) [73]. “YouTube-8M” is the largest multi-label video classification dataset, consisting of up to 8 million videos (500,000 hours of video), annotated with a vocabulary of 4800 visual entities [74]. The Moments in Time Dataset is a large human-annotated collection of one million short videos corresponding to dynamic events that unfold in less than three seconds [75]. HACS is a dataset for human action recognition using a taxonomy of 200 action classes, which is identical to that of the ActivityNet-v1.3 dataset [76]. It has 504K videos retrieved from YouTube and each one is strictly shorter than 4 minutes, and the average length is 2.6 minutes. A large-scale “Holistic Video Understanding Dataset” (HVU) is organized hierarchically in a semantic taxonomy that focuses on multi-label and multi-task video understanding [77]. AViD is a large-scale video dataset which has 467k videos and 887 basic action classes. The AViD dataset consists of similiar actions to those in Kinetics, plus some additional actions such as talking, explosion, boating [78]. HDMB is the largest action video database with 51 basical action categories, which in total contain around 7,000 manually annotated clips extracted from a variety of sources ranging from digitized movies to YouTube [79]. “something something” is a large collection of labeled video clips that show humans performing pre-defined basic 174 actions with everyday objects [80]. “Charades” is composed of 9,848 annotated videos with an average length of 30 seconds involving 157 action classes such as “pouring into cup” [58]. “Kinetics-600” is a large-scale action recognition dataset which consists of around 480K videos from 600 action categories [58]. Each video in the dataset is a 10-second clip of action moment annotated from raw YouTube video [81]. “Kinetic-700” is an extensions of the Kinetics-600 dataset [82].
Table 2

Comparison of video datasets.

Dataset nameFixed camera?Multi-label&time detection?Action/WorkContentTime range
Charades [58]noyesgeneral workgeneralseconds
Kinetics [59, 60]nonoactiongeneralseconds
EPIC-Kitchen [61]noyesspecific workcookingminutes
Something-Something [62]nonoactiongeneralseconds
UCF101 [63]nonoactiongeneralseconds
HMDB51 [64]nonoactiongeneralseconds
Sports-1M [65]nonoactiongeneralseconds
AVA [66]nonoactiongeneralseconds
Ntu rgb+ d [67]nonoactiongeneralseconds
Egogesture [68]nonoactiongeneralseconds
THUMOS14 [69]nonoactiongeneralseconds
Activitynet [70]nonoactiongeneralseconds
Volleyball [71]nonoactiongeneralseconds
Win-Fail [72]nonoactiongeneralseconds
HAA500 [73]nonoactiongeneralseconds
Youtube-8M [74]nonoactiongeneralseconds
Hollywood in Homes [58]nonoactiongeneralseconds
Moments in Times [75]nonoactiongeneralseconds
Hacks Clip [76]nonoactiongeneralseconds
HVU [77]nonoactiongeneralseconds
AViD [78]nonoactiongeneralseconds
Oursyesyesspecific workfactoryseconds to minutes
Charades [58] and EPIC-Kitchen [61] contain multiple actions per video, and, like our dataset, they are intended for action detection tasks. Charades is a dataset of videos consisting of common actions such as “Holding a laptop” or “Watching TV.” Each of these actions ranges from a few seconds to tens of seconds in length; the length of time required for each action is small. EPIC-Kitchen is a video dataset with clips of first-person work activities in the kitchen; it differs from general action datasets such as Charades in that it consists of complex work activities in the specialized domain of cooking. In addition, the duration of each work activity is on the order of minutes, and it includes a wide range of work durations, although not as wide as our dataset. Given these characteristics, EPIC-Kitchen is the dataset most similar to ours. There are three important differences, however, between our dataset and EPIC-Kitchen. First, our dataset covers work in the factory. To the best of our knowledge, there is no other video dataset for work detection in the factory. Therefore, it has been difficult to automate the operation process analysis using video; we believe that the proposed dataset will contribute to the progress of future research. In addition, unlike conventional general action datasets, this dataset is intended for the analysis of specialized work activities consisting of series of complex body movements. Second, our dataset includes a wider range of time periods than EPIC-Kitchen, from a few seconds to several tens of minutes per work activity. In general, having a wider range of work times increases the difficulty of work detection because the model needs to capture both micro and macro changes in the video. As mentioned above, the range of work times has been limited in the past, and therefore models that address this problem have not been well studied. We believe that this dataset will contribute to the development of research on work detection over long periods. Third, the videos in our dataset were captured by a fixed-point camera. Considering actual application to the work site, the method whereby a camera is affixed to the operator and acquires video, such as was done for the videos in EPIC-Kitchen, is not appropriate from the standpoint of the burden on the operator; it is also costly because it is necessary to prepare a camera for each worker. In addition, it has been pointed out that in cases where the background information varies, the work content may be estimated using only the features of the background [83]. An example that is mentioned is that if the background is a tennis court, it is possible to infer that the action in the video is “Tennis Swing” even if the person is removed from the video. If a first-person camera is used for task detection as in EPIC-Kitchen, we can expect that each work activity will have its own characteristic background (e.g., a sink when the activity is washing dishes). Because our dataset uses a fixed-point camera, the background is the same throughout the video. Therefore, it is possible to avoid the problems pointed out by [83] and set up a task that will infer the work content from the worker’s movements rather than the background.

2.3 Action recognition methods

Until the use of CNNs became widespread, manually defined features were proposed as the analysis model for use with the above action recognition datasets. Examples include Oreifej and Liu et al.’s HOND4 feature [84], the HOG feature-based model [85], STIP [86], Gradient and Histogram of Optical Flow [87], 3D Histogram of Gradient [88], SIFT-3D [89], and SURF [90]. The method using Fisher vector representation [91] was also mainstream. Since the publication of the study by [92], CNN methods have become mainstream in the field of computer vision, and in the video processing field, the SOTA method is a model based on CNN. Video processing requires not only spatial computation, as in the case of image processing, but also temporal computation. There are two main methods for computing the temporal dimension in CNN-based methods (see Table 3 for a summarized comparison). The first is a 3D convolutional neural network (3D-CNN) such as C3D [93]. This is a method for computing both spatial and temporal convolutions simultaneously; it has high representational capability. Following the proposal of C3D, 3D-CNN has been actively studied, such as in I3D [60], which employs Inception, and in work by [94], which is based on ResNet. Qui et al. proposed the Local and Global Diffusion (LGD) model. This model is equipped with LGD blocks that learn local and global representations in parallel [95]. D3D is a method for spatio-temporal recognition by distillation. Specifically, it is based on a teacher network that grasps the features of the optical flow and learns from the features of RGB only with a student network [96]. Tran et al. devised a model consisting of 3D ResNet by searching for the optimal architecture of convolutional neural networks [97]. PERF-Net, which uses the rendering information of poses in RGB frames to capture the motion features, has also been proposed [98]. In addition to the two streams, Hong et al. propose to use additional pose and pairwise streams which can be extracted contextual action cues to improve the performance of action recognition [99]. Shuiwang et al. have proposed a method for action recognition using 3D convolutional neural networks [100]. FstCN [101] also decompose a 3-dimensional convolution into 2-dimensional and 1-dimensional convolutions the model to reduce the amount of computation. ResNet [102] and DenseNet [103] has also been proposed as a model to reduce computational complexity. X3D is a model that searches for the optimal network structure [104]. A method to reduce the number of channels, called R(2+1)D [105], has been proposed, and the computation efficiency is being improved. Many other methods based on 3D-CNN such as P3D [106], S3D [107], and CSN [108] have been proposed and are SOTA for Kinetics and other datasets [109].
Table 3

Comparison of 3D-CNN and CNN-LSTM.

Model typeSegmentation?Inference memoryTime range
3D-CNNnolargeshort
CNN-LSTMyessmalllong
The second method of CNN-based video processing is one that first calculates spatial directions with an ordinary 2D-CNN and extracts features of the frame and then calculates temporal direction with LSTM or CNN [110]. This method is inferior to the 3D-CNN method in terms of representational capability because the spatial and temporal calculations are separate from each other, but it is advantageous in terms of computational cost. In the case of a large model such as 3D-CNN, the computer specifications are too high to introduce it to a factory, and it might not be possible to infer the actual environment. Therefore, there is a strong need to develop a relatively lightweight model such as CNN-LSTM. In addition, in order to create a work detection model for operation process analysis, it is necessary to create a video dataset with work annotations for each work site. As the cost of creating such a dataset is very high, it is difficult to create a large dataset. With consideration of this background, we investigated ways of creating a better model by training a conventional basic CNN-LSTM model on a small dataset.

3 Task definition

In this study, we wished to analyze the work time of the main worker by categorizing his or her work content by the second from video recordings of factory work. The particular focus was to develop a method of factory work analysis that could be used in real-world applications. However, we did not find any video datasets available in the market that met the requirements for video content and labeling. Therefore, a dataset was created for this study by recording three days’ work content in an actual factory using a common video camera. Fig 1 shows an example of one frame of video recorded for the dataset. The task was then to classify the content of the main worker’s work by the second to annotate this dataset.
Fig 1

Sample frame of proposed factory video dataset.

The proposed work video dataset has several features that differ from public video datasets that are conventionally used: The videos are captured by a fixed camera from a bird’s eye view in an actual printing factory. The videos are untrimmed and have a length of about 10 h per day. There is approximately 30 h of labeled video for the entire dataset. Videos show the full body of the worker and include complex movements. The background environment may also vary independently of the work. The worker uses multiple machine tools and small tools to perform the work. Twelve work labels were defined to categorize the work content. (These are not the same as action labels. Therefore, even if two physical actions are similar, they will be assigned different work labels if they differ in terms of work content.) With consideration of this background, we investigated ways of creating a better model by training a conventional basic CNN-LSTM model on a small dataset. The length of time to perform the same work consecutively varies considerably, ranging from a few seconds to tens of minutes, and tasks depend on one other over a very long period.

4 Method

The data were collected after prior explanation to the company (Echigo Fudagami Inc) that the data would be used for research purposes and verbal consent was obtained from the company manager and the subject workers. Permission to conduct the research was obtained from the company manager and subject workers. Our university’s ethics committee confirmed that ethical approval was waived. And no third-party ethical oversight was provided. In this section, we first describe the work detection model used in our method, that is, the series of processes for classifying the operator’s work by the second based on a series of input images, as shown in Fig 2.
Fig 2

CNN-LSTM model architecture (baseline and ours).

All CNN-Modules and FCs share weights with each other. The number marked @ is the output channel. Each layer component corresponds to the class diagram in Fig 3. Hyper-parameters not listed use the default values from Fig 3. CNN-Module: It is described later in Fig 4.

CNN-LSTM model architecture (baseline and ours).

All CNN-Modules and FCs share weights with each other. The number marked @ is the output channel. Each layer component corresponds to the class diagram in Fig 3. Hyper-parameters not listed use the default values from Fig 3. CNN-Module: It is described later in Fig 4.
Fig 3

Class diagram of the layers used in the model diagrams in Figs 2 and 4.

This figure describes the hyper-parameters that can be set for each class and their default values.

Fig 4

Our CNN-Module architecture.

All layers of blue boxes are the encoder of DeepLab v3+ [113] (MobileNetV2 [115]-backbone). Vectorizer is a CNN+GAP layer that we added to convert feature maps to vectors. Each layer component corresponds to the class diagram in Fig 3. Hyper-parameters not listed use the default values from Fig 3.

Then, we explain our approach for improving the generalization performance by pre-training this model on a task to segment workers. Even with pre-training on the segmentation task, the model structure in the inference process is almost identical to that of the work detection model described above. Therefore, we expected the pre-training to improve performance considerably without increasing the computational cost of inference. For the pre-training, an encoder–decoder segmentation model [111-113] is used; this makes it possible for the model to effectively learn important information such as the body parts of the worker from a small quantity of data. During the training of the work detection model, only the encoder module is used, to enable the feature extraction to focus on the important information. The dataset videos were filmed using a fixed camera for three days at a specific work site in a printing factory (Echigo Fudagami Inc: Uenoyama 1-2-8, Ojiya, Niigata, Japan.), with a data size of 10 h per day, i.e., 30 h for the entire dataset.

4.1 CNN-LSTM model

The CNN-LSTM model for work detection is shown in Fig 2. The algorithm of this step is explained in Algorithm 1. In the hyper-parameters for each layer in Fig 2, the default values described in Fig 3 are used unless a value is specified. The hyper-parameters of the LSTM were determined through grid search by comparing validation scores. Each frame image of the video data is entered into the same CNN model [92]; the CNN model computes the images in a convolutional calculation and converts them into feature vectors. Then, a series of these feature vectors are input to LSTM [114]. LSTM performs feature extraction by considering the temporal patterns of the feature vectors of the frames. After that, the fully connected layer and Softmax infer the work content for each frame.

Class diagram of the layers used in the model diagrams in Figs 2 and 4.

This figure describes the hyper-parameters that can be set for each class and their default values. For video-based action detection, let X = {x, …, x} denote a series of frames starting from time n, where T is the length of the input sequence. The CNN-Module performs the following calculation: where x is the t-th frame of X, and CNN-Module(⋅) is CNN feature extractor that transforms a 2D image into a vector. The CNN-Module used in our experiment is shown in Fig 4. As the CNN module, we adopted the Encoder of DeepLab v3+ [113] that we utilize for pre-training, as described in the following section. Since the output of DeepLab v3+ Encoder is a feature map, we added Vectorizer right after Encoder to convert it into a vector. Hyper-parameters for each layer were determined as follows: the hyper-parameters for Encoder of DeepLab v3+ were the values recommended in the original paper. To effectively reduce the feature resolution before performing GAP, we added three additional layers of CNNs with stride = 2 as Vectorizer. The hyper-parameters of these CNNs were determined through grid search by comparing validation scores. The mean and standard deviation parameters used in the standardization function were calculated from the entire image of training set. Therefore, the result of the calculation, f, is a d-dimensional feature vector that represents the spatial information of x.

Our CNN-Module architecture.

All layers of blue boxes are the encoder of DeepLab v3+ [113] (MobileNetV2 [115]-backbone). Vectorizer is a CNN+GAP layer that we added to convert feature maps to vectors. Each layer component corresponds to the class diagram in Fig 3. Hyper-parameters not listed use the default values from Fig 3. The LSTM module performs the following calculation: where F = {f, …, f} denotes a series of output feature vectors from CNN(X), which are computed independently. The result, Y = {y, …, y}, denotes a corresponding series of output action labels. The loss function is computed as where gt is the ground truth at time t, and CE(⋅) is a cross-entropy loss function. Algorithm 1 Training factory segmentation pre-trained CNN-LSTM Require: X = {x, x, …, x}, gt ∈ GT (GT: work ground truth set) Ensure: Y = {y, y, …, y} (y = predicted work label) 1: while Until Convergence do 2:  for NUMBER_of_IMAGES / BATCH_SIZE do 3:   L ← 0 4:   for BATCH_SIZE do 5:    feature_maps ← Encoder(X) # Parallel, Freeze weight 6:    feature_vectors ← Vectorizer(feature_maps) # Parallel 7:    Y ← LSTM(feature_vectors) 8: 9:   end for 10:   Adam(L) then update params 11:  end for 12: end while

4.2 Factory segmentation pre-training

The factory segmentation pre-training of the CNN-Module is shown in Fig 5. The algorithm of this step is explained in Algorithm 2.
Fig 5

Our segmentation heuristics factory pre-training architecture.

Pre-training [116] is a learning method in which models are trained previously using a different dataset than the target task uses; the use of pre-trained models can be expected to reduce over-fitting in the target task. Traditionally, large-scale general image datasets such as ImageNet [117] and MSCOCO [118] are often used as pre-training datasets in neural net models for image processing. The CNN-Module used in the baseline model in this study was also pre-trained on ImageNet. However, pre-training with general image datasets such as ImageNet is only useful for capturing general features in images and is not suitable for extracting task-specific features. [119] successfully improved model classification performance in a study of pattern classification of interstitial lung diseases from lung CT images by pre-training the model on general texture datasets in addition to lung CT images to address the lack of training data. Thus, a more domain-specific form of pre-training may be more effective than pre-training on a general dataset alone for applied tasks for which training data are limited. There are heuristics for identifying the important features in a frame image (e.g., position of a worker’s body part, position and state of a machine tool) for work detection in practical applications. Because the CNN-Module used in this study is only an image feature extractor (the so-called encoder mechanism), the CNN-Module is generally pre-trained for an image recognition task that uses only the encoder. However, the image recognition task only determines whether or not there is a target in the image, which makes it difficult to utilize the heuristics. Therefore, in this study, we constructed an encoder–decoder model by adding a decoder mechanism after the encoder, which is not used in the work detection model but only during pre-training, so that the object segmentation task can be learned. Because segmentation performs pixel-level object classification, it is possible to extract more highly representative features. In this study, we trained the model for the segmentation of the worker’s body part, which is the most common heuristic in factory work detection. More advanced heuristics, such as the segmentation of tools and specific machine parts, can be included in applications in the real-world. It is expected that this pre-training will allow the encoder to extract more domain-specific features with an emphasis on heuristics. We used DeepLab v3+ [113] as a segmentation model for pre-training because it is a simple encoder–decoder model consisting of a single stage and is thus naturally applicable. We randomly sampled 337 frames from the training set of our factory video dataset so that work labels were distributed as evenly as possible. We annotated the worker’s head, upper body, lower body, and arm segments in these images to create a dataset for the segmentation task. We used this dataset to train the ImageNet pre-trained DeepLab v3+ model. The dataset was holdout validated with train = 269 images and test = 68 images, and the evaluation result on the test set was mIoU = 0.86. After the above pre-training, learning is transferred to the task of work detection. For this, we remove the decoder in the segmentation-pre-trained model because the decoder is not needed for work detection. In addition, because the output of the encoder is a feature map, it is transformed into a vector by a CNN consisting of three layers and GAP (Global Average Pooling) [120]. Algorithm 2 Pretraining Encoder with factory segmentation Require: x ∈ X, gt ∈ GT (GT: segmentation ground truth set) Ensure: y ∈ Y (Y: predicted segmentation set) 1: while Until Convergence do 2:  for NUMBER_of_IMAGES/ BATCH_SIZE do 3:   L ← 0 4:   for BATCH_SIZE do 5:    feature_map ← Encoder(x) 6:    y ← Decoder(feature_map) 7:    L ← L + Loss(gt, y) 8:   end for 9:   Adam(L) 10:  end for 8: end while

5 Experiment

To ascertain the effect of the segmentation task pre-training on work detection performance, we compared the work detection performance of two methods: the CNN-LSTM model (baseline), in which the CNN-Module was pre-trained with ImageNet only, and the CNN-LSTM model (ours), which was pre-trained with the segmentation task. In the first section, we describe the details of our dataset. Then, we describe the experiment implementation details and the results for the two methods.

5.1 Dataset

Here, we provide details about our factory operations video dataset. The videos were filmed using a fixed camera for three days at a specific work site in a printing factory, with a data size of 10 h per day, i.e., 30 h for the entire dataset. The dataset is annotated by the second with 12 labels according to the work content. Because this dataset has a time-series feature, it is appropriate to divide it into the training, validation, and test sets along the time series. Thus, we divided the dataset such that the training set consists of all of day 1 and the morning of day 2, the validation set consists of the afternoon of day 2, and the test set consists of all of day 3. Table 4 shows the distribution of the labels in the dataset. It can be seen that the frequency of class occurrences is biased. This is a common situation in real-world applications of the work detection model, and it is important to train the model so that the tasks that appear less frequently can be correctly inferred.
Table 4

Distribution of labels in our factory video dataset.

Class IDLabelProportion [%]
Class 1Chores7.66
Class 2Paperwork6.42
Class 3Cylinder preparation11.26
Class 4Material preparation8.96
Class 5Ink roller adjustment3.35
Class 6Winding preparation (top)3.73
Class 7Winding preparation (bottom)3.15
Class 8Creation of sample9.11
Class 9Ink adjustment1.75
Class 10Workbench0.85
Class 11Product check11.00
Class 12Nonhuman work32.75

4.2 Implementation details

In our experiments, we sampled a fixed length T at a sampling rate FPS from each frame sequence as the input. We set T to 100 frames and FPS to 1 fps. Hence, the frame sequence we utilized had a time length of 100 s. The output channel of the CNN-Module was set to 256. During training, we used the Adam optimizer [121] to optimize the network. The initial learning rate was set to 10−3. The batch size was 128. Table 5 shows the equipment hardware and software infrastructure details of our experiment. Computational resource of AI Bridging Cloud Infrastructure (ABCI) provided by National Institute of Advanced Industrial Science and Technology (AIST) was used.
Table 5

The equipment hardware and software infrastructure details.

HardwareProductSpec
CameraPLANEX CS-QR2015FPS, 640×480px
ComputerABCICPU:Intel Xeon Gold 6148 Processor
CPU Memory:360GB
GPU:NVIDIA Tesla V100 SXM2 (16GB HBM2)×4
Software version
Python3.6.5
CUDA10.1.243
cudnn7.6.4
PyTorch1.8.0

5.3 Results and comparisons

In this section, we compare the results from our proposed segmentation heuristics pre-training method with those of the ImageNet-pre-trained model. Fig 6 shows a loss curve of our factory-segmentation-pre-trained CNN-LSTM. For prediction, we used the weights of the epochs with the highest class accuracy in the validation set. In this experiment, the best validation class accuracy was achieved at 9 epochs (training took 21 minutes), so the weights at 9 epochs were used for the experiment. In the loss curve, the validation loss and training loss both show a decreasing trend, and there is no significant overfitting. Although there is a gap between these two, it is considered to be within the acceptable range for learning progress.
Fig 6

Loss curve of our factory-segmentation-pre-trained CNN-LSTM training.

Table 6 summarizes our experimental results. In the dataset used in this study, the number of data in each class differs considerably, and the micro average is not suitable because the influence of some classes is large. For this reason, we used the macro average as the average score in the results shown in the table.
Table 6

Comparison of the two CNN-LSTM models.

The baseline is ImageNet-pre-trained CNN-LSTM; ours is the factory-segmentation-pre-trained CNN-LSTM. Scores were calculated using the macro average. The highest scores are expressed in a bold font.

ModelAccuracyF1RecallPrecision
Baselinetraining 0.9954 0.9946 0.9947 0.9945
validation0.63640.39400.38460.4578
testing0.36850.32670.35280.4955
Ourstraining0.89910.86910.86140.8823
validation 0.7152 0.5528 0.5678 0.6096
testing 0.7125 0.6173 0.6221 0.6407

Comparison of the two CNN-LSTM models.

The baseline is ImageNet-pre-trained CNN-LSTM; ours is the factory-segmentation-pre-trained CNN-LSTM. Scores were calculated using the macro average. The highest scores are expressed in a bold font. From the table, we can see that the proposed method achieved an F1 performance of 0.6173, better than that of the baseline (ImageNet pre-training) method of 0.3267. This result shows that the baseline model is in a quite over-fitted state. In contrast, our method has a smaller performance gap between training and testing, which demonstrates that our segmentation heuristics pre-training method contributes to the reduction in over-fitting. In addition, we performed a Welch’s t-test to confirm the significance of the difference between the output of each method. As a specific procedure, we randomly divided the dataset into 100 parts, calculated the F1 score for each class in each part, and performed Welch’s t-test on each class. Consequently, the null hypothesis on all classes except class 7 F1 score can be rejected at the 1% significance level for the difference between baseline and ours. Tables 7–9 show the recall, precision, and F1 score, respectively, for each class. In these tables, the “Diff” column was calculated as Diff = Ours−Baseline, which indicates the amount by which the performance of our method differed from that of the baseline method.
Table 7

Comparison of label-by-label recall between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font.

ClassLabelBaselineOursDiff
Class 1Chores0.2138 0.5481 +0.3343
Class 2Paperwork 0.8271 0.6530−0.1741
Class 3Cylinder preparation0.0775 0.7797 +0.7022
Class 4Material preparation0.3680 0.8647 +0.4967
Class 5Ink roller adjustment0.2642 0.4786 +0.2144
Class 6Winding preparation (top)0.2651 0.5336 +0.2685
Class 7Winding preparation (bottom)0.3839 0.7198 +0.3359
Class 8Creation of sample0.1891 0.4427 +0.2536
Class 9Ink adjustment0.1641 0.4660 +0.3019
Class 10Workbench0.2335 0.3684 +0.1349
Class 11Product check 0.8273 0.6775−0.1498
Class 12Nonhuman work0.4204 0.9324 +0.5120
Table 9

Comparison of label-by-label F1 score between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font.

ClassLabelBaselineOursDiff
Class 1Chores0.1987 0.4301 +0.2314
Class 2Paperwork0.3120 0.6991 +0.3871
Class 3Cylinder preparation0.1414 0.7859 +0.6445
Class 4Material preparation0.5225 0.8747 +0.3522
Class 5Ink roller adjustment0.3605 0.6148 +0.2543
Class 6Winding preparation (top)0.3184 0.5596 +0.2412
Class 7Winding preparation (bottom)0.4281 0.4430 +0.0149
Class 8Creation of sample0.2454 0.4931 +0.2476
Class 9Ink adjustment0.2675 0.4496 +0.1821
Class 10Workbench0.0400 0.4038 +0.3638
Class 11Product check0.5002 0.7030 +0.2028
Class 12Nonhuman work0.5851 0.9505 +0.3654

Comparison of label-by-label recall between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font.

Comparison of label-by-label precision between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font.

Comparison of label-by-label F1 score between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font. Our method improved F1 in all classes. The most notable improvement was in class 3, where F1 was 0.7859, an increase of 0.6445 points from 0.1414. Class 7, which showed the least improvement, still improved by 0.0149 points, and the other classes generally improved by 0.18 to 0.4 points. Class 3, which showed the greatest improvement in F1, is the cylinder preparation process; this work content is similar to that of classes 2 and 10 and requires a detailed analysis of the worker’s behavior. Because our method significantly improved the detection accuracy of these work activities, we believe that our proposed factory segmentation pre-training method is particularly effective for the detailed analysis of these work activities. Besides, the occurrence rate of classes 9 and 10 in the dataset is extremely low (1.75% for class 9 and 0.85% for class 10), and it is difficult to extract features from the work content dataset alone, as the baseline model showed. However, our method improves F1 scores of classes 9 and 10 by 0.2036 and 0.3693 points, respectively. This suggests that pre-training of heuristics can generalize the model even for tasks with a low frequency of occurrence.

6 Discussion

6.1 Analysis of model architecture

In the dataset we used, the training set and the test set have different work days, and the worker’s environment can vary throughout the dataset. These aspects reflect the reality of the situation, and there would be considerable academic value in being able to generalize the model for application to a setting other than that of training. As the baseline model uses ImageNet for pre-training, there is a high degree of flexibility in the worker’s movement in the training of the task detection model. In our factory segmentation pre-training, by contrast, the heuristics are included in the pre-training so that the direction of learning can be specified at the time the work detection model is trained. For example, we pre-trained for the segmentation of the worker’s body part, and this provided the direction during the training of the work detection model that the worker’s pose be considered. If necessary, by pre-training for other heuristics such as segmentation of work tools or materials, we can adjust the direction of learning to focus on these movements. As described in the previous section, our method significantly reduced over-fitting. In cases such as the present one, in which the quantity of data is limited, we believe that giving the model a heuristic direction for learning may contribute effectively to reducing over-fitting. The annotation of the worker segmentation performed for pre-training was much less costly than the annotation of the work content for work detection. The work detection performance improvement of 0.2906 points for macro F1 was obtained simply by pre-training on a segmentation task with low annotation costs; this suggests that our method can improve performance whilkke keeping annotation costs low.

6.2 Failure cases

Here we discuss the reasons for the cases on which detection failed using this method. Table 10 shows the confusion matrix of result of our CNN-LSTM model of the test set. The rows show the classes of the groundtruths (GT). The columns show the classes predicted by the model (PR). The value of each element is a count of the pairs of the GT and the PR. The model predicted once per second.
Table 10

Confusion matrix of our model’s result (test set).

The rows show the classes of the groundtruths (GT). The columns show the classes predicted by the model (PR). The value of each element is a count of the pairs of the GT and the PR. The diagonal line is the true positive and is expressed in a blue cell. The class numbers in the table correspond to the class IDs in Table 4.

PR
1 2 3 4 5 6 7 8 9 10 11 12
GT 1 946671613842377162414911840
2 23212272533162793747388
3 1601230803012443311314776
4 1171113219922905606668
5 144247756836124044001490
6 311548861793124352403911
7 18010121236681970080
8 4292462871942722011931409418
9 29001002681637001091
10 7694250000084012
11 369521107927482310175427
12 12215174390136411196052

Confusion matrix of our model’s result (test set).

The rows show the classes of the groundtruths (GT). The columns show the classes predicted by the model (PR). The value of each element is a count of the pairs of the GT and the PR. The diagonal line is the true positive and is expressed in a blue cell. The class numbers in the table correspond to the class IDs in Table 4. Fig 7 shows the output feature vectors of factory segmentation pre-trained CNN-LSTM dimensionally reduced using t-SNE and mapped onto a two-dimensional graph. The 64-dimensional feature vector, which is the output of the LSTM prior to the FC+Softmax layer in Fig 2, was compressed into a 2-dimensional vector by t-SNE. The parameters of t-SNE were set to perplexity = 30 and n_iter = 1000. t-SNE dimensionality reduction was performed on the entire test set, from which 12,000 samples were randomly sampled so that all classes were evenly distributed, and a scatter plot was created.
Fig 7

t-SNE of factory segmentation pre-trained CNN-LSTM outputs. (test set).

We consider the causes for the classes having an F1 score of less than 0.5 (Table 9, namely, classes 1, 7, 8, 9, and 10. First, let us consider classes 1 and 10. These are the labels having low F1 scores with the baseline method as well and are presumably difficult to classify in the first place. As shown in Table 4, class 10 is a rare class, constituting only 0.85% of the dataset. In addition, our review of the dataset suggests that the physical movements of the workers in classes 1 and 10 are particularly diverse, making it difficult to capture their work patterns. This can be seen from the distribution of the output feature vectors in Fig 7. Class 1 (red) and class 10 (greenyellow) are widely and finely distributed around the upper right area in the figure and are mixed with other classes such as class 2 and 3. All of the remaining classes—7, 8, and 9—are work activities performed around the rotary press. As can be seen from Table 10, these work activities are frequently misclassified for each other. In Fig 7, these classes are widely distributed from the center to the lower left of the figure. However, they differ from classes 1 and 10 in that they form a rough clump. In many cases, even humans cannot classify these work activities correctly unless they carefully observe the operation of the rotary press and the time series of the work content. The operation of the rotary press in particular is closely related to the content of the work, and so it is quite important to take this into account. Overall, there are sparse and dense regions in the feature space of Fig 7. Classes 3, 4, and 12 occupy a relatively large space, while the other classes are clustered in a close space. In order to deal with this, it is necessary to be able to perform feature extraction that can separate the classes that are densely clustered. For this purpose, two approaches can be considered: 1. improvement of spatial feature extraction, and 2. improvement of temporal feature extraction. For the improvement of spatial feature extraction, as mentioned above, it is possible to use segmentation models to pre-train the machines and tools that are important for work identification. As for the improvement of temporal feature extraction, we can expect that it can be improved by using a model with higher expressive power than LSTM (Bi-LSTM, Transformer, etc.) or by using a new model that can consider the flow of work over a longer span.

6.3 Limitations

In this dataset, the same person performs works throughout the entire dataset. Since our method includes abstraction of the body parts of the worker by segmentation, it is considered to be robust to changes in the workers. However, the method may not be able to infer correctly when the procedures and motions of the workers change significantly due to differences in the workers. In this case, we need to add new training data. There were cases where the working time for the same task varied from several minutes to several tens of minutes. This is thought to be because the work procedures may differ even for the same work. For example, “Product check” work is performed continuously for several tens of minutes, but it may be interrupted irregularly when defective products are found, so the work may last from a few seconds to several minutes. Since the training data contains multiple work order patterns even for the same task, the model can predict when the work order changes to some extent. However, if the order of the tasks is highly irregular, or if the tasks are completely new and unknown, the inference may not be correct. In such cases, it is necessary to add training data or review the contents of the work labels. Our method is robust to changes in background information that is irrelevant to the task by using the segmentation of critical elements. However, since the method assumes a fixed camera, the trained model cannot be used as-is when the layout of the machines and the work environment related to the work changes. In these cases, it is considered necessary to re-collect the training data.

7 Conclusion

In this paper, we have proposed a factory segmentation pre-training method for video-based work detection, which is the first attempt using our proposed printing factory video dataset. This method does not interfere with the worker’s work operations because it detects his or her work solely from the video. Compared with conventional methods that involve attaching sensors to workers, this system enables the detection of work activities in environments where workers have a higher degree of flexibility. The proposed pre-training method can not only learn to capture heuristics that are found important for distinguishing among similar work classes but also substantially reduce over-fitting with only a few additional segmentation annotations. Furthermore, our factory segmentation pre-training method does not result in a detection model so large that it would seriously impair the inference speed. In addition, we proposed a printing factory video dataset, which is 30 h in length and contains annotations (12 classes) for every second. On this dataset, the proposed factory segmentation pre-training method outperforms the baseline (ImageNet pre-training) method, with an improvement of 0.2906 points in the macro F1 score. As future research, we would like to improve our spatial and temporal feature extraction methods respectively. As for spatial features, we are interested in being able to take into account the state of machines and tools (e.g., rotation speed of the rotary press). As for temporal features, we are interested in time-series analysis methods that can extract and take into account the flow of factory operations over a longer time range. 20 Sep 2021
PONE-D-21-26196
Improving work detection by segmentation heuristics pre-training on factory operations video
PLOS ONE Dear Dr. Kataoka, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Specifically, the authors should better discuss the background of their study and its placement within the body of existing literature on the topic of this manuscript, improve technical presentation, present more details from experiment for replicability, and discuss the limitations of their methodology. Please submit your revised manuscript by Nov 04 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Robertas Damaševičius Academic Editor PLOS ONE Additional Editor Comments: The authors are invited to revise their paper following the suggestions and comments of the reviewers. Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Thank you for stating the following financial disclosure: “Unfunded studies” At this time, please address the following queries: a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution. b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” c) If any authors received a salary from any of your funders, please state which authors and which funders. d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.” Please include your amended statements within your cover letter; we will change the online submission form on your behalf. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Paper aims to present an improvement (from performance perspective) of a classic CNN-LSTM model for computing spatio-temporal information. Introduction promotes "a method for considering heuristics in a model that uses an image segmentation task as a pre-training method" as a key novelty of the paper in addition to the dataset provided at Google's server. I'm not so sure that this method is unique, as similar approaches have been done before in a different context. This requires a clarification. Authors must provide an in-depth mathematical apparatus (missing completely, only generic statements were made) and full UMLs of the implementation, including the architecture along with necessary hyper parameters of the network. Pseudo codes provided are too shallow, failing to showcase what exactly is different from other approaches on computing spatio-temporal information. Must be expanded. Provide loss curves (prove that your solution is not over-fitting) and tsne. Add explanation to figure 4. What each class means. Experimental part does not detail the equipment HW and SW infrastructure used. Performance part (this method was supposedly much faster than accurate vs others) was not covered sufficiently. Literature review is organized, but contains too few references. There has been a lot done in this area. Estimate around 20+ papers per section easily... Image (video) data is available on the google drive. Yet the aggregated raw results were not provided. Please back up this statement "manufacturing industry, the use of local computers to make inferences is often required for practicality and confidentiality reasons, necessitating a low computational cost, and so the latter, a lightweight model, needs to have improved performance". Computing costs are negligible compared to manufacturing costs and equipment. Paper focuses on analyzing workers (live human subjects) yet provides no ethical permit. Image (video) data is available on the google drive. Yet the aggregated raw results were not provided. Reviewer #2: This paper presents an issue: improving work detection based on the video-analysis and on the use an image segmentation model. 1. Introduction – a very good research question and a significant problem of working time analysis. However please express the area of recognition of employee activities in the production process. So please extend your introduction this area and add the sources: https://doi.org/10.3390/sym11091151; https://www.mdpi.com/2071-1050/13/10/5495. 2. Related work – the authors assume that background is the same throughout the video. Is the environment at the adopted position really not changing? And if so, how can this problem be solved? 3. In the video data set – the authors state, that the length of time to perform the same work consecutively varies considerably, ranging from a few seconds to tens of minutes, and tasks depend on one other over a very long period. Is the same activity performed by the same employee then? Are there no time standards for individual activities? Are the steps always performed in the same sequence? 4. Please add an explanation why DeepLab v3 was chosen as a segmentation model for pre-training - description: "it is a simple encoder – decoder model consisting of a single stage" is not sufficient. 5. How the findings can be exploited by future similar works? 6. Please add the limitations of the work. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 25 Nov 2021 Manuscript: PONE-D-21-26196 Response to Reviewers Editor 1. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. Response: We did not use “Protocols.io” due to the time limitation, since the source code was newly published and the flow was defined in detail. We would like to consider it in the future works. Reviewer 1 1. Paper aims to present an improvement (from performance perspective) of a classic CNN-LSTM model for computing spatio-temporal information. Introduction promotes "a method for considering heuristics in a model that uses an image segmentation task as a pre-training method" as a key novelty of the paper in addition to the dataset provided at Google's server. I'm not so sure that this method is unique, as similar approaches have been done before in a different context. This requires a clarification. Authors must provide an in-depth mathematical apparatus (missing completely, only generic statements were made) and full UMLs of the implementation, including the architecture along with necessary hyper parameters of the network. Pseudo codes provided are too shallow, failing to showcase what exactly is different from other approaches on computing spatio-temporal information. Must be expanded. Response: We thank the reviewer for the recommendation. We have added the following sentences. Uniqueness: we added a sentence to the introduction section. “There has been no research on the application of such a method to work videos in factories. In particular, the video proposed in this study has completely new characteristics, as described in Table 2, and it would be of great academic value to confirm the effectiveness of heuristic pre-training on this video.” Implementation details: we added or fixed figures and sentences to the method section. Figure2(fixed), Figure3(added) “The CNN-Module used in our experiment is shown in Figure 3. As the CNN module, we adopted the Encoder of DeepLab v3+ [116] that we utilize for pretraining, as described in the following section. Since the output of DeepLab v3+ Encoder is a feature map, we added Vectorizer right after Encoder to convert it into a vector” We published the CNN-LSTM code and added the following sentence to the introduction. “The code is available at https://github.com/ShotaroKataoka/work_detection_factory_pretrain.” 2. Provide loss curves (prove that your solution is not over-fitting) and tsne. Response: We thank the reviewer for the comment. We have added the following sentences and figures. Loss curve: we added a sentence and a figure to the experiment section. Figure5(added) “Figure 5 shows a loss curve of our factory-segmentation-pre-trained CNN-LSTM. For prediction, we used the weights of the epochs with the highest class accuracy in the validation set. In this experiment, the best validation class accuracy was achieved at 9 epochs (training took 21 minutes), so we used the weights at 9 epochs for the experiment. In the loss curve, the validation loss and training loss both show a decreasing trend, and there is no significant overfitting. Although there is some gap between these two, it is considered to be within the acceptable range for learning progress.” t-SNE: we added a sentence and a figure to the discussion section. Figure6(added) “Figure 6 shows the output feature vectors of factory segmentation pre-trained CNN-LSTM dimensionally reduced using t-SNE and mapped onto a two-dimensional graph. The 64-dimensional feature vector, which is the output of the LSTM prior to the FC+Softmax layer in Figure 2, was compressed into a 2-dimensional vector by t-SNE. The parameters of t-SNE were set to perplexity = 30 and n_iter = 1000. t-SNE dimensionality reduction was performed on the entire test set, from which 12,000 samples were randomly sampled so that all classes were evenly distributed, and a scatter plot was created.” 3. Add explanation to figure 4. What each class means. Response: We thank the reviewer for the recommendation. We have fixed figure4 and its explanation. - Figure 4 has been changed to Table 10. “Table 10 shows the confusion matrix of result of our CNN-LSTM model of test set. The rows show the classes of the groundtruths (GT). The columns show the classes predicted by the model (PR). The value of each element is a count of the pairs of the GT and the PR. The model predicted once per second.” “Table 10: Confusion matrix of our model’s result (test set). The rows show the classes of the groundtruths (GT). The columns show the classes predicted by the model (PR). The value of each element is a count of the pairs of the GT and the PR. The diagonal line is the true positive and is expressed in a bold font. The class numbers in the table correspond to the class IDs in Table 4.” 4. Experimental part does not detail the equipment HW and SW infrastructure used. Response: We thank the reviewer for the comment. We have added the Figure 5 and a sentence to the experiment section. “Table 5 shows the equipment hardware and software infrastructure details of our experiment. Computational resource of AI Bridging Cloud Infrastructure (ABCI) provided by National Institute of Advanced Industrial Science and Technology (AIST) was used.” 5. Performance part (this method was supposedly much faster than accurate vs others) was not covered sufficiently. Response: We thank the reviewer for the comment. We have added the following sentence to the experiment section. “training took 21 minutes” 5. Literature review is organized, but contains too few references. There has been a lot done in this area. Estimate around 20+ papers per section easily Response: We thank the reviewer for the comment. As the reviewer recommended, we have added total 76 references and the detailed explanation written in red in the manuscripts. And we have also updated the “Table 2; Comparison of video datasets”. 6. Image (video) data is available on the google drive. Yet the aggregated raw results were not provided. Response: We thank the reviewer for the comment. We have uploaded raw results to the google drive and added the following sentence to the introduction section. “The raw result is available along with the dataset described above.” 7. Please back up this statement "manufacturing industry, the use of local computers to make inferences is often required for practicality and confidentiality reasons, necessitating a low computational cost, and so the latter, a lightweight model, needs to have improved performance". Computing costs are negligible compared to manufacturing costs and equipment. Response: Thank you for your comment. It is true that the cost of a computer may be small compared to the cost of equipment. However, this system is intended for small and medium-sized companies. However, this system is designed for small and medium-sized enterprises (SMEs), which do not have ample funds and cannot easily afford to purchase computers. Therefore, it is important to have a simple model that does not require a lot of machine power. 8. Paper focuses on analyzing workers (live human subjects) yet provides no ethical permit. Response: We thank the reviewer for this comment. As described in the paper, the dataset videos were filmed using a fixed camera for three days at a specific work site in a printing factory (Echigo Fudagami Inc : Uenoyama 1-2-8, Ojiya, Niigata, Japan.), with a data size of 10 h per day, i.e., 30 h for the entire dataset. The data were collected after prior explanation to the company (Echigo Fudagami Inc) that the data would be used for research purposes and verbal consent was obtained from the company manager and the subject workers. Permission to conduct the research was obtained from the company manager and subject workers. Our university's ethics committee confirmed that ethical approval was waived. And no third-party ethical oversight was provided. Reviewer 2 1. Introduction – a very good research question and a significant problem of working time analysis. However please express the area of recognition of employee activities in the production process. So please extend your introduction this area and add the sources: https://doi.org/10.3390/sym11091151; https://www.mdpi.com/2071-1050/13/10/5495. Response: Thank you for your detailed response and valuable insights. We add the following sentence written in red into “Introduction” and 17 references included “https://www.mdpi.com/2071-1050/13/10/5495”. “Recognition of employee activities in the production process is gaining attention \\cite{Bell, E.:2020}-\\cite{Graessley, S. :2019}. Among them, employee behavior recognition in the manufacturing industry is attracting increasing attention from both industry and academia \\cite{Meyers, T. D.:2019}. The importance of real-time process monitoring and analysis of employee behavior based on various sensors and other devices is increasing from both industry and academia nowadays.\\cite{White, T.:2020} - \\cite{Davis, R.:2020}” 2. Related work – the authors assume that background is the same throughout the video. Is the environment at the adopted position really not changing? And if so, how can this problem be solved? Response: We thank the reviewer for the important question. Our target is a production line with many routine tasks. In this type of line, the background rarely changes, such as when the machines change. On the other hand, when the background changes, it is when the machine layout of the factory changes drastically. In such a factory where the layout changes, the work procedures and contents also change, so they are not the subject of the work analysis in this research. We will address this issue in the future. 3. In the video data set – the authors state, that the length of time to perform the same work consecutively varies considerably, ranging from a few seconds to tens of minutes, and tasks depend on one other over a very long period. Is the same activity performed by the same employee then? Are there no time standards for individual activities? Are the steps always performed in the same sequence? Response: We thank the reviewer for the question. In many companies, there are no individual time standards, so they have to be identified using this kind of method. Although work procedures may change, they are included in the scope of work process analysis, and this research is able to handle them. 4. Please add an explanation why DeepLab v3 was chosen as a segmentation model for pre-training - description: "it is a simple encoder – decoder model consisting of a single stage" is not sufficient. Response: We thank the reviewer for the valuable comment. Deep Lab V3 is one of the best performers in the field of segmentation. For example, according to the ranking of https://paperswithcode.com/, it ranks 2nd in Pascal 2012 and 1st in Anatomical Tracings of Lesions After Stroke (ATLAS). In addition, it is widely used in practical applications and has stable operation. 5. How the findings can be exploited by future similar works? Response: As described in the paper, unlike the conventional image recognition task, the analysis of manufacturing tasks is difficult because the work time varies greatly for each task. Since this research has solved this problem, we believe that we have achieved a significant result in that we have developed a standard model for future work analysis in manufacturing and a dataset that can be used for pre-training. This research can be used to analyze various tasks in the manufacturing industry, including the detection of abnormalities in work, and will contribute to the diversification of work analysis tasks. 6. Please add the limitations of the work. Response: We thank the reviewer for the important comment. Our proposed method cannot be used when the factory layout changes drastically. If another task is added, relearning is required. Submitted filename: Response to Reviwers_final_publish.docx Click here for additional data file. 7 Dec 2021
PONE-D-21-26196R1
Improving work detection by segmentation heuristics pre-training on factory operations video
PLOS ONE Dear Dr. Kataoka, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Specifically, we authors need to improve statistical analysis, discuss the limitations of the methodology and address other issues noted by the reviewers. Please submit your revised manuscript by Jan 21 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Robertas Damaševičius Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: N/A ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: English language still requires a rework. Statistical analysis still requires more effort to prove the validity. Reviewer #2: The authors did not make all corrections I had proposed. I still have comments to the improved manuscript. Why were 17 new literature references added and indicated in the recommendations not added? Moreover, the most important problem that has not been explained in the article is the incorrect assumption that "In many companies, there are no individual time standards, so they have to be identified using this kind of method". In the production process, each workstation has a different duration of activity. You stated, that "Although work procedures may change, they are included in the scope of work process analysis". So you can use the method without any requirements? This aspect requires a thorough explanation. Also, my suggestion to present the limitations of the work was as also explained laconic. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
24 Jan 2022 Manuscript: PONE-D-21-26196 Response to Reviewers We have corrected the points pointed out by the reviewers. The red letters are the corrections made in the previous revision, and the blue letters are the corrections made in this revision. Underlines are grammatical corrections in the previous revision. Reviewer 1 1. English language still requires a rework. Response: We thank the reviewer for the recommendation. We have made some grammatical corrections to our manuscript. We have underlined the corrections. 2. Statistical analysis still requires more effort to prove the validity. Response: We thank the reviewer for the recommendation. We performed a Welch's t-test on the results. We have added the following sentences to the experiment section. “In addition, we performed a Welch’s t-test to confirm the significance of the difference between the output of each method. As a specific procedure, we randomly divided the dataset into 100 parts, calculated the F1 score for each class in each part, and performed Welch's t-test on each class. Consequently, the null hypothesis on all classes F1 score can be rejected at the 1\\% significance level for the difference between baseline and ours.” Reviewer 2 1. Why were 17 new literature references added and indicated in the recommendations not added? Response: We are very sorry about it. The number of citations has been reduced by mistake. We have added one reference that was not cited previously, and one more citation ([13],[14]). Therefore, we have added 18 citations from the first edition. [13] Andronie, M., Lăzăroiu, G., Ștefănescu, R., Uță, C., & Dijmărescu, I. (2021). Sustainable, Smart, and Sensing Technologies for Cyber-Physical Manufacturing Systems: A Systematic Literature Review. Sustainability, 13(10), 5495. [14] Patalas-Maliszewska, J., & Halikowski, D. (2019). A model for generating workplace procedures using a CNN-SVM architecture. Symmetry, 11(9), 1151. The text has been changed to the following sentence. “Recognition of employee activities in the production process is gaining 2 attention [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12]. Among them, employee behavior 3 recognition in the manufacturing industry is attracting increasing attention from both 4 industry and academia [13]. For example, to improve management and effectiveness of 5 employees ’learning processes, Patalas-Maliszewska et.al [14] used CNN-SVM to 6 extract features from video material concerning each work activity, and comparison 7 with the features of the instruction picture. The importance of real-time process 8 monitoring and analysis of employee behavior based on various sensors and other 9 devices has been increasing from both industry and academia [15] [16] [17] [18].” 2. Moreover, the most important problem that has not been explained in the article is the incorrect assumption that "In many companies, there are no individual time standards, so they have to be identified using this kind of method". In the production process, each workstation has a different duration of activity. You stated, that "Although work procedures may change, they are included in the scope of work process analysis". So you can use the method without any requirements? This aspect requires a thorough explanation. Also, my suggestion to present the limitations of the work was as also explained laconic. Response: We thank the reviewer for the recommendation. In our previous reply, we misinterpreted your question and gave a wrong answer. Each individual task has its own time standards. Throughout the dataset, the worker is the same, but there seems to be a time variance of a few minutes to a few tens of minutes when the work procedures are different. An explanation of this has been added to the manuscript with an example. Also, as you pointed out, the description of limitations in the paper was insufficient. We have added the following sentences to the limitations sub-section in the discussion section. “In this dataset, the same person performs works throughout the entire dataset. Since our method includes abstraction of the body parts of the worker by segmentation, it is considered to be robust to changes in the workers. However, the method may not be able to infer correctly when the procedures and motions of the workers change significantly due to differences in the workers. In this case, we need to add new training data. There were cases where the working time for the same task varied from several minutes to several tens of minutes. This is thought to be because the work procedures may differ even for the same work. For example, "Product check" work is performed continuously for several tens of minutes, but it may be interrupted irregularly when defective products are found, so the work may last from a few seconds to several minutes. Since the training data contains multiple work order patterns even for the same task, the model can predict when the work order changes to some extent. However, if the order of the tasks is highly irregular, or if the tasks are completely new and unknown, the inference may not be correct. In such cases, it is necessary to add training data or review the contents of the work labels. Our method is robust to changes in background information that is irrelevant to the task by using the segmentation of critical elements. However, since the method assumes a fixed camera, the trained model cannot be used as-is when the layout of the machines and the work environment related to the work changes. In these cases, it is considered necessary to re-collect the training data.” Submitted filename: Response to Reviwers_revision_3_2_publish.docx Click here for additional data file. 31 Jan 2022
PONE-D-21-26196R2
Improving work detection by segmentation heuristics pre-training on factory operations video
PLOS ONE Dear Dr. Kataoka, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Specifically, the authors need to address the reviewer comments seriously and provide a major revision of the manuscript according to every suggestion and comment of the reviewers. Please submit your revised manuscript by Mar 17 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Robertas Damaševičius Academic Editor PLOS ONE Additional Editor Comments: I urge the authors to address the reviewer comments seriously and revise the manuscript accordingly. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: No ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Literature review requires a major rework. Address each reference individually not as a group without any comment on contents of each. 1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12]; 20] [21] [22] [23] [24; [25] [26] [27] [2; ] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [5;. [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [80] [81] [82] [83]; 101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114]. A lot of the references seem to be inserted at random to inflate the overall number and does not bring much added value. Dedicated text must be added to each reference explaining results and relation to the paper. Figure 3 should be a proper UML. Still not all hyper parameters are clearly listed, nor reasoning exists on how they were determined. Some are in figure 3. I suggest also adding a dedicated table. Draw confusion matrix in color. Figure are low res and hard to see. Raw output data still not provided (data you draw all your results from) Reviewer #2: The authors have significantly improved the text given in the manuscrip, but the website: https://github.com/ShotaroKataoka/work%20detection%20factory%20pretrain is not available ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 16 Mar 2022 Reviewer 1 1. Literature review requires a major rework. Address each reference individually not as a group without any comment on contents of each. 1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12]; 20] [21] [22] [23] [24; [25] [26] [27] [2; ] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [5;. [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [80] [81] [82] [83]; 101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114]. A lot of the references seem to be inserted at random to inflate the overall number and does not bring much added value. Dedicated text must be added to each reference explaining results and relation to the paper. Response: Thank you for your suggestion. We have added comments on each reference. 2.  Figure 3 should be a proper UML. Response: We thank the reviewer for this comment. A new UML class diagram of our proposed system has been added as Figure 3. Therefore, the model diagram that was previously Figure 3 has been changed to Figure 4 which is a conceptual diagram to illustrate the layer structure of a complex model, so we have provided the simplest illustration. 3. Still not all hyper parameters are clearly listed, nor reasoning exists on how they were determined. Some are in figure 3. I suggest also adding a dedicated table. Response: Thank you for your comment. A new explanation of hyperparameters has been added as Figure 3. In addition, the followings were added to "3.1 CNN-LSTM model". “The hyper-parameters of the LSTM were determined through grid search by comparing validation scores.” “The hyper-parameters of these CNNs were determined through grid search by comparing validation scores. The mean and standard deviation parameters used in the standardization function were calculated from the entire image of training set.” 4. Draw confusion matrix in color. Response: Thank you for your comment. We have redrawn confusion matrix in color. 5. Figure are low res and hard to see. Response: Thank you for your comment. Figures 2, 4, and 5 have been changed to vector images. These have improved the resolution of the figures. 6. Raw output data still not provided (data you draw all your results from) Response: Thank you for your comment. All raw data has been redistributed at the following URL. This URL is provided in the introduction. In addition, programs for score calculation, t-SNE, and T-test are now available at this URL. https://drive.google.com/drive/folders/1yqKXnDFhRCyUmO7x2-wFZG9oLnBiB8L1?usp=sharing Reviewer 2 1. The website: https://github.com/ShotaroKataoka/work%20detection%20factory%20pretrain is not available Response: Thank you for your comment. It appears that the PDF link-click feature was incompatible with the underbar in the URL. Therefore, the URL was modified to be clickable. Submitted filename: Response to Reviwers_revision_4_1.docx Click here for additional data file. 11 Apr 2022 Improving work detection by segmentation heuristics pre-training on factory operations video PONE-D-21-26196R3 Dear Dr. Kataoka, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Robertas Damaševičius Academic Editor PLOS ONE Additional Editor Comments (optional): The reviewers presented a favourable opinion regarding the revised version of the article. However, the authors should check and resolve language errors. Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: N/A ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: My remarks have been addressed in this iteration, therefor I have no objection in further processing of this paper. I would still recommend a thorough review of English wording and style, or a dedicated service. Reviewer #2: The authors have improved the text given in the manuscript. I appreciate the change you made. I suggest to accept the paper. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No 11 May 2022 PONE-D-21-26196R3 Improving work detection by segmentation heuristics pre-training on factory operations video Dear Dr. Kataoka: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Professor Robertas Damaševičius Academic Editor PLOS ONE
Table 8

Comparison of label-by-label precision between the two LSTM-CNN models (test set).

The highest scores are expressed in a bold font.

ClassLabelBaselineOursDiff
Class 1Chores0.1856 0.3539 +0.1683
Class 2Paperwork0.1923 0.7523 +0.5600
Class 3Cylinder preparation 0.8037 0.7922−0.0115
Class 4Material preparation 0.9007 0.8849−0.0158
Class 5Ink roller adjustment0.5671 0.8591 +0.2920
Class 6Winding preparation (top)0.3986 0.5883 +0.1897
Class 7Winding preparation (bottom) 0.4836 0.3199−0.1637
Class 8Creation of sample0.3496 0.5564 +0.2069
Class 9Ink adjustment 0.7222 0.4343−0.2879
Class 10Workbench0.0219 0.4468 +0.4249
Class 11Product check0.3584 0.7305 +0.3721
Class 12Nonhuman work0.9623 0.9694 +0.0071
  10 in total

1.  3D convolutional neural networks for human action recognition.

Authors:  Shuiwang Ji; Ming Yang; Kai Yu
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2013-01       Impact factor: 6.226

2.  A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer.

Authors:  Adil Mehmood Khan; Young-Koo Lee; Sungyoung Y Lee; Tae-Seong Kim
Journal:  IEEE Trans Inf Technol Biomed       Date:  2010-06-07

3.  Moments in Time Dataset: One Million Videos for Event Understanding.

Authors:  Mathew Monfort; Alex Andonian; Bolei Zhou; Kandan Ramakrishnan; Sarah Adel Bargal; Tom Yan; Lisa Brown; Quanfu Fan; Dan Gutfreund; Carl Vondrick; Aude Oliva
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2019-02-25       Impact factor: 6.226

4.  Applying value stream mapping techniques to eliminate non-value-added waste for the procurement of endovascular stents.

Authors:  Ulf K Teichgräber; Maximilian de Bucourt
Journal:  Eur J Radiol       Date:  2011-02-11       Impact factor: 3.528

5.  Long short-term memory.

Authors:  S Hochreiter; J Schmidhuber
Journal:  Neural Comput       Date:  1997-11-15       Impact factor: 2.026

6.  A new application of value-stream mapping in new drug development: a case study within Novartis.

Authors:  Mareike Heinzen; Samuel Mettler; Annina Coradi; Roman Boutellier
Journal:  Drug Discov Today       Date:  2014-10-30       Impact factor: 7.851

7.  SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.

Authors:  Vijay Badrinarayanan; Alex Kendall; Roberto Cipolla
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2017-01-02       Impact factor: 6.226

8.  A deep convolutional neural network architecture for interstitial lung disease pattern classification.

Authors:  Sheng Huang; Feifei Lee; Ran Miao; Qin Si; Chaowen Lu; Qiu Chen
Journal:  Med Biol Eng Comput       Date:  2020-01-22       Impact factor: 2.602

9.  Contextual Action Cues from Camera Sensor for Multi-Stream Action Recognition.

Authors:  Jongkwang Hong; Bora Cho; Yong Won Hong; Hyeran Byun
Journal:  Sensors (Basel)       Date:  2019-03-20       Impact factor: 3.576

10.  Classification of sporting activities using smartphone accelerometers.

Authors:  Edmond Mitchell; David Monaghan; Noel E O'Connor
Journal:  Sensors (Basel)       Date:  2013-04-19       Impact factor: 3.576

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.