Next Article in Journal
Potential of a New, Flexible Electrode sEMG System in Detecting Electromyographic Activation in Low Back Muscles during Clinical Tests: A Pilot Study on Wearables for Pain Management
Previous Article in Journal
Recent Advances in Trajectory Planning and Object Recognition for Robot Sensing and Control
Previous Article in Special Issue
Industry 5 and the Human in Human-Centric Manufacturing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

System Design for Sensing in Manufacturing to Apply AI through Hierarchical Abstraction Levels

by
Georgios Sopidis
1,2,*,
Michael Haslgrübler
1,
Behrooz Azadi
1,
Ouijdane Guiza
1,
Martin Schobesberger
2,
Bernhard Anzengruber-Tanase
1 and
Alois Ferscha
2
1
Pro2Future GmbH, Altenberger Strasse 69, 4040 Linz, Austria
2
Institute of Pervasive Computing, Johannes Kepler University, Altenberger Straße 69, 4040 Linz, Austria
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4508; https://doi.org/10.3390/s24144508
Submission received: 4 June 2024 / Revised: 8 July 2024 / Accepted: 11 July 2024 / Published: 12 July 2024
(This article belongs to the Special Issue Human-Centred Smart Manufacturing - Industry 5.0)

Abstract

:
Activity recognition combined with artificial intelligence is a vital area of research, ranging across diverse domains, from sports and healthcare to smart homes. In the industrial domain, and the manual assembly lines, the emphasis shifts to human–machine interaction and thus to human activity recognition (HAR) within complex operational environments. Developing models and methods that can reliably and efficiently identify human activities, traditionally just categorized as either simple or complex activities, remains a key challenge in the field. Limitations of the existing methods and approaches include their inability to consider the contextual complexities associated with the performed activities. Our approach to address this challenge is to create different levels of activity abstractions, which allow for a more nuanced comprehension of activities and define their underlying patterns. Specifically, we propose a new hierarchical taxonomy for human activity abstraction levels based on the context of the performed activities that can be used in HAR. The proposed hierarchy consists of five levels, namely atomic, micro, meso, macro, and mega. We compare this taxonomy with other approaches that divide activities into simple and complex categories as well as other similar classification schemes and provide real-world examples in different applications to demonstrate its efficacy. Regarding advanced technologies like artificial intelligence, our study aims to guide and optimize industrial assembly procedures, particularly in uncontrolled non-laboratory environments, by shaping workflows to enable structured data analysis and highlighting correlations across various levels throughout the assembly progression. In addition, it establishes effective communication and shared understanding between researchers and industry professionals while also providing them with the essential resources to facilitate the development of systems, sensors, and algorithms for custom industrial use cases that adapt to the level of abstraction.

1. Introduction

Industrial manufacturing and assembly processes are fundamental to modern society, driving innovation, economic growth, and technological advancement [1]. These procedures rely heavily on human expertise and involve manual assembly tasks and activities [2,3]. Despite the prevailing focus on locomotion activities in human activity recognition (HAR), it is necessary to recognize that human industrial activities involve complexities far beyond locomotion activities, which need to be explored. In this field, the efficient execution of activities is essential for achieving production goals, ensuring product quality, and maximizing operational efficiency. However, understanding the numerous activities involved in industrial processes poses significant challenges, requiring sophisticated approaches for analysis, classification, and optimization.
At the core of industrial operations lie the activities performed by individuals or individuals in combination with machines or automated systems [4]. According to the authors of [5], manual assembly tasks have a clear distinction regarding complexity. Objective assembly complexity refers to inherent properties of the assembly process, such as the number and dependencies of components. In contrast, perceived assembly complexity is subjective and influenced by personal capabilities and experience. From the manipulation of individual components to the orchestration of complex production lines, each activity plays a crucial role in shaping the outcome of the manufacturing process. Yet, the complexity and diversity of these activities present difficulties for researchers, engineers, and workers in the field who want to understand them better to enhance performance.
Traditional approaches for activity analysis have often relied on recognizing few activities and in simplistic categorizations, such as simple versus complex activities [6,7]. Simple activities, such as walking, running, sitting, or standing, are extensively researched, whereas everything else is categorized as complex. Consequently, the recognition of complex activities remains relatively unexplored, as highlighted by the authors in [8]. While these frameworks provide a basic understanding of activity complexity, they prove ineffective in capturing the nuanced structure and hierarchical organization of industrial tasks or miss the depth and detail needed to analyze and categorize industrial activities. Activities can be categorized along a scale of increasing complexity, as discussed by Schneider et al. [9], where simple activities occur for a short period, while complex activities may occur for more prolonged periods. Furthermore, Peng et al. [10] state that the features designed for simple activities are poor at representing complex activities.
Our work proposes a novel hierarchical activity abstraction framework tailored to the specific level of abstraction, ensuring that the activity model fits the task in industrial contexts. Within this framework, distinguishing between activity abstractions becomes critical since, e.g., distinct statistical, temporal, and spatial properties appear at different levels of abstraction, which could lead to more efficient sensor placement, information capture, feature extraction, resource utilization, and architecture design.
Our approach contributes to industrial activity analysis and optimization by examining the structure of industrial production and assembly processes to enable an understanding of the wide range of involved activities and their role in complex operational environments. In that regard, we introduce an activity abstraction framework that acknowledges factors such as complexity, granularity, and hierarchical organization, providing a structured approach to activity analysis. Additionally, we recommend suitable AI-guided solutions for optimizing industrial processes, ranging from real-time activity monitoring, which forms the core of our analysis, to broader areas such as predictive maintenance and adaptive control strategies. Finally, our work intends to enable the development of recognition technologies and the awareness of industrial assembly in uncontrolled environments, allowing researchers and practitioners to design systems, sensors, and algorithms aligned with the required level of abstraction.
The following sections outline the structure of the paper: Section 2 provides a review of related work. Section 3 introduces an industrial assembly process and proposes our taxonomy framework. In Section 4, we compare our model with the state of the art by applying our and the existing approaches to a real industrial use case. Section 5 presents the application of our taxonomy to develop AI systems for custom industrial use cases, recommending necessary factors and variables to consider during the design process in addition to the AI model itself. Section 6 provides a detailed discussion of the proposed taxonomy, including its limitations, and outlines the future directions of the study. Finally, Section 7 presents a summary of the discussed points throughout the paper.

2. Related Work

In industrial assembly, human activity recognition requires significant attention due to the critical role humans play in the process [11]. Assembly tasks, characterized by intricate processes and workflows in different fields, are the focus of HAR, which aims to develop models for understanding and classifying these activities, ranging from simple actions to complex, multi-step processes [6]. However, while this categorization assists in assessing complexity, it may overlook factors such as context and granularity. By acknowledging the connection between industrial assembly and HAR, we can explore how insights from activity recognition research can enhance the optimization of manufacturing and assembly operations.
Two main approaches prevail in the literature: binary class and non-binary multiclass approaches, as shown in Table 1, which groups research publications from various domains, such as manufacturing, into recognized activity abstraction categories.

2.1. Binary Class Approaches

As mentioned by the authors in [12,13,14,33], compared to simple activities, complex activities are composed of actions and are much more complex and semantically consistent with a human’s real life. Ramanujam et al. [48] attempted a division of public HAR datasets that apply deep learning techniques into simple and complex activities, categorizing them into conventional and hybrid models. Nevertheless, the problem remains without the additional distinctions because any task that is more complicated than simple is automatically labeled as complex. Meanwhile, Bouton et al. [47] investigate complex daily activities in sedentary settings such as remote work or study environments, concluding that gaining insights into an individual’s daily activities can help develop applications that improve their well-being and overall health. The work from [37] exhibits higher detection accuracy and less ‘intersubject variability’ as mentioned by the authors, since simple types of activities are performed similarly across users. Additionally, they observed confusion between simple and complex activities such as “cooking and standing” or “sitting on sofa and lying on sofa”. Chen et al. in [44] separate human activities into simple human activities (SHAs) and complex human activities (CHAs), where SHAs may be recognized with an accelerometer, whereas CHAs need multimodal sensor data.
The distinction between activities is crucial in HAR; however, existing taxonomies for human activity abstraction levels are often limited to simple and ambiguous categories that cannot capture the complexity and diversity of human activities across various domains [38,42,43,45,51,52,53,54,55,56,62,63,64,65,66,67]. Zhang et al. in [68] explored complex activities such as eating, which involves a variety of movements, and reported that they increase the challenge and difficulty for HAR methods. For example, an activity such as “cooking” involves actions like “cut”, “take”, or “mix” in different order and frequency. The authors in [16,39] refer to these actions as “micro-activities” or micro-motions and characterize the complex activities as “macro activities” or macro-motions. In their work, they discuss how micro-motions serve several purposes in understanding macro-motions. This includes (i) confirmation of the execution of all necessary micro-motions; (ii) facilitating evaluation of their sequence correctness; and (iii) recognizing differences among macro-activities or their execution variations.
Different types of activities require different granularity levels, as mentioned by the authors in [34]. In their study, they utilized the terminology “complex activities” of daily living (ADLs), which are built on top of simple activities and convey more specific contextual information, and “simple or coarse-grained activities”, such as walking, sitting, and cycling, which may be directly assessed from an inertial sensor unit.
The terms “coarse-grained and fine-grained activities” and “low-level and high-level activities” are also used to describe different aspects of activity abstraction. The division between coarse-grained or gross and fine-grained activities,which reflects the level of detail or granularity in activities, appears in industrial manufacturing [15] but also in various other domains [57] including daily activities [40]. Coarse-grained activities are broad categories or high-level actions, while fine-grained activities are more specific and detailed. Fine-grained activities provide a richer understanding of the activity by capturing smaller sub-actions or variations within a broader category. However, this division may not explicitly address complexity or context.
On the other hand, the distinction between low-level and high-level activities refers to the hierarchical organization of activities [17,18]. Low-level activities are typically microscopic or elementary actions that form the building blocks for higher-level activities. High-level activities encompass a collection of lower-level actions or represent more abstract concepts. This division emphasizes the hierarchical relationship between activities and can be useful for understanding the composition and structure of complex activities. Investigating this topic is interesting since, in most settings, only a few activities are considered (less than 10), and only a few of them discuss hierarchical dependencies between activities on lower and higher levels [69].
Previous research has proposed various levels of activity abstractions, ranging from activities such as walking and sitting to motions such as lifting and grasping. The hierarchical approach to human activity recognition involves recognizing simpler activities initially and using them to recognize higher-level activities. The representation of high-level activities is based on the sub-events or sub-activities that serve as observations derived from the higher-level activity [41]. The use of sub-events not only makes the recognition process computationally tractable and conceptually understandable but also reduces redundancy in the recognition process by reusing recognized sub-events multiple times. An example given by the authors in [41] is that the high-level activity of “fighting” may involve detecting a sequence of sub-events such as punching and kicking interactions.
By distinguishing between activity abstractions, researchers can develop methods or approaches that are tailored to the specific level of abstraction [70,71], thereby improving activity recognition models.
One main finding derived from the authors in [32] is that “activities involving several body parts are more easily recognizable and allow for shorter window sizes”. Yet, they state that the recognition accuracy for such complex activities is still low. This is due to several factors like (i) the inter-class similarity, (ii) the difficulty in defining each activity and its boundaries, and (iii) the lack of open datasets. Furthermore, activity abstractions provide means of interpretability that enable comprehension of the behavior of the assisted user or system. The ability to identify the appropriate level of abstraction is critical for (i) developing effective activity recognition models, (ii) better understanding and gaining insights into human behavior, and (iii) modeling human behavior in a given context.
Several ongoing challenges persist, including issues such as incomplete information within activity datasets, insufficient contextual details accompanying activity data, and the complexities involved in modeling composite activities [6]. Composite activities, as highlighted in the literature by [40,59], pose particular challenges due to their composition of multiple shorter activities and the associated temporal decomposition required for their recognition. Kulsoom et al., in [60], extend the concept of composite activities and present additional categorizations based on operation types, namely concurrent, sequential, and interleaved activities. Moreover, despite the growing interest in interaction recognition, the recognition of activities involving groups of people [30] has received relatively less focus, as identified by Morshed et al. [61].
In their work, the authors in [72] state that “the existing taxonomies in the field of activity recognition, while valuable, exhibit limitations in their categorization by not encompassing sufficient distinct activity categories, thereby indicating the need for further refinement and improvement”. The presented studies provide evidence of the prevalence of binary distinctions and the limitations in existing methodologies. Thus, describing activities as simple or complex, high or low level, and fine or coarse-grained hinders a detailed enough understanding of the underlying components and structures of human activities, the different sub-components of activity, and how they relate to each other. Therefore, there is a research gap in exploring beyond the simple–complex dichotomy, which can be addressed by incorporating additional levels of activity abstraction.

2.2. Multiclass (Non-Binary) Approaches

The hierarchical framework provided by Kuutti et al. [73] offers insights to address the limitations of binary classifications in human activity recognition. In this, the authors explain how activities consist of chains of actions, with each action comprising individual operations, where individual actions become comprehensible only when viewed within the larger context of the activity. Recent findings by Miranda et al. [46] suggest that a promising strategy for dealing with complex HAR is to model activities as a series of dependent atomic actions. In this context, Saguna et al. [50] focus on the semantics of the application and explain it as a fundamental activity unit that cannot be further decomposed.
Following the previous explanation and using comparable methods in the existing literature, we refer to the framework of Atomic Actions, Primitive Tasks, Tasks that comprises three hierarchical levels: task, primitive task, and atomic action [19,20,21,22,23,74]. An assembly task is decomposable into multiple continuous primitive tasks, which can be executed sequentially or in parallel. Similarly, each primitive task can be further broken down into continuous atomic actions, which are also capable of sequential or parallel execution. While the proposed framework addresses the binarization encountered in the literature regarding activities, our approach proposes different classifications for levels or actions involved in certain activity stages, respectively. The proposed taxonomy provides a more refined classification for some actions than existing frameworks to cover the complexities of human activity in industrial contexts. We believe that certain actions, as defined in their context, can be further decomposed or integrated into a broader hierarchy of activities.
The challenge in activity recognition lies in the generalization of models across diverse contexts [65] through recognizing patterns in sensory input. Conventional approaches often oversimplify activities as either simple or complex without accounting for contextual variations. Furthermore, human activity recognition tends to be associated with locomotion activities, and often instead of employing categorizations based on factors like complexity or granularity, many studies opt for a broad approach to activities without specific classification [25,27,28,29,31,58]. An additional aspect was addressed by the authors in [36] where they provide insight into the characteristics of complex activities, highlighting that complex activities exhibit longer duration, comprise a combination of simple activities, and encompass multiple behaviors. These complex activities often possess high-level semantics, such as daily activities like cooking, cleaning, or industrial assembly tasks.
Examining binary and multiclass approaches revealed that certain categories of activities often receive less attention, as they are often processed using the same models designed for other categories. However, these activities are important in the recognition and classification process and require customized methods due to their distinct attributes and complexities. Hence, it is critical to distinguish the nuanced differences between activities across various contexts and categorize them appropriately to leverage context-specific characteristics for understanding user behavior to improve activity recognition systems in real-world settings.

3. Proposed Taxonomy Model

We propose a taxonomy using a hierarchical activity abstraction structure that combines the simplicity of the simple–complex division with the clarity of the non-binary approaches to study activities at different levels of granularity and abstraction along with their respective sub-divisions. Additionally, our work extends beyond taxonomy development to offer practical applicability and relevance in real-world industrial scenarios.

3.1. Industrial Assembly Process

In industrial manufacturing, assembly processes are essential for joining multiple components to produce functional products [11]. Whether it is assembling automobiles, electronic devices, or machinery, the efficiency and effectiveness of the assembly process impact the success of the manufacturing operation. Workflow optimization depends on an understanding of the hierarchical structure and relationships between different phases of assembly. Every stage of the assembly process on a manufacturing floor, typically involving a frame (e.g., chassis) where modules are systematically assembled to construct complex products, requires coordination, precision, and adherence to defined work instructions. In [75], the authors describe the hierarchical nature of complex product assembly data, highlighting three granularity levels of product, assembly, and part. They state that the refined management of assembly processes and hierarchical organization of assembly data can be achieved by decomposing complex assembly activities into more detailed activities.
Building upon this framework, our analysis of real-world assembly processes revealed the hierarchical structure observed in complex assembly scenarios [11,75,76,77]. Moreover, our analysis aligns with findings from the literature regarding the roles of modular assembly and units or sub-assemblies in the assembly process. By breaking down complex assembly tasks into modular units, manufacturers can enhance efficiency and flexibility in their production workflows, separate design tasks into distinct units, and simplify the design process while prioritizing product customization [78,79] and responsiveness [80]. The authors in [81] highlighted the importance of the number of modules, joining sequences between modules, and tolerance management issues in car body design. In assembly processes, modules are self-contained subsystems designed to be constructed, examined, manufactured, and developed independently from the overall system for independent integration. This ensures interchangeability, standardization, and re-usability, with clear interfaces and relative independence from other modules [82].
In addition, our work recommends adding the post-assembly processes [83] as a level to the assembly process hierarchy where the concept of the final product is contextualized within the scope of the specific manufacturing line. This indicates that what constitutes the final product may vary depending on the manufacturing line. For instance, what serves as the final product on one manufacturing line might be considered a sub-assembly or component in another line. With this feature, industrial processes are observed, covering every stage of the manufacturing process, from components to finished products, by performing the required manual activities for each stage. Therefore, the assembly chain example presented in Figure 1, for understanding various industrial workflows, does not extend to subsequent stages of product development or distribution but focuses on the assembly process of a product, which is completed on the production line.

3.2. Abstraction Levels of the Taxonomy Framework

The proposed taxonomy presented in Figure 2 introduces five stages—atomic, micro, meso, macro, and mega—providing a hierarchical framework for analyzing and categorizing activities within the assembly process of discrete manufacturing products to understand the progression of assembly tasks. In our framework, we employ the term “activities” for the different levels to ensure consistency, adopting a unified terminology throughout our analysis. With the systematic use of the terms atomic, micro, meso, macro, and mega, we denote the distinctions of activities at various stages of the assembly workflow.
Atomic: This level includes the smallest self-contained operations or steps that can be performed by a human in a manual product assembly scenario involving the basic components or tools for discrete or singular manipulations within a broader action or activity. Examples of such atomic operations or sub-tasks that contribute to completing a particular action could be grasping the screw or tool, positioning the screw or tool, turning the screw or tool, and releasing the screw or tool.
Micro: This level includes the smallest recognizable actions within a task or process that serve a specific goal, detectable using sensors or observation, requiring tools to execute certain tasks like joining or attaching components. They are composed of a sequence of atomic steps, generating singular activities, including operations that can be executed by a single individual without extensive planning or coordination. They are characterized by their relatively short duration and repetitive pattern, and they can be performed independently or as part of a larger activity as an individual work step. At the micro-level of assembly, individual tasks are executed by integrating multiple atomic operations, each involving distinct components, tools, or parts. The sensor data required to detect screwing as an activity can be limited to the movement and orientation of the tool, as this information alone contains indicative signals of the screwing action. An example of a micro-activity is the entire process of picking up a tool, a screw, and a component, positioning the screw in the part, and using the tool with rotating movements to tighten a screw.
Meso: This level includes a collection of coordinated actions and operations from the micro-level and/or atomic level, forming a coherent sequence to achieve an objective driven by a particular motive. This level bridges the gap between fundamental atomic and micro-level processes toward increasingly complex processes, representing a crucial stage in assembly workflows to accomplish intermediate goals typically undertaken by a single individual or a small group of individuals. At the meso-level, sub-assemblies, and modules of the product are prepared through the aggregating micro-level activities, which are each designed with specific functions and aims. They are self-contained units that meet intermediate ends, which are essential steps toward the final assembly. An example of a meso-level task would be to prepare the sub-assemblies of a module, such as the cash box for an ATM.
Macro: At the macro-level, tasks evolve into broader and more complex tasks, encompassing multiple steps and components from the previous meso-level. This process involves coordinating multiple meso-level activities, which include the assembly of sub-assemblies and modules prepared in earlier stages. Macro-activities can involve sequences of actions, including those demanding higher-level cognitive functions like decision making and planning. These tasks might require coordination between individuals, groups, or machines (robots) where the outcome of this coordination results in the accomplishment of a complex goal, such as the full assembly of a product (e.g., the ATM) depicted in Figure 1.
Mega: This level comprises a series of coordinated macro-level processes to ensure the smooth functioning and quality output of industrial operations. In a broader sense, mega-level activities may involve various tasks conducted by humans beyond the assembly line operation yet included in the industrial workflow. These could include packaging products for safe transport and inspection, testing functionality, or conducting quality-control checks. These activities contribute to the overall goal of achieving efficient and effective production processes on a large scale. Examples include the post-assembly operations that exist in mass production lines of individual products, such as cars or other complex items assembled from multiple modules or sub-assemblies.
Following the presented hierarchical structure of complex processes in industrial manufacturing, focusing on the assembly of products, we connect these concepts with the industrial example illustrated in Figure 1. This example provides an overview or a reference to our approach rather than a comprehensive representation of all applications included in an industrial setting. In the assembly of automated teller machines (ATMs), the hierarchical structure of assembly processes is evident across separate abstraction levels. At the component level, where atomic activities are involved, screws, tools, and individual parts like display panels form the foundational building blocks of the ATM. These components are then manipulated and used to create the required sequential work steps that contribute to developing units at the micro-level. They often involve repetitive tasks such as tightening motions with rotating movements using tools. While the micro-level focuses on one specific single action every time, the meso-level relies on the coordination of multiple actions to complete larger modules like cash-handling systems or user interface systems that are assembled independently during that level. All modules, units, and the remaining components are assembled into a complete ATM system at the product or macro-level. Subsequently, in the post-assembly stage, which represents the mega-level, the final product is integrated into the manufacturing line, undergoing additional processes such as quality control, software installation, functionality testing, and packaging.
In addition to the hierarchical structure demonstrated within the ATM assembly, our activity recognition taxonomy extends beyond this specific application to cover various industrial contexts. To validate its versatility, we have developed a hierarchical formulation that represents assembly processes using equations. This conceptual representation, through equations, serves as an initial tool for our activity recognition taxonomy, providing insights into the intricacies of assembly tasks to understand the relationships between activities within industrial assembly workflows. In this model for assembly processes, we utilize a set of symbols to quantify various attributes at different hierarchical levels. The specific step of the assembly activity, performed at a given level, is indicated by the symbol A level , and the time spent on each step of the activity is indicated by the symbol t level . Moreover, a level represents the kinematic characteristics associated with the assembly step or operations at that level, like acceleration or angular velocity. The equation P level expresses the overall occurred activities at a specific level within the taxonomy framework. Throughout the model, summations aggregate the contributions of individual activities, providing an analysis of assembly workflows.
  • Hierarchical formulation of our taxonomy in an assembly process
  • Attributes:
  • P level : Activity on level of taxonomy.
  • A level : Step of assembly activity.
  • t level : Time spent on step of assembly activity.
  • a level : Kinematics properties on step of assembly activity
P atomic i = f ( A i , t i , a i )
P micro j = { P atomic j } j = 1 n
P meso k = { P micro k } { P atomic k } k = 1 m
P macro l = { P meso l } { P micro l } { P atomic l } l = 1 p
P mega q = { P macro q } { P micro q } { P PA q } { P atomic q } q = 1 r

4. Comparative Analysis with SOTA

In this section, we compare our introduced taxonomy to the current state-of-the-art (SOTA) approaches that exist in the literature, and in the following section, we present the application of our taxonomy to real-world industrial scenarios. After reviewing relevant publications outlined in Section 2, we generated a categorization methodology for an assembly scenario to visualize the differences between binary and non-binary approaches and the proposed one. In Figure 3, we present a simplified process assembly of an ATM illustrating some key steps that capture a subset of the entire assembly, as the complete one typically involves numerous additional steps and complexities. This example is derived from original scenarios and real-world use cases documented in our prior work [66,70,71,76,84], discussing ATM and digger assemblies in more detail.
In simple/complex, fine-grained/coarse-grained, low-level/high-level classifications, we observe a two-stage categorization method between detailed and broader classifications as outlined in our related work. These approaches emphasize the distinction between the initial assembly tasks as sub-actions and their combination as broader categories. For instance, in the simple/complex approach, a screwing process already belongs to the complex category, since it is considered much more complicated compared to lifting a hand or grabbing an object. Additionally, in the fine-grained/coarse-grained, low-level/high-level categories, lower-level classifications consist of activities similar to screwing, as the primary stage, and continue afterward to more demanding activities.
At the atomic level, activities involve manipulating individual tools and components through discrete or singular manipulations. These actions serve as the building blocks for more complex tasks. For instance, lifting a tool or grasping a single screw constitutes an atomic action. Moving to the micro-level, tasks involve the execution of specific actions that combine interactions between at least two elements from the atomic level to form a single type of action. For example, screw tightening involves grasping, lifting, and turning a screw with a tool. The micro-level is characterized by repetitive actions, such as repeatedly rotating a screw to fix it tight. At this level, a difference can be noticed between the proposed methodology and the non-binary approaches reported in the literature. In our methodology, the screwing process occupies a different level compared to lifting or grasping because it is considered a combination of those actions.
Transitioning to the meso-level, activities become more comprehensive, involving the combination of micro-level tasks to accomplish intermediate goals. For instance, assembling an ATM’s front panel system requires the integration of micro-level activities. At the macro-level, larger components and subsystems, previously prepared at the meso-level, are brought together to construct the final product. This step involves coordinating multiple meso-level activities, each contributing essential components or modules designed with specific functions and aims. For example, assembling the ATM at the macro-level entails integrating pre-assembled sub-parts, such as the cash-handling system and front panel to the chassis into a cohesive structure. In our methodology, we view the assembly process as an integral part of a larger manufacturing ecosystem rather than an isolated event. In this regard, at the mega-level, the focus extends beyond assembly to encompass broader activities such as quality control, inspection, and overall process optimization, as opposed to existing approaches that often overlook such human tasks. In the second example of an industrial assembly process, the focus shifts from screwing processes to welding processes within the context of car assembly. This hierarchical framework presented in Figure 4 remains the same, including the five levels: atomic, micro, meso, macro, and mega, illustrating the progression of tasks.

5. Application of the Taxonomy for Guiding AI System Design

In this section, we proceed to the practical application of our taxonomy by presenting its deployment in real-world industrial scenarios. In our analysis, we identify and propose key categories to guide or support practitioners in the research and implementation of AI applications and systems in industrial assembly processes, as previously presented in Figure 5. We provide recommendations for designing AI systems customized to meet the specific requirements of individual use cases and address the design of an AI system as a collective contribution of multiple elements beyond just the AI model itself. These elements encompass but are not limited to (i) sensor placement, sensor types, and sensor mobility [7,60,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102], (ii) sampling rate, duration, frequency of actions [66,76,103,104,105,106,107,108,109], (iii) preprocessing techniques, window size, models to use [8,24,26,35,94,97,103,109,110,111,112,113,114,115,116,117,118], and (iv) interaction and feedback mechanisms [93,119,120,121,122,123].
At the atomic level, where activities involve individualized tool interactions and component manipulations, sensors are typically mounted on the parts, components, or body, ensuring the capture of object movements and interactions. These identification sensors (ID) encompass a diverse range, including RFID, weight sensors, proximity-aware, pick-by-light systems, or EMG, allowing for multifaceted data collection. Preprocessing techniques applied to raw sensor data involve signal filtering, feature extraction, peak detection, and PCA approaches, which, in the scope of our work, are selected for refining the data for subsequent analysis. Machine learning models, such as basic signal processing and classifiers like SVMs and decision trees, are employed to recognize basic manipulation tasks characteristic of this level, sometimes without the need for temporal windows.
Transitioning to the micro-level, where activities entail more specific tasks involving interactions with parts from the previous atomic-level, a similar sensor setup is maintained with an increased focus on hand manipulations and sequences of atomic-level tasks. Preprocessing techniques, if applied, become more refined, incorporating signal segmentation and feature engineering within time-series data to capture sequential actions. Machine learning models evolve to include CNNs and RNNs capable of analyzing more complex interactions, while short or adaptive windows facilitate the analysis of sequential actions and interactions. Despite these advancements, the focus remains usually on single users engaging in short-duration tasks with high-frequency actions.
As we progress to the meso-level, activities involve performing specific tasks using larger components, which are typically occurring at the main assembly station or sub-assembly stations. Sensor placement may extend to encompassing the assembly workstation, capturing interactions between larger components or modules. In order to facilitate comprehensive data capture, vision sensors, depth cameras, and Mocap systems are introduced. In addition, preprocessing techniques become more sophisticated and memory extensive, incorporating feature extraction, windowing, augmentation techniques, image filtering, and object recognition. Computer vision and deep learning models, such as CNNs and LSTMs, are employed to recognize complex assembly patterns with longer or adaptive windows enabling the analysis of entire assembly sequences and sub-tasks.
At the macro-level, where activities encompass complete assembly processes and integration tasks, the sensor setup may cover the entire assembly workstation or multiple adjacent workstations, combining sensor systems [102] to capture interactions across multiple sub-assemblies or components. This level requires advanced AI models, ensemble models, and rule-based systems to analyze complex assembly workflows effectively. Long and adaptive windows, depending on the type of employed data, facilitate the analysis of entire assembly processes and integration tasks, while the subject number may extend to multiple users or groups engaged in longer assembly processes and system integration.
Finally, at the mega-level, activities involve comprehensive manufacturing processes and the optimization of production workflows across the entire factory floor or assembly facility. Sensor placement becomes more varied, capturing interactions across the entire facility, while preprocessing techniques involve global context analysis, advanced machine learning, and data fusion techniques. Advanced AI models, deep learning architectures, and custom ensemble models are employed to analyze complex manufacturing processes and optimize production workflows. Variable window sizes and sampling rates are tailored to capture diverse manufacturing activities and system performance metrics across the facility with infrastructure supporting real-time data processing and the optimization of manufacturing operations.

6. Discussion

The employment of our multiple-level activity abstraction scheme offers several compelling benefits. Firstly, it facilitates a better understanding of complex systems by breaking them into smaller-level activities. This breakdown allows researchers and industry professionals to gain insights into the individual components that constitute the larger system, leading to a more detailed understanding of workflows, training programs, and automation systems. Consequently, it becomes easier to identify areas for improvement, troubleshoot problems, and optimize overall system performance. Apart from that, recognizing such complex activities is also important for daily living activities because it can enable tracking digital well-being, providing context-aware user experiences and notifications, and allowing better content recommendations [34].
In our understanding of assembly processes, certain activities may appear similar to atomic tasks but differ in nature and purpose. For instance, while activities like walking between workstations or reading assembly instructions are essential, they are either locomotion activities or cognitive tasks rather than physical manipulations of components. Although locomotion activities are inherently part of the overall activity framework in assembly tasks, individuals typically maintain a static posture while executing them, focusing on manipulation and precise task execution over physical movement. On the other hand, operating machinery or inspecting finished products involves complex activities beyond the scope of atomic actions. These tasks encompass a range of actions, such as controlling machinery or evaluating product quality, that require cognitive judgment and coordination. While they contribute to the overall process, they are distinct from the discrete singular actions associated with atomic tasks that would be under the latter stage in our hierarchy.
Within the domain of human activity recognition in industrial assembly tasks, the ability to recognize micro-activities, such as screwing, without the requirement to detect separately every underlying sub-action emerges as an interesting aspect. This is due to the goal-oriented nature and foundational role that micro-activities play in the hierarchical structure of assembly processes. Even though a micro-activity can be divided into smaller steps, recognizing the indicative pattern of the micro-activity reduces the need to understand separately the independent atomic actions except when tasks specifically require identifying manipulations at the atomic-level. The rationale for this is that detecting all the sub-actions would require more sensors and computing capacity, which might not essentially provide more meaningful information about the task at hand. Instead, the focus is on detecting the overall activity and its characteristics (such as duration, frequency, etc.), which can be captured using a smaller number of sensors and analyzed more efficiently while indicative of the underlying atomic-level components.
Regarding the duration of activities, each level exhibits a distinct timescale in a particular scenario. Micro-activities, which are characterized by repetitive actions, typically have shorter duration. As activities progress to higher levels, incorporating multiple previous activity levels, the duration increases proportionally, reflecting the cumulative complexity and scale of tasks involved. For instance, a meso-level activity comprising 10 micro-activities has a longer duration, which is calculated by aggregating the duration of each micro-level and atomic-level activity. This temporal progression also highlights the hierarchical nature of assembly tasks and underscores the incremental development of products across different levels. Nevertheless, there may be some slight overlap in the activity duration due to the variety of activities in the industry. Therefore, while the timescale can indicate each level, scenario analysis is required for interpretation.
Additionally, atomic, micro, meso, macro, and mega-level activity abstractions contribute to a common understanding that bridges the gap among people of diverse roles and expertise in the assembly process, facilitating effective communication and knowledge transfer to achieve greater productivity, execution accuracy, and scalability. Each successive abstraction level builds upon actions that exhibit variations, such as the number of individuals involved, the execution station utilized, the overarching goals they serve, and the complexity of the tasks they encompass. For instance, at the atomic or micro-levels, activities may involve individual actions such as grasping a component or tightening a screw. These actions are relatively simple and executed by individual workers at specific workstations. As we move to higher levels such as the meso or macro and mega-levels, micro-activities become more complex, involving coordination among multiple workers or machines across different stations to achieve larger production goals. Furthermore, the end goals of micro-activities differ across levels. At the micro-level, the focus may be on completing discrete assembly tasks, whereas at the macro and mega-levels, micro-activities contribute to broader objectives such as optimizing production efficiency or meeting customer demand.
The hierarchy of assembly stages can also benefit both the arrangement of sensors and data analysis. By grouping the assembly process into distinct stages, we can deploy wearable sensors, such as IMUs or visual sensors, to gather relevant information. These stages can provide real-time data awareness of worker movements and interactions with components or data collection tools at each level. In addition, researchers can identify and consider more factors related to specific cases. For instance, experiments at the atomic level may occur in controlled lab environments, ensuring precise data collection. However, as assembly complexity evolves at the macro and mega-levels, experiments transition to real-world factory settings, introducing factors that may include limited lighting, occlusion effects, noise, and other environmental variables along with the system’s obtrusiveness. The complexity of tasks also impacts the frequency of actions, with activities occurring more frequently at lower assembly levels, such as the atomic and micro-levels, compared to the macro and mega-levels, where actions occur less frequently. Additionally, privacy concerns rise with increasing complexity reflecting the diverse and extensive nature of data collection across the factory floor while underscoring ethical and practical considerations in system development. Addressing potential data privacy implications led us to prioritize the selection of less invasive and privacy-friendly sensors for the applicable levels, considering the individual’s privacy rights.
These differences underline the need for tailored development measures and regulations compliance, as assembly activities progress from lower to higher levels of complexity, which plays a significant role in user acceptance, system deployment, and overall effectiveness in real-world assembly environments. Moreover, considerations surrounding data processing and power consumption grow across assembly levels. While at the atomic and micro-levels, data processing and sensor cost are low, progressing to the macro and mega-levels, there is an increasing requirement for computational and finance resources, especially if the generated data will be stored.
Overall, it is reasonable that while we try to create a versatile and robust framework, there may be limitations to covering each case in all different domains of HAR. Consequently, we present our method for industrial assembly processes, but it may need adjustments to describe activities in other domains. Additionally, some specific use cases may fall within the boundaries of the described levels. In such cases, we recommend drawing recommendations from both levels and addressing with a hybrid approach the complexities of the case. Although extensive, we acknowledge that these recommendations cannot serve as the sole method or rigid rules due to the individual requirements, challenges, and outcomes inherent to each unique case. However, they provide a valuable starting point and a preliminary framework for navigating the complexities of designing and implementing AI technologies in industrial settings.
Our future work focuses on the emerging trends and paradigms shaping the landscape of human interaction with advanced technologies beyond the manufacturing sector. Addressing the complexity of activities within HAR while extending hierarchical analysis to diverse domains will be a priority to create a more general framework that can include multiple activities. Moreover, these categorizations of activities may support the research and development of activity datasets and models that specifically address the complexities of each level. Additionally, it is mandatory to recognize the importance of safety within industrial environments [124,125], while improving the system’s interoperability and the user’s experience, making assembly activity recognition systems more human-responsive. This means creating systems that understand and respond to human needs and behaviors better. One example would be to develop technologies that detect when a worker needs help, is feeling stressed, or is feeling fatigued, and assist or adjust the workload accordingly. It is therefore important to investigate how individuals interact with assistive technology (e.g., smart products, collaborative robots) for assembly tasks. Investigation of this socio-technical aspect is already present, e.g., in [126,127]. By exploring these insights in future research and development, we can advance the capabilities of assembly activity recognition systems, improve communication between humans and machines, promote teamwork, and facilitate more efficient and effective operations in industrial environments, while manufacturing processes can become more efficient.

7. Conclusions

Our work aims to enhance the understanding and support of human activities within industrial contexts. To achieve this, we developed a comprehensive taxonomy, that spans from the atomic to the mega-levels, for categorizing human activities in assembly processes for discrete product manufacturing. This taxonomy offers a hierarchical structure that facilitates better decision making, process optimization, and system design that addresses the binarization of <<simple>> and <<complex>> activities. We demonstrate the differences in the approaches with a comparison between the proposed taxonomy and the existing categorization schemes through an example of a real assembly scenario. By breaking down tasks into distinct levels, we enable a more granular analysis, identifying areas for improvement, troubleshooting problems, and optimizing the overall system performance.
Building upon this taxonomy, we provide specific recommendations for designing AI systems tailored for activity recognition in industrial assembly tasks. Although each use case may be unique with specific requirements and goals, the fundamental human activities and physical manipulations of the tools and components across assemblies typically remain similar. These recommendations examine, among others, sensor placement, preprocessing techniques, and model selection across different levels of activity abstraction.
Hence, it is important to align the crucial role of sensor systems attributes with specific application requirements to optimize the design of an AI system while ensuring both performance and functionality are effectively maintained. By leveraging the taxonomy, we aim to support the development of reliable, robust, and suitable for real-world deployment AI systems besides facilitating effective communication between experts and non-experts.
Overall, our research contributes to advancing the capabilities of AI-driven systems in industrial settings, fostering more efficient, safe, and human-responsive manufacturing processes. By providing a structured framework for analyzing and understanding human activities and offering recommendations for AI system design, we seek to empower researchers and professionals in various industrial fields to develop more effective methods and tools for studying, modeling, and supporting human activities.

Author Contributions

Conceptualization, G.S. and M.H.; methodology, G.S.; software, G.S.; validation, G.S.; formal analysis, G.S.; investigation, G.S.; resources, G.S.; data curation, G.S.; writing—original draft preparation, G.S.; writing—review and editing, G.S., M.H., B.A., O.G., M.S., B.A.-T. and A.F.; visualization, G.S.; supervision, A.F.; project administration, G.S. and M.H.; funding acquisition, G.S, M.H. and A.F. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by Johannes Kepler University Open Access Publishing Fund. This work has been supported by the FFG, Contract No. 881844: “Pro²Future is funded within the Austrian COMET Program Competence Centers for Excellent Technologies under the auspices of the Austrian Federal Ministry for Climate Action, Environment, Energy, Mobility, Innovation and Technology, the Austrian Federal Ministry for Digital and Economic Affairs and of the Provinces of Upper Austria and Styria. COMET is managed by the Austrian Research Promotion Agency FFG”. This work has been supported by the FFG, Contract No. 892220.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Skoda public dataset for Human activity recognition.

Acknowledgments

Supported by Johannes Kepler University Open Access Publishing Fund.

Conflicts of Interest

The authors were employed by the company Pro2Future GmbH and the Johannes Kepler University Linz (JKU Linz). The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be constructed as a potential conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Franke, J.; Wang, L.; Bock, K.; Wilde, J. Electronic module assembly. CIRP Ann. 2021, 70, 471–493. [Google Scholar] [CrossRef]
  2. Abdul Hadi, M.; Kraus, D.; Kajmakovic, A.; Suschnigg, J.; Guiza, O.; Gashi, M.; Sopidis, G.; Vukovic, M.; Milenkovic, K.; Haslgruebler, M.; et al. Towards flexible and cognitive production—Addressing the production challenges. Appl. Sci. 2022, 12, 8696. [Google Scholar] [CrossRef]
  3. Falck, A.C.; Örtengren, R.; Rosenqvist, M.; Söderberg, R. Criteria for assessment of basic manual assembly complexity. Procedia CIRP 2016, 44, 424–428. [Google Scholar] [CrossRef]
  4. Hassan, M.A.; Zardari, S.; Farooq, M.U.; Alansari, M.M.; Nagro, S.A. Systematic Analysis of Risks in Industry 5.0 Architecture. Appl. Sci. 2024, 14, 1466. [Google Scholar] [CrossRef]
  5. Capponi, M.; Gervasi, R.; Mastrogiacomo, L.; Franceschini, F. Assessing perceived assembly complexity in human-robot collaboration processes: A proposal based on Thurstone’s law of comparative judgement. Int. J. Prod. Res. 2023, 62, 5315–5335. [Google Scholar] [CrossRef]
  6. Islam, M.M.; Nooruddin, S.; Karray, F.; Muhammad, G. Human activity recognition using tools of convolutional neural networks: A state of the art review, data sets, challenges, and future prospects. Comput. Biol. Med. 2022, 149, 106060. [Google Scholar] [CrossRef] [PubMed]
  7. Gu, F.; Chung, M.H.; Chignell, M.; Valaee, S.; Zhou, B.; Liu, X. A survey on deep learning for human activity recognition. ACM Comput. Surv. (CSUR) 2021, 54, 1–34. [Google Scholar] [CrossRef]
  8. Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep learning for sensor-based human activity recognition: Overview, challenges, and opportunities. ACM Comput. Surv. (CSUR) 2021, 54, 1–40. [Google Scholar] [CrossRef]
  9. Schneider, B.; Banerjee, T. Bridging the Gap between Atomic and Complex Activities in First Person Video. In Proceedings of the 2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Luxembourg, 11–14 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  10. Peng, L.; Chen, L.; Wu, X.; Guo, H.; Chen, G. Hierarchical complex activity representation and recognition using topic model and classifier level fusion. IEEE Trans. Biomed. Eng. 2016, 64, 1369–1379. [Google Scholar] [CrossRef]
  11. Malik, A.A.; Bilberg, A. Complexity-based task allocation in human-robot collaborative assembly. Ind. Robot. Int. J. Robot. Res. Appl. 2019, 46, 471–480. [Google Scholar] [CrossRef]
  12. Roitberg, A.; Somani, N.; Perzylo, A.; Rickert, M.; Knoll, A. Multimodal human activity recognition for industrial manufacturing processes in robotic workcells. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 259–266. [Google Scholar]
  13. Wu, H.; Li, H.; Fang, X.; Luo, X. A survey on teaching workplace skills to construction robots. Expert Syst. Appl. 2022, 205, 117658. [Google Scholar] [CrossRef]
  14. Lucci, N.; Monguzzi, A.; Zanchettin, A.M.; Rocco, P. Workflow modelling for human–robot collaborative assembly operations. Robot. Comput.-Integr. Manuf. 2022, 78, 102384. [Google Scholar] [CrossRef]
  15. Kubota, A.; Iqbal, T.; Shah, J.A.; Riek, L.D. Activity recognition in manufacturing: The roles of motion capture and sEMG+ inertial wearables in detecting fine vs. gross motion. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 6533–6539. [Google Scholar]
  16. Park, S.; Lee, H.; Kim, S.; Baek, J.; Jang, K.; Kim, H.C.; Kim, M.; Park, J. Robotic furniture assembly: Task abstraction, motion planning, and control. Intell. Serv. Robot. 2022, 15, 441–457. [Google Scholar] [CrossRef]
  17. Xia, K.; Sacco, C.; Kirkpatrick, M.; Saidy, C.; Nguyen, L.; Kircaliali, A.; Harik, R. A digital twin to train deep reinforcement learning agent for smart manufacturing plants: Environment, interfaces and intelligence. J. Manuf. Syst. 2021, 58, 210–230. [Google Scholar] [CrossRef]
  18. Sosa-Ceron, A.D.; Gonzalez-Hernandez, H.G.; Reyes-Avendaño, J.A. Learning from Demonstrations in Human–Robot Collaborative Scenarios: A Survey. Robotics 2022, 11, 126. [Google Scholar] [CrossRef]
  19. Slama, R.; Slama, I.; Tlahig, H.; Slangen, P.; Ben-Ammar, O. An overview on human-centred technologies, measurements and optimisation in assembly systems. Int. J. Prod. Res. 2023, 62, 5336–5358. [Google Scholar] [CrossRef]
  20. Guo, P.; Zhang, Z.; Liu, Y.; Liu, Y.; Zhu, D.; Shao, C. A skill programming method based on assembly motion primitive for modular assembly system. IEEE Access 2021, 9, 101369–101380. [Google Scholar] [CrossRef]
  21. Cao, Y.; Lee, C. Robot Behavior-Tree-Based Task Generation with Large Language Models. arXiv 2023, arXiv:2302.12927. [Google Scholar]
  22. Wang, X.; Wang, S.; Menassa, C.C.; Kamat, V.R.; McGee, W. Automatic high-level motion sequencing methods for enabling multi-tasking construction robots. Autom. Constr. 2023, 155, 105071. [Google Scholar] [CrossRef]
  23. Suárez-Ruiz, F.; Pham, Q.C. A framework for fine robotic assembly. In Proceedings of the 2016 IEEE international conference on robotics and automation (ICRA), Stockholm, Sweden, 16–21 May 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 421–426. [Google Scholar]
  24. Motrenko, A.; Simchuk, E.; Khairullin, R.; Inyakin, A.; Kashirin, D.; Strijov, V. Continuous physical activity recognition for intelligent labour monitoring. Multimed. Tools Appl. 2022, 81, 4877–4895. [Google Scholar] [CrossRef]
  25. Akkaladevi, S.C.; Plasch, M.; Chitturi, N.C.; Hofmann, M.; Pichler, A. Programming by interactive demonstration for a human robot collaborative assembly. Procedia Manuf. 2020, 51, 148–155. [Google Scholar] [CrossRef]
  26. Zhang, J.; Wang, P.; Gao, R.X. Hybrid machine learning for human action recognition and prediction in assembly. Robot. Comput.-Integr. Manuf. 2021, 72, 102184. [Google Scholar] [CrossRef]
  27. He, F.; You, X.; Wang, W.; Bai, T.; Xue, G.; Ye, M. Recent progress in flexible microstructural pressure sensors toward human–machine interaction and healthcare applications. Small Methods 2021, 5, 2001041. [Google Scholar] [CrossRef]
  28. Nguyen, D.C.; Pham, Q.V.; Pathirana, P.N.; Ding, M.; Seneviratne, A.; Lin, Z.; Dobre, O.; Hwang, W.J. Federated learning for smart healthcare: A survey. ACM Comput. Surv. (CSUR) 2022, 55, 1–37. [Google Scholar] [CrossRef]
  29. Ali, F.; El-Sappagh, S.; Islam, S.R.; Ali, A.; Attique, M.; Imran, M.; Kwak, K.S. An intelligent healthcare monitoring framework using wearable sensors and social networking data. Future Gener. Comput. Syst. 2021, 114, 23–43. [Google Scholar] [CrossRef]
  30. Host, K.; Ivašić-Kos, M. An overview of Human Action Recognition in sports based on Computer Vision. Heliyon 2022, 8, e09633. [Google Scholar] [CrossRef]
  31. Shao, D.; Zhao, Y.; Dai, B.; Lin, D. Finegym: A hierarchical video dataset for fine-grained action understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2616–2625. [Google Scholar]
  32. Banos, O.; Galvez, J.M.; Damas, M.; Pomares, H.; Rojas, I. Window size impact in human activity recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef]
  33. Liu, Y.; Nie, L.; Han, L.; Zhang, L.; Rosenblum, D.S. Action2Activity: Recognizing complex activities from sensor data. arXiv 2016, arXiv:1611.01872. [Google Scholar]
  34. Assi, K.; Meegahapola, L.; Droz, W.; Kun, P.; De Götzen, A.; Bidoglia, M.; Stares, S.; Gaskell, G.; Chagnaa, A.; Ganbold, A.; et al. Complex Daily Activities, Country-Level Diversity, and Smartphone Sensing: A Study in Denmark, Italy, Mongolia, Paraguay, and UK. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, Hamburg, Germany, 23–28 April 2023; pp. 1–23. [Google Scholar]
  35. Peng, L.; Chen, L.; Ye, Z.; Zhang, Y. Aroma: A deep multi-task learning based simple and complex human activity recognition method using wearable sensors. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–16. [Google Scholar] [CrossRef]
  36. Huan, R.; Jiang, C.; Ge, L.; Shu, J.; Zhan, Z.; Chen, P.; Chi, K.; Liang, R. Human complex activity recognition with sensor data using multiple features. IEEE Sens. J. 2021, 22, 757–775. [Google Scholar] [CrossRef]
  37. Bharti, P.; De, D.; Chellappan, S.; Das, S.K. HuMAn: Complex activity recognition with multi-modal multi-positional body sensing. IEEE Trans. Mob. Comput. 2018, 18, 857–870. [Google Scholar] [CrossRef]
  38. Mekruksavanich, S.; Jitpattanakul, A.; Sitthithakerngkiet, K.; Youplao, P.; Yupapin, P. Resnet-se: Channel attention-based deep residual network for complex activity recognition using wrist-worn wearable sensors. IEEE Access 2022, 10, 51142–51154. [Google Scholar] [CrossRef]
  39. Lago, P.; Takeda, S.; Alia, S.S.; Adachi, K.; Bennai, B.; Charpillet, F.; Inoue, S. A dataset for complex activity recognition withmicro and macro activities in a cooking scenario. arXiv 2020, arXiv:2006.10681. [Google Scholar]
  40. Rohrbach, M.; Rohrbach, A.; Regneri, M.; Amin, S.; Andriluka, M.; Pinkal, M.; Schiele, B. Recognizing fine-grained and composite activities using hand-centric features and script data. Int. J. Comput. Vis. 2016, 119, 346–373. [Google Scholar] [CrossRef]
  41. Aggarwal, J.K.; Ryoo, M.S. Human activity analysis: A review. Acm Comput. Surv. (Csur) 2011, 43, 1–43. [Google Scholar] [CrossRef]
  42. Dernbach, S.; Das, B.; Krishnan, N.C.; Thomas, B.L.; Cook, D.J. Simple and complex activity recognition through smart phones. In Proceedings of the 2012 Eighth International Conference on Intelligent Environments, Guanajuato, Mexico, 26–29 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 214–221. [Google Scholar]
  43. Sanhudo, L.; Calvetti, D.; Martins, J.P.; Ramos, N.M.; Meda, P.; Goncalves, M.C.; Sousa, H. Activity classification using accelerometers and machine learning for complex construction worker activities. J. Build. Eng. 2021, 35, 102001. [Google Scholar] [CrossRef]
  44. Chen, L.; Liu, X.; Peng, L.; Wu, M. Deep learning based multimodal complex human activity recognition using wearable devices. Appl. Intell. 2021, 51, 4029–4042. [Google Scholar] [CrossRef]
  45. Hnoohom, N.; Jitpattanakul, A.; You, I.; Mekruksavanich, S. Deep learning approach for complex activity recognition using heterogeneous sensors from wearable device. In Proceedings of the 2021 Research, Invention, and Innovation Congress: Innovation Electricals and Electronics (RI2C), Bangkok, Thailand, 1–3 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 60–65. [Google Scholar]
  46. Miranda, L.; Viterbo, J.; Bernardini, F. A survey on the use of machine learning methods in context-aware middlewares for human activity recognition. Artif. Intell. Rev. 2022, 55, 3369–3400. [Google Scholar] [CrossRef]
  47. Bouton-Bessac, E.; Meegahapola, L.; Gatica-Perez, D. Your Day in Your Pocket: Complex Activity Recognition from Smartphone Accelerometers. In Pervasive Computing Technologies for Healthcare, Proceedings of the PervasiveHealth 2022, Thessaloniki, Greece, 12–14 December 2022; Tsanas, A., Triantafyllidis, A., Eds.; Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering; Springer: Cham, Switzerland, 2022; pp. 247–258. [Google Scholar] [CrossRef]
  48. Ramanujam, E.; Perumal, T.; Padmavathi, S. Human activity recognition with smartphone and wearable sensors using deep learning techniques: A review. IEEE Sens. J. 2021, 21, 13029–13040. [Google Scholar] [CrossRef]
  49. Chen, C.; Wang, T.; Li, D.; Hong, J. Repetitive assembly action recognition based on object detection and pose estimation. J. Manuf. Syst. 2020, 55, 325–333. [Google Scholar] [CrossRef]
  50. Saguna, S.; Zaslavsky, A.; Chakraborty, D. Complex activity recognition using context-driven activity theory and activity signatures. ACM Trans. Comput.-Hum. Interact. (TOCHI) 2013, 20, 1–34. [Google Scholar] [CrossRef]
  51. Omolaja, A.; Otebolaku, A.; Alfoudi, A. Context-aware complex human activity recognition using hybrid deep learning models. Appl. Sci. 2022, 12, 9305. [Google Scholar] [CrossRef]
  52. Martínez-Villaseñor, L.; Ponce, H. A concise review on sensor signal acquisition and transformation applied to human activity recognition and human–robot interaction. Int. J. Distrib. Sens. Netw. 2019, 15, 1550147719853987. [Google Scholar] [CrossRef]
  53. Mekruksavanich, S.; Tancharoen, D.; Jitpattanakul, A. Human Activity Recognition in Logistics Using Wearable Sensors and Deep Residual Network. In Proceedings of the TENCON 2023-2023 IEEE Region 10 Conference (TENCON), Chiang Mai, Thailand, 31 October–3 November 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 194–198. [Google Scholar]
  54. Alexan, A.I.; Alexan, A.R.; Oniga, S. Real-Time Machine Learning for Human Activities Recognition Based on Wrist-Worn Wearable Devices. Appl. Sci. 2023, 14, 329. [Google Scholar] [CrossRef]
  55. Liu, L.; Peng, Y.; Wang, S.; Liu, M.; Huang, Z. Complex activity recognition using time series pattern dictionary learned from ubiquitous sensors. Inf. Sci. 2016, 340, 41–57. [Google Scholar] [CrossRef]
  56. Challa, S.K.; Kumar, A.; Semwal, V.B. A multibranch CNN-BiLSTM model for human activity recognition using wearable sensor data. Vis. Comput. 2022, 38, 4095–4109. [Google Scholar] [CrossRef]
  57. Chung, J.; Wuu, C.h.; Yang, H.r.; Tai, Y.W.; Tang, C.K. Haa500: Human-centric atomic action dataset with curated videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 13465–13474. [Google Scholar]
  58. Khaire, P.; Kumar, P. Deep learning and RGB-D based human action, human–human and human–object interaction recognition: A survey. J. Vis. Commun. Image Represent. 2022, 86, 103531. [Google Scholar] [CrossRef]
  59. Dang, L.M.; Min, K.; Wang, H.; Piran, M.J.; Lee, C.H.; Moon, H. Sensor-based and vision-based human activity recognition: A comprehensive survey. Pattern Recognit. 2020, 108, 107561. [Google Scholar] [CrossRef]
  60. Kulsoom, F.; Narejo, S.; Mehmood, Z.; Chaudhry, H.N.; Butt, A.; Bashir, A.K. A review of machine learning-based human activity recognition for diverse applications. Neural Comput. Appl. 2022, 34, 18289–18324. [Google Scholar] [CrossRef]
  61. Morshed, M.G.; Sultana, T.; Alam, A.; Lee, Y.K. Human Action Recognition: A Taxonomy-Based Survey, Updates, and Opportunities. Sensors 2023, 23, 2182. [Google Scholar] [CrossRef]
  62. Amjad, F.; Khan, M.H.; Nisar, M.A.; Farid, M.S.; Grzegorzek, M. A comparative study of feature selection approaches for human activity recognition using multimodal sensory data. Sensors 2021, 21, 2368. [Google Scholar] [CrossRef] [PubMed]
  63. Liu, Y.; Nie, L.; Liu, L.; Rosenblum, D.S. From action to activity: Sensor-based activity recognition. Neurocomputing 2016, 181, 108–115. [Google Scholar] [CrossRef]
  64. Azadi, B.; Haslgrübler, M.; Anzengruber-Tanase, B.; Grünberger, S.; Ferscha, A. Alpine skiing activity recognition using smartphone’s IMUs. Sensors 2022, 22, 5922. [Google Scholar] [CrossRef] [PubMed]
  65. Azadi, B.; Haslgrübler, M.; Anzengruber-Tanase, B.; Sopidis, G.; Ferscha, A. Robust Feature Representation Using Multi-Task Learning for Human Activity Recognition. Sensors 2024, 24, 681. [Google Scholar] [CrossRef] [PubMed]
  66. Anzengruber-Tanase, B.; Sopidis, G.; Haslgrübler, M.; Ferscha, A. Determining Best Hardware, Software and Data Structures for Worker Guidance during a Complex Assembly Task. In Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu, Greece, 29 June–1 July 2022; pp. 63–72. [Google Scholar]
  67. Laube, M.; Sopidis, G.; Anzengruber-Tanase, B.; Ferscha, A.; Haslgrübler, M. Analyzing Arc Welding Techniques improves Skill Level Assessment in Industrial Manufacturing Processes. In Proceedings of the 16th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu, Greece, 5–7 July 2023; pp. 177–186. [Google Scholar]
  68. Zhang, S.; Li, Y.; Zhang, S.; Shahabi, F.; Xia, S.; Deng, Y.; Alshurafa, N. Deep learning in human activity recognition with wearable sensors: A review on advances. Sensors 2022, 22, 1476. [Google Scholar] [CrossRef] [PubMed]
  69. Mannhardt, F.; Bovo, R.; Oliveira, M.F.; Julier, S. A taxonomy for combining activity recognition and process discovery in industrial environments. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Madrid, Spain, 21–23 November 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 84–93. [Google Scholar]
  70. Azadi, B.; Haslgrübler, M.; Sopidis, G.; Murauer, M.; Anzengruber, B.; Ferscha, A. Feasibility analysis of unsupervised industrial activity recognition based on a frequent micro action. In Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments, Island of Rhodes, Greece, 5–7 June 2019; pp. 368–375. [Google Scholar]
  71. Ahmad, A.; Haslgrübler, M.; Sopidis, G.; Azadi, B.; Ferscha, A. Privacy Preserving Workflow Detection for Manufacturing Using Neural Networks based Object Detection. In Proceedings of the 11th International Conference on the Internet of Things, St. Gallen, Switzerland, 8–12 November 2021; pp. 126–133. [Google Scholar]
  72. Qiu, S.; Zhao, H.; Jiang, N.; Wang, Z.; Liu, L.; An, Y.; Zhao, H.; Miao, X.; Liu, R.; Fortino, G. Multi-sensor information fusion based on machine learning for real applications in human activity recognition: State-of-the-art and research challenges. Inf. Fusion 2022, 80, 241–265. [Google Scholar] [CrossRef]
  73. Kuutti, K. Activity theory as a potential framework for human-computer interaction research. Context Conscious. Act. Theory Hum.-Comput. Interact. 1996, 1744, 9–22. [Google Scholar]
  74. Lee, R.K.J.; Zheng, H.; Lu, Y. Human-Robot Shared Assembly Taxonomy: A step toward seamless human-robot knowledge transfer. Robot. Comput.-Integr. Manuf. 2024, 86, 102686. [Google Scholar] [CrossRef]
  75. Zhuang, C.; Gong, J.; Liu, J. Digital twin-based assembly data management and process traceability for complex products. J. Manuf. Syst. 2021, 58, 118–131. [Google Scholar] [CrossRef]
  76. Sopidis, G.; Haslgrübler, M.; Azadi, B.; Anzengruber-Tánase, B.; Ahmad, A.; Ferscha, A.; Baresch, M. Micro-activity recognition in industrial assembly process with IMU data and deep learning. In Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu, Greece, 29 June–1 July 2022; pp. 103–112. [Google Scholar]
  77. Kyratsis, P.; Gabis, E.; Tzotzis, A.; Tzetzis, D.; Kakoulis, K. CAD based product design: A case study. Int. J. Mod. Manuf. Technol. 2019, 11, 110–115. [Google Scholar]
  78. Maddikunta, P.K.R.; Pham, Q.V.; Prabadevi, B.; Deepa, N.; Dev, K.; Gadekallu, T.R.; Ruby, R.; Liyanage, M. Industry 5.0: A survey on enabling technologies and potential applications. J. Ind. Inf. Integr. 2022, 26, 100257. [Google Scholar] [CrossRef]
  79. Miqueo, A.; Torralba, M.; Yagüe-Fabra, J.A. Lean manual assembly 4.0: A systematic review. Appl. Sci. 2020, 10, 8555. [Google Scholar] [CrossRef]
  80. Yin, Y.; Stecke, K.E.; Li, D. The evolution of production systems from Industry 2.0 through Industry 4.0. Int. J. Prod. Res. 2018, 56, 848–861. [Google Scholar] [CrossRef]
  81. Pandremenos, J.; Paralikas, J.; Salonitis, K.; Chryssolouris, G. Modularity concepts for the automotive industry: A critical review. CIRP J. Manuf. Sci. Technol. 2009, 1, 148–152. [Google Scholar] [CrossRef]
  82. BuchmÃller, M.; EidmÃller, T.; Mussmann, A.; Kopia, J. The Status of Modular Sourcing Compared to Other Procurement Strategies. Ecoforum J. 2018, 7, 9. [Google Scholar]
  83. Niemann, F.; Reining, C.; Moya Rueda, F.; Nair, N.R.; Steffens, J.A.; Fink, G.A.; Ten Hompel, M. Lara: Creating a dataset for human activity recognition in logistics using semantic attributes. Sensors 2020, 20, 4083. [Google Scholar] [CrossRef]
  84. Sopidis, G.; Ahmad, A.; Haslgruebler, M.; Ferscha, A.; Baresch, M. Micro Activities Recognition and Macro Worksteps Classification for Industrial IoT Processes. In Proceedings of the 11th International Conference on the Internet of Things, St. Gallen, Switzerland, 8–12 November 2021; pp. 185–188. [Google Scholar]
  85. Mark, B.G.; Rauch, E.; Matt, D.T. Worker assistance systems in manufacturing: A review of the state of the art and future directions. J. Manuf. Syst. 2021, 59, 228–250. [Google Scholar] [CrossRef]
  86. Guo, L.; Lu, Z.; Yao, L. Human-machine interaction sensing technology based on hand gesture recognition: A review. IEEE Trans. Hum.-Mach. Syst. 2021, 51, 300–309. [Google Scholar] [CrossRef]
  87. Bortolini, M.; Faccio, M.; Gamberi, M.; Pilati, F. Motion Analysis System (MAS) for production and ergonomics assessment in the manufacturing processes. Comput. Ind. Eng. 2020, 139, 105485. [Google Scholar] [CrossRef]
  88. Valarezo Añazco, E.; Han, S.J.; Kim, K.; Lopez, P.R.; Kim, T.S.; Lee, S. Hand gesture recognition using single patchable six-axis inertial measurement unit via recurrent neural networks. Sensors 2021, 21, 1404. [Google Scholar] [CrossRef]
  89. Vandevoorde, K.; Vollenkemper, L.; Schwan, C.; Kohlhase, M.; Schenck, W. Using Artificial Intelligence for Assistance Systems to Bring Motor Learning Principles into Real World Motor Tasks. Sensors 2022, 22, 2481. [Google Scholar] [CrossRef]
  90. Digo, E.; Pastorelli, S.; Gastaldi, L. A narrative review on wearable inertial sensors for human motion tracking in industrial scenarios. Robotics 2022, 11, 138. [Google Scholar] [CrossRef]
  91. Bortolini, M.; Faccio, M.; Galizia, F.G.; Gamberi, M.; Pilati, F. Adaptive automation assembly systems in the industry 4.0 era: A reference framework and full–scale prototype. Appl. Sci. 2021, 11, 1256. [Google Scholar] [CrossRef]
  92. Borghetti, M.; Bellitti, P.; Lopomo, N.F.; Serpelloni, M.; Sardini, E. Validation of a modular and wearable system for tracking fingers movements. Acta IMEKO 2020, 9, 157–164. [Google Scholar] [CrossRef]
  93. Riedel, A.; Gerlach, J.; Dietsch, M.; Herbst, S.; Engelmann, F.; Brehm, N.; Pfeifroth, T. A deep learning-based worker assistance system for error prevention: Case study in a real-world manual assembly. Adv. Prod. Eng. Manag. 2021, 16, 393–404. [Google Scholar] [CrossRef]
  94. Kurata, T.; Harada, M.; Nakahira, K.; Maehata, T.; Ito, Y.; Aso, H. Analyzing operations on a manufacturing line using geospatial intelligence technologies. In Advances in Production Management Systems. Smart Manufacturing and Logistics Systems: Turning Ideas into Action, Proceedings of the APMS 2022, Gyeongju, Republic of Korea, 25–29 September 2022; Kim, D.Y., von Cieminski, G., Romero, D., Eds.; IFIP Advances in Information and Communication Technology; Springer: Cham, Switzerland, 2022; pp. 69–76. [Google Scholar]
  95. Dallel, M.; Havard, V.; Baudry, D.; Savatier, X. Inhard-industrial human action recognition dataset in the context of industrial collaborative robotics. In Proceedings of the 2020 IEEE International Conference on Human-Machine Systems (ICHMS), Rome, Italy, 7–9 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  96. Sener, F.; Chatterjee, D.; Shelepov, D.; He, K.; Singhania, D.; Wang, R.; Yao, A. Assembly101: A large-scale multi-view video dataset for understanding procedural activities. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 21096–21106. [Google Scholar]
  97. Mastakouris, A.; Andriosopoulou, G.; Masouros, D.; Benardos, P.; Vosniakos, G.C.; Soudris, D. Human worker activity recognition in a production floor environment through deep learning. J. Manuf. Syst. 2023, 71, 115–130. [Google Scholar] [CrossRef]
  98. Hayward, S.; van Lopik, K.; Hinde, C.; West, A.A. A survey of indoor location technologies, techniques and applications in industry. Internet Things 2022, 20, 100608. [Google Scholar] [CrossRef]
  99. Benmessabih, T.; Slama, R.; Havard, V.; Baudry, D. Online human motion analysis in industrial context: A review. Eng. Appl. Artif. Intell. 2024, 131, 107850. [Google Scholar] [CrossRef]
  100. Xu, J.; Li, Z.; Zhang, K.; Yang, J.; Gao, N.; Zhang, Z.; Meng, Z. The principle, methods and recent progress in RFID positioning techniques: A review. IEEE J. Radio Freq. Identif. 2023, 7, 50–63. [Google Scholar] [CrossRef]
  101. Guiza, O.; Mayr-Dorn, C.; Weichhart, G.; Mayrhofer, M.; Zangi, B.B.; Egyed, A.; Fanta, B.; Gieler, M. Monitoring of human-intensive assembly processes based on incomplete and indirect shopfloor observations. In Proceedings of the 2021 IEEE 19th International Conference on Industrial Informatics (INDIN), Palma de Mallorca, Spain, 21–23 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
  102. Yadav, S.K.; Tiwari, K.; Pandey, H.M.; Akbar, S.A. A review of multimodal human activity recognition with special emphasis on classification, applications, challenges and future directions. Knowl.-Based Syst. 2021, 223, 106970. [Google Scholar] [CrossRef]
  103. Günther, L.C.; Kärcher, S.; Bauernhansl, T. Activity recognition in manual manufacturing: Detecting screwing processes from sensor data. Procedia CIRP 2019, 81, 1177–1182. [Google Scholar] [CrossRef]
  104. Faccio, M.; Ferrari, E.; Gamberi, M.; Pilati, F. Human Factor Analyser for work measurement of manual manufacturing and assembly processes. Int. J. Adv. Manuf. Technol. 2019, 103, 861–877. [Google Scholar] [CrossRef]
  105. Hernandez, J.; Valarezo, G.; Cobos, R.; Kim, J.W.; Palacios, R.; Abad, A.G. Hierarchical Human Action Recognition to Measure the Performance of Manual Labor. IEEE Access 2021, 9, 103110–103119. [Google Scholar] [CrossRef]
  106. Shilkrot, R.; Narasimhaswamy, S.; Vazir, S.; Hoai, M. WorkingHands: A hand-tool assembly dataset for image segmentation and activity mining. 2019. In Proceedings of the 30th British Machine Vision Conference, Cardiff, Wales, 9–12 September 2019. [Google Scholar]
  107. Büsch, L.; Koch, J.; Schoepflin, D.; Schulze, M.; Schüppstuhl, T. Towards recognition of human actions in collaborative tasks with robots: Extending action recognition with tool recognition methods. Sensors 2023, 23, 5718. [Google Scholar] [CrossRef]
  108. Gkournelos, C.; Konstantinou, C.; Angelakis, P.; Tzavara, E.; Makris, S. Praxis: A framework for AI-driven human action recognition in assembly. J. Intell. Manuf. 2023, 1–15. [Google Scholar] [CrossRef]
  109. Reining, C.; Niemann, F.; Moya Rueda, F.; Fink, G.A.; ten Hompel, M. Human activity recognition for production and logistics—A systematic literature review. Information 2019, 10, 245. [Google Scholar] [CrossRef]
  110. Ferrari, A.; Micucci, D.; Mobilio, M.; Napoletano, P. Deep learning and model personalization in sensor-based human activity recognition. J. Reliab. Intell. Environ. 2023, 9, 27–39. [Google Scholar] [CrossRef]
  111. Gomes, E.L.; Fonseca, M.; Lazzaretti, A.E.; Munaretto, A.; Guerber, C. Clustering and Hierarchical Classification for High-Precision RFID Indoor Location Systems. IEEE Sens. J. 2022, 22, 5141–5149. [Google Scholar] [CrossRef]
  112. Al-Amin, M.; Tao, W.; Doell, D.; Lingard, R.; Yin, Z.; Leu, M.C.; Qin, R. Action recognition in manufacturing assembly using multimodal sensor fusion. Procedia Manuf. 2019, 39, 158–167. [Google Scholar] [CrossRef]
  113. Gjeldum, N.; Mladineo, M.; Crnjac, M.; Veza, I.; Aljinovic, A. Performance analysis of the RFID system for optimal design of the intelligent assembly line in the learning factory. Procedia Manuf. 2018, 23, 63–68. [Google Scholar] [CrossRef]
  114. Bordel, B.; Alcarria, R.; Robles, T. Recognizing human activities in Industry 4.0 scenarios through an analysis-modeling-recognition algorithm and context labels. Integr. Comput.-Aided Eng. 2022, 29, 83–103. [Google Scholar] [CrossRef]
  115. Moutinho, D.; Rocha, L.F.; Costa, C.M.; Teixeira, L.F.; Veiga, G. Deep learning-based human action recognition to leverage context awareness in collaborative assembly. Robot. Comput.-Integr. Manuf. 2023, 80, 102449. [Google Scholar] [CrossRef]
  116. Ahmad, H.M.; Rahimi, A. Deep learning methods for object detection in smart manufacturing: A survey. J. Manuf. Syst. 2022, 64, 181–196. [Google Scholar] [CrossRef]
  117. Feradov, F.; Markova, V.; Ganchev, T. Automated detection of improper sitting postures in computer users based on motion capture sensors. Computers 2022, 11, 116. [Google Scholar] [CrossRef]
  118. Zhou, L.; Zhang, L.; Konz, N. Computer vision techniques in manufacturing. IEEE Trans. Syst. Man, Cybern. Syst. 2022, 53, 105–117. [Google Scholar] [CrossRef]
  119. Baroroh, D.K.; Chu, C.H.; Wang, L. Systematic literature review on augmented reality in smart manufacturing: Collaboration between human and computational intelligence. J. Manuf. Syst. 2021, 61, 696–711. [Google Scholar] [CrossRef]
  120. Bollinger, S.; Stich, V.; Holst, L.; Defèr, F.; Schuldt, F. Evaluation of Potential Benefits of Augmented Reality for Industrial Services. In Advances in Production Management Systems. Smart Manufacturing and Logistics Systems: Turning Ideas into Action, Proceedings of the APMS 2022, Gyeongju, Republic of Korea, 25–29 September 2022; Kim, D.Y., von Cieminski, G., Romero, D., Eds.; IFIP Advances in Information and Communication Technology; Springer: Cham, Switzerland, 2022; pp. 135–144. [Google Scholar] [CrossRef]
  121. Rueckert, P.; Birgy, K.; Tracht, K. Image Based Classification of Methods-Time Measurement Operations in Assembly Using Recurrent Neuronal Networks. In Advances in System-Integrated Intelligenc, Proceedings of the SYSINT 2022, Genova, Italy, 7–9 September 2022; Valle, M., Lehmhus, D., Gianoglio, C., Ragusa, E., Seminara, L., Bosse, S., Ibrahim, A., Thoben, K.-D., Eds.; Lecture Notes in Networks and Systems; Springer: Cham, Switzerland, 2022; pp. 53–62. [Google Scholar] [CrossRef]
  122. Selvaraj, V.; Al-Amin, M.; Tao, W.; Min, S. Intelligent assembly operations monitoring with the ability to detect non-value-added activities as out-of-distribution (OOD) instances. CIRP Ann. 2023, 72, 413–416. [Google Scholar] [CrossRef]
  123. Mabkhot, M.M.; Al-Ahmari, A.M.; Salah, B.; Alkhalefah, H. Requirements of the smart factory system: A survey and perspective. Machines 2018, 6, 23. [Google Scholar] [CrossRef]
  124. Schobesberger, M.; Huber, J.; Grünberger, S.; Haslgrübler, M.; Ferscha, A. Designing Proactive Safety Systems for Industrial Workers Using Intelligent Mechanisms. In Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu, Greece, 29 June–1 July 2022; pp. 480–485. [Google Scholar]
  125. Huber, J.; Haslgrübler, M.; Schobesberger, M.; Ferscha, A.; Malisa, V.; Effenberger, G. Addressing Worker Safety and Accident Prevention with AI. In Proceedings of the 11th International Conference on the Internet of Things, St. Gallen, Switzerland, 8–12 November 2021; pp. 150–157. [Google Scholar]
  126. Makarius, E.E.; Mukherjee, D.; Fox, J.D.; Fox, A.K. Rising with the machines: A sociotechnical framework for bringing artificial intelligence into the organization. J. Bus. Res. 2020, 120, 262–273. [Google Scholar] [CrossRef]
  127. Sony, M.; Naik, S. Industry 4.0 integration with socio-technical systems theory: A systematic review and proposed theoretical model. Technol. Soc. 2020, 61, 101248. [Google Scholar] [CrossRef]
Figure 1. This figure presents a hierarchy of an exemplary assembly process: components, units, modules, products, and post-assembly. Additionally, it demonstrates how these stages are interconnected and how activities and tasks flow inside a real assembly scenario from components to the final product. Each stage builds upon the previous one, with components being assembled into units, units into modules, modules into the final product, and finally, the product being integrated into the production line.
Figure 1. This figure presents a hierarchy of an exemplary assembly process: components, units, modules, products, and post-assembly. Additionally, it demonstrates how these stages are interconnected and how activities and tasks flow inside a real assembly scenario from components to the final product. Each stage builds upon the previous one, with components being assembled into units, units into modules, modules into the final product, and finally, the product being integrated into the production line.
Sensors 24 04508 g001
Figure 2. The figure presents a visualization for the proposed taxonomy. At the atomic level, individual assembly activities are considered as singular tasks involving basic operations or manipulations on discrete components or tools. The micro-level aggregates multiple atomic operations into coherent sequences, representing actions within the assembly process. Larger assembly tasks are formed at the meso-level by combining multiple micro-level activities, often involving the assembly of sub-components or partial assemblies. The macro-level encompasses entire assembly processes, including stages such as the assembly of major components or modules. Finally, the mega-level represents the overall assembly process, incorporating post-assembly activities such as quality control checks, packaging, or final inspection.
Figure 2. The figure presents a visualization for the proposed taxonomy. At the atomic level, individual assembly activities are considered as singular tasks involving basic operations or manipulations on discrete components or tools. The micro-level aggregates multiple atomic operations into coherent sequences, representing actions within the assembly process. Larger assembly tasks are formed at the meso-level by combining multiple micro-level activities, often involving the assembly of sub-components or partial assemblies. The macro-level encompasses entire assembly processes, including stages such as the assembly of major components or modules. Finally, the mega-level represents the overall assembly process, incorporating post-assembly activities such as quality control checks, packaging, or final inspection.
Sensors 24 04508 g002
Figure 3. The table illustrates a simplified ATM assembly process, derived from a real industrial use case [76], showcasing activities across different assembly levels: atomic, micro, meso, macro, and mega. It serves as a comparative analysis with existing approaches for activity categorization, highlighting how each level contributes to the overall process. Specific activities are provided for clarity, offering insights into the hierarchical organization of assembly tasks. The color coding highlights differences in categorization when distinguishing tasks across levels.
Figure 3. The table illustrates a simplified ATM assembly process, derived from a real industrial use case [76], showcasing activities across different assembly levels: atomic, micro, meso, macro, and mega. It serves as a comparative analysis with existing approaches for activity categorization, highlighting how each level contributes to the overall process. Specific activities are provided for clarity, offering insights into the hierarchical organization of assembly tasks. The color coding highlights differences in categorization when distinguishing tasks across levels.
Sensors 24 04508 g003
Figure 4. The table illustrates welding processes in car assembly, presenting the hierarchical framework of tasks, and showcasing activities across different assembly levels: atomic, micro, meso, macro, and mega. It serves as a comparative analysis with existing approaches for activity categorization, highlighting how each level contributes to the overall process and showing how individual actions aggregate into more complex tasks across the assembly line. Specific activities are provided for clarity, offering insights into the hierarchical organization of assembly tasks. The color coding highlights differences in categorization when distinguishing tasks across levels.
Figure 4. The table illustrates welding processes in car assembly, presenting the hierarchical framework of tasks, and showcasing activities across different assembly levels: atomic, micro, meso, macro, and mega. It serves as a comparative analysis with existing approaches for activity categorization, highlighting how each level contributes to the overall process and showing how individual actions aggregate into more complex tasks across the assembly line. Specific activities are provided for clarity, offering insights into the hierarchical organization of assembly tasks. The color coding highlights differences in categorization when distinguishing tasks across levels.
Sensors 24 04508 g004
Figure 5. This figure illustrates key characteristics across atomic, micro, meso, macro, and mega-levels of assembly activity recognition systems. Each group of related elements is color-coded, and each line represents a different category, ensuring distinctions between aspects. The figure highlights variations that are important in the overall design of an AI system, such as sensor placement, types of sensors used, system mobility, sampling rate, duration of experiments, frequency of actions, preprocessing techniques, models employed for activity recognition, window size for data processing, and feedback mechanisms. Associated recommendations are provided for each category and level to serve as a starting point for the development of AI models under the “Models to Use” category, which is related to industrial assembly.
Figure 5. This figure illustrates key characteristics across atomic, micro, meso, macro, and mega-levels of assembly activity recognition systems. Each group of related elements is color-coded, and each line represents a different category, ensuring distinctions between aspects. The figure highlights variations that are important in the overall design of an AI system, such as sensor placement, types of sensors used, system mobility, sampling rate, duration of experiments, frequency of actions, preprocessing techniques, models employed for activity recognition, window size for data processing, and feedback mechanisms. Associated recommendations are provided for each category and level to serve as a starting point for the development of AI models under the “Models to Use” category, which is related to industrial assembly.
Sensors 24 04508 g005
Table 1. This table presents an overview of the existing literature in the domain of human activity recognition, presenting a quantitative distribution of papers across different abstraction levels and domains. The dominance of publications at the simple/complex level across various domains suggests a significant focus on understanding activities in this distinction. On the other hand, the comparatively small number of publications in some areas, such as group activities, suggests possible topics for more investigation and study. Furthermore, the existence of publications at various levels of abstraction highlights the complexity of human activity recognition research and emphasizes the necessity for sophisticated methods of activity analysis and classification.
Table 1. This table presents an overview of the existing literature in the domain of human activity recognition, presenting a quantitative distribution of papers across different abstraction levels and domains. The dominance of publications at the simple/complex level across various domains suggests a significant focus on understanding activities in this distinction. On the other hand, the comparatively small number of publications in some areas, such as group activities, suggests possible topics for more investigation and study. Furthermore, the existence of publications at various levels of abstraction highlights the complexity of human activity recognition research and emphasizes the necessity for sophisticated methods of activity analysis and classification.
Abstraction Levels
Approaches:BinaryNon-Binary
DomainAtomic-Simple, Complex-CompositeComposite-Gross, Fine-GrainedMicro, MacroLow, High-LevelGestures, Actions, Interactions, Group ActivitiesAtomic action, Primitive Task, TaskVarious, Other Terminology
Manufacturing, Robotics, Construction[12,13,14][15][16][17,18] [19,20,21,22,23,24][25,26]
Healthcare [27,28,29]
Sports [30] [31,32]
ADLs[9,10,33,34,35,36,37,38] [33]
IADL[37,39][40][39]
Group [41]
Other Domain, Unspecific HAR[6,8,36,38,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56][57] [58,59,60,61]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sopidis, G.; Haslgrübler, M.; Azadi, B.; Guiza, O.; Schobesberger, M.; Anzengruber-Tanase, B.; Ferscha, A. System Design for Sensing in Manufacturing to Apply AI through Hierarchical Abstraction Levels. Sensors 2024, 24, 4508. https://doi.org/10.3390/s24144508

AMA Style

Sopidis G, Haslgrübler M, Azadi B, Guiza O, Schobesberger M, Anzengruber-Tanase B, Ferscha A. System Design for Sensing in Manufacturing to Apply AI through Hierarchical Abstraction Levels. Sensors. 2024; 24(14):4508. https://doi.org/10.3390/s24144508

Chicago/Turabian Style

Sopidis, Georgios, Michael Haslgrübler, Behrooz Azadi, Ouijdane Guiza, Martin Schobesberger, Bernhard Anzengruber-Tanase, and Alois Ferscha. 2024. "System Design for Sensing in Manufacturing to Apply AI through Hierarchical Abstraction Levels" Sensors 24, no. 14: 4508. https://doi.org/10.3390/s24144508

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop