Next Article in Journal
Long-Term Assessment of the Properties of Load Sensors Applied in Weigh-in-Motion Systems
Previous Article in Journal
Reinforcement Learning-Based Time-Slotted Protocol: A Reinforcement Learning Approach for Optimizing Long-Range Network Scalability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Design of a Vision-Assisted Dynamic Antenna Positioning Radio Frequency Identification-Based Inventory Robot Utilizing a 3-Degree-of-Freedom Manipulator

by
Abdussalam A. Alajami
* and
Rafael Pous
Department of Information and Communication Technologies Engineering, Pompeu Fabra University, 08002 Barcelona, Spain
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(8), 2418; https://doi.org/10.3390/s25082418
Submission received: 17 February 2025 / Revised: 8 April 2025 / Accepted: 9 April 2025 / Published: 11 April 2025

Abstract

:
In contemporary warehouse logistics, the demand for efficient and precise inventory management is increasingly critical, yet traditional Radio Frequency Identification (RFID)-based systems often falter due to static antenna configurations that limit tag detection efficacy in complex environments with diverse object arrangements. Addressing this challenge, we introduce an advanced RFID-based inventory robot that integrates a 3-degree-of-freedom (3DOF) manipulator with vision-assisted dynamic antenna positioning to optimize tag detection performance. This autonomous system leverages a pretrained You Only Look Once (YOLO) model to detect objects in real time, employing forward and inverse kinematics to dynamically orient the RFID antenna toward identified items. The manipulator subsequently executes a tailored circular scanning motion, ensuring comprehensive coverage of each object’s surface and maximizing RFID tag readability. To evaluate the system’s efficacy, we conducted a comparative analysis of three scanning strategies: (1) a conventional fixed antenna approach, (2) a predefined path strategy with preprogrammed manipulator movements, and (3) our proposed vision-assisted dynamic positioning method. Experimental results, derived from controlled laboratory tests and Gazebo-based simulations, unequivocally demonstrate the superiority of the dynamic positioning approach. This method achieved detection rates of up to 98.0% across varied shelf heights and spatial distributions, significantly outperforming the fixed antenna (21.6%) and predefined path (88.5%) strategies, particularly in multitiered and cluttered settings. Furthermore, the approach balances energy efficiency, consuming 22.1 Wh per mission—marginally higher than the fixed antenna (18.2 Wh) but 9.8% less than predefined paths (24.5 Wh). By overcoming the limitations of static and preprogrammed systems, our robot offers a scalable, adaptable solution poised to elevate warehouse automation in the era of Industry 4.0.

1. Introduction

Efficient inventory management constitutes a cornerstone of modern supply chain logistics, where the imperative for rapid, accurate, and autonomous product tracking has escalated alongside the demands of globalized commerce [1,2]. The evolution of robotic technology has been pivotal in addressing these challenges [3], transitioning from rudimentary automated guided vehicles (AGVs) reliant on fixed tracks and magnetic tapes to sophisticated Autonomous Mobile Robots (AMRs) empowered by advanced localization, mapping, and artificial intelligence (AI) systems. Early robotic solutions, constrained by static navigation paths and minimal environmental adaptability, have given way to a new generation of intelligent platforms capable of dynamic decision making, real-time spatial reasoning, and collaborative operation within complex warehouse ecosystems [4].
A critical breakthrough in this progression emerged with the integration of Simultaneous Localization and Mapping (SLAM) algorithms and high-precision sensors such as LiDAR, inertial measurement units (IMUs), and depth cameras. These technologies enabled robots to construct real-time maps of unstructured environments while concurrently estimating their position within them [5,6]. For inventory robots, SLAM eliminated dependence on preconfigured infrastructure, allowing navigation through dynamic aisles cluttered with pallets, personnel, and transient obstacles. Enhanced by probabilistic filtering techniques (e.g., Kalman and particle filters), modern SLAM frameworks achieve centimeter-level localization accuracy [7], ensuring reliable traversal and repeatable scanning operations even in densely populated warehouses.
The advent of AI and computer vision has further revolutionized robotic inventory systems, infusing them with unprecedented cognitive depth [8]. Deep learning architectures, such as convolutional neural networks (CNNs) and transformer models, empower robots to interpret visual data with human-like acuity. For instance, YOLO (You Only Look Once) object detection algorithms [9] enable real-time identification of items, enabling the possibility of RFID tag locations, while semantic segmentation models distinguish between product categories and storage configurations. These capabilities synergize with SLAM to optimize path planning, prioritize scan regions, and adapt to fluctuating inventory layouts [10].
Radio Frequency Identification (RFID)-enabled robots epitomize the convergence of these advancements, offering automated solutions to transcend the inefficiencies of manual inventory processes [11]. However, persistent challenges in tag detection accuracy, environmental adaptability, and operational endurance underscore the need for innovative solutions that further leverage these technological strides. While contemporary robots harness SLAM for robust navigation and AI for basic rfid-based inventory task execution, the next frontier lies in unifying these technologies into cohesive, context-aware platforms capable of dynamic sensor reconfiguration, adaptability to different warehouse configurations, and increasing the efficiency and performance of these machines.
This paper is structured as follows. Section 2 reviews existing state-of-the-art RFID-based inventory management approaches, including ground-based AMRs, aerial UAVs, hybrid systems, and advanced sensor-equipped AMRs, discussing their limitations. Section 3 details the hardware and software architecture of the proposed RFID-HAND robot, focusing on its manipulator, sensor suite, mobility system, and ROS2 framework. Section 5 presents comparative experiments that evaluate the performance of different scanning strategies, demonstrating the superiority of dynamic positioning. Section 6 compares the RFID-HAND robot with a state-of-the-art robot in a virtual warehouse, highlighting the advantages of dynamic antenna positioning. Section 7 outlines potential research directions, including algorithmic optimizations and expanded capabilities. Section 8 summarizes the key findings, emphasizing the potential of vision-assisted dynamic antenna positioning for enhanced RFID-based inventory management.

2. Related Work

Recent advancements in autonomous warehouse robotics for inventory management have converged around two critical axes: improving the mobility/navigation of warehouse inventory robots and the fusion of location/proximity sensors with RFID to increase the accuracy of inventory perception. Traditional systems, while effective in structured environments, struggle to reconcile the competing demands of dynamic obstacle avoidance, real-time inventory tracking, and energy efficiency. Five principal paradigms dominate the literature, each offering unique tradeoffs:
Ground-based fixed RFID antennas on AMRs equipped with static RFID antennas have been widely deployed and investigated by researchers for warehouses and retail environments [12,13,14,15]. These systems leverage ground mobility to navigate aisles, yet their fixed antenna configurations inherently limit spatial coverage, particularly in cluttered or densely stacked environments. Occluded tags and vertical blind spots necessitate labor-intensive manual verification, undermining the autonomy these systems aim to achieve.
Aerial inventory unmanned aerial vehicles (UAVs) address spatial limitations by deploying RFID sensors from elevated vantage points, as shown in various research [16,17,18]. While aerial platforms enhance accessibility to high-altitude storage areas and can navigate in hard-to-access zones for AMRs, they face critical constraints, including limited flight endurance (usually less than 30 min), navigational complexity for indoor environments, and reduced tag detection accuracy due to variable antenna–tag distances.
Hybrid ground–aerial systems: Recent research to harmonize aerial and terrestrial mobility has yielded hybrid robots capable of multimodal ground and areal operations in [19,20]. For example, Ref. [21] introduced a hybrid UAV-wheeled platform with custom brakes on board for 3D warehouse inventory tasks. Despite improved versatility, such systems grapple with energy inefficiency, mode-switching latency, and the need for intricate control architectures to manage heterogeneous locomotion dynamics.
Advanced sensor-equipped AMRs have been investigated by scientists developing cutting-edge platforms like the Dextory robot [22] that utilizes extendable sensor towers (e.g., 12 m masts) for vertical movement of RFID antennas and coverage, although at the cost of mechanical complexity and persistent blind spots to occlusion. In contrast, the research in [23] adopts a minimalist design, leveraging rotating antennas and LiDAR-guided navigation to achieve high-resolution 3D localization without bulky infrastructure. Its integration of visual odometry and depth cameras enables robust navigation in shop-like scenarios with metal racks and wooden shelves, although its static rotational patterns lack the real-time adaptability required for irregular tag distributions, especially tags placed at high shelves. AI-enhanced navigation systems recently have been investigated and adapted by scientists to increase the performance of these robots. The work in [24] exemplifies the shift toward AI-driven adaptability. By replacing conventional Q-learning with Deep Q-Networks (DQNs), their framework achieved a 40% improvement in learning efficiency, enabling robotic agents to dynamically optimize paths in response to fluctuating inventory layouts. While their YOLO-based object detection (95% accuracy) supplants legacy QR code scanning, the continued dependence on visual markers, a technology increasingly eclipsed by RFID in modern warehouses, limits scalability in retrofitted environments.
This paper introduces the RFID-HAND robot, a vision-assisted autonomous platform that features a 3-degree-of-freedom (3DOF) robotic manipulator for dynamic antenna positioning. Unlike prior systems reliant on static or preprogrammed configurations, our design uniquely synergizes real-time visual perception, adaptive manipulator kinematics, and RFID signal optimization to address the limitations of existing solutions.
The proposed RFID-HAND robot transcends these paradigms by unifying SLAM navigation, AI-vision-guided antenna manipulation and scanning, and adaptive RFID sensing. Unlike the robot developed in [24], which employs decision trees for post-detection analysis, our system eliminates intermediary processing through closed-loop control: YOLO detections directly inform inverse kinematic calculations, enabling precision antenna positioning without QR code dependencies. This integration resolves the perceptual latency inherent in hybrid architectures while surpassing the spatial constraints of rotating antenna systems. Furthermore, our ROS2-based navigation stack—augmented by SLAM and dynamic costmaps—parallels the policy gradient advantages of DQNs, achieving collision-free traversal in cluttered environments without sacrificing RFID scanning efficacy.

3. Robot Design

3.1. RFID-HAND Robot Hardware Description

The hardware description of the RFID-HAND robot can be categorized and explained into three blocks, as shown in Figure 1 and described in detail below:
  • Robot arms block: The robot is designed with a robotic hand offering 3 DOF, which is powered by three cost-effective servo motors equipped with feedback capabilities for precise and accurate movement. An RFID antenna is mounted on the end effector, allowing for efficient spatial manipulation. The joints of the robotic hand are interconnected with an elastic rubber band, capable of supporting weights up to 30 kg, which not only reduces pressure on the servos but also enhances the stall torque limit, thus optimizing the performance of each joint.
  • Sensors block: The sensor array of the robot is meticulously arranged, featuring a 2D LiDAR sensor mounted on top of the chassis for critical proximity detection and 2D environmental mapping. Additionally, an RGB-D camera is integrated to capture both color images and depth information, enabling the robot to accurately perceive the shape, size, and distance of objects in its vicinity. The computational backbone of the robot is a cost-effective and energy-efficient single-board computer (SBC), specifically the n100, which processes sensor data, executes control algorithms, and supports advanced functionalities such as neural network-based object recognition, navigation, and localization.
  • Mobility block: For mobility, the robot is equipped with two 24 V high-torque motors, each fitted with hall sensors and a motion controller, ensuring precise control over the robot’s movement, velocity, and acceleration. The power system is designed for sustained operation, either via a rechargeable battery pack or an external power source, ensuring the robot’s continuous functionality.

3.2. RFID-HAND Robot Software Description

The software architecture deployed on the robotic platform is centered around the Robot Operating System 2 (ROS2) framework [25], which provides a robust and scalable foundation for complex robotic operations. The integration of the differential drive motor controller with ROS2’s control software [26] facilitates precise communication, enabling accurate odometry and controlled locomotion.
The robot’s navigation system utilizes the ROS2 Nav2 stack [27], a multilayered software that enables the creation of both local and global costmaps. These costmaps are derived from sensor data gathered by the robot’s comprehensive perception layer, allowing for sophisticated interaction with the environment. The perception layer is a critical component, comprising the data obtained from the array of sensors. This sensor suite includes 360-degree LiDAR data for comprehensive proximity data, an RGB-D depth camera point cloud data for enhanced environmental awareness through depth perception, and localization sensor data, such as from the hall sensors integrated into the motors and an inertial measurement unit (IMU), which contribute to the robot’s spatial orientation and map generation. Collectively, these sensory inputs provide the necessary data to construct and refine maps, which are essential for effective navigation and obstacle avoidance. The navigation and path-planning module leverages data from the perception layer to compute optimal trajectories toward predefined goals while dynamically avoiding obstacles. This capability is achieved through path-planning algorithms that process real-time sensor data to assess the environment, analyze current conditions, and determine the most efficient and safe paths for the robot’s movement. In addition to navigation, the software system includes a custom-designed controller for the robotic arms mounted on the platform. This controller acts as an intermediary, integrating inverse kinematics algorithms from ROS2 MoveIt [28] with essential data on joint angles, velocities, and feedback for precise control of the robot’s arms. Through its interface with ROS2 MoveIt, the custom controller translates high-level motion plans into executable commands, enabling the robot’s arms to operate with accuracy and responsiveness. Moreover, the robot’s software system incorporates an algorithm that utilizes detection information from a pretrained YOLO model [9] to identify potential warehouse items within its environment. Upon detecting an item, the algorithm computes the optimal kinematic movements for the robot’s 3DOF arm, which is equipped with an RFID antenna at its end effector. The algorithm precisely orients the arm toward the detected item, ensuring accurate targeting. Once aligned, the arm performs a controlled circular motion around the item’s perimeter, effectively covering its entire surface area to maximize RFID scanning efficiency and ensure comprehensive data capture of the item.

4. Technical Overview

The robot employs a pretrained YOLO9 model to detect boxes and similar warehouse products that are attached to RFID tags. YOLO9 provides bounding boxes and pixel positions of detected objects in the camera’s image plane. To interact with these objects, the robot translates the 2D pixel positions into 3D world coordinates utilizing the information of the depth point supplied by the camera, enabling the robotic arm to orient its end effector accurately. The depth value g z is obtained from the RGB-D depth camera corresponding to the center pixel ( x c , y c ) . The camera’s intrinsic matrix K is defined using Equation (1):
K = f x 0 c x 0 f y c y 0 0 1
where f x and f y are the focal lengths, and ( c x , c y ) is the optical center. For a detected object’s bounding box with the top-left corner ( x min , y min ) and the bottom-right corner ( x max , y max ) , the center of the bounding box in pixel coordinates and the 3D world coordinates ( g x , g y , g z ) of the detected object can be derived in Equations (2) and (3):
x c = x min + x max 2 , y c = y min + y max 2
g x = ( x c c x ) · g z f x , g y = ( y c c y ) · g z f y
The accuracy of object detection using YOLO can be affected by varying lighting conditions and occlusions commonly encountered in warehouse environments. Such factors may lead to missed detections or inaccurate bounding boxes, subsequently impacting antenna positioning accuracy. To mitigate these issues, we employed data augmentation techniques during YOLO training that specifically included varied lighting conditions and partial occlusions to enhance robustness against real-world environmental variations. A position of the center of a detected object is therefore extracted with coordinates g x , g y , g z . Using the trigonometry equations in Equations (4)–(8) and as illustrated in Figure 2, we compute the position h x , h y , h z along the line r . h is the desired point position to move the end effector (RFID antenna) to a point at the center of the detected object.
n = l 2 2 h z 2
tan θ = m o = h z l 1 + n
h z = m ( l 1 + l 2 2 h z 2 ) o
= h z = ( m 2 · l 2 2 + m · l 1 o 2 + m 2 )
tan ϕ = h y t = p r
h x = l 1 + n
where l 1 is the distance between the first joint in M1 (called the arm base) and the second joint at M2 (called the shoulder joint). l 2 is the distance between the second joint in M2 and the third joint in M3 (called the elbow joint). l 3 is the distance between the third joint in M3 and the end effector (the antenna).
θ 1 , θ 2 , and θ 3 are the angles of the y-axis plane joints M1, M2, and M3, respectively, as shown in Figure 2.
ϕ 1 , ϕ 2 , and ϕ 3 are the angles of the joint of the x-axis planes of M1, M2, and M3, respectively, as shown in Figure 2. We assume that θ 3 , which is the angle of the elbow joint shown in Figure 2, is 0 deg , as the robotic arm should be extended to the maximum for forming a pointing shape toward a target.
For further manipulation of the robotic arm, a forward and inverse kinematics model is used to move the frames as M 1 M 2 M 3 , as shown in Figure 2. To represent the kinematic chain, we use the Denavit–Hartenberg (DH) [29] convention to define each joint’s transformation matrix. Using the DH parameters, the transformation matrices for each joint T 1 , T 2 , and T 3 are respectively represented as Equations (9)–(11):
T 1 = cos ( θ 1 ) sin ( θ 1 ) 0 l 1 sin ( θ 1 ) cos ( θ 1 ) 0 0 0 0 1 0 0 0 0 1
T 2 = cos ( θ 2 ) 0 sin ( θ 2 ) l 2 cos ( θ 2 ) 0 1 0 0 sin ( θ 2 ) 0 cos ( θ 2 ) l 2 sin ( θ 2 ) 0 0 0 1
T 3 = cos ( θ 3 ) sin ( θ 3 ) 0 l 3 cos ( θ 3 ) sin ( θ 3 ) cos ( θ 3 ) 0 l 3 sin ( θ 3 ) 0 0 1 0 0 0 0 1
The overall transformation from the base to the end effector is given by Equation (12):
T = T 1 · T 2 · T 3
Inverse kinematics involves determining the joint angles ( θ 1 , θ 2 , θ 3 ) given the desired position ( x , y , z ) in the Cartesian coordinate system and can be extracted using the previous equations.
From the obtained transformation matrix T, the position of the end effector is defined as R 1 × 3 .
The computed joint angles ( θ 1 , θ 2 , θ 3 ) are fed into the robotic arm controller, which adjusts the arm’s position dynamically. To optimize RFID tag scanning and inventory management, the robot ensures that the RFID antenna at the end effector points toward and entirely covers the detected object. To move the end effector in a circular motion around the product, the robot must adjust the angles θ 1 , θ 2 , and θ 3 dynamically. The circle’s radius must be appropriate to cover the product surface, with the center at g x , g y , g z . From the given boundary box of the detected item, the distance between the maximum edge of the boundary box to the center would be considered as the radius r. The desired circular motion path that the end effector would need to follow is a set of consecutive points defined as x , y , z . The parametric equations for the circular path are shown in Equation (13):
x ( t ) = x + r cos ( t ) , y ( t ) = y + r sin ( t )
where t ranges from 0 to 2 π .

5. Experiments

5.1. Short Aile Low Shelves Scanning

The objective of the experiment described here in Section 5.1 is to assess the comparative effectiveness of three RFID scanning approaches for inventory management. This study specifically evaluates the performance of the RFID-HAND robot using a static RFID antenna versus a predefined path strategy, where the manipulator moves the antenna across preset spatial points in addition to an intelligent detection and dynamic positioning method, where a pretrained YOLO model identifies probable products and utilizes forward and inverse kinematics to precisely position the manipulator’s end effector (the antenna) to perform a tailored circular motion around the object, ensuring comprehensive RFID tag scanning. The primary metric of comparison is the number of RFID tags detected while scanning shelves containing tagged boxes in a controlled environment. The experiment was carried out in a controlled laboratory with 15 m × 5 m layout dimensions. Low shelves with (0.2 m height) and high shelves (up to 1.5 m height) were placed along a straight line in the laboratory. A total of 500 RFID tags were distributed across shelf levels. The distribution of RFID tags were evenly placed inside boxes arranged systematically along shelves. The optimal RFID antenna reading range is (0–3 m), and the specific RFID reader power settings specified (30 dBm). The robot navigated a given path while placed 0.8 m away from the shelves to simulate a real warehouse aisle. The robot conducted two distinct experiments in the scenario illustrated in Figure 3.

5.1.1. Fixed Antenna

In the experiment detailed here in Section 5.1.1, the robot was programmed to navigate in a straight line along the aisle, maintaining a constant speed. The antenna was fixed in one position oriented perpendicular to the shelves to scan for RFID tags, as shown in Figure 4. The robot made a single pass navigating along the aisle, recording the number of RFID tag reads.

5.1.2. Articulated Antenna

In the experiment described here in Section 5.1.2, the robot was programmed to navigate the same straight line path along the aisle at the same speed as in experiment Section 5.1.1. As the robot moved, the robotic arm articulated in 3D space, moving the antenna in a sweeping motion to cover a broader spatial plane. The motion pattern of the arm was preprogrammed to ensure consistent coverage, as shown in Figure 5. The robot made a single pass along the aisle, recording the number of RFID tag reads.
Upon completion of both experiments, the RFID read data were analyzed to determine the number of unique RFID tags detected in each scenario. The results in Figure 6 demonstrate a significant difference in the number of RFID tags detected between the two scanning methods. The fixed antenna position resulted in a limited detection range, with a total of 117 tags read; however, the articulated robotic arm significantly increased the coverage area, resulting in 438 tags read from 500 total placed tags in the environment.
The articulated antenna setup showed a higher number of RFID reads, indicating improved spatial coverage and detection capability. This improvement is attributed to the dynamic motion of the antenna, which allowed better orientation and positioning relative to the tags.

5.1.3. Dynamic Movement Articulated Antenna

In this experiment, the RFID-HAND robot utilized the vision-assisted dynamic antenna positioning method to evaluate its performance in reading RFID tags. The experimental setup was identical to that of the experiments described in Section 5.1.1 and Section 5.1.2.
The vision-assisted dynamic antenna positioning method integrated a pretrained YOLO model for object detection, which identified tagged boxes in real time. Using forward and inverse kinematics, the robot dynamically adjusted its 3DOF manipulator to align the RFID antenna with each detected box, as shown in Figure 7. The antenna then performed a tailored circular scanning motion around the object’s surface to maximize RFID tag detection coverage, as described in Section 4.
This experiment resulted in the detection of 479 unique tags out of the 500 placed tags, achieving a detection rate of 95.8%. The results are illustrated in Figure 8 alongside those from the experiments described in Section 5.1.1 and Section 5.1.2 for comparison.
The results demonstrate that incorporating vision-based dynamic positioning significantly enhances RFID tag detection performance compared to static or preprogrammed scanning methods. The ability to read 479 out of 500 tags demonstrates that this method effectively compensates for challenges such as occlusions, varied object heights, and complex spatial arrangements often encountered in warehouse environments.

5.2. Tall Ailes High Shelves Scanning

The objective of this experiment described in Section 5.2 is to compare the effectiveness and performance of the two different RFID scanning methods for inventory management as in the experiments described in Section 5.1.1, Section 5.1.2 and Section 5.1.3. The experiment was carried out in a controlled laboratory with 15 m × 5 m layout dimensions. Low shelves with (0.2 m height) and high shelves (up to 2.5 m height) were placed along a straight line in the laboratory. A total of 350 RFID tags were distributed across the shelf levels. The distribution of RFID tags were evenly placed inside and on the boxes arranged systematically along shelves. The optimal RFID antenna reading range is (0–3 m), and the specific RFID reader power settings specified (30 dBm).
The robot navigated a given path while placed 0.8 m away from the shelves to simulate a real warehouse aisle, as shown in Figure 9. The robot conducted two separate experiments.

5.2.1. Fixed Antenna

As in experiment Section 5.1.1, the robot also was programmed to navigate in a straight line along the aisle, maintaining a constant speed. The antenna was fixed in one position oriented perpendicular to the shelves to scan for RFID tags. The robot made a single pass along the aisle, recording the number of RFID tag reads from both high and low shelves.

5.2.2. Articulated Antenna

As in the experiment described in Section 5.2.1, the robot was programmed to navigate the same path and with the same speed along the aisle at the same speed as in the experiment in Section 5.2.1. As the robot moved, the robotic arm articulated in 3D space, moving the antenna in a sweeping motion to cover both high and low shelves as shown in Figure 10. The same preprogrammed motion pattern of the arm was used to ensure consistent coverage of all shelf levels. The robot made a single pass along the aisle, recording the number of RFID tag reads from both high and low shelves, as shown in Figure 11.
The fixed antenna position as in the experiment described in Section 5.2.1 resulted in a limited detection range, with a total of 67 tags read from high and low shelves. The programmed motion of the articulated arm increased the coverage area, resulting in a total of 314 RFID tags read from shelves from a total of 350 RFID tags placed, which shows a remarkable difference.

5.2.3. Dynamic Movement Articulated Antenna

For the experiment described here in Section 5.2.3, the robot was programmed to navigate the same straight-line path along the aisle at the same speed as in previous experiments. As the robot moved, utilizing the information from the detections from the YOLO model, the algorithm dynamically guided the robotic arm, adjusting the antenna’s main radiation lobe toward the detected boxes with RFID tags in real time and then performing a controlled circular motion around the item’s perimeter, as shown in Figure 12, effectively covering its entire surface area, as explained in Section 3.2. The robot made a single pass along the aisle, recording the number of RFID tag reads from both high and low shelves.
The results demonstrated in Figure 13 show significant differences in the number of RFID tags detected between the three scanning methods, particularly for the high shelves.
The vision-assisted dynamic antenna positioning further increased the coverage area. The approach outperformed the others, resulting in 343 tags read from all shelves.
Table 1 summarizes the RFID tag detection rates and energy consumption among the scanning methods. The dynamic positioning approach achieved detection rates of 95.8% (479/500) and 98.0% (343/350) for low and high shelves, respectively, outperforming the fixed (23.4%, 19.1%) and predefined path (87.6%, 89.7%) methods. Notably, while the dynamic method consumed 22.1 Wh per mission and was marginally higher than the fixed antenna (18.2 Wh), it reduced energy use by 9.8% compared to the predefined paths method (24.5 Wh), demonstrating a balance between efficiency and performance.

6. Simulation Experiment: Dynamic vs. Fixed Antenna Performance

To quantitatively evaluate the advantages of dynamic antenna positioning, we conducted a comparative simulation of the proposed robot in this paper vs. a SOTA RFID-based inventory robot [30] in a Gazebo-based virtual warehouse environment.
The simulated workspace comprised three industrial pallet racks (2.4 m height × 5.2 m depth), each containing three horizontal tiers (0.8 m vertical spacing) replicating common warehouse storage configurations. A total of 600 passive UHF RFID tags were distributed pseudo-randomly across pallet surfaces using a uniform spatial distribution model, with tag orientations randomized to emulate real-world product placement scenarios. The tags were simulated using the RFID Gazebo plugin [31], which implements a probabilistic model specially designed for environments with large constellations of RFID tags, as described in [32]. Undetected tags were visually represented as gray cuboids (5 cm × 5 cm × 5 cm) to facilitate post-experiment analysis, as shown in Figure 14.
The experiment compared our RFID-HAND robot with the Robin50 platform, which is an RFID-based inventory robot employing a four static antenna configuration. To ensure equitable comparison, both systems operated with two active antennas: RFID-HAND utilized its dual-arm manipulators (one antenna per end effector), while Robin50’s left and right chassis-mounted antennas were activated, deactivating its front/rear arrays to match the antenna count. As illustrated in Figure 14, robots traversed a 78 m predefined path (0.8 m aisle clearance) at 0.2 m/s, adhering to ISO 3691-4 safety standards for warehouse AMRs.
Key performance metrics focused on tag detection rate versus temporal progression, with results shown in Figure 15. When the RFID-HAND robot detects an RFID tag, the tag changes its color to a red cube, as shown in Figure 16. The RFID-HAND achieved 97.1% detection accuracy (582/600 tags), demonstrating superior performance through its vision-guided adaptive scanning. In contrast, when the Robin50 robot detects an RFID tag, the tag changes its color to a yellow cube, as shown in Figure 16.
Robin50’s fixed antennas detected only 70.6% (424/600), which were primarily limited by the following:
  • Vertical Coverage Constraints: Static antennas failed to maintain optimal read zones for upper-tier tags (1.6–2.4 m height), exhibiting 58.2% detection rate above 1.2 m versus 96.4% for RFID-HAND.
  • Occlusion Sensitivity: Fixed beam patterns could not circumvent pallet obstructions, whereas the manipulator actively reoriented antennas to exploit RF propagation paths.
  • Angular Coverage Limitations: The circular scanning motion increased antenna–tag polarization alignment opportunities, raising read likelihood by around 26.5% compared to fixed orientations.
The results demonstrated in Figure 13, show significant differences in the number of RFID tags detected between the three scanning methods, particularly for high shelves. These results empirically validate that dynamic antenna positioning substantially outperforms static configurations in cluttered, multitiered environments. The 26.5% performance gap highlights the critical importance of spatial adaptability in RFID-based inventory systems, particularly when scanning irregularly arranged stock-keeping units. This experiment’s methodology and dataset have been open sourced to facilitate reproducibility and benchmarking efforts in autonomous inventory robotics.

7. Future Work

Future research will focus on further refinement of the robot’s object detection and antenna positioning algorithms to enhance accuracy and reduce processing time. Additionally, we aim to explore the integration of machine learning techniques to adapt the scanning patterns based on real-time feedback, potentially improving performance in highly cluttered environments. From the experiments, we observed that an optimal robot velocity of approximately 0.2 m/s ensures both accurate RFID tag readings and safe interactions within dynamic environments. At this speed, the processing time required by the YOLO model for object detection and subsequent real-time kinematic calculations is sufficient to maintain robust performance without noticeable delays. Nonetheless, we recognize that computational latency could pose limitations if higher operational speeds were desired. Therefore, future research will explore algorithmic optimization techniques such as hardware acceleration (e.g., GPU integration) or lightweight neural network architectures to further reduce computational delays and potentially enable higher speed operations without compromising accuracy or safety. Another area of interest is the expansion of the robot’s capabilities to handle a broader range of object shapes and materials, as well as the development of multirobot systems for cooperative inventory tasks. Ultimately, although our current experiments demonstrated significant advantages of dynamic antenna positioning under simplified controlled scenarios, future research will involve rigorous testing within real warehouse environments and more complex controlled settings featuring densely packed shelves, dynamic obstacles such as moving personnel or vehicles, partial occlusions, varied lighting conditions, and realistic operational constraints.

8. Conclusions

In this study, we presented the design and implementation of a vision-assisted dynamic antenna positioning RFID-based inventory robot that utilizes a 3DOF manipulator. The robot integrates a 3DOF arm with an RFID antenna in its end effector, making it capable of detecting and scanning objects within a warehouse environment. Our approach leverages both forward and inverse kinematics to precisely position the antenna, allowing for a comprehensive scan of RFID tags through circular motion tailored to the object’s size. The comparative analysis demonstrated that our proposed method outperforms traditional fixed antenna configurations and predefined movement strategies in terms of tag detection efficiency and coverage. Specifically, the dynamic positioning and targeted scanning enabled by YOLO-based object detection significantly increased the number of RFID tags read across various object sizes and heights, marking a substantial advancement in inventory management robotics. Furthermore, the dynamic method achieved superior energy efficiency, consuming 22.1 Wh per mission compared to 24.5 Wh for predefined paths, thus optimizing detection performance while maintaining sustainable operational costs in inventory tasks. The simulation results demonstrate that dynamic antenna positioning significantly enhances RFID detection efficacy, with the proposed system achieving 97.1% tag detection versus 74.0% for static configurations. This 26.5% improvement highlights the value of adaptive spatial reconfiguration in addressing vertical coverage gaps, occlusions, and polarization mismatches in multitiered environments. Vision-guided manipulation enables precise alignment, particularly for elevated (>1.2 m) or obstructed tags. While these findings suggest substantial potential for real-world applications, further validation through physical deployments in operational warehouses is necessary to confirm practical scalability and robustness. Although our experiments demonstrated significant improvements under controlled laboratory conditions, future work will involve rigorous testing in more complex warehouse environments featuring densely packed shelves, dynamic obstacles such as moving personnel or vehicles, partial occlusions, varied lighting conditions, and realistic operational constraints to thoroughly validate our system’s robustness.

Author Contributions

Conceptualization, A.A.A.; Methodology, A.A.A.; Software, A.A.A.; Validation, A.A.A.; Formal analysis, A.A.A.; Investigation, A.A.A.; Resources, A.A.A.; Data curation, A.A.A.; Writing—original draft, A.A.A.; Writing—review & editing, A.A.A.; Visualization, A.A.A.; Supervision, R.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work is part of Project StoreTwin, funded by the “Interempresas Internacional” program of the Spanish CDTI (Ref. CIIP-20221007) under the Eurostars program (Ref. CIIP-20221007).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krishnan, R.; Perumal, E.; Govindaraj, M.; Kandasamy, L. Enhancing Logistics Operations Through Technological Advancements for Superior Service Efficiency. In Innovative Technologies for Increasing Service Productivity; IGI Global: Hershey, PA, USA, 2024; pp. 61–82. [Google Scholar]
  2. Lai, Y. Innovative Strategies in Logistics and Supply Chain Management: Navigating Modern Challenges. In Proceedings of the SHS Web of Conferences, Funchal, Portugal, 11–12 November 2024; EDP Sciences: Les Ulis, France, 2024; Volume 183, p. 02020. [Google Scholar]
  3. Mohan Banur, O.; Patle, B.; Pawar, S. Integration of robotics and automation in supply chain: A comprehensive review. Robot. Syst. Appl. 2024, 4, 1–19. [Google Scholar] [CrossRef]
  4. Liu, C.; Liu, Y.; Xie, R.; Li, Z.; Bai, S.; Zhao, Y. The evolution of robotics: Research and application progress of dental implant robotic systems. Int. J. Oral Sci. 2024, 16, 28. [Google Scholar] [CrossRef] [PubMed]
  5. Li, Z.; Fan, C.; Ding, W.; Qian, K. Robot Navigation and Map Construction Based on SLAM Technology. 2024. Available online: https://adwenpub.com/index.php/wjimt/article/view/291 (accessed on 15 May 2024).
  6. Yarovoi, A.; Cho, Y.K. Review of simultaneous localization and mapping (SLAM) for construction robotics applications. Autom. Constr. 2024, 162, 105344. [Google Scholar] [CrossRef]
  7. Ress, V.; Zhang, W.; Skuddis, D.; Haala, N.; Soergel, U. Slam for indoor mapping of wide area construction environments. arXiv 2024, arXiv:2404.17215. [Google Scholar] [CrossRef]
  8. Khan, I.; Blake, G. The Future of SME Manufacturing: Computer Vision and AI for Enhanced Quality Control and Inventory Accuracy. 2024. Available online: https://www.researchgate.net/publication/386409991_The_Future_of_SME_Manufacturing_Computer_Vision_and_AI_for_Enhanced_Quality_Control_and_Inventory_Accuracy (accessed on 16 February 2025).
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  10. Zhang, K.; Dong, C.; Guo, H.; Ye, Q.; Gao, L.; Xiang, S.; Chen, X.; Wu, Y. A semantic visual SLAM based on improved mask R-CNN in dynamic environment. Robotica 2024, 42, 3570–3591. [Google Scholar] [CrossRef]
  11. Bandara, I.; Simpson, O.; Sun, Y. Optimizing Efficiency Using a Low-Cost RFID-Based Inventory Management System. In Proceedings of the 2024 International Wireless Communications and Mobile Computing (IWCMC), Ayia Napa, Cyprus, 27–31 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1729–1733. [Google Scholar]
  12. Gareis, M.; Parr, A.; Trabert, J.; Mehner, T.; Vossiek, M.; Carlowitz, C. Stocktaking Robots, Automatic Inventory, and 3D Product Maps: The Smart Warehouse Enabled by UHF-RFID Synthetic Aperture Localization Techniques. IEEE Microw. Mag. 2021, 22, 57–68. [Google Scholar] [CrossRef]
  13. Sharma, R.; Patange, A.D.; Padalghare, R.; Kale, R.C. Development of LiDAR operated inventory control and assistance robot. Proc. Inst. Mech. Eng. Part E 2024, 238, 192–202. [Google Scholar] [CrossRef]
  14. Zhang, J.; Lyu, Y.; Roppel, T.; Patton, J.; Senthilkumar, C. Mobile robot for retail inventory using RFID. In Proceedings of the 2016 IEEE International Conference on Industrial Technology (ICIT), Taipei, Taiwan, 14–17 March 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 101–106. [Google Scholar]
  15. Morenza-Cinos, M.; Casamayor-Pujol, V.; Soler-Busquets, J.; Sanz, J.L.; Guzmán, R.; Pous, R. Development of an RFID inventory robot (AdvanRobot). In Robot Operating System (ROS); The Complete Reference; Springer: Cham, Switzerland, 2017; Volume 2, pp. 387–417. [Google Scholar]
  16. Alajami, A.A.; Moreno, G.; Pous, R. Design of a UAV for Autonomous RFID-Based Dynamic Inventories Using Stigmergy for Mapless Indoor Environments. Drones 2022, 6, 208. [Google Scholar] [CrossRef]
  17. Beul, M.; Droeschel, D.; Nieuwenhuisen, M.; Quenzel, J.; Houben, S.; Behnke, S. Fast Autonomous Flight in Warehouses for Inventory Applications. IEEE Robot. Autom. Lett. 2018, 3, 3121–3128. [Google Scholar] [CrossRef]
  18. Gago, R.M.; Pereira, M.Y.; Pereira, G.A. An aerial robotic system for inventory of stockpile warehouses. Eng. Rep. 2021, 3, e12396. [Google Scholar] [CrossRef]
  19. Alajami, A.A.; Santa Cruz, L.D.; Pous, R. Design of an Energy-Efficient Self-Heterogeneous Aerial-Ground Vehicle. In Proceedings of the 2023 9th International Conference on Automation, Robotics and Applications (ICARA), Abu Dhabi, United Arab Emirates, 10–12 February 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 213–218. [Google Scholar]
  20. Cao, M.; Xu, X.; Yuan, S.; Cao, K.; Liu, K.; Xie, L. DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels. In Proceedings of the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; pp. 6962–6969. [Google Scholar] [CrossRef]
  21. Alajami, A.A.; Perez, F.; Pous, R. The Design of an RFID-Based Inventory Hybrid Robot for Large Warehouses. In Proceedings of the 2024 9th International Conference on Control and Robotics Engineering (ICCRE), Osaka, Japan, 10–12 May 2024; pp. 50–54. [Google Scholar] [CrossRef]
  22. Dexory. Dexory Inventory Robot. 2024. Available online: https://www.dexory.com/ (accessed on 9 August 2024).
  23. Bernardini, F.; Motroni, A.; Nepa, P.; Tripicchio, P.; Buffi, A.; Del Col, L. The MONITOR Project: RFID-based Robots enabling real-time inventory and localization in warehouses and retail areas. In Proceedings of the 2021 6th International Conference on Smart and Sustainable Technologies (SpliTech), Bol and Split, Croatia, 8–11 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  24. Vivek, N.; Suriya, A.; Kamala, V.; Chinnasamy, N.V.; Akshya, J.; Sundarrajan, M.; Choudhry, M.D. Integrating Deep Q-Networks and YOLO with Autonomous Robots for Efficient Warehouse Management. In Proceedings of the 2024 5th International Conference on Recent Trends in Computer Science and Technology (ICRTCST), Jamshedpur, India, 9–10 April 2024; pp. 419–424. [Google Scholar] [CrossRef]
  25. Macenski, S.; Foote, T.; Gerkey, B.; Lalancette, C.; Woodall, W. Robot Operating System 2: Design, architecture, and uses in the wild. Sci. Robot. 2022, 7, eabm6074. [Google Scholar] [CrossRef] [PubMed]
  26. Fröhlich, C. Ros2 Control Framework. 2022. Available online: https://github.com/ros-controls/ros2_control (accessed on 21 November 2024).
  27. Macenski, S.; Moore, T.; Lu, D.V.; Merzlyakov, A.; Ferguson, M. From the desks of ROS maintainers: A survey of modern & capable mobile robotics algorithms in the robot operating system 2. Robot. Auton. Syst. 2023, 168, 104493. [Google Scholar]
  28. Görner, M.; Haschke, R.; Ritter, H.; Zhang, J. Moveit! task constructor for task-level motion planning. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
  29. Balasubramanian, R. The Denavit Hartenberg Convention; Robotics Insitute Carnegie Mellon University: Pittsburgh, PA, USA, 2011. [Google Scholar]
  30. Casamayor-Pujol, V.; Morenza-Cinos, M.; Gastón, B.; Pous, R. Autonomous stock counting based on a stigmergic algorithm for multi-robot systems. Comput. Ind. 2020, 122, 103259. [Google Scholar] [CrossRef]
  31. Alajami, A.; Moreno, G.; Pous, R. RFID Sensor Gazebo Plugin. 2025. Available online: http://wiki.ros.org/RFIDsensor_Gazebo_plugin (accessed on 3 April 2025).
  32. Alajami, A.A.; Moreno, G.; Pous, R. A ROS Gazebo Plugin Design to Simulate RFID Systems. IEEE Access 2022, 10, 93921–93932. [Google Scholar] [CrossRef]
Figure 1. RFID-HAND hardware description.
Figure 1. RFID-HAND hardware description.
Sensors 25 02418 g001
Figure 2. (a) Joint configurations and circular scanning motion around detected product. (b) Trigonometric parameters used for end effector positioning.
Figure 2. (a) Joint configurations and circular scanning motion around detected product. (b) Trigonometric parameters used for end effector positioning.
Sensors 25 02418 g002
Figure 3. Controlled warehouse scenario illustrating shelf arrangement with clearly defined shelf heights (0.2 m low shelves; 1.5 m high shelves), aisle dimensions (15 m length × 5 m width), and systematic RFID tag distribution.
Figure 3. Controlled warehouse scenario illustrating shelf arrangement with clearly defined shelf heights (0.2 m low shelves; 1.5 m high shelves), aisle dimensions (15 m length × 5 m width), and systematic RFID tag distribution.
Sensors 25 02418 g003
Figure 4. A descriptive illustration of the RFID-HAND robot performing inventory with a fixed antenna in experiment Section 5.1.1.
Figure 4. A descriptive illustration of the RFID-HAND robot performing inventory with a fixed antenna in experiment Section 5.1.1.
Sensors 25 02418 g004
Figure 5. A descriptive illustration of the RFID-HAND robot performing inventory with an articulated antenna in experiment Section 5.1.2.
Figure 5. A descriptive illustration of the RFID-HAND robot performing inventory with an articulated antenna in experiment Section 5.1.2.
Sensors 25 02418 g005
Figure 6. A graph of RFID tag readings from the robot with fixed antenna and articulated antenna experiments.
Figure 6. A graph of RFID tag readings from the robot with fixed antenna and articulated antenna experiments.
Sensors 25 02418 g006
Figure 7. A descriptive illustration of the RFID-HAND robot performing inventory with an dynamic moving articulated antenna in experiment in Section 5.1.3.
Figure 7. A descriptive illustration of the RFID-HAND robot performing inventory with an dynamic moving articulated antenna in experiment in Section 5.1.3.
Sensors 25 02418 g007
Figure 8. A graph of the resulting RFID tag readings from the robot with fixed antenna in experiment in Section 5.1.1, articulated antenna in experiment in Section 5.1.2, and articulated antenna in experiment in Section 5.1.3.
Figure 8. A graph of the resulting RFID tag readings from the robot with fixed antenna in experiment in Section 5.1.1, articulated antenna in experiment in Section 5.1.2, and articulated antenna in experiment in Section 5.1.3.
Sensors 25 02418 g008
Figure 9. A descriptive illustration of the RFID-HAND robot performing inventory with a fixed antenna in experiment in Section 5.2.1.
Figure 9. A descriptive illustration of the RFID-HAND robot performing inventory with a fixed antenna in experiment in Section 5.2.1.
Sensors 25 02418 g009
Figure 10. A descriptive illustration of the RFID-HAND robot performing inventory with an articulated antenna in experiment in Section 5.2.2.
Figure 10. A descriptive illustration of the RFID-HAND robot performing inventory with an articulated antenna in experiment in Section 5.2.2.
Sensors 25 02418 g010
Figure 11. A graph of the resulting RFID tag readings from the robot with fixed antenna in experiment in Section 5.2.1 and articulated antenna in experiment in Section 5.2.2.
Figure 11. A graph of the resulting RFID tag readings from the robot with fixed antenna in experiment in Section 5.2.1 and articulated antenna in experiment in Section 5.2.2.
Sensors 25 02418 g011
Figure 12. A descriptive illustration of the RFID-HAND robot performing inventory with an dynamic moving articulated antenna in experiment in Section 5.2.3.
Figure 12. A descriptive illustration of the RFID-HAND robot performing inventory with an dynamic moving articulated antenna in experiment in Section 5.2.3.
Sensors 25 02418 g012
Figure 13. A graph of the resulted RFID tag readings from the robot with fixed antenna in experiment in Section 5.2.1, articulated antenna in experiment in Section 5.2.2, and dynamic movement antenna in experiment in Section 5.2.3.
Figure 13. A graph of the resulted RFID tag readings from the robot with fixed antenna in experiment in Section 5.2.1, articulated antenna in experiment in Section 5.2.2, and dynamic movement antenna in experiment in Section 5.2.3.
Sensors 25 02418 g013
Figure 14. (a) Simulated robots in gazebo simulator. (b) Simulated warehouse in Gazebo simulator.
Figure 14. (a) Simulated robots in gazebo simulator. (b) Simulated warehouse in Gazebo simulator.
Sensors 25 02418 g014
Figure 15. The plot of the detected RFID tags vs. time obtained by both robots from the simulation experiment.
Figure 15. The plot of the detected RFID tags vs. time obtained by both robots from the simulation experiment.
Sensors 25 02418 g015
Figure 16. (a) Illustration of the RFID-HAND robot with the detected RFID tags (in red boxes) after successfully navigating the given path. (b) Illustration of the Robin50 with the detected RFID tags (in yellow boxes) after successfully navigating the given path.
Figure 16. (a) Illustration of the RFID-HAND robot with the detected RFID tags (in red boxes) after successfully navigating the given path. (b) Illustration of the Robin50 with the detected RFID tags (in yellow boxes) after successfully navigating the given path.
Sensors 25 02418 g016
Table 1. Comparative performance and energy efficiency of scanning methods.
Table 1. Comparative performance and energy efficiency of scanning methods.
Scanning MethodTags ReadTotal Read (%)Energy Used (Wh)
Low ShelvesHigh Shelves
Fixed Antenna117/500 (23.4%)67/350 (19.1%)21.618.2
Predefined Path438/500 (87.6%)314/350 (89.7%)88.524.5
Dynamic Positioning479/500 (95.8%)343/350 (98.0%)96.722.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alajami, A.A.; Pous, R. The Design of a Vision-Assisted Dynamic Antenna Positioning Radio Frequency Identification-Based Inventory Robot Utilizing a 3-Degree-of-Freedom Manipulator. Sensors 2025, 25, 2418. https://doi.org/10.3390/s25082418

AMA Style

Alajami AA, Pous R. The Design of a Vision-Assisted Dynamic Antenna Positioning Radio Frequency Identification-Based Inventory Robot Utilizing a 3-Degree-of-Freedom Manipulator. Sensors. 2025; 25(8):2418. https://doi.org/10.3390/s25082418

Chicago/Turabian Style

Alajami, Abdussalam A., and Rafael Pous. 2025. "The Design of a Vision-Assisted Dynamic Antenna Positioning Radio Frequency Identification-Based Inventory Robot Utilizing a 3-Degree-of-Freedom Manipulator" Sensors 25, no. 8: 2418. https://doi.org/10.3390/s25082418

APA Style

Alajami, A. A., & Pous, R. (2025). The Design of a Vision-Assisted Dynamic Antenna Positioning Radio Frequency Identification-Based Inventory Robot Utilizing a 3-Degree-of-Freedom Manipulator. Sensors, 25(8), 2418. https://doi.org/10.3390/s25082418

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop