Next Article in Journal
Biplanar Nulling Coil System for OPM-MEG Using Printed Circuit Boards
Previous Article in Journal
Low Signal-to-Noise Ratio Optoelectronic Signal Reconstruction Based on Zero-Phase Multi-Stage Collaborative Filtering
Previous Article in Special Issue
A Multi-Modal Light Sheet Microscope for High-Resolution 3D Tomographic Imaging with Enhanced Raman Scattering and Computational Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Precision Pose Measurement of Containers on the Transfer Platform of the Dual-Trolley Quayside Container Crane Based on Machine Vision

1
Logistics Engineering College, Shanghai Maritime University, Shanghai 201306, China
2
Institute of Logistics Science and Engineering, Shanghai Maritime University, Shanghai 201306, China
3
School of Technology and Architecture, ISCTE-Instituto Universitário de Lisboa, 1649-026 Lisbon, Portugal
4
Shanghai SMU Vision Co., Ltd., Shanghai 201306, China
5
Instituto de Telecomunicações, ISCTE-Instituto Universitário de Lisboa, 1649-026 Lisbon, Portugal
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(9), 2760; https://doi.org/10.3390/s25092760
Submission received: 27 March 2025 / Revised: 17 April 2025 / Accepted: 25 April 2025 / Published: 27 April 2025
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems)

Abstract

:
To address the high-precision measurement requirements for container pose on dual-trolley quayside crane-transfer platforms, this paper proposes a machine vision-based measurement method that resolves the challenges of multi-scale lockhole detection and precision demands caused by complex illumination and perspective deformation in port operational environments. A hardware system comprising fixed cameras and edge computing modules is established, integrated with an adaptive image-enhancement preprocessing algorithm to enhance feature robustness under complex illumination conditions. A multi-scale adaptive frequency object-detection framework is developed based on YOLO11, achieving improved detection accuracy for multi-scale lockhole keypoints in perspective-distortion scenarios (mAP@0.5 reaches 95.1%, 4.7% higher than baseline models) through dynamic balancing of high–low-frequency features and adaptive convolution kernel adjustments. An enhanced EPnP optimization algorithm incorporating lockhole coplanar constraints is proposed, establishing a 2D–3D coordinate transformation model that reduces pose-estimation errors to millimeter level (planar MAE-P = 0.024 m) and sub-angular level (MAE- θ = 0.11°). Experimental results demonstrate that the proposed method outperforms existing solutions in container pose-deviation-detection accuracy, efficiency, and stability, proving to be a feasible measurement approach.

1. Introduction

With the rapid development of automated container terminals, dual-trolley quayside container cranes, as core equipment for efficient loading/unloading operations, rely on their critical connecting component—the container-transfer platform—to facilitate container handover between the main trolley and gantry trolley. However, during container-unloading operations, mechanical clearance caused by long-term wear of container guide frames and spreader oscillations frequently result in actual container poses deviating from preset positions on the transfer platform. Figure 1a illustrates standard poses of a 40-foot general purpose container and dual 20-foot general purpose containers on the transfer platform; Figure 1b demonstrates horizontal deviation of containers; Figure 1c depicts rotational deviation; Figure 1d presents combined horizontal–rotational deviations. These deviation patterns necessitate frequent spreader pose adjustments during secondary grasping operations, which critically depend on perception systems’ precise calculation of 3D positions and deflection angles. The accuracy and speed of container pose estimation thus become crucial factors affecting operational efficiency. Enhancing the measurement accuracy and reliability of container positions and orientations on transfer platforms remains a significant research focus. Traditional manual measurement methods suffer from low efficiency and inconsistent precision, while existing vision-based systems face multiple challenges in complex port environments: dynamic lighting conditions (e.g., intense glare, rain/fog interference) causing image feature degradation and target-recognition failures; perspective-induced multi-scale deformations of container top-surface lockholes under fixed camera top-view configurations that hinder traditional image algorithms; and stringent robustness requirements for millimeter-level pose measurement.
This paper introduces a high-precision pose-measurement method for containers on the dual-trolley quayside container crane-transfer platform. The method employs a fixed camera-based pure vision detection approach, utilizing an adaptive image-enhancement preprocessing algorithm to flexibly adapt to complex illumination environments in ports. A multi-scale adaptive frequency object-detection framework based on YOLO11 is developed. The framework combines image characterization at different scales and dynamically adjusts the frequency domain information-processing strategy to more accurately identify and localize targets of different sizes in complex scenes. It can effectively address the issue of excessive scale variations in container lockhole keypoints caused by perspective transformation. Subsequently, the three-dimensional coordinates of container lockhole keypoints are calculated by combining the coplanar characteristics of container top lockholes with an improved EPnP optimization method, achieving millimeter-level horizontal deviation estimation and sub-angle-level deflection angle measurement. The proposed method provides a feasible solution for high-precision container pose measurement on dual-trolley quayside container crane-transfer platforms.
The contributions of this paper are summarized as follows:
  • To address the limitations of low efficiency and unstable accuracy in traditional manual operations for field engineering applications, a machine vision-based high-precision pose-measurement system for containers on the dual-trolley quayside container crane-transfer platform is proposed;
  • To mitigate interference from complex illumination and meteorological interference in port environments on operational site images, an adaptive image-enhancement preprocessing algorithm is designed to strengthen image features;
  • To resolve the challenge of large-scale variations in lockhole keypoints on container tops caused by perspective transformation in operational scenarios, a multi-scale adaptive frequency object-detection framework is developed based on the YOLO11 architecture, enabling robust target recognition and keypoint detection;
  • To overcome the low precision of traditional pose-estimation algorithms, an improved EPnP optimization method is proposed to achieve high-accuracy measurement of 3D container positions and orientations.

2. Related Work

Container pose-measurement technology is a core component of automated port operations. Early research primarily relied on manual measurement and sensor-assisted positioning techniques, such as LiDAR-based [1] or inertial navigation-based pose-estimation systems [2]. Although these methods achieve high accuracy, they suffer from high equipment costs, complexity of deployment, and susceptibility to mechanical vibration interference [3]. With the advancement of machine vision, monocular or binocular camera-based visual measurement methods have emerged as research hotspots. Kuo et al. [4] proposed a container damage-detection method based on machine vision, and Ji et al. [5] developed a vision-based truck-lifting accident-detection approach. However, these methods generally face dual constraints of calibration sensitivity and nonlinear computational complexity growth. Notably, monocular vision-based pose estimation has gained significant attention in the last five years due to its deployment flexibility and cost-effectiveness, yet challenges remain in overcoming dynamic illumination disturbances and target scale variations in actual port operations.
The complex optical conditions in port operational scenarios, including strong reflections and rain/fog blurring, pose significant challenges to image feature extraction. Traditional image-enhancement methods (e.g., histogram equalization) exhibit unstable feature performance under non-stationary illumination due to environmental adaptability limitations. Degradation-aware image enhancement is an intelligent enhancement technique that incorporates the analysis of image-degradation factors. Its core idea is to explicitly identify and model the types of degradation in the image (e.g., noise, blur, low resolution, etc.) when improving the image quality (e.g., denoising, deblurring, contrast enhancement, etc.), and adaptively adjust the enhancement strategy based on this information, so as to achieve more accurate and efficient restoration and enhancement. Researchers have proposed various degradation-aware image-enhancement strategies, such as Retinex theory-based dynamic contrast adjustment algorithms [6], which tend to introduce noise under extreme illumination. Recent deep learning-based feature-enhancement methods [7,8,9] demonstrate stronger robustness, exemplified by Li et al.’s [10] attention-guided residual blocks for real-time low-light image enhancement in smart ports. Lin et al. [11] achieved container number recognition with surface contamination and damage through deep learning and low-light enhancement algorithms. Nevertheless, existing methods still struggle with composite interference from large-scale variations and degradation coupling in port scenarios, requiring further optimization of feature fidelity and computational efficiency.
Under fixed camera configurations, container terminal operations frequently encounter drastic target scale variations. Traditional object-detection models like Faster R-CNN [12] face limitations in handling scale diversity due to fixed receptive field designs. YOLO-series algorithm [13] improvements enhance multi-scale detection capabilities through Feature Pyramid Networks [14], yet still suffer from missed detections under extreme scale differences. Zhou et al. [15] integrated SRCNN [16] and Resblock [17] to develop a contour feature-enhancement module, combining DConv [18] for cross-scale feature-enhancement networks. However, this approach incurs substantial computational overhead. Recent studies explore frequency-domain analysis with spatial feature extraction [19], enhancing detail features through high-frequency components while maintaining structural integrity via low-frequency components [20,21], offering new paradigms for multi-scale detection. However, their generalization capability in dynamic port scenarios requires further validation.
The core of 3D pose estimation lies in solving camera extrinsic parameters through 2D–3D point correspondences. Traditional PnP algorithms like DLT and EPnP minimize reprojection errors for pose estimation but tend to fall into local optima under noise interference. The EPnP algorithm reduces computational complexity through virtual control points [22]. Some studies integrate RANSAC mechanisms to enhance noise resistance [23], yet iterative processes compromise real-time performance. Moreover, existing methods rarely exploit practical constraints from real operational environments, leading to redundant degrees of freedom and limited accuracy in pose estimation.
Despite progress in visual detection and pose estimation, significant gaps remain between current technical capabilities and operational requirements for dual-trolley quayside container crane-transfer platforms. Existing image-enhancement methods lack joint modeling capabilities for illumination, meteorological interference, and geometric distortion when addressing dynamic degradation coupling in complex port environments. Traditional convolutional networks struggle with extreme scale variations of container top lockhole keypoints during port operations. The accuracy of the pose estimation remains constrained by detection precision and algorithmic limitations, leaving substantial room for improvement. Current dual-trolley quayside container crane-transfer platform operations still lack effective and reliable solutions for container position and orientation measurement.

3. Three-Dimensional Positioning and Pose-Measurement System

3.1. Hardware System

This paper proposes a hardware system based on visual measurement for the measurement of three-dimensional positioning and poses of containers on the transfer platform of a dual-trolley quayside container crane. The system consists of fixed cameras and edge computing modules. Fixed visual cameras are installed on the land side column of the quay crane above the transfer platform, as shown in Figure 2. These cameras are tilted downward to capture images, with their field of view fully covering the container-transfer platform to record the operational workflow of container placement.
The edge computing module processes visual data captured from containers on the transfer platform. In actual operations, the fixed cameras continuously capture container images and transmit them to the edge computing module. Through advanced image-processing algorithms and pose-detection algorithms, the module analyzes and processes these images to ultimately obtain accurate container position and attitude information.

3.2. Algorithm Design

The workflow of the 3D positioning and pose-measurement algorithm for transfer platform containers based on visual measurement proposed in this paper is illustrated in Figure 3. First, the video stream captured by fixed cameras is input frame-by-frame as raw images. Image preprocessing is performed using the proposed adaptive image-enhancement preprocessing algorithm, which dynamically adjusts enhancement strategies for complex illumination and meteorological interference in port environments to efficiently enhance image features.
After dynamic image feature enhancement, container recognition/classification, and lockhole keypoint detection are required. To address the large-scale variations caused by perspective transformations in containers and their top lockholes captured by fixed cameras on dual-trolley quayside container cranes, the proposed multiscale adaptive frequency object-detection method is implemented. Based on the YOLO11 framework, this method identifies 20-foot and 40-foot general purpose containers while obtaining two-dimensional image coordinates of lockhole keypoints.
Finally, using the output container-recognition information and 2D image coordinates of lockhole keypoints, the proposed 2D–3D keypoint coordinate-conversion algorithm calculates the three-dimensional coordinates of container lockhole keypoints. These 3D coordinates are then input into the pose analysis algorithm to ultimately determine the container’s offset direction/distance and rotation direction/angle relative to baseline positions on the transfer platform.

3.2.1. Adaptive Enhancement Image Feature Preprocessing Method

To address interference issues caused by high-contrast complex illumination and rain fog blur in complex environments at container terminals that affect container images on transfer platforms, we designed an adaptive enhancement image feature preprocessing algorithm. This algorithm serves as a preprocessing component before image-detection algorithms to mitigate the impacts of complex illumination and meteorological interference. As illustrated in Figure 4, inspired by Zhang et al. [24], we propose a novel method leveraging the characteristics of the Chain-of-Thought Prompt Generation Module (CGM) and Content-Driven Prompt Block (CPB) to enhance port container images under degradation conditions such as complex illumination and environmental challenges.
CGM Module: The Chain-of-Thought prompting mechanism constructs a multilevel degradation-aware semantic encoding framework, as illustrated in Figure 5. Its core employs a transposed convolution sequence to generate resolution-increasing prompts, progressively refining from low-resolution global semantics (Such as rain blur, fog blur, dark light attenuation and other types of degradation categories) to high-resolution local features (Degradation intensity distribution for rain and fog blurring, dark light weakening, etc.). Initial prompts undergo multi-stage upsampling and channel compression, with information filtering achieved through the Hardswish activation function [25]. This establishes inter-level dependencies. The design breaks through the static limitations of traditional independent prompts, guiding the model through chain-of-thought reasoning to parse degradation patterns from coarse to fine. Driven by the training data, CGM automatically learns the features of different degradation types such as rain blur and dark light weakening. Combined with the multi-scale features of the decoder, it significantly enhances adaptive capability for degradation types.
Specifically, the construction of this module primarily consists of following steps. First, an initial prompt is constructed by initializing a tensor of a learnable prompt P 3 R H ^ × W ^ × C ^ in the third layer of the decoder, where H ^ × W ^ denotes the initial spatial resolution and C ^ represents the dimension of the channel. This prompt learns global representations of degradation patterns through backpropagation. Subsequently, multiscale prompt sequences are progressively generated via stacked transposed convolutions—a 3 × 3 transposed convolution operation upsamples P 3 to produce the second-layer prompt P 2 R 2 H ^ × 2 W ^ × C ^ / 2 , following which the first-layer prompt P 1 R 4 H ^ × 4 W ^ × C ^ / 4 is generated similarly. After each transposed convolution layer, the Hardswish activation function is used to suppress irrelevant information flow, selectively propagating degradation-related characteristics as formulated in Equation (1).
P i = Hardswish ( TC 3 × 3 ( P i + 1 ) ) , i { 1 , 2 }
where TC 3 × 3 represents the 3 × 3 transposed convolution operation. The CGM module establishes hierarchical dependency relationships between prompts, enabling coarse-to-fine progressive reasoning of degradation patterns through multiscale feature interactions.
CPB Module: Achieves degradation-aware feature enhancement through dual-path hybrid attention and parallel Transformers as illustrated in Figure 6. First, the channel-spatial attention jointly models feature importance to generate content-sensitive weight distributions. Subsequently, the prompt information undergoes interpolation alignment and concatenation with features. The fused features are then fed into multiple parallel Transformer sub-blocks via a channel-splitting strategy, where cross-channel attention is computed individually and gating mechanisms control information flow. Prompts generated by the CGM dynamically adjust the enhancement strategy by interacting with image features at each layer of the decoder through the CPB module. At the shallow decoder, the prompt may guide the removal of rain noise (e.g., suppressing rain line artifacts). Whereas at the deep decoder, prompt may enhance object-critical features (e.g., edges of container lock holes) to improve detection robustness. This design reduces computational complexity through a divide-and-conquer principle while achieving adaptive degradation context enhancement via fine-grained feature-prompt interactions.
The specific implementation process of the CPB Module is as follows. Firstly, generate channel attention weights W i c R 1 × 1 × C i and spatial attention weights W i s R H i × W i × C i . to capture the key information of the input features, the calculation equation for W i c is shown in Equation (2), and the calculation equation for W i s is shown in Equation (3).
W i c = C 1 × 1 ReLU C 1 × 1 GAP c ( F i )
W i s = C 7 × 7 [ GAP s ( F i ) , GMP s ( F i ) ]
where C k × k denotes k × k convolution. ReLU represents the ReLU activation function. GAP c is the global average pooling operation across the spatial dimensions. GMP s is the global max pooling operation across channel attention. · denotes the channel-wise concatenation operation. Subsequently, the attention weights are fused with the input features, as shown in Equations (4)–(6).
F i w = [ ( W i c W i s ) F i , F i ]
F i s = σ DC 7 × 7 CS ( F i w )
F i p = C 1 × 1 [ F i , Rescale ( P i ) F i s ]
where ⊙ and ⊕, respectively, represent element-wise multiplication and element-wise addition. DC k × k is the depthwise separable convolution with a stride of k × k . CS denotes the channel shuffle operation.
F i p is split into n blocks along the channel dimension in Equation (7).
F p i , j = F i p : , : , ( j 1 ) C i n : j C i n , j { 1 , 2 , , n }
Each sub-block is input into an independent Transformer Block. Finally, the enhanced feature is obtained by concatenating the results of all sub-blocks:
F i g = [ F g i , 1 , , F g i , j , , F g i , n ] , j { 1 , 2 , , n }
In the design of the enhancement module, we employ LDConv [26] (Linear Deformable Convolution) to generate convolution kernels of arbitrary sizes and diverse initial sampling positions. By adaptively adjusting sampling points through offsets, the convolution operation can better accommodate shape variations of the targets. This flexibility enables the model to efficiently extract critical information and enhance feature representation capabilities when processing multi-scale image data.
The approach combines hierarchical reasoning and dynamic coupled synergetic architecture construction. It can effectively deal with the multiple challenges of complex illumination, meteorological interference and geometric distortion. This approach significantly improves both semantic adaptability and structural fidelity in container image enhancement for transfer platforms.

3.2.2. Multi-Scale Adaptive Frequency Object Recognition and Keypoint-Detection Method

Traditional object-detection methods exhibit poor performance in the complex environments of container terminals, particularly in scenarios with significant scale variations of container top lockhole targets caused by perspective transformation effects from fixed cameras during container operations at transfer platforms. To address this issue, our recognition framework improves upon the conventional YOLO11 and proposes a multi-scale object-detection and keypoint-detection method, as illustrated in Figure 7.
The original bottleneck module employs fixed-size convolution kernels, leading to insufficient information capture or excessive smoothing when processing features at different scales, thereby failing to meet the demands of diversified feature representation. Simultaneously, traditional convolution operations extract features solely in the spatial domain, exhibiting limited capacity to capture fine details under large-scale structures. Although stacking multiple layers can expand the receptive field, the fixed kernel size inherently restricts the local feature extraction capacity of single-layer convolution operations. This limitation becomes particularly pronounced in high-level semantic feature extraction or multi-scale tasks, where long-range dependency information in input data remains underutilized. To address this, we introduce Frequency-Adaptive Dilated Convolution (FADConv) into the C3k2 module for the first time, with its network architecture illustrated in Figure 8.
Through a frequency-adaptive mechanism, the convolutional weights are decomposed into high-frequency components and low-frequency components. As shown in Figure 9, by dynamically adjusting the convolution dilation rate and convolution kernel weight, the model can adaptively adjust its receptive field according to different local frequency characteristics, thereby enhancing its ability to capture high-frequency detail information. Additionally, the model effectively expands the receptive field of convolutional layers by balancing high- and low-frequency components. This enables the model to perform more stably when processing inputs containing both high-frequency details (e.g., textures of small-scale objects) and low-frequency structures (e.g., contours of large-scale objects), particularly in multi-scale adaptive frequency object-detection scenarios.
Specifically, this method balances Effective Bandwidth and Receptive Field through Adaptive Dilation Rate (AdaDR). The spatial dynamic dilation rate is shown in Equation (9).
Y ( p ) = i = 1 K × K W i X p + Δ p i × D ^ ( p )
where D ^ ( p ) is the dynamic dilation rate at position P, predicted by a lightweight convolution. We first transform the feature map X R H × W into the frequency domain using the Discrete Fourier Transform (DFT), it can be represented as:
X F ( u , v ) = 1 H W h = 0 H 1 w = 0 W 1 X ( h , w ) e 2 π j ( u h + v w )
High-Frequency Power is defined as:
H P ( p ) = ( u , v ) H D ^ ( p ) + | X F ( p , s ) ( u , v ) | 2
where H D ^ ( p ) + = { ( u , v ) | u | > 1 2 D ^ ( p ) or | v | > 1 2 D ^ ( p ) } represents the high-frequency region that cannot be captured by the dilation rate D ^ ( p ) . Then dynamically adjust the proportion of low-frequency and high-frequency components of the convolution kernel to improve the effective bandwidth.
W ¯ = 1 K × K i = 1 K × K W i , W ^ = W W ¯
where W ¯ represents the kernel-wise averaged W, W ^ captures high-frequency details. Through global pooling and convolutional layers to generate channel-level dynamic weights λ l and λ h , reconstructing the adaptive convolutional kernel:
W = λ l W ¯ + λ h W ^
Finally, the feature spectrum is balanced through frequency band decomposition and reweighting to expand the receptive field.
X b = F 1 ( M b X F )
M b is the binary mask that extracts the frequency band [ ϕ b , ϕ b + 1 ) . A spatially variant attention weight A b R H × W is applied to each frequency band X b :
X ^ ( i , j ) = b = 0 B 1 A b ( i , j ) X b ( i , j )
This method can suppress high-frequency components in background regions, encouraging FADConv to select larger dilation rates to expand the receptive field.
Based on the fundamental framework of YOLO11, combined with the multi-scale adaptive frequency object-detection method, the preprocessed images undergo precise recognition to obtain identification results for 20-foot and 40-foot general purpose containers. Simultaneously, a keypoint-detection branch is added through an inherited detection head to accurately detect the lockhole keypoints of different containers in the image and acquire their two-dimensional image coordinates. The final recognition and detection results are illustrated in Figure 10.

3.2.3. Three-Dimensional Position and Pose-Measurement Method for Containers

After obtaining the container-identification results and the 2D image coordinates of the lockhole keypoints, it is necessary to calculate the position and pose of 20-foot and 40-foot general purpose containers separately. This allows determining the offset and rotation angle of the container relative to the standard position on the transfer platform. This paper proposes a three-dimensional position and pose-measurement method for containers, which consists of two parts: (1) 2D–3D Lockhole Keypoint Coordinate Conversion; (2) Container Pose Calculation.
  • Step 1: 2D–3D Lockhole Keypoint Coordinate Conversion.
Assuming the coordinates of the container lockhole in the 3D world coordinate system are denoted as P w = [ X , Y , Z , 1 ] T , and its projected coordinates in the image coordinate system are p = [ u , v , 1 ] T , the mapping relationship between them can be expressed as Equation (16) according to the fixed camera model.
s p = K [ R t ] P w
Among them, s is the scale factor, K is the intrinsic matrix of the camera, R R 3 × 3 is the rotation matrix, and t R 3 is the translation vector. The intrinsic matrix is defined by Equation (17).
K = f x 0 c x 0 f y c y 0 0 1
In the formula, f x , f y represents the focal length parameter, and ( c x , c y ) denotes the principal point coordinates.
Given n sets of 3D-2D correspondences { P w ( i ) , p ( i ) } i = 1 n , this work employs the EPnP algorithm (Efficient PnP) to solve the extrinsic parameters R and t . The core idea involves representing 3D points as weighted combinations of four virtual control points { C j } j = 1 4 , as shown in Equation (18).
P w ( i ) = j = 1 4 α i j C j , j = 1 4 α i j = 1
By substituting Equation (18) into the projection Equation (16), Equation (19) can be derived.
s i p ( i ) = K j = 1 4 α i j ( RC j + t )
Through algebraic elimination of the scale factor s i , a linear system of equations concerning control point coordinates is constructed. After solving this system using Singular Value Decomposition (SVD), the rotation matrix R and translation vector t are recovered via Orthogonal Procrustes Analysis. The specific steps are as follows:
  • Control Point Initialization: Select four non-coplanar control points, typically choosing the centroid of the 3D point set and principal component directions;
  • Weight Coefficient Calculation: Solve for α i j using the least squares method to minimize the residual error in Equation (18);
  • Camera Coordinate System Control Point Solution: Construct an overdetermined system of equations using Equation (19) and solve it via SVD;
  • Extrinsic Parameter Recovery: Align control points in the world coordinate system with those in the camera coordinate system, minimizing registration errors as shown in Equation (20).
    arg min R , t j = 1 4 C j c a m ( RC j w o r l d + t ) 2
The containers on the transfer platform are positioned horizontally, with their upper surface lockholes also located on the same horizontal plane ( Z = h ) . This plane constraint enhances the stability of solver computations. In this configuration, only four pairs of initial point correspondences between 2D image coordinates and 3D world coordinates are required to obtain a unique solution for the solver.
Given the 2D coordinates p k = [ u k , v k ] T of the lockholes in the image and their height constraint Z k = h , the 3D coordinates A are computed through back projection calculation. Equation (16) is expanded into Equation (21).
s u k = f x ( r 11 X k + r 12 Y k + r 13 h + t x ) + c x ( r 31 X k + r 32 Y k + r 33 h + t z ) s v k = f y ( r 21 X k + r 22 Y k + r 23 h + t y ) + c y ( r 31 X k + r 32 Y k + r 33 h + t z ) s = r 31 X k + r 32 Y k + r 33 h + t z
By eliminating the scale factor s , we obtain the linear equation system Equation (22) concerning X k , Y k .
f x r 11 + ( c x u k ) r 31 f x r 12 + ( c x u k ) r 32 f y r 21 + ( c y v k ) r 31 f y r 22 + ( c y v k ) r 32 X k Y k = u k ( r 33 h + t z ) f x ( r 13 h + t x ) c x ( r 33 h + t z ) v k ( r 33 h + t z ) f y ( r 23 h + t y ) c y ( r 33 h + t z )
Solving Equation (22) uniquely determines the three-dimensional coordinates ( X k , Y k , h ) of the lockhole.
  • Step 2: Container Pose Calculation.
Through the acquired three-dimensional coordinates of lockholes on the container’s top surface, we analyze the container’s pose to calculate its position deviation and deflection angle on the transfer platform. Given the three-dimensional coordinates { P k = ( X k , Y k , h ) } k = 1 4 of four lockholes on the container’s top surface, their geometric center coordinate C = ( C x , C y , h ) can be calculated using the spatial point set centroid formula, as shown in Equation (23).
C x = 1 4 k = 1 4 X k , C y = 1 4 k = 1 4 Y k
For the estimation of the container’s principal direction vectors, the long-axis direction vector v L and short-axis direction vector v W are defined on the container’s top surface. By selecting long-edge container lockhole pairs, two sets of vectors are calculated as shown in Equation (24).
v L 1 = P 3 P 1 , v L 2 = P 4 P 2
Since each container lockhole keypoint contains unique numbering, the vector direction will not be reversed due to incorrect lockhole numbering sequence, eliminating the need for direction consistency correction. A weighted average of the two vector sets yields the final long-axis direction estimation as demonstrated in Equation (25).
v L = w 1 v L 1 + w 2 v L 2 w 1 + w 2
In this scenario, the weight allocation strategy adopts equal-weight averaging, w 1 = w 2 = 1 .
The resolved container position and pose information are calculated with the standard position center point C 0 = ( C x 0 , C y 0 , h ) and standard direction vector v L 0 = ( cos θ 0 , sin θ 0 , 0 ) (preset angle: θ 0 ) to obtain the positional offset and directional deviation.
The container positional deviation Δ C is shown in Equation (26).
Δ C x = C x C x 0 , Δ C y = C y C y 0
By integrating dot product and cross product information, the four-quadrant arctangent function arctan 2 ( y , x ) is employed to calculate the signed deviation angle as shown in Equation (27).
Δ θ = arctan 2 v L × v L 0 · e z , v L · v L 0
After expansion, it is equivalent to Equation (28).
Δ θ = arctan 2 v L x sin θ 0 v L y cos θ 0 , v L x cos θ 0 + v L y sin θ 0
This method ensures that Δ θ ( π , π ] precisely reflects the rotational direction and magnitude of container deviation. The final solution yields the container position deviation Δ C and the rotational direction and magnitude Δ θ of the container.

4. Experiments

4.1. Experimental Setup

To validate the effectiveness of the machine vision-based pose-measurement algorithm for containers on the transfer platform of dual-trolley quayside container cranes, a series of related experiments were conducted.

4.1.1. Experimental Environment

The training environment parameters for these experiments are shown in Table 1.
The camera used in this study is a fixed camera with adjusted and fixed shooting angles and focal lengths. This fixed camera has a pixel resolution of 1920 × 1080 and an fps of 30. The actual installation position of the fixed camera is shown in Figure 11. The red rectangle indicates the transfer platform of the dual-trolley quayside container crane, while the green rectangle marks the actual installation location of the fixed camera.

4.1.2. Datasets and Evaluation Metrics

The container samples from the transfer platform were collected using fixed cameras, comprising a dataset of 4250 images. These include 20-foot general purpose containers and 40-foot general purpose containers, with samples captured under complex illumination and diverse port environmental conditions. The sample contains 827 rainy day images, 1273 sunny day images, 516 foggy day images, and 302 low light images, as illustrated in Figure 12. The richness of image samples will continue to increase with subsequent engineering deployments.
The measurement system proposed in this study primarily depends on three aspects: container-recognition and lockhole keypoint-detection accuracy, container-offset-detection accuracy, and container rotation angle-detection accuracy. To evaluate these, three core metrics are designed: detection model inference accuracy, container-offset-detection accuracy, and container rotation angle-detection accuracy.
(a) Detection Model Inference Accuracy
The evaluation metrics include precision, recall, and mean Average Precision (mAP). Precision refers to the ratio of true positive predictions among all positive predictions in the test data, defined as Equation (29).
Precision = TP TP + FP
In this context, TP (True Positive) represents the number of true positive instances, while FP (False Positive) denotes the number of false positive instances.
Recall measures the proportion of correctly identified positive class samples relative to all actual positive class samples. This is expressed in Equation (30).
Recall = TP TP + FN
where TP (True Positive) indicates the number of correctly predicted positive class samples, and FN (False Negative) represents the number of samples that are actually positive but were erroneously predicted as negative.
mAP (mean Average Precision) is the average of the Average Precision (AP) values computed for each detection category. This experimental section employs two evaluation metrics: mAP@0.5 and mAP@0.5:0.95. Specifically, mAP@0.5 is defined as the mean average precision calculated with an Intersection over Union (IoU) threshold of 0.5, where a detection is considered correct if the IoU between the predicted bounding box and the ground truth bounding box is ≥0.5. To comprehensively evaluate model performance under varying localization accuracy requirements, particularly focusing on detection capability at high IoU thresholds, this study additionally calculates the average mAP across 10 distinct IoU thresholds ranging from 0.5 to 0.95 with a step size of 0.05.
(b) The detection accuracy of container horizontal deviation
In the experimental evaluation, to quantify the translation-detection accuracy of containers along coordinate axes within the planar coordinate system, this study conducts independent error analyses on the offset deviations along the X-axis and Y-axis within the horizontal plane. For the offset deviations Δ C x in the X-axis direction and Δ C y in the Y-axis direction, the evaluation metrics are defined as the deviations from manually measured ground truth values Δ C x 0 and Δ C y 0 .
Single-axis Mean Absolute Deviation (MAD): To characterize systematic errors in single-axis offset deviations, this metric calculates the mean absolute deviation across all samples, as shown in Equation (31).
MAD X = 1 N i = 1 N Δ C x ( i ) Δ C x 0 ( i ) , MAD Y = 1 N i = 1 N Δ C y ( i ) Δ C y 0 ( i )
Mean Absolute Error in the Plane (MAE-P): To further evaluate the overall translation accuracy within the horizontal plane (XY plane), we define the mean value of two-dimensional projection absolute errors, as shown in Equation (32).
MAE P = 1 N i = 1 N Δ C x ( i ) Δ C x 0 ( i ) 2 + Δ C y ( i ) Δ C y 0 ( i ) 2
(c) The detection accuracy of container rotational deviation
In experimental evaluations, for the calculated container rotation angle Δ θ and manually measured rotation angle Δ θ 0 , we employ the Mean Absolute Error of Rotation Angle (MAE- θ ) to establish an angular error evaluation system, shown in Equation (33).
MAE θ = 1 N i = 1 N Δ θ ( i ) Δ θ 0 ( i )

4.2. Experimental Results

4.2.1. Model Accuracy Testing

To validate the effectiveness of the adaptive enhanced image feature preprocessing method and multi-scale adaptive frequency object detection with keypoint detection for container recognition and lockhole detection on container roofs, comparative experiments were conducted using traditional algorithms, the original YOLO11 algorithm, and our improved algorithm. The visual comparison results are shown in Figure 13.
The comparison of training results between our proposed adaptive enhanced image feature preprocessing method and the multis-cale adaptive frequency object-detection algorithm versus the original YOLO11 network is illustrated in Figure 14. The blue line represents our method, while the red line denotes the original YOLO11 network. The horizontal axis indicates the number of training epochs.
From the training comparison graph, we observe that both methods exhibit rapid increases in precision curves at similar rates during the initial training phase. However, our method demonstrates significantly smaller oscillation amplitudes compared to the YOLO11 model, reflecting stronger robustness against noisy samples in the optimization process. Concurrently, our method achieves faster convergence speed than YOLO11, with higher accuracy and reduced oscillations between 200–400 epochs. In the recall curve, our model shows superior capability in recalling true targets compared to YOLO11, with smoother convergence and smaller late-stage oscillations, indicating enhanced localization ability for occluded and multi-scale targets. The mAP50 curve reveals that our method stabilizes after 200 epochs while significantly outperforming YOLO11. The mAP50-95 curve further confirms the consistent superiority of our proposed model. These results demonstrate substantial improvements in target-recognition accuracy, localization precision, and overall algorithm performance. The experimental results are summarized in Table 2.
As can be seen from the table, our method demonstrates significant advantages over traditional approaches in terms of precision (P), recall (R), mean Average Precision (mAP@0.5), and mAP@0.5:0.95. The reason for this construction is mainly due to the fact that the traditional SIFT algorithm is based on localized features and is sensitive to changes in lighting, viewing angle, etc., which can easily lead to matching failures in this scenario. Our approach is also significantly improved compared to YOLO11. Specifically, the precision is improved by 3.7% and recall by 4.2% compared to YOLO11, indicating dual optimization of the model’s capability in handling large-scale variation target classification and target-localization completeness under complex illumination and challenging port environments. Concurrently, the mAP@0.5 shows a 4.7% enhancement and mAP@0.5:0.95 an 8.7% improvement, reflecting substantial progress in basic detection capabilities and enhanced multi-scale target-recognition and -localization abilities. Compared with the multi-scale target-detection algorithm HRNet, all evaluation metrics are improved but not significantly. However, the larger number of parameters in HRNet requires more computational resources, and the lighter YOLO algorithm is more suitable for deployment in port operation environments where computational resources are limited. These results validate the effectiveness and superior recognition accuracy of the proposed vision-based container-identification and lockhole keypoint-localization methodology. At the same time, we noticed a complete loss of distinguishable structure of the lock hole in the image under extreme low light and extreme high light reflections. There are some limitations of our algorithm in that case.

4.2.2. The Detection Accuracy of Container Horizontal Deviation

To verify the precision of container positioning calculation, this paper compares the proposed three-dimensional position-measurement method with traditional manual measurements through a complete quayside container crane operation cycle. Independent error analyses are conducted for horizontal deviations along both the X-axis and Y-axis directions. Figure 15 is the schematic diagram of horizontal deviation.
The method is scientifically evaluated through Mean Absolute Deviation (MAD) for single-axis analysis and Mean Absolute Error in the Plane (MAE-P), providing comprehensive assessment from both individual axes and integrated planar perspectives.
The comparison results between the method proposed in this paper and manual operations in Table 3 demonstrate that the proposed method achieves a precision error of 0.012 m on the x-axis under the Mean Absolute Deviation (MAD), outperforming the manual operation’s precision error of 0.013 m. However, its y-axis precision error of 0.018 m is slightly higher than the manual operation’s 0.016 m. In terms of the Mean Absolute Error in the Plane (MAE-P), the proposed method achieves a precision error of 0.024 m, reaching the level of manual operation (0.023 m). Additionally, the proposed method reduces the average operation time by 0.68 s compared to manual operations.

4.2.3. The Detection Accuracy of Container Rotational Deviation

To verify the accuracy of container deflection angle calculation, this paper conducted comparative experiments between the proposed three-dimensional container pose-measurement method and traditional manual measurement results through a quayside crane operation cycle. The comparison was established using the Mean Absolute Error of Rotation Angle (MAE- θ ) as the angular error evaluation metric. Figure 16 is the schematic diagram of deflection angles.
From the comparative experimental results between the proposed method and manual operations in Table 4, it can be observed that the proposed method achieves a MAE- θ of 0.11°, outperforming the manual operation’s average error of 0.15°. Additionally, it demonstrates superior average operation time by 1.15 s compared to manual operations. The smaller MAE- θ indicates higher alignment precision, while manual operations tend to rely on spreader guide plates sliding into containers. Therefore, the proposed method not only saves operation time but also potentially reduces wear between spreader guide plates and containers to some extent.

5. Conclusions

As the core equipment for efficient container handling operations, the dual-trolley quayside container crane relies on its critical connecting component—the container-transfer platform—to perform essential container-transfer functions between the main trolley and gantry trolley. To address the technical challenge of container pose measurement on the transfer platform, this study proposes a high-precision vision-based measurement system. The hardware system integrates fixed cameras with edge computing modules. An adaptive image-enhancement preprocessing algorithm enhances image feature robustness under complex illumination conditions. A multi-scale adaptive frequency object-detection framework is developed based on YOLO11, achieving significant improvement in multi-scale lockhole keypoint-detection accuracy through dynamic balance of high–low frequency features and adjustable deformable convolution kernels in perspective-distortion scenarios. An improved EPnP optimization algorithm incorporating lockhole coplanarity constraints establishes a 2D–3D coordinate transformation model, reducing pose solution errors to millimeter-level positional accuracy and sub-degree angular precision. Experimental validation confirms the effectiveness of this algorithm for container pose measurement on dual-trolley quayside container crane-transfer platforms. This method provides automated ports with an efficient and cost-effective solution for container pose measurement, effectively reducing spreader adjustment time and mechanical wear while enhancing operational efficiency and safety of dual-trolley quayside container cranes. The demonstrated practicality and effectiveness highlight its engineering application value.

Author Contributions

Conceptualization, J.W. and M.H.; methodology, J.W. and M.H.; Software, Y.Z., Z.Z. and O.P.; Validation, M.H. and Y.Z.; Formal analysis, J.W. and C.M.; Investigation, M.H. and Y.Z.; Resources, C.M., O.P. and Y.Z.; Data curation, J.W. and Y.Z.; Writing—original draft preparation, J.W., M.H. and Y.Z.; Writing—review and editing, J.W. and O.P.; Visualization, J.W. and C.M.; Supervision, J.W., M.H. and Z.Z.; Project administration, J.W., Z.Z. and O.P.; Funding acquisition, Y.Z., O.P. and C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by National Natural Science Foundation of China (No. 52472435), the Education Science Research Project of Shanghai Municipality (No. B2023003), the Science and Technology Commission of Shanghai Municipality (No. 22ZR1427700), the Fundação para a Ciência e Tecnologia (UIDB/50008/2020). It is also funded by Iscte-Instituto Universitario de Lisboa, in the form of a merit scholarship awarded to the 3rd author.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Authors Zhiwei Zhang and Chao Mi were employed by the company Shanghai SMU Vision Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Shen, Y.; Man, X.; Wang, J.; Zhang, Y.; Mi, C. Truck Lifting Accident Detection Method Based on Improved PointNet++ for Container Terminals. J. Mar. Sci. Eng. 2025, 13, 256. [Google Scholar] [CrossRef]
  2. Jiang, Z. Frame Definition, Pose Initialization, and Real-Time Localization in a Non-Stationary Reference Frame with LiDAR and IMU: Application in Cooperative Transloading Tasks. Ph.D. Thesis, University of California, Los Angeles, CA, USA, 2022. [Google Scholar]
  3. Shen, Z.; Wang, J.; Zhang, Y.; Zheng, L.; Mi, C.; Shen, Y. Identification and Positioning Method of Bulk Cargo Terminal Unloading Hopper Based on Monocular Vision Three-Dimensional Measurement. J. Mar. Sci. Eng. 2024, 12, 1282. [Google Scholar] [CrossRef]
  4. Kuo, S.Y.; Lin, P.C.; Huang, X.R.; Huang, G.Z.; Chen, L.B. Cad-Transformer: A CNN–Transformer Hybrid Framework for Automatic Appearance Defect Classification of Shipping Containers. IEEE Trans. Instrum. Meas. 2025, 74, 1–21. [Google Scholar] [CrossRef]
  5. Ji, Z.; Zhao, K.; Liu, Z.; Hu, H.; Sun, Z.; Lian, S. A Novel Vision-Based Truck-Lifting Accident Detection Method for Truck-Lifting Prevention System in Container Terminal. IEEE Access 2024, 12, 42401–42410. [Google Scholar] [CrossRef]
  6. Xu, H.; Zhang, H.; Yi, X.; Ma, J. CRetinex: A progressive color-shift aware Retinex model for low-light image enhancement. Int. J. Comput. Vis. 2024, 132, 3610–3632. [Google Scholar] [CrossRef]
  7. Liu, Q.; Liu, M.; Jonathan, Q.; Shen, W. A real-time anchor-free defect detector with global and local feature enhancement for surface defect detection. Expert Syst. Appl. 2024, 246, 123199. [Google Scholar] [CrossRef]
  8. Wang, Z.; Guo, J.; Zhang, S.; Xu, N. Marine object detection in forward-looking sonar images via semantic-spatial feature enhancement. Front. Mar. Sci. 2025, 12, 1539210. [Google Scholar] [CrossRef]
  9. Nou, S.; Lee, J.S.; Ohyama, N.; Obi, T. Human pose feature enhancement for human anomaly detection and tracking. Int. J. Inf. Technol. 2025, 17, 1311–1320. [Google Scholar] [CrossRef]
  10. Li, L.; Xu, W.; Gao, Y.; Lu, Y.; Yang, D.; Liu, R.W.; Zhang, R. Attention-oriented residual block for real-time low-light image enhancement in smart ports. Comput. Electr. Eng. 2024, 120, 109634. [Google Scholar] [CrossRef]
  11. Lin, Z.; Dong, C.; Wan, Y. Research on Intelligent Recognition Algorithm of Container Numbers in Ports Based on Deep Learning. In Proceedings of the International Conference on Intelligent Computing, Tianjin, China, 5–8 August 2024; pp. 184–196. [Google Scholar]
  12. Bandong, S.; Nazaruddin, Y.Y.; Joelianto, E. Container detection system using CNN based object detectors. In Proceedings of the 2021 International Conference on Instrumentation, Control, and Automation (ICA), Bandung, Indonesia, 25–27 August 2021; pp. 106–111. [Google Scholar]
  13. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo algorithm developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  14. Wu, G.; Wu, Q. Enhancing Steel Surface Defect Detection: A Hyper-YOLO Approach with Ghost Modules and Hyper FPN. IAENG Int. J. Comput. Sci. 2024, 9, 1321–1330. [Google Scholar]
  15. Zhou, R.; Gu, M.; Hong, Z.; Pan, H.; Zhang, Y.; Han, Y.; Wang, J.; Yang, S. SIDE-YOLO: A Highly Adaptable Deep Learning Model for Ship Detection and Recognition in Multisource Remote Sensing Imagery. IEEE Geosci. Remote Sens. Lett. 2025, 22, 1–5. [Google Scholar] [CrossRef]
  16. Xiao, Y.; Li, Z.; Zhang, Z. Enhancing Image Perception Quality: Exploring Loss Function Variants in SRCNN. In Proceedings of the 2024 7th International Conference on Computer Information Science and Artificial Intelligence, Shaoxing, China, 13–15 September 2024; pp. 352–356. [Google Scholar]
  17. Qian, L.; Zheng, Y.; Cao, J.; Ma, Y.; Zhang, Y.; Liu, X. Lightweight ship target detection algorithm based on improved YOLOv5s. J. Real-Time Image Process. 2024, 21, 3. [Google Scholar] [CrossRef]
  18. Cao, J.; Li, Y.; Sun, M.; Chen, Y.; Lischinski, D.; Cohen-Or, D.; Chen, B.; Tu, C. Do-conv: Depthwise over-parameterized convolutional layer. IEEE Trans. Image Process. 2022, 31, 3726–3736. [Google Scholar] [CrossRef] [PubMed]
  19. Ren, C.; Wang, A.; Yang, C.; Wu, J.; Wang, M. Frequency Domain-Based Cross-Layer Feature Aggregation Network for Camouflaged Object Detection. IEEE Signal Process. Lett. 2025. [Google Scholar] [CrossRef]
  20. Weng, W.; Wei, M.; Ren, J.; Shen, F. Enhancing Aerial Object Detection with Selective Frequency Interaction Network. IEEE Trans. Artif. Intell. 2024, 5, 6109–6120. [Google Scholar] [CrossRef]
  21. Chen, L.; Gu, L.; Zheng, D.; Fu, Y. Frequency-adaptive dilated convolution for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 22–24 June 2024; pp. 3414–3425. [Google Scholar]
  22. Wang, B.; Bai, A.; Ma, F.; Ji, P. A 3D object localization method based on EPNP and dual-view images. In Proceedings of the 2024 IEEE 7th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 15–17 March 2024; Volume 7, pp. 1349–1353. [Google Scholar]
  23. Xie, X.; Zou, D. Depth-Based Efficient PnP: A Rapid and Accurate Method for Camera Pose Estimation. IEEE Robot. Autom. Lett. 2024, 9, 9287–9294. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Wu, Y.; Liu, Y.; Peng, X. CPA-Enhancer: Chain-of-Thought Prompted Adaptive Enhancer for Downstream Vision Tasks Under Unknown Degradations. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Suzhou, China, 23–25 May 2025; pp. 1–5. [Google Scholar]
  25. Alhassan, A.M.; Zainon, W.M.N.W. Brain tumor classification in magnetic resonance image using hard swish-based RELU activation function-convolutional neural network. Neural Comput. Appl. 2021, 33, 9075–9087. [Google Scholar] [CrossRef]
  26. Zhang, X.; Song, Y.; Song, T.; Yang, D.; Ye, Y.; Zhou, J.; Zhang, L. LDConv: Linear deformable convolution for improving convolutional neural networks. Image Vis. Comput. 2024, 149, 105190. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of horizontal and rotational deviations of containers on transfer platform.
Figure 1. Schematic diagram of horizontal and rotational deviations of containers on transfer platform.
Sensors 25 02760 g001
Figure 2. Fixed camera installation location schematic diagram.
Figure 2. Fixed camera installation location schematic diagram.
Sensors 25 02760 g002
Figure 3. Algorithm flowchart.
Figure 3. Algorithm flowchart.
Sensors 25 02760 g003
Figure 4. Network structure diagram of adaptive image-enhancement preprocessing method.
Figure 4. Network structure diagram of adaptive image-enhancement preprocessing method.
Sensors 25 02760 g004
Figure 5. CGM module network architecture diagram.
Figure 5. CGM module network architecture diagram.
Sensors 25 02760 g005
Figure 6. CPB module network structure diagram.
Figure 6. CPB module network structure diagram.
Sensors 25 02760 g006
Figure 7. Network architecture diagram of multi-scale adaptive frequency object recognition and keypoint detection.
Figure 7. Network architecture diagram of multi-scale adaptive frequency object recognition and keypoint detection.
Sensors 25 02760 g007
Figure 8. C3k2_FADC network architecture.
Figure 8. C3k2_FADC network architecture.
Sensors 25 02760 g008
Figure 9. Schematic diagram of high–low frequency features in container images.
Figure 9. Schematic diagram of high–low frequency features in container images.
Sensors 25 02760 g009
Figure 10. Detection results.
Figure 10. Detection results.
Sensors 25 02760 g010
Figure 11. Installation position of fixed camera.
Figure 11. Installation position of fixed camera.
Sensors 25 02760 g011
Figure 12. Schematic diagram of 20-foot and 40-foot general purpose container samples.
Figure 12. Schematic diagram of 20-foot and 40-foot general purpose container samples.
Sensors 25 02760 g012
Figure 13. Algorithm comparison results.
Figure 13. Algorithm comparison results.
Sensors 25 02760 g013
Figure 14. Training results comparison between our method and YOLO11 algorithm.
Figure 14. Training results comparison between our method and YOLO11 algorithm.
Sensors 25 02760 g014
Figure 15. Schematic diagram of horizontal deviation.
Figure 15. Schematic diagram of horizontal deviation.
Sensors 25 02760 g015
Figure 16. Schematic diagram of deflection angles.
Figure 16. Schematic diagram of deflection angles.
Sensors 25 02760 g016
Table 1. Training environment parameters for this experiment.
Table 1. Training environment parameters for this experiment.
Hardware/SoftwareConfiguration Parameters
CPUIntel(R) Xeon(R) CPU E5-2690
GPUNVIDIA GeForce RTX 3090
Memory64 GB
Operating SystemUbuntu 20.04
Programming LanguagePython = 3.10
Deep Learning FrameworkPyTorch = 2.0
Table 2. Algorithm comparison experimental results.
Table 2. Algorithm comparison experimental results.
MethodsP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)
Traditional24.112.6//
HRNet90.388.692.188.1
YOLO1189.788.390.480.9
OURS93.492.595.189.6
Table 3. Experimental results of the detection accuracy of container horizontal deviation.
Table 3. Experimental results of the detection accuracy of container horizontal deviation.
MethodsMean Absolute Deviation, MADMAD-P (m)Average Operation Time (s)
MAD X (m) MAD X (m)
Manual operation0.0130.0160.0239.36
Ours0.0120.0180.0248.68
Table 4. Experimental results of container deflection angle-detection accuracy.
Table 4. Experimental results of container deflection angle-detection accuracy.
MethodsMAE- θ (°)Average Operation Time (s)
Manual Operation0.159.86
OURS0.118.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; He, M.; Zhang, Y.; Zhang, Z.; Postolache, O.; Mi, C. High-Precision Pose Measurement of Containers on the Transfer Platform of the Dual-Trolley Quayside Container Crane Based on Machine Vision. Sensors 2025, 25, 2760. https://doi.org/10.3390/s25092760

AMA Style

Wang J, He M, Zhang Y, Zhang Z, Postolache O, Mi C. High-Precision Pose Measurement of Containers on the Transfer Platform of the Dual-Trolley Quayside Container Crane Based on Machine Vision. Sensors. 2025; 25(9):2760. https://doi.org/10.3390/s25092760

Chicago/Turabian Style

Wang, Jiaqi, Mengjie He, Yujie Zhang, Zhiwei Zhang, Octavian Postolache, and Chao Mi. 2025. "High-Precision Pose Measurement of Containers on the Transfer Platform of the Dual-Trolley Quayside Container Crane Based on Machine Vision" Sensors 25, no. 9: 2760. https://doi.org/10.3390/s25092760

APA Style

Wang, J., He, M., Zhang, Y., Zhang, Z., Postolache, O., & Mi, C. (2025). High-Precision Pose Measurement of Containers on the Transfer Platform of the Dual-Trolley Quayside Container Crane Based on Machine Vision. Sensors, 25(9), 2760. https://doi.org/10.3390/s25092760

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop