Next Article in Journal
An Innovative Fusion-Based Scenario for Improving Land Crop Mapping Accuracy
Next Article in Special Issue
CamNuvem: A Robbery Dataset for Video Anomaly Detection
Previous Article in Journal
Characterization of Infants’ General Movements Using a Commercial RGB-Depth Sensor and a Deep Neural Network Tracking Processing Tool: An Exploratory Study
Previous Article in Special Issue
Gaze Estimation Approach Using Deep Differential Residual Network
 
 
Article
Peer-Review Record

ArbGaze: Gaze Estimation from Arbitrary-Sized Low-Resolution Images

Sensors 2022, 22(19), 7427; https://doi.org/10.3390/s22197427
by Hee Gyoon Kim and Ju Yong Chang *
Reviewer 1: Anonymous
Reviewer 2: Anonymous
Reviewer 3: Anonymous
Sensors 2022, 22(19), 7427; https://doi.org/10.3390/s22197427
Submission received: 6 September 2022 / Revised: 23 September 2022 / Accepted: 27 September 2022 / Published: 30 September 2022

Round 1

Reviewer 1 Report

This paper proposes the use of knowledge distillation and feature adaption to allow the use of arbitrary resolution images to estimate gaze direction. This is worthwhile research with intriguing results; however some limitations of the work need to be more transparently addressed and discussed.

Detailed comments:

1.      Line (Ln) 45-51: The teacher-student analogy in not explained well upon introduction

2.      Ln 95-104: More details on the related work, issues and solutions would be appreciated, particularly how they are relevant/related to your problem and solution

3.      Ln 133-140: Why aren’t these methods good enough? Is your method clearly superior? A comparison of results may be in order.

4.      Ln 151-152: Up until now, you said your method would work with any arbitrary resolution, but now you introduce the limitation that the aspect ratios must be the same. This reduces the general applicability and means your input resolution can no longer be truly arbitrary. This should be addressed up front and as a limitation of the method.

5.      Ln 205: Before jumping into the results, a subsection of your method section could explain your intended method of testing, and its rational. This would help prep the reader for the upcoming tests, and interpreting the results.

6.      Ln 220-224: given your source dataset, you do not test on true arbitrary resolution images, but (artificial) estimations of them. This should be noted.

7.      Ln 250: Ablation studies allow us to see the effect of a method/algorithm comparted to baseline (nothing). This alone does not prove the method is worthwhile in the real world. A small increase may not be worthwhile, especially if better methods exist. So be careful not to over-sell the significance of the ablation study.

8.      Table 1: Did you determine the statistical significance of these results?

9.      Figure 7: the MPII Gaze results are too small to interpret well; a larger image would be helpful

10.  Line 352: A section discussing the limitations of the proposed method is warranted.

 

Author Response

Please see the attachment.

Author Response File: Author Response.pdf

Reviewer 2 Report

Well written paper, few suggestions for authors

-Introduce the motivation of the study in the light of the literature review

-Please introduce the Application section where authors can celebrate the application of this project 

-Include research questions and hypothesis

-Baseline results should be compared with statistical testing with a specific confidence 

-Please use discussion sections before the conclusion and discuss your results in light of the literature review. 

 

 

 

Author Response

Please see the attachment.

Author Response File: Author Response.pdf

Reviewer 3 Report

The organization of the paper is good. However some improvements are required:

1)     The authors mentioned 2 contributions of their works. They need to write clearly in discussion sections or at the end of experimental results section that they have achieved 2 contributions throughout the different dimension of experiments.

2)     ResNet and VGG are used for feature extractions in this paper. They need to justify why they have used these two for feature extractions. What is the reason behind changing the existing ResNet-18-based teacher and student structure.

3)     Other than the ResNet and VGG, the authors should compare their model with others network classifiers.

4)     Authors should compare their results with some other existing recent works. The result should be presented in tabular form for better understanding,

5)     Write the paper in passive voice. Don’t write : I, we and so on.

6)     Please read and cite some quality works in this area : “Eye gaze estimation: A survey on deep learning-based approaches (2022)”, “A Human-Robot Interaction System Calculating Visual Focus of Human’s Attention Level (2021)”, “When i look into your eyes: A survey on computer vision contributions for human gaze estimation and tracking (2020)”, “Appearance-based gaze estimation with deep learning: A review and benchmark (2021)”,.

Author Response

Please see the attachment.

Author Response File: Author Response.pdf

Round 2

Reviewer 3 Report

All comments are addressed successfully

Back to TopTop