Next Article in Journal
ML Approach to Improve the Costs and Reliability of a Wireless Sensor Network
Next Article in Special Issue
Efficient Fine Tuning for Fashion Object Detection
Previous Article in Journal
Recent Development of Mechanical Stimuli Detectable Sensors, Their Future, and Challenges: A Review
Previous Article in Special Issue
Design Principle and Proofing of a New Smart Textile Material That Acts as a Sensor for Immobility in Severe Bed-Confined Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Echo-ID: Smartphone Placement Region Identification for Context-Aware Computing

Department of Computer Science and Technology, Ocean University of China, Qingdao 266100, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(9), 4302; https://doi.org/10.3390/s23094302
Submission received: 3 April 2023 / Revised: 17 April 2023 / Accepted: 25 April 2023 / Published: 26 April 2023
(This article belongs to the Special Issue Human-Centric Sensing Technology and Systems)

Abstract

:
Region-function combinations are essential for smartphones to be intelligent and context-aware. The prerequisite for providing intelligent services is that the device can recognize the contextual region in which it resides. The existing region recognition schemes are mainly based on indoor positioning, which require pre-installed infrastructures or tedious calibration efforts or memory burden of precise locations. In addition, location classification recognition methods are limited by either their recognition granularity being too large (room-level) or too small (centimeter-level, requiring training data collection at multiple positions within the region), which constrains the applications of providing contextual awareness services based on region function combinations. In this paper, we propose a novel mobile system, called Echo-ID, that enables a phone to identify the region in which it resides without requiring any additional sensors or pre-installed infrastructure. Echo-ID applies Frequency Modulated Continuous Wave (FMCW) acoustic signals as its sensing medium which is transmitted and received by the speaker and microphones already available in common smartphones. The spatial relationships among the surrounding objects and the smartphone are extracted with a signal processing procedure. We further design a deep learning model to achieve accurate region identification, which calculate finer features inside the spatial relations, robust to phone placement uncertainty and environmental variation. Echo-ID requires users only to put their phone at two orthogonal angles for 8.5 s each inside a target region before use. We implement Echo-ID on the Android platform and evaluate it with Xiaomi 12 Pro and Honor-10 smartphones. Our experiments demonstrate that Echo-ID achieves an average accuracy of 94.6% for identifying five typical regions, with an improvement of 35.5% compared to EchoTag. The results confirm Echo-ID’s robustness and effectiveness for region identification.

1. Introduction

Context-aware computing refers to the ability of computing systems to automatically recognize and respond to the user’s scenario, providing intelligent and personalized services [1,2,3,4,5]. The global market size of context-aware services exceeded USD 36.3 billion in 2020 and is projected to reach over USD 318.6 billion by 2030 [6]. This growth is driven by the increasing adoption of mobile devices and the need for personalized and intelligent services.
With the prevalence of smartphones, there is a growing demand for smartphones to be more intelligent and context-aware. By capturing and utilizing contextual information, smartphones can automatically adjust their behavior based on the user’s current location, time of day, and activity level [1,2]. An example of context-aware computing with smartphones includes automatically activating the silent mode when a smartphone is placed on the nightstand to prevent disturbing the user’s sleep during the night. Similarly, when a smartphone is placed on a desk, it can enter a focus mode for a certain duration to ensure work quality, as shown in Figure 1. All these user scenarios require that the phone recognize the region where it resides and provide a specific function depending on the region. Note that users cannot place their phone on the exact location in the above scenarios, so context-aware services should be provided by identifying regions.
Despite the growing demand for a lightweight region-function combination for smartphones, such solutions are not yet pervasive. Traditionally, passive sensing solutions require pre-installed infrastructures or tedious calibration efforts. Solutions based on Wi-Fi [7], and Radio Frequency (RF) [8] require pre-installed equipment, such as wireless transmitters or beacons. Solutions that use indoor background lighting [9], sound [10,11], or mobile signal patterns [12] as fingerprints require laborious calibrations, which limits their practicality for personal use. Compared to other types of wireless sensing, acoustic sensing offers two advantages, (1) the widespread application of speakers and microphones in commercial electronic products, such as smartphones, smartwatches, and tablets, eliminates the need for additional hardware support; and (2) the lower propagation speed of sound in air compared to RF signal favors achieving higher distance measuring accuracy. Recently, active acoustic sensing solutions, such as EchoTag [13], enable phones to tag and remember locations by actively transmitting a sound signal with the phone’s speaker and sensing its reflections with the microphones. While EchoTag can provide a high 1cm resolution location-function combination, it requires users to place their phone in the exact location where they tagged it before, creating a new burden for users to remember and place their phone at the exact position and angle where it has been tagged.
To overcome these limitations, we propose Echo-ID, a low-cost region identification system that enables smartphones to identify regions and provide context-aware services without requiring any additional sensors or pre-installed infrastructure. Echo-ID actively generates an acoustic signature for target regions by using the phone’s speaker to transmit ultrasound and microphones to capture the reflections. Echo-ID requires users only to put their phone at two orthogonal angles for 8.5 s each inside a target region and bind a function with the region. The user can then place their phone freely in the target region and automatically obtain the context-aware functions in the future. Echo-ID only use the built-in speakers and microphones available in commodity smartphones and has low deployment and memory overhead for users.
However, it is non-trivial to identify regions for a smartphone only through ultrasound sensing. To ensure accuracy and robustness, we have to address two new challenges. (1) There is uncertainty in the phone placement inside a target region, which includes variations in position and angle. Asking users to place their phones at every possible position and angle to collect training data before using Echo-ID is impractical. (2) The reflective objects surrounding the target area may change over time. Therefore, Echo-ID must be able to accommodate placement uncertainty and environmental variations.
To address these challenges, Echo-ID applies Frequency Modulated Continuous Wave (FMCW) signals as a sensing medium and design a deep learning network to identify the target region. Echo-ID applies FMCW signals with carefully selected parameters. After removing the random delay by detecting the starting timestamp of the received sequence, the received signal is orthogonal demodulated to the baseband. Secondly, two types of coarse features are extracted from the baseband signals through Fast Fourier Transform (FFT) and Short Time Fourier Transform (STFT) guided by the FMCW ranging model, representing the coarse features of the spatial relationships among the surrounding reflective objects and the smartphone. Echo-ID’s deep learning model integrates residual network and attention mechanisms. The convolution operations inside residual network aim to dig out the invariant inside the coarse features, which is originally impacted by the phone placement uncertainty of position and angle. The attention mechanism may further focus on the core of spatial relationships among the surrounding reflective objects and the smartphone, reducing the impact of environmental changes for region identification. Though these two mechanisms, Echo-ID extracts finer features from the coarse features that emphasize the characteristics of the target region and are robust to phone placement uncertainty and environmental variation. Finally, the model identify the target region based on the finer features extracted.
We implement Echo-ID using the LibAS framework [14] on the Android platform and evaluate its performance on two smartphone models, Xiaomi 12 Pro and Honor-10, for identifying five typical regions, namely placing the phone on a nightstand, a kitchen operating table, a desk, a shelf, and a TV cabinet. The size of the Xiaomi 12 Pro device is 16.3 cm × 7.5 cm, and to allow users to place their phones easily and freely within the region, we set the target region at 35.6 cm × 35.6 cm, which is larger than twice of the phone’s longer edge. Theoretically, it would enhance Echo-ID’s ability to identify regions by obtaining training data from all conceivable positions and angles, but this imposes an impossible burden on users. Since Echo-ID represents the spatial relationship among surrounding objects in two-dimensional space under acoustic sensing, collecting training data by placing the smartphone at two orthogonal angles within the region may be possible to extract spatial features for classifying regions. Thus, the user only needs to place their phone at the region center’s two orthogonal angles for 8.5 s each to collect training data.
We use a Chirp of 42.6 ms as the sensing medium. For collecting training data, a Chirp is interleaved with a 42.6 ms idle slot, ensuring the elimination of reflection from distant objects in the received signals. Thus, we can obtain 100 spectrograms to train Echo-ID’s identification model from the sample collection of 8.5 s for each angle. To evaluate the impact of phone placement uncertainty, we evaluate the most varied 16 positions and 12 angles for phone placement inside each region. We also intentionally added or removed objects near the target regions to evaluate the system’s performance under environmental variations.
Our experiments demonstrate that Echo-ID achieved an average accuracy of 94.6% for identifying the five regions. We also thoroughly evaluated the system parameters that affect Echo-ID’s performance, including feature selection, residual module numbers, multi-angle combinations in the training set, smartphone model, permanence over time, and identification for 10 regions. The results confirm Echo-ID’s robustness and effectiveness for region identification.
The main contributions in Echo-ID are summarized as follows.
  • We propose Echo-ID, a low-cost region identification system for providing contextual services that exploits FMCW as a sensing medium. Echo-ID is not limited to placement position and angle, adapts to environmental changes, and without requiring any additional sensors or pre-installed infrastructure.
  • Based on the extraction of spatial properties of surrounding reflective objects of the target region, we design a region identification model incorporating residual networks, attention mechanisms, and multi-layer perceptron to enable Echo-ID to achieve high accuracy in region recognition under the constraints of arbitrary placement of smartphones and dynamic nature of the surrounding environment.
  • We implement Echo-ID on the Android platform and evaluate it with Xiaomi 12 Pro and Honor-10 smartphones, demonstrating its usability and robustness through extensive experiments.
The remainder of this paper is organized as follows: Section 2 describes the preliminary of FMCW signals and the ranging model. Section 3 provides a detailed description of Echo-ID design. Experimental details and the performance of Echo-ID are presented in Section 4. Section 5 discusses the related works. Finally, Section 6 summarizes the paper.

2. Preliminaries

In this section, we introduce the basics and the range model of FMCW signals used in acoustic sensing.

2.1. FMCW Fundamentals

FMCW is a radar technology that employs modulated continuous waves, where the transmitted signal is a sine wave whose frequency sweeps linearly over time, known as a Chirp signal. Compared to Continuous Wave (CW), the Chirp signal allows us to determine the absolute distance of a target and separate reflections from different distances. The transmitted signal S T t [15] can be obtained from Equation (1)
S T t = c o s T t = c o s 2 π f 0 t + B t 2 2 T s
where f 0 is the starting frequency, B is the bandwidth, T s is the sweep time. The signal frequency can be expressed as f t = f 0 + B T s t . In Figure 2, the blue line represents S T t .

2.2. Ranging Model

Chirps are commonly used in distance measurements between a target and a transceiver. In the case of acoustic sensing, a high-frequency Chirp is emitted from the speaker, and the signal reflected by surrounding objects is captured by the microphone. Consider a simple scenario where there is just a single reflection from the surrounding object. When the signal encounters a target at a distance of R, the reflected signal S R t is a time-delayed version of S T t , which can be represented as
S R t = A c o s T t τ = A c o s 2 π f 0 t τ + B t τ 2 2 T s ,
where A is the amplitude attenuation factor, and τ is the Time-of-Flight (ToF) of the signal in the air.
To calculate the distance between the transceiver and the object, the received signal needs to be demodulated by multiplying it with c o s T t and s i n T t . After applying the product-to-sum and product-to-difference conversion (i.e., c o s α · cos β = 1 2 c o s α β + c o s α + β and c o s α · sin β = 1 2 sin α β sin α + β ), the demodulated signal S D 1 t and S D 2 t are calculated according to Equations (3) and (4).
S D 1 t = A 2 c o s 2 π B τ T s t + f 0 τ B τ 2 2 T s + c o s 2 π B T s t 2 2 f 0 t + B τ T s t + f 0 τ B τ 2 2 T s ,
S D 2 t = A 2 sin 2 π B τ T s t + f 0 τ B τ 2 2 T s s i n 2 π B T s t 2 2 f 0 t + B τ T s t + f 0 τ B τ 2 2 T s .
The demodulated signal contains low-frequency and high-frequency components. The low-frequency component corresponds to the frequency difference between the received and transmitted signals, which is constant, labeled as f d = B τ T s . On the other hand, the high-frequency component represents the sum of their phases and varies with time, which can be filtered out by a low-pass filter. Figure 2 also shows an example of the received signals. The bottom part of Figure 2 shows the frequencies of the low- and high-frequency components of the demodulated signals. The low-frequency components of the demodulated signals S D 1 t and S D 2 t can be treated as the in-phase component, denoted as I, and the quadrature component, denoted as Q, respectively. The complex baseband signal S B t is generated from Equation (5).
S B t = A 2 e j 2 π B τ T s t + f 0 τ B τ 2 2 T s .
As shown in Figure 2, the baseband signal has a constant frequency, and, thus, S B t in Equation (5) can be simplified to Equation (6).
S B t = A 2 e j 2 π f d t + φ d ,
where f d denotes the beat frequency, and φ d = 2 π f 0 τ π B τ 2 T s is the phase. By calculating the beat frequency of the baseband signal, the propagation delay τ can be determined, and, thus, the distance between the device and the object can be calculated by R = c τ 2 , where c represents the speed of sound in air. Therefore, the radar using Chirp signals can find the distance between transceiver and the major reflected object, which is called the FMCW range model.

2.3. Multi-Target Range

In real-world scenarios, acoustic signals are reflected by multiple surrounding objects, leading to the superposition of reflected signals from different propagation paths at the microphone, called multipath effect. The received signal is a combination of signals from L paths, each with different delay and attenuation factors. The baseband signal can then be rewritten as Equation (7).
S B t = l = 1 L A l 2 e j 2 π f d l t + φ d l ,
where A l , f d l , φ d l represents the attenuation factor, beat frequency, and phase of the signal received from the l-th path, respectively.
The time delays and beat frequencies of signals reflected from various distances create a one-to-one correspondence between propagation path length and baseband signal frequency. The spectrum of the baseband signal reveals the effect of reflections from different distances on the signal’s frequency, which can represent the spatial features of surrounding objects for a target region. The distance resolution of the FMCW system determines its ability to differentiate between two targets based on distance, with higher resolution allowing the radar to distinguish more closely spaced targets. When appropriate parameters for the FMCW acoustic signal are selected, a minimum resolvable distance of d ^ = c 2 f s = 3.57 mm, which will be discussed in Section 4.2.
The acoustic characteristics of an environment is also represented by the uneven attenuation of the signal at different frequencies, as shown in Figure 3. This phenomenon can be attributed to three reasons, as reported in previous studies [16,17]. First, the imperfection in speakers and microphones will cause changes in the signals. Secondly, when sound reaches the surface of the reflected objects, the surface material may absorb signals at a particular frequency. Different materials have varying absorption frequencies. The combination of these two factors causes the received signal to amplify at certain frequencies and attenuate at others. Thirdly, the propagation of acoustic signals is subject to multipath effects when encountering reflections from surrounding objects, leading to different propagation paths and phase differences that impact the signal’s strength at different frequencies as discussed above. Therefore, Echo-ID may leverages the spectrum to recognize the region where the phone resides, because the spectrum presents the difference in the amplitudes on frequencies of the received signals.

3. System Design

This section first presents the design overview and then illustrates the design of each module.

3.1. System Overview

Echo-ID consists of four components, namely signal transmission, signal processing, coarse feature extraction, and a deep learning model, as illustrated in Figure 4. The signal transmission module transmits FMCW Chirps with appropriate parameters to obtain reflections from the surrounding objects. In the signal processing module, auto-correlation is employed to detect the Chirp start timestamp inside the received signals. Subsequently, orthogonal demodulation is performed to derive the baseband signal, followed by the elimination of the direct propagated component. Time-frequency domain analysis is used to extract spectrogram from the baseband signal, which serve as the coarse features for classification. Finally, the deep learning model employs a deep learning network that integrates convolution and residual networks with attention mechanisms to compute finer features and identify region.

3.2. Signal Transmission

Echo-ID transmits the Chirps to capture reflections from surrounding objects for region identification. The Chirp parameters include the frequency band and sweep time. As the sampling rate of the smartphone’s speakers is typically 48 kHz, the highest detectable sound frequency is limited to 24 kHz according to the Nyquist-Shannon sampling theorem. To avoid audible noise and ensure compatibility with most commercial smartphones, the scanning frequency band of 18 kHz–22 kHz is selected since humans can hear sounds within the range of 20 Hz to 18 kHz [15].
The sweep time is closely related to the Signal-to-Noise Ratio (SNR) of the received signal. Although a longer sweep time can increase the reflecting range and frequency resolution, it also consumes more energy and may introduce interference from distant objects. Since Echo-ID is designed to recognize specific regions, reflected signals from distant objects far from the target area are not necessary. Hence, a sweep time of 42.6 ms is chosen for each Chirp sequence, resulting in 2048 samples for each sweep, to obtain sufficient frequency resolution and avoid interference caused by distant reflections. The time-domain and spectrogram of the transmitted Chirp sequence are depicted in Figure 5.

3.3. Signal Processing

This module performs three operations, including removing the random delay by auto-correlation, orthogonal demodulation, and direct propagated component removal. To identify the start of the received Chirp, we leverage the strong correlation property of the Chirps. The Chirp is a CAZAC sequence [18] with a constant envelope and zero auto-correlation characteristics, making it an ideal candidate for detecting the start with auto-correlation. By correlating the transmitted Chirp with the received signal, the timestamp corresponding to the maximum correlation value marks the start of the received Chirp. An example of the auto-correlation result is illustrated in Figure 6, where the peak time (42.6 ms) indicates the start of the received Chirp.
After obtaining the received Chirp, demodulation is performed to transform it into the baseband. The orthogonal demodulation process is illustrated in Figure 7, where the received signal is multiplied by the Chirp and passed through a low-pass filter to separate the in-phase component I and quadrature component Q. Echo-ID employs an equiripple low-pass filter, which is a linear-phase finite impulse response filter, to retains the phase spectrum of the received signals. The filter’s pass-band cut-off frequency is set to 4 kHz, fully preserving the effective signal.
To avoid significant interference from directly propagated signals that travel from the speaker to the microphones, it is necessary to measure the direct propagation path and remove the low-frequency components that correspond to this distance, according to the FMCW range model. This is because the distances of the directly propagated signals are shorter than those of the reflected signals.

3.4. Coarse Feature Extraction

Context-aware computing refers to the ability of computing systems to automatically recognize and respond to the user’s scenario, providing intelligent and personalized services. An example of context-aware computing with smartphones includes automatically activating the silent mode when a smartphone is placed on the nightstand to prevent disturbing the user’s sleep during the night. Similarly, when a smartphone is placed on a desk or shelf, we hope it can enter a focus mode for a certain duration to ensure work quality. When a smartphone is on a TV cabinet, the phone volume is automatically turned up to avoid missing calls and messages. When a smartphone is placed on a kitchen operating table, the sound is transmitted to the speaker for playback, or a predefined timer is started when it is near a device, such as a microwave oven.
All these user scenarios require that the phone recognize the region where it resides and provide a specific function depending on the region. Additionally, the surrounding objects within the kitchen and desk scenes change frequently, which helps to verify the validity of Echo-ID. These scenarios reflect how users can benefit from region–function combinations of context-aware services. Therefore, we positioned the smartphone in these five regions for experiments, on a nightstand, a kitchen operating table, a desk, a shelf, and a TV cabinet, as shown in Figure 8.
Since the reflections from multiple objects are combined within the received signals, it is crucial to extract features from the baseband signals. The frequency spectrum of the baseband signal corresponds to different reflection paths from objects at various distances, as discussed in the FMCW range model, and can be obtained through FFT. Figure 9 displays the frequency spectrum of five regions, exhibiting differences, particularly at their peak values. The distributions of peak frequencies represent the major reflected signal paths, corresponding to the locations of surrounding reflective objects. For example, the two peaks in the amplitude of the nightstand marks the frequency of 102 Hz and 177 Hz, which labels two reflective objects, i.e., the glasses and lamp, at a distance of about 20.4 cm and 35.4 cm, respectively.
Echo-ID requires a feature that portrays the spatial relationships between the surrounding objects and the smartphone. According to the FMCW range model, the peak in the amplitude spectrum labels a reflected path of the surrounding object with the peak corresponding to the distance between the object and the smartphone. However, FFT does not account for the arrival times of the reflected signals along multiple reflection paths, so it majorly treats a reflective object as a particle.
Considering that actual objects have a continuous reflective area, we apply STFT on aggregated signal sequences to extract spatial features. Through the fine-grained movement of the time window (0.1 ms), STFT can capture a more detailed description of the multi-path reflected signals by the same object surface. In other words, an extreme point of the frequency corresponding to the amplitude spectrum will be a hot zone in the spectrogram around the same frequency with a time length representing the difference between the shortest and longest reflected path. Moreover, other transforms, such as Wavelet Transform (WT), Hilbert-Huang Transform (HHT) [19], and Discrete Tchebichef polynomials (DTPs) [20], may also provide different features from the received acoustic sequences, but since they do not represent the distance, like Fourier transform and STFT, we will try it in future work.
Because Echo-ID achieves region identification based on the absolute distance from the object around the device to the microphone and its reflective surface size, the signal variation caused by objects around the device is more important for Echo-ID. Therefore, we only focus on the frequency from 0 to 500 Hz in the spectrograms, corresponding to a range of 90 cm around the device. Figure 10 shows variations in spectrograms among five regions, especially in the frequency and time corresponding to the hot zones. The distribution of the highlight areas represents the spatiotemporal distribution of the arrival of the reflected signals reflected by the surrounding objects. In Figure 10a, the two hot zones also mark the two reflective objects (the lamp and glasses). The area difference in the two hot zones further indicates different sizes of the continuous reflective area of two objects, which can serve more clues about the surrounding objects. We call the spectrum and spectrogram features as coarse features, because the noises are still buried inside. In our evaluation, we will compare the spectrum and spectrogram feature for region identification.

3.5. Deep Learning Model

Echo-ID leverages deep learning networks to extract finer features from the coarse features and recognize region in which a user places their smartphone, taking into account the variations of each displacement. For example, a user may place their phone on a desk while working, but the precise location of the phone may vary within a given region each time. To accommodate these variations in device placement, Echo-ID uses deep learning to capture the unique characteristics of each region. The network architecture is depicted in Figure 11.
To extract fine-grained features from spectrograms, the learning network employs a convolution neural network (CNN) [21,22] and a residual network (ResNet) [23] as its backbone, which take 256 × 256 spectrograms of amplitude and phase as input. The CNN’s convolution operation extracts local features while preserving spatial information, and pooling operations reduce the dimension of the feature map, enhancing the model’s computational efficiency. As the CNN depth increases, the model’s spatial receptive field expands, capturing global spatial features across the frequency band and time domain. However, excessively deep networks can cause issues such as gradient vanishing or exploding, as well as smaller gradient updates for shallow CNN layers, which can increase training difficulty. To address these issues, ResNet incorporates a shortcut path from input to output, allowing the model to learn the residual between the input and output directly during back propagation. By using a deep learning network for feature extraction, fine-grained features inside the spectrograms can be effectively extracted.
To emphasize the differences among these fine-grained features, a lightweight convolution block attention module (CBAM) [24] is integrated into the residual network. CBAM adds attention in the channel and spatial dimensions by performing pooling operations while maintaining the feature map’s channel or spatial dimension unchanged. A Multi-Layer Perceptron (MLP) generates attention values for each channel or spatial dimension, and the attention values for both the channel and spatial dimensions are multiplied and added to the feature map. By calculating the attention weight for each channel and spatial dimension, the feature map is weighted, enhancing key features, suppressing noise and irrelevant features, and enabling the model to focus more on specific features in a certain time and frequency range. CBAM can visualize the model’s attention regions using the classification activation map (CAM) [25], which will be presented in the evaluation section.
Finally, the extracted fine-grained features are concatenated into one feature vector, which is then classified using a MLP. Table 1 provides the parameter configurations for each module as well as the corresponding input and output dimensions.
Figure 12a–e shows the spectrogram of five placement modes in the kitchen operating table region and Figure 12f shows the concatenated features from the well-trained model. When the positions and angles of the phone and environmental changes, the deep learning model can still extract the fine-grained features and, thus, gives similar feature values before the classifier. Figure 12g,h shows that the CBAM helps the model pay more attention to specific features in a certain time and frequency range. We input the spectrogram in Figure 10 to the model and show the concatenated features from the well-trained model in Figure 13a. Compared to Figure 13f, the deep learning model is able to magnify the features difference in distinct regions. Figure 13b,c also demonstrate that our model can enhance key features in distinct regions.
In principle, obtaining training data from all conceivable positions and angles would enhance Echo-ID’s ability to identify regions, but this would also impose a burden on users. Since Echo-ID represents the spatial relationship among surrounding objects in two-dimensional space under acoustic sensing, collecting training data by placing the smartphone at two orthogonal angles at the region center is sufficient. Thus, the user only needs to place their phone at the region center’s two orthogonal angles for 8.5 s each to collect training data. During the 8.5 s, a Chirp of 42.6 ms is interleaved with a 42.6 ms idle slot, ensuring the elimination of reflection from distant objects in the received signals. The deep learning model design compensates for position and angle differences through convolution operations and accommodates the surrounding changes through the attention mechanism in region identification. As Section 4.4 demonstrates, collecting training data at more than two orthogonal angles only slightly improves region identification accuracy.

4. Experiment and Evaluation

In this section, we conduct comprehensive experiments to demonstrate the effectiveness of Echo-ID.

4.1. Experiment

4.1.1. Settings

We implemented Echo-ID on the Android platform based on LibAS [14] and conducted experiments to evaluate its performance using Xiaomi 12 Pro and Honor-10 smartphones in five different regions, as shown in Figure 8. The target regions are 35.6 cm × 35.6 cm, which accommodates the size of the Xiaomi 12 Pro device (16.3 cm × 7.5 cm), allowing users to place their phones easily and freely within the region. The deep learning model is completely implemented by PyTorch with a RTX 3060 Ti GPU. The network is trained using Adam with a learning rate = 1 × 10 4 .

4.1.2. Dataset

Echo-ID uses the speaker at the bottom to send out FMCW signals and the two microphones at the top and bottom to receive signals. During the experiment, individuals in the environment are not restricted in their activities, such as typing on a keyboard or walking.
To reduce the burden for users before using the Echo-ID, we collect the training set by placing the smartphones inside the target region with two orthogonal angles of 0° and 90°, as shown in Figure 14a, after which the system can recognize the region in the future. Echo-ID requires only 8.5 s at each angle to collect 100 FMCW traces of 42.6 ms with a 42.6 ms idle slot inserted between two Chirps to eliminate the distant reflection on the subsequent trace. The total time cost of training data collection is 17 s, which is easy for users to accept. The total size of the training set obtained for the five regions is 1000 traces.
To fully evaluate the impact of phone placement uncertainty on region identification, we select 16 different positions and 12 angles, as illustrated in Figure 14b,c. Additionally, we simulate environmental variations by adding or removing reflective objects, such as cups, bowls, and books, to assess Echo-ID’s adaptability.
To evaluate the effect of different coarse features extracted from the received signal on the recognition capability, we compare different features from the received signals and baseband signals, including received signal spectrum, baseband signal spectrum, baseband signal spectrogram, and spectrum and spectrogram of baseband signal.
To fully assess the impact of different numbers of residual blocks on the network classification performance, we evaluate the network identification accuracy with 2, 3, 4, and 5 residual blocks.
To determine whether the training set should include data from multiple angles, we train models with data collected from 1 to 6 different angles at the center position and tested them with data from other angles and the changed environment.
To verify the effect of differences in microphones and speakers used by different smartphones, we evaluate the usability of Echo-ID with Xiaomi 12 Pro and Honor-10.
To evaluate the permanence of region identification, the experiments last for 12 weeks across the summer, fall, and winter seasons. We collect 56,000 traces every week, resulting in a total of 672,000 traces. The traces in the first week are used for training and the remaining traces are used for testing.
At the end of this section, we choose five regions where smartphones are placed less frequently to test the applicability of Echo-ID. The dataset is available at https://github.com/JiangXueting1225/Echo-ID, accessed on 27 March 2023.

4.2. Identification Accuracy

The identification accuracy of Echo-ID for five regions is shown in Figure 15a. Considering noise, imperfect hardware, and accommodating environmental changes in practice, it is difficult for this solution to achieve 100% recognition accuracy. Moreover, the different scenario may require different acceptable accuracy according to their special region-combination. For example, the nightstand scenario may require high accuracy in order to ensure the quality of the user’s sleep. The results in Figure 15 show that the nightstand has the highest accuracy of 100% to ensure that the user’s preset mode, such as mute, can be activated. This indicates that reflecting objects in this environment were highly discriminable, resulting in the least path interference. The kitchen operating table and TV cabinet have lower identification accuracy, possibly due to the presence of multiple objects with continuous reflecting surfaces, like the walls. In the desk region, users usually need to turn on focus mode. However, the desk region is partially recognized as the nightstand, likely due to the concentrated objects on one side. The lowest accuracy of 88% exists for the desk region, attributed to the high degree of dynamic factors in the office/lab environment that affected the extraction of time-frequency domain features. When the user finds that the mode is not turned on, Echo-ID can be performed again. The whole process only takes only 42.6 ms, which is shorter than manually turning on functions such as switching to focus mode. The overall accuracy of 94.6% ensures the probability of correct identification for continuous 12 times is still over 50%. This indicates that Echo-ID can automatically identify usage regions in a user’s daily life and adapt to various placement conditions and variations in the surrounding environment. And this only requires the user to collect training data of 17 s in five regions each. Hence, the identification accuracy in terms of user experience is acceptable.
Furthermore, we compare Echo-ID with EchoTag in Figure 15b which uses the received signal spectrum as features, and Support Vector Machines (SVM) for region classification. The identification accuracy of EchoTag was 69.8%, lower than Echo-ID, especially for the nightstand region. EchoTag does not account for placement ambiguity and environmental variations, whereas Echo-ID demonstrated compatibility with phone placement positions, angles, and environmental variations by applying residual networks with attention modules to accommodate the differences inside one region. Moreover, we combine Echo-ID features with a SVM classifier, whose results are shown in Figure 15c with an average accuracy of 75%. The result is higher than EchoTag, confirming that the Echo-ID’s features are superior to represent the region’s characteristics.
To evaluate the compatibility of Echo-ID, we present the identification results for various phone placement positions, angles, and environmental variations in Figure 16. Figure 16a indicates that positions closer to the center have higher accuracy, while positions further away have lower accuracy. Nevertheless, the identification accuracy of 90.2% is sufficient to demonstrate compatibility with phone placement positions. Figure 16b reveals that accuracy is higher for specific angles (30°, 60°, 120°, and 330°), likely due to their similarity in energy distribution with training angles, but decreases for other angles. Accuracy is relatively high for 180° and 270° due to the use of two pairs of microphones at the top and bottom of the device to receive the signals. When the phone is placed at 180° and 270°, the dual microphones receive signals similar to those received when the device is placed at 0° and 90°. Figure 16c displays identification results for environmental variations, with the nightstand having the highest accuracy due to that most of its reflection paths are from walls and lamps. The movement of cups does not significantly affect the features, while the accuracy of the desk is the lowest since the main reflection object after removing the cup and tissue box is the computer monitor. Nevertheless, the accuracy of 85% for the desk is sufficient to indicate that Echo-ID can be compatible with dynamic environmental variations.
EchoTag achieves centimeter-level spatial position recognition. If the user is willing to clearly mark the location where they place their phone, EchoTag will provide contextual services with more spatial granularity than Echo-ID. For example, drawing the outline of the phone with a pen at the target location and carefully placing the phone in that position with the same placement angle each time. When the two locations are spaced 5 cm apart on a desk, Echo-Tag can recognize and activate two different contextual services for each location. In addition, EchoTag puts so much effort into centimeter-level location discrimination that it deviates when there is a small change of surrounding objects positions.
The major concern here is whether the user is willing to tag out every exact phone location and put their phone carefully at these positions with the same placement angles. Moreover, the users rarely require two contextual services when their phone locations are only 5 cm different. In contrast, Echo-ID aims to provide region-function combinations for users, giving them more freedom to place their phones. The target to provide context services is a region with an area with each edge twice than the longer edge of the phone. In addition, Echo-ID puts no restrictions on the phone placement angle, and it can adapt to small changes in the surrounding environment over time. To collect training data, Echo-ID only requires the user to place their phone at the two orthogonal angles of the region center for 8.5 s each. The spectrogram feature and the designed deep learning network model make Echo-ID robust against phone placement uncertainty and environmental changes.
Figure 17 shows the attention areas of the magnitude and phase spectrogram for 5 regions, deduced from the last layer of CBAM. The low-frequency portion of the amplitude variation within 20.5 cm from the smartphone microphone was the focus of the kitchen operating table and TV cabinet, while the higher frequency amplitude variation in the distant position was of more concern to the nightstand, desk, and shelf. Regarding the phase domain attention areas, the nightstand concentrated on the phase variation at around 26 ms and 42 ms, while the kitchen operating table paid attention to the phase variation throughout the period from 8.3 ms to 40 ms. The desk only showed attentiveness to the phase variation within 8 ms, while the kitchen operating table and TV cabinet were attentive to the phase variation in three different periods. The variation of the attention areas of magnitude and phase across different regions suggests that Echo-ID can distinguish these 5 regions.

4.3. Acoustic Feature

The ability of region identification is directly affected by the features extracted from the received signals. The features compared are shown in Table 2, and Figure 18a presents the results of region identification using different features. The highest classification accuracy is achieved using the spectrogram feature for all regions. This is because it expresses the spatial distribution of the environment’s reflective objects through the spectrum and also captures the signal delay differences of different reflection paths through time-domain window division.
The spectrum feature of the baseband signal achieves the second-highest classification accuracy, with similar results for the nightstand and slightly lower results for the other four regions. This validates the effectiveness of the FMCW’s spatial ranging principle, which reflects the spatial distribution of reflective objects in the environment to a certain extent. The spectrum feature’s classification accuracy is lower than that of the spectrogram feature because the entire spectrum of the reception period is directly input to the identification network, whereas the spectrogram provides multi-level spectra divided by the continuous time windows, increasing the spatial distribution details for different periods.
The joint feature of the baseband achieves lower results than the first two features because the network structure becomes more complex, leading to overfitting due to the small size of the training dataset. Using the spectrum feature of the received signal without demodulation is also less accurate than the first two features, as the dominant feature in the original signal spectrum is the modulation frequency band, submerging the spatial distribution spectrum performance. The accuracy of using the received signal’s spectrum feature for identification is lower than that of EchoTag, particularly in the kitchen operating table, shelf, and TV cabinet regions, because the region identification network in this paper cannot obtain spatial distribution information from the spectrum of the received signal, making it difficult to distinguish environments with more reflective objects.
The nightstand environment consistently achieves perfect accuracy (100%) in both the spectrum and spectrogram characteristics of the baseband signal across all regions. This is due to the high discrimination of reflection distribution, which results in the least amount of confusion in spatial distribution for signals reaching the receiving device, and, therefore, the region can be correctly identified as long as the spatial distribution features are effectively extracted. The shelf and kitchen operating table regions have lower accuracy in the frequency spectrum characteristics due to the continuous reflective surfaces of objects confusing spatial distribution. Although time-domain segmentation improves the recognition rate of these regions, the desk region has the lowest identification rate equivalent to the frequency spectrum characteristics due to its high degree of dynamics.
Figure 18b displays the comparison of the average accuracy of each characteristic, and the identification accuracy of the spectrogram feature is the highest at 94.6%. This is approximately 12% higher than the frequency spectrum characteristic, which ranks second, indicating the efficacy of Echo-ID’s use of the time-frequency spectrum characteristic for region identification.

4.4. Number of Residual Blocks

The classification performance of the network is affected by the number of residual blocks used. We evaluate the influence of different numbers of residual blocks on Echo-ID’s recognition outcomes and assess network identification accuracy with 2, 3, 4, and 5 residual blocks, shown in Figure 19.
The highest identification accuracy for each region is achieved with three residual blocks, resulting in an average identification accuracy of 94.6%. Using 2, 4, or 5 residual blocks results in a decrease in accuracy by 18.8%, 12.8%, and 16.2%, respectively. This is because when only two residual blocks are utilized, the model may not be able to capture the complexity of the input features, and the network cannot fully learn the features for classification from the training data, resulting in underfitting. When there are too many residual blocks, the model may be too complex, leading to overfitting and reduced generalization ability on the test set.
Regardless of the number of residual blocks used, the nightstand consistently achieved the highest accuracy in all regions, indicating that the reflectivity distribution on the nightstand is the most distinguishable and the features are more distinct. Conversely, the identification accuracy of other regions decreased significantly with different numbers of residual blocks. The kitchen operating table, desk, and shelf regions had particularly low recognition accuracy due to the continuous reflective surfaces, such as walls, nearby. Figure 19b displays the average identification accuracy for different numbers of residual blocks, indicating that the network with three residual blocks used in Echo-ID is effective for region identification.

4.5. Size of Training Set

To determine whether the training set should include data from multiple angles, we train models with data collected from 1 to 6 different angles at the center position, including {0°}, {0°, 90°}, {0°, 120°, 240°}, {0°, 90°, 150°, 300°}, {0°, 90°, 150°, 210°, 300°}, and {0°, 60°, 120°, 180°, 240°, 300°}, and test them with data from other angles and the changed environment. The identification accuracy reaches 94.6% when only two angles were used for training, as shown in Figure 20a, and slightly improves of about 0.9% with additional training data. When the training data are increasing to six angles, the identification accuracy only increases by 0.9%, but it adds three times burden for data collection.
Moreover, we train models with data collected from 1 to 8 different position combinations with two orthogonal angles, including position {6}, {1, 6}, {1, 6, 11}, {1, 6, 11, 16}, {1, 4, 6, 11, 16}, {1, 4, 6, 11, 13, 16}, {1, 4, 6, 7, 11, 13, 16}, and {1, 4, 6, 7, 10, 11, 13, 16}. Figure 20b shows that adding the training data of more positions with two orthogonal angles also does not increase the identification accuracy greatly, and the accuracy only increases by 1.0% with training data of eight positions.
These results can be attributed to the fact that all placements in the target region can be represented by two orthogonal angles at the region center as the spatial base. Therefore, to reduce the burden on users under the premise of ensuring accuracy, Echo-ID only requires users to place their phones horizontally and vertically at the center of the target region for 8.5 s each to collect training data.

4.6. Mobile Phone Model

To assess the robustness of Echo-ID across different phone models, we also collect the signal traces using Honor-10 smartphones. We divide all the collected traces into four categories, Dataset I uses the samples from two angles of the Xiaomi 12 Pro for training and all other samples for testing; Dataset II uses the samples from two angles of the Honor-10 for training and all other samples for testing; Dataset III uses data samples from two angles of the Xiaomi 12 Pro for training and all samples from the Honor-10 for testing; and Dataset IV uses a mixture of samples from both Honor-10 and Xiaomi 12 Pro for training and testing.
As shown in Figure 20c, Echo-ID achieves accuracy rates of 94.6% for Dataset I, 90.2% for Dataset II, 71.8% for Dataset III, and 79.6% for Dataset IV. This first confirms that Echo-ID can also be used for Honor-10 smartphones. It also suggests that the highest accuracy is achieved when training and testing are performed on the same phone model. The accuracy of mixed training and testing using two different smartphones is the second highest, while the lowest accuracy is achieved when training on one phone and testing on another. Therefore, when deploying Echo-ID on a new smartphone, it is better to collect training data specific to the new model. We will study whether transfer learning will help accommodate different phone models in the future.

4.7. Permanence

To examine the permanence of region identification, we divide the six-week traces according to their weeks, using only the 8.5 s traces for two orthogonal angles in the first week for training and the remaining data for testing. Figure 20d illustrates the identification accuracy of the five regions for each week, showing that the identification accuracy is stable over time.

4.8. Ten Regions Identification

To validate the effectiveness of Echo-ID, we expand the region variable to ten. The deployment of the remaining five regions is illustrated in Figure 21. Figure 22 displays the confusion matrix for the identification of ten indoor regions using Echo-ID, with region index referencing Table 3. The identification accuracy for all ten regions is above 69%, with corridors achieving the highest accuracy (98%) due to their spaciousness and lack of reflective objects, which distinguish them from other regions. Classrooms have the lowest accuracy (69%) because their layout is similar to that of bedside tables, desks, and washbasins, all of which have a prominent reflective object on one side. More than half of the regions have an accuracy of over 85%, and the overall accuracy for all ten regions is 82.9%. These findings suggest that Echo-ID can effectively perceive regions.

5. Related Works

Currently, the main target related to the region identification of Echo-ID is the indoor positioning research. Global Positioning System (GPS) sensors in mobile devices has been widely used for high-precision outdoor positioning [26]. However, accurate indoor location sensing remains challenging due to the inability of GPS signals to penetrate buildings. In recent years, research efforts have emerged to address indoor location using various signal processing techniques, including Global System for Mobile Communications (GSM) [12], FM radio [27], Wi-Fi [7], and acoustic signals [13]. This section provides an overview of existing indoor location systems, which are categorized into three types based on their perceptual modalities, dedicated equipment, RF signal, and acoustic signal.

5.1. Dedicated Equipment

Specialized devices are previously used for indoor location sensing. For instance, WALRUS [10] uses ultrasonic transmitters and receivers to locate targets in a room by measuring propagation time and Angle of Arrival (AoA). Wang et al. [8] track employees’ indoor locations by attaching battery-powered radio transmitters to their work badges. Addlesee et al. [7] use various sensors and devices to collect environmental samples and automate responses to environmental changes through machine learning and intelligent algorithms. Liu et al. [28] implement smartphone location sensing by placing ultrasonic transmitters with unique identity tags indoors.
The dedicated equipment requires pre-installation of beacons throughout the environment, with higher installation and maintenance costs for infrastructure deployment, imposing an additional burden on the user.

5.2. RF Signal

The use of RF signals for indoor location sensing relies on the existing wireless infrastructure, such as GSM and Wi-Fi, to determine the location of mobile devices. This is achieved by measuring the Received Signal Strength Indicator (RSSI) from cellular or Wi-Fi stations with known locations. Haeberlen et al. [29] collected RSSI measurements from access points in a large-scale 802.11 Wi-Fi network to achieve location sensing. Hightower et al. [12] established a geographic information database using Wi-Fi, cellular, and geomagnetic signals to identify the user’s location. Horus [30] utilized multiple Wi-Fi receivers to determine the wireless device’s position through the analysis of parameters, such as RSSI, arrival time, and AoA. Vasisht et al. [31] and Boney et al. [32] implemented indoor location sensing using RSSI from a single Wi-Fi access point.
The localization accuracy of such schemes depends on the density of cellular or Wi-Fi stations in the environment, and the RSSI is susceptible to fluctuations due to multi-path fading in complex indoor environments.

5.3. Acoustic

Acoustic sensing has been widely adopted in various smart sensing scenarios, such as identity verification [33,34,35,36,37,38,39,40], health monitoring [41,42,43,44,45], input interfaces [46,47,48,49,50], keystroke inference [51,52], and gesture recognition [53,54,55]. Compared to other types of wireless sensing, acoustic sensing offers two advantages, (1) the widespread application of speakers and microphones in commercial electronic products, such as smartphones, smartwatches, and tablets eliminates the need for additional hardware support; and (2) the lower propagation speed of sound in air compared to RF signal favors achieving higher distance measuring accuracy.
Passive acoustic location sensing captures environmental sounds as fingerprints for indoor positioning [9,11,56]. However, these solutions are susceptible to the influence of environmental noise, such as music or conversation. In addition, passive acoustic location solutions are greatly affected when the environment changes, for example, Batphone [11] notes that when the Heating, Ventilation, and Air Conditioning (HVAC) is turned off for maintenance, the characteristics of the background sound can change significantly, which can interfere with the accuracy of locating single or multiple sound sources. To address this issue, some studies [57,58] have proposed using microphone arrays to locate sound sources more precisely.
Active acoustic location sensing emits pre-designed acoustic signals through the speaker and capture their echoes using the microphone to estimate the device’s position [16,59,60,61,62,63]. Tachikawa et al. [59] and IndoLabel [63] combine active and passive localization techniques to obtain impulse response features or MFCC coefficients by transmitting a Chirp sequence, and they combine features from multiple other sensors to achieve room-level accuracy for indoor localization. The closest to Echo-ID is EchoTag [13], which successfully employed active acoustic sensing to tag and remember indoor locations by enriching acoustic features with different delays Chirp signals. Recent studies [64,65] have proposed using active acoustic sensing to estimate room shapes.
The existing acoustic-based location sensing systems mentioned above consider the entire reflected signal as an acoustic fingerprint and necessitate the sound source to follow a predefined path or require additional microphone arrays to be installed in the room. Furthermore, these systems are restricted by the placement location and angle of smartphones, posing a challenge for users to remember and place their phones precisely, and their performance degrades considerably when environmental conditions change. In contrast, Echo-ID overcomes the limitations of existing acoustic-based location sensing systems by utilizing FMCW chirps to generate acoustic features for regions with the built-in speakers and microphones, and a deep learning model that is capable of handling placement uncertainty and environmental variations.

6. Conclusions and Future Work

The paper proposes Echo-ID, a low-cost region identification system that uses the built-in sensors available in commodity smartphones to identify regions and provide context-aware services without requiring additional sensors or pre-installed infrastructure. Echo-ID applies FMCW signals as a sensing medium and designs a deep learning network to recognize the target region. We implement Echo-ID on the Android platform and evaluate it with Xiaomi 12 Pro and Honor-10 smartphones, demonstrating its usability and robustness through extensive experiments. Echo-ID achieves an average accuracy of 94.6% for identifying the five regions which only requires users placing their phones in each target region at two orthogonal angles for 8.5 s.
We only explore the feasibility and robustness of region identification with STFT to extract the spatial features. We will investigate other types of discrete transforms for spatial feature extraction in the future. Moreover, the users need to place their phones on purpose at the region center’s two orthogonal angles for 8.5 s each to collect training data. Our future study will propose an online learning scheme that lets Echo-ID learn the placement regions automatically and only requires the user to decide whether make a combination between the learned region and contextual services.

Author Contributions

Writing—original draft preparation, X.J., Z.Z. and Z.L.; Writing—review and editing, F.H.; Funding acquisition, F.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC), grant no. 41976185.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in Section 4.1.2.

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable comments and helpful suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yan, Z.; Chakraborty, D.; Parent, C.; Spaccapietra, S.; Aberer, K. Semantic trajectories: Mobility data computation and annotation. ACM Trans. Intell. Syst. Technol. 2013, 4, 49:1–49:38. [Google Scholar] [CrossRef]
  2. Li, F.; Chen, H.; Song, X.; Zhang, Q.; Li, Y.; Wang, Y. CondioSense: High-quality context-aware service for audio sensing system via active sonar. Pers. Ubiquitous Comput. 2017, 21, 17–29. [Google Scholar] [CrossRef]
  3. Khanum, A.; Lee, C.Y.; Yang, C.S. Deep-Learning-Based Network for Lane Following in Autonomous Vehicles. Electronics 2022, 11, 3084. [Google Scholar] [CrossRef]
  4. Faisal, M.M.; Mohammed, M.S.; Abduljabar, A.M.; Abdulhussain, S.H.; Mahmmod, B.M.; Khan, W.; Hussain, A. Object Detection and Distance Measurement Using AI. In Proceedings of the 2021 14th International Conference on Developments in eSystems Engineering (DeSE), Sharjah, United Arab Emirates, 7–10 December 2021; pp. 559–565. [Google Scholar]
  5. Mohammed, M.S.; Abduljabar, A.M.; Faisal, M.M.; Mahmmod, B.M.; Abdulhussain, S.H.; Khan, W.; Liatsis, P.; Hussain, A. Low-cost autonomous car level 2: Design and implementation for conventional vehicles. Results Eng. 2023, 17, 100969. [Google Scholar] [CrossRef]
  6. Mayabrahmma, A.; Beesetty, Y.; Shadaab, K.; Vineet, K. Location-Based Services Market. 2021. Available online: https://www.alliedmarketresearch.com/location-based-services-market (accessed on 1 March 2023).
  7. Addlesee, M.; Curwen, R.W.; Hodges, S.; Newman, J.; Steggles, P.; Ward, A.; Hopper, A. Implementing a Sentient Computing System. Computer 2001, 34, 50–56. [Google Scholar] [CrossRef]
  8. Want, R.; Hopper, A.; Falcao, V.; Gibbons, J. The active badge location system. ACM Trans. Inf. Syst. 1992, 10, 91–102. [Google Scholar] [CrossRef]
  9. Azizyan, M.; Constandache, I.; Choudhury, R.R. SurroundSense: Mobile phone localization via ambience fingerprinting. In Proceedings of the ACM/IEEE International Conference on Mobile Computing and Networking, Beijing, China, 20–25 September 2009. [Google Scholar]
  10. Borriello, G.; Liu, A.L.; Offer, T.; Palistrant, C.; Sharp, R. WALRUS: Wireless acoustic location with room-level resolution using ultrasound. In Proceedings of the ACM SIGMOBILE International Conference on Mobile Systems, Applications, and Services, Singapore, 25–30 June 2005. [Google Scholar]
  11. Tarzia, S.P.; Dinda, P.A.; Dick, R.P.; Memik, G. Indoor localization without infrastructure using the acoustic background spectrum. In Proceedings of the ACM SIGMOBILE International Conference on Mobile Systems, Applications, and Services, Bethesda, MD, USA, 28 June–1 July 2011. [Google Scholar]
  12. Hightower, J.; Consolvo, S.; LaMarca, A.; Smith, I.E.; Hughes, J. Learning and Recognizing the Places We Go. In Proceedings of the Ubiquitous Computing, Nagasaki, Japan, 6–9 December 2005. [Google Scholar]
  13. Tung, Y.C.; Shin, K.G. EchoTag: Accurate Infrastructure-Free Indoor Location Tagging with Smartphones. In Proceedings of the 21st Annual International Conference on Mobile Computing and Networking, Paris, France, 7–11 September 2015. [Google Scholar]
  14. Tung, Y.C.; Bui, D.V.; Shin, K.G. Cross-Platform Support for Rapid Development of Mobile Acoustic Sensing Applications. In Proceedings of the 16th Annual International Conference on Mobile Systems, Applications, and Services, Munich, Germany, 10–15 June 2018. [Google Scholar]
  15. Cheng, H.; Lou, W. Push the Limit of Device-Free Acoustic Sensing on Commercial Mobile Devices. In Proceedings of the IEEE INFOCOM 2021—IEEE Conference on Computer Communications, Vancouver, BC, Canada, 10–13 May 2021; IEEE Press: Piscataway, NJ, USA, 2021; pp. 1–10. [Google Scholar]
  16. Kunze, K.; Lukowicz, P. Symbolic Object Localization Through Active Sampling of Acceleration and Sound Signatures. In Proceedings of the Ubiquitous Computing, Innsbruck, Austria, 16–19 September 2007. [Google Scholar]
  17. Zhou, Z.; Diao, W.; Liu, X.; Zhang, K. Acoustic Fingerprinting Revisited: Generate Stable Device ID Stealthily with Inaudible Sound. In Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security, Scottsdale, AZ, USA, 3–7 November 2014. [Google Scholar]
  18. Horai, M.; Kobayashi, H.; Nitta, T.G. Chirp Signal Transform and Its Properties. J. Appl. Math. 2014, 2014, 161989:1–161989:8. [Google Scholar] [CrossRef]
  19. Yang, Y.; Cao, J.; Liu, X.; Liu, X. Door-Monitor: Counting In-and-Out Visitors with COTS WiFi Devices. IEEE Internet Things J. 2020, 7, 1704–1717. [Google Scholar] [CrossRef]
  20. Abdulhussain, S.H.; Mahmmod, B.M.; Baker, T.; Al-Jumeily, D. Fast and accurate computation of high-order Tchebichef polynomials. Concurr. Comput. Pract. Exp. 2022, 34, e7311. [Google Scholar] [CrossRef]
  21. Heaton, J. Ian Goodfellow, Yoshua Bengio, and Aaron Courville: Deep learning. Genet. Program. Evolvable Mach. 2017, 19, 305–307. [Google Scholar] [CrossRef]
  22. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. arXiv 2015, arXiv:1512.07108. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  25. Zhou, B.; Khosla, A.; Lapedriza, À.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  26. GPS: The Global Positioning System. 2022. Available online: https://www.gps.gov/ (accessed on 20 January 2023).
  27. Chen, Y.; Lymberopoulos, D.; Liu, J.; Priyantha, B. FM-based indoor localization. In Proceedings of the ACM SIGMOBILE International Conference on Mobile Systems, Applications, and Services, Low Wood Bay Lake District, UK, 25–29 June 2012. [Google Scholar]
  28. Liu, K.; Liu, X.; Li, X. Guoguo: Enabling Fine-Grained Smartphone Localization via Acoustic Anchors. IEEE Trans. Mob. Comput. 2016, 15, 1144–1156. [Google Scholar] [CrossRef]
  29. Haeberlen, A.; Flannery, E.; Ladd, A.M.; Rudys, A.; Wallach, D.S.; Kavraki, L.E. Practical robust localization over large-scale 802.11 wireless networks. In Proceedings of the ACM/IEEE International Conference on Mobile Computing and Networking, Philadelphia, PA, USA, 15–18 September 2004. [Google Scholar]
  30. Youssef, M.; Agrawala, A.K. The Horus WLAN location determination system. In Proceedings of the ACM SIGMOBILE International Conference on Mobile Systems, Applications, and Services, Seattle, WA, USA, 6–8 June 2005. [Google Scholar]
  31. Vasisht, D.; Kumar, S.; Katabi, D. Decimeter-Level Localization with a Single WiFi Access Point. In Proceedings of the Symposium on Networked Systems Design and Implementation, Santa Clara, CA, USA, 16–18 March 2016. [Google Scholar]
  32. Labinghisa, B.A.; Lee, D.M. Indoor localization system using deep learning based scene recognition. Multimed. Tools Appl. 2022, 81, 28405–28429. [Google Scholar] [CrossRef]
  33. Lu, L.; Yu, J.; Chen, Y.; Wang, Y. VocalLock: Sensing Vocal Tract for Passphrase-Independent User Authentication Leveraging Acoustic Signals on Smartphones. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 51:1–51:24. [Google Scholar] [CrossRef]
  34. Zhou, M.; Wang, Q.; Yang, J.; Li, Q.; Xiao, F.; Wang, Z.; Chen, X. PatternListener: Cracking Android Pattern Lock Using Acoustic Signals. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018. [Google Scholar]
  35. Rathore, A.S.; Zhu, W.; Daiyan, A.; Xu, C.; Wang, K.; Lin, F.; Ren, K.; Xu, W. SonicPrint: A generally adoptable and secure fingerprint biometrics in smart devices. In Proceedings of the 18th International Conference on Mobile Systems, Applications, and Services, Toronto, ON, Canada, 15–19 June 2020. [Google Scholar]
  36. Dai, H.; Wang, W.; Liu, A.X.; Ling, K.; Sun, J. Speech Based Human Authentication on Smartphones. In Proceedings of the 2019 16th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON), Boston, MA, USA, 10–13 June 2019; pp. 1–9. [Google Scholar]
  37. Tan, J.; Nguyen, C.T.; Wang, X. SilentTalk: Lip reading through ultrasonic sensing on mobile phones. In Proceedings of the IEEE INFOCOM 2017—IEEE Conference on Computer Communications, Atlanta, GA, USA, 1–4 May 2017; pp. 1–9. [Google Scholar]
  38. Lu, L.; Yu, J.; Chen, Y.; Liu, H.; Zhu, Y.; Kong, L.; Li, M. Lip Reading-Based User Authentication Through Acoustic Sensing on Smartphones. IEEE/ACM Trans. Netw. 2019, 27, 447–460. [Google Scholar] [CrossRef]
  39. Tan, J.; Wang, X.; Nguyen, C.T.; Shi, Y. SilentKey: A New Authentication Framework through Ultrasonic-based Lip Reading. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 36:1–36:18. [Google Scholar] [CrossRef]
  40. Zhou, B.; Lohokare, J.; Gao, R.; Ye, F. EchoPrint: Two-factor Authentication using Acoustics and Vision on Smartphones. In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, New Delhi, India, 29 October–2 November 2018. [Google Scholar]
  41. Wang, X.; Huang, R.; Mao, S. SonarBeat: Sonar Phase for Breathing Beat Monitoring with Smartphones. In Proceedings of the 2017 26th International Conference on Computer Communication and Networks (ICCCN), Vancouver, BC, Canada, 31 July–3 August 2017; pp. 1–8. [Google Scholar]
  42. Wang, T.; Zhang, D.; Zheng, Y.; Gu, T.; Zhou, X.; Dorizzi, B. C-FMCW Based Contactless Respiration Detection Using Acoustic Signal. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 1, 1–20. [Google Scholar] [CrossRef]
  43. Qian, K.; Wu, C.; Xiao, F.; Zheng, Y.; Zhang, Y.; Yang, Z.; Liu, Y. Acousticcardiogram: Monitoring Heartbeats using Acoustic Signals on Smart Devices. In Proceedings of the IEEE INFOCOM 2018—IEEE Conference on Computer Communications, Honolulu, HI, USA, 16–19 April 2018; pp. 1574–1582. [Google Scholar]
  44. Xu, X.; Yu, J.; Chen, Y.; Zhu, Y.; Kong, L.; Li, M. BreathListener: Fine-grained Breathing Monitoring in Driving Environments Utilizing Acoustic Signals. In Proceedings of the 17th Annual International Conference on Mobile Systems, Applications, and Services, Seoul, Republic of Korea, 17–21 June 2019. [Google Scholar]
  45. Nandakumar, R.; Gollakota, S.; Watson, N.F. Contactless Sleep Apnea Detection on Smartphones. GetMobile Mob. Comput. Commun. 2015, 19, 22–24. [Google Scholar] [CrossRef]
  46. Du, H.; Li, P.; Zhou, H.; Gong, W.; Luo, G.; Yang, P. WordRecorder: Accurate Acoustic-based Handwriting Recognition Using Deep Learning. In Proceedings of the IEEE INFOCOM 2018—IEEE Conference on Computer Communications, Honolulu, HI, USA, 16–19 April 2018; pp. 1448–1456. [Google Scholar]
  47. Kim, H.; Byanjankar, A.; Liu, Y.; Shu, Y.; Shin, I. UbiTap: Leveraging Acoustic Dispersion for Ubiquitous Touch Interface on Solid Surfaces. In Proceedings of the 16th ACM Conference on Embedded Networked Sensor Systems, Shenzhen, China, 4–7 November 2018. [Google Scholar]
  48. Chen, M.; Lin, J.; Zou, Y.; Ruby, R.; Wu, K. SilentSign: Device-free Handwritten Signature Verification through Acoustic Sensing. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications (PerCom), Austin, TX, USA, 23–27 March 2020; pp. 1–10. [Google Scholar]
  49. Zou, Y.; Yang, Q.; Ruby, R.; Han, Y.; Wu, S.; Li, M.; Wu, K. EchoWrite: An Acoustic-based Finger Input System Without Training. In Proceedings of the 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS), Dallas, TX, USA, 7–9 July 2019; pp. 778–787. [Google Scholar]
  50. Zou, Y.; Yang, Q.; Han, Y.; Wang, D.; Cao, J.; Wu, K. AcouDigits: Enabling Users to Input Digits in the Air. In Proceedings of the 2019 IEEE International Conference on Pervasive Computing and Communications (PerCom 2019), Kyoto, Japan, 11–15 March 2019; pp. 1–9. [Google Scholar]
  51. Xiao, Z.; Chen, T.; Liu, Y.; Li, Z. Mobile Phones Know Your Keystrokes through the Sounds from Finger’s Tapping on the Screen. In Proceedings of the 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS), Singapore, 29 November–1 December 2020; pp. 965–975. [Google Scholar]
  52. Lu, L.; Yu, J.; Chen, Y.; Zhu, Y.; Xu, X.; Xue, G.; Li, M. KeyLiSterber: Inferring Keystrokes on QWERTY Keyboard of Touch Screen through Acoustic Signals. In Proceedings of the IEEE INFOCOM 2019—IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; pp. 775–783. [Google Scholar]
  53. Wang, Y.; Shen, J.; Zheng, Y. Push the Limit of Acoustic Gesture Recognition. In Proceedings of the IEEE INFOCOM 2020—IEEE Conference on Computer Communications, Virtual, 6–9 July 2020; pp. 566–575. [Google Scholar]
  54. Ling, K.; Dai, H.; Liu, Y.; Liu, A.X.; Wang, W.; Gu, Q. UltraGesture: Fine-Grained Gesture Sensing and Recognition. IEEE Trans. Mob. Comput. 2022, 21, 2620–2636. [Google Scholar]
  55. Sun, K.; Zhao, T.; Wang, W.; Xie, L. VSkin: Sensing Touch Gestures on Surfaces of Mobile Devices Using Acoustic Signals. In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, New Delhi, India, 29 October–2 November 2018. [Google Scholar]
  56. Wirz, M.; Roggen, D.; Tröster, G. A wearable, ambient sound-based approach for infrastructureless fuzzy proximity estimation. In Proceedings of the International Symposium on Wearable Computers (ISWC) 2010, Seoul, Republic of Korea, 10–13 October 2010; pp. 1–4. [Google Scholar]
  57. Shen, S. Voice localization using nearby wall reflections. In Proceedings of the 26th Annual International Conference on Mobile Computing and Networking, London, UK, 21–25 September 2020. [Google Scholar]
  58. Wang, W.; Li, J.; He, Y.; Liu, Y. Symphony: Localizing multiple acoustic sources with a single microphone array. In Proceedings of the 18th Conference on Embedded Networked Sensor Systems, Virtual Event, Japan, 16–19 November 2020. [Google Scholar]
  59. Tachikawa, M.; Maekawa, T.; Matsushita, Y. Predicting location semantics combining active and passive sensing with environment-independent classifier. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016. [Google Scholar]
  60. Song, Q.; Gu, C.; Tan, R. Deep Room Recognition Using Inaudible Echos. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 135. [Google Scholar] [CrossRef]
  61. Chen, C.; Ren, Y.; Liu, H.; Chen, Y.; Li, H. Acoustic-Sensing-Based Location Semantics Identification Using Smartphones. IEEE Internet Things J. 2022, 9, 20640–20650. [Google Scholar] [CrossRef]
  62. Rossi, M.; Seiter, J.; Amft, O.; Buchmeier, S.; Tröster, G. RoomSense: An indoor positioning system for smartphones using active sound probing. In Proceedings of the International Conference on Adaptive Hypermedia and Adaptive Web-Based Systems, Stuttgart, Germany, 7–8 March 2013. [Google Scholar]
  63. Dissanayake, T.; Maekawa, T.; Hara, T.; Miyanishi, T.; Kawanabe, M. IndoLabel: Predicting Indoor Location Class by Discovering Location-Specific Sensor Data Motifs. IEEE Sens. J. 2022, 22, 5372–5385. [Google Scholar] [CrossRef]
  64. Pradhan, S.; Baig, G.; Mao, W.; Qiu, L.; Chen, G.; Yang, B. Smartphone-based Acoustic Indoor Space Mapping. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 75. [Google Scholar] [CrossRef]
  65. Zhou, B.; Elbadry, M.; Gao, R.; Ye, F. BatMapper: Acoustic Sensing Based Indoor Floor Plan Construction Using Smartphones. In Proceedings of the 15th Annual International Conference on Mobile Systems, Applications, and Services, Niagara Falls, NY, USA, 19–23 June 2017. [Google Scholar]
Figure 1. Examples of Echo-ID user scenarios. (a) Automatic activation of silent mode when placed on a nightstand and (b) automatic initiation of focus mode when placed on a work desk.
Figure 1. Examples of Echo-ID user scenarios. (a) Automatic activation of silent mode when placed on a nightstand and (b) automatic initiation of focus mode when placed on a work desk.
Sensors 23 04302 g001
Figure 2. Sketch of transmission and reception of FMCW signal.
Figure 2. Sketch of transmission and reception of FMCW signal.
Sensors 23 04302 g002
Figure 3. Uneven signal attenuation at different frequencies and the corresponding reasons.
Figure 3. Uneven signal attenuation at different frequencies and the corresponding reasons.
Sensors 23 04302 g003
Figure 4. System overview of Echo−ID.
Figure 4. System overview of Echo−ID.
Sensors 23 04302 g004
Figure 5. Transmitted Chirp. (a) Time domain and (b) Spectrogram.
Figure 5. Transmitted Chirp. (a) Time domain and (b) Spectrogram.
Sensors 23 04302 g005
Figure 6. The start time of the received Chirp.
Figure 6. The start time of the received Chirp.
Sensors 23 04302 g006
Figure 7. Orthogonal demodulation.
Figure 7. Orthogonal demodulation.
Sensors 23 04302 g007
Figure 8. Five typical regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Figure 8. Five typical regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Sensors 23 04302 g008
Figure 9. Frequency spectrum of five regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Figure 9. Frequency spectrum of five regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Sensors 23 04302 g009
Figure 10. Spectrogram of five regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Figure 10. Spectrogram of five regions. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet.
Sensors 23 04302 g010
Figure 11. Finer feature extraction and region identification network structure.
Figure 11. Finer feature extraction and region identification network structure.
Sensors 23 04302 g011
Figure 12. Spectrogram of five placement modes in the kitchen operating table region and visualization of features. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet. (f) Concatenated feature values of different placement modes. (g) Attention area of (b). (h) Attention area of (c).
Figure 12. Spectrogram of five placement modes in the kitchen operating table region and visualization of features. (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet. (f) Concatenated feature values of different placement modes. (g) Attention area of (b). (h) Attention area of (c).
Sensors 23 04302 g012
Figure 13. Visualization of features after input Figure 10 into the model. (a) Concatenated feature values of different regions. (b) Attention area of Figure 10b. (c) Attention area of Figure 10d.
Figure 13. Visualization of features after input Figure 10 into the model. (a) Concatenated feature values of different regions. (b) Attention area of Figure 10b. (c) Attention area of Figure 10d.
Sensors 23 04302 g013
Figure 14. Phone placement for experimental data collection. (a) Collecting training traces at 2 orthogonal angles. (b) Collecting test traces at 16 positions. (c) Collecting test traces at 12 angles.
Figure 14. Phone placement for experimental data collection. (a) Collecting training traces at 2 orthogonal angles. (b) Collecting test traces at 16 positions. (c) Collecting test traces at 12 angles.
Sensors 23 04302 g014
Figure 15. Confusion matrices. (a) Echo-ID. (b) EchoTag. (c) Echo-ID’s coarse feature + SVM.
Figure 15. Confusion matrices. (a) Echo-ID. (b) EchoTag. (c) Echo-ID’s coarse feature + SVM.
Sensors 23 04302 g015
Figure 16. Identification accuracy (a) of 16 positions (b) of 12 angles (c) with environmental changes.
Figure 16. Identification accuracy (a) of 16 positions (b) of 12 angles (c) with environmental changes.
Sensors 23 04302 g016
Figure 17. Attention results: amplitude of (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet; phase of (f) Nightstand. (g) Kitchen operating table. (h) Desk. (i) Shelf. (j) TV cabinet.
Figure 17. Attention results: amplitude of (a) Nightstand. (b) Kitchen operating table. (c) Desk. (d) Shelf. (e) TV cabinet; phase of (f) Nightstand. (g) Kitchen operating table. (h) Desk. (i) Shelf. (j) TV cabinet.
Sensors 23 04302 g017
Figure 18. Identification accuracy of different features. (a) Accuracy of each region. (b) Average accuracy.
Figure 18. Identification accuracy of different features. (a) Accuracy of each region. (b) Average accuracy.
Sensors 23 04302 g018
Figure 19. Identification accuracy of different numbers of residual blocks. (a) Accuracy of each region. (b) Average accuracy.
Figure 19. Identification accuracy of different numbers of residual blocks. (a) Accuracy of each region. (b) Average accuracy.
Sensors 23 04302 g019
Figure 20. Identification accuracy. (a) Using traces collected from multiple angles for training. (b) Using traces collected from multiple positions for training. (c) Different phone model. (d) Permanence.
Figure 20. Identification accuracy. (a) Using traces collected from multiple angles for training. (b) Using traces collected from multiple positions for training. (c) Different phone model. (d) Permanence.
Sensors 23 04302 g020
Figure 21. Experimental scenarios for other five regions. (a) Corridor. (b) Washbasin. (c) Washing machine. (d) Classroom. (e) Conference room.
Figure 21. Experimental scenarios for other five regions. (a) Corridor. (b) Washbasin. (c) Washing machine. (d) Classroom. (e) Conference room.
Sensors 23 04302 g021
Figure 22. Confusion matrix of ten regions identification.
Figure 22. Confusion matrix of ten regions identification.
Sensors 23 04302 g022
Table 1. Configuring the parameters and data flow dimensions for every module within the finer feature extraction and region identification network.
Table 1. Configuring the parameters and data flow dimensions for every module within the finer feature extraction and region identification network.
Model NameInput DimensionParameter SettingsOutput Dimension
Conv 11 × 256 × 25632 × 3 × 332 × 256 × 256
ResBlock 132 × 256 × 25632 × 3 × 332 × 256 × 256
MaxPooling32 × 256 × 2562 × 232 × 128 × 128
Conv 232 × 128 × 12832 × 3 × 332 × 128 × 128
ResBlock 232 × 128 × 12832 × 3 × 332 × 128 × 128
MaxPooling32 × 128 × 1282 × 232 × 64 × 64
Conv 332 × 64 × 6432 × 3 × 332 × 64 × 64
ResBlock 332 × 64 × 6432 × 3 × 332 × 64 × 64
MaxPooling32 × 64 × 642 × 232 × 32 × 32
Conv 432 × 32 × 3232 × 3 × 332 × 32 × 32
AdaptiveMaxPool + Flatten32 × 32 × 32-32
Feature Fusion32 (Amplitude feature)
32 (Phase feature)
-64
MLP6464 × 512
512 × 1024
1024 × 512
512 × 5
5 (Probability of classifying 5 regions)
Total number of parameters180M
Table 2. Acoustic Feature.
Table 2. Acoustic Feature.
Feature IFeature IIFeature IIIFeature IV
received signal spectrumbaseband signal spectrumbaseband signal spectrogramspectrum and spectrogram of baseband signal
Table 3. Index of 10 regions.
Table 3. Index of 10 regions.
IndexRegionIndexRegion
1nightstand6corridor
2kitchen operating table7washbasin
3desk8washing machine
4shelf9classroom
5TV cabinet10wonference room
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, X.; Zhao, Z.; Li, Z.; Hong, F. Echo-ID: Smartphone Placement Region Identification for Context-Aware Computing. Sensors 2023, 23, 4302. https://doi.org/10.3390/s23094302

AMA Style

Jiang X, Zhao Z, Li Z, Hong F. Echo-ID: Smartphone Placement Region Identification for Context-Aware Computing. Sensors. 2023; 23(9):4302. https://doi.org/10.3390/s23094302

Chicago/Turabian Style

Jiang, Xueting, Zhongning Zhao, Zhiyuan Li, and Feng Hong. 2023. "Echo-ID: Smartphone Placement Region Identification for Context-Aware Computing" Sensors 23, no. 9: 4302. https://doi.org/10.3390/s23094302

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop