Next Article in Journal
Genetic Algorithms to Maximize the Relevant Mutual Information in Communication Receivers
Next Article in Special Issue
Embedded System-Based Sticky Paper Trap with Deep Learning-Based Insect-Counting Algorithm
Previous Article in Journal
STERapp: Semiautomatic Software for Stereological Analysis. Application in the Estimation of Fish Fecundity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Smart-Object-Based Reasoning System for Indoor Acoustic Profiling of Elderly Inhabitants

1
School of Computing Sciences, University of East Anglia, Norwich NR4 7TJ, UK
2
Department of Electronic Engineering, Royal Holloway, University of London, Egham TW20 0EX, UK
3
School of Computer Science and Electronic Engineering, Universtiy of Essex, Colchester CO4 3SQ, UK
4
Hong Kong Applied Science and Technology Research Institute, Hong Kong, China
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(12), 1433; https://doi.org/10.3390/electronics10121433
Submission received: 7 May 2021 / Revised: 6 June 2021 / Accepted: 11 June 2021 / Published: 15 June 2021
(This article belongs to the Special Issue AI for Embedded Systems)

Abstract

:
Many countries are facing significant challenges in relation to providing adequate care for their elderly citizens. The roots of these issues are manifold, but include changing demographics, changing behaviours, and a shortage of resources. As has been witnessed in the health sector and many others in society, technology has much to offer in terms of supporting people’s needs. This paper explores the potential for ambient intelligence to address this challenge by creating a system that is able to passively monitor the home environment, detecting abnormal situations which may indicate that the inhabitant needs help. There are many ways that this might be achieved, but in this paper, we will describe our investigation into an approach involving unobtrusively ’listening’ to sound patterns within the home, which classifies these as either normal daily activities, or abnormal situations. The experimental system we built was composed of an innovative combination of acoustic sensing, artificial intelligence (AI), and the Internet-of-Things (IoT), which we argue in the paper that it provides a cost-effective approach to alerting care providers when an elderly person in their charge needs help. The majority of the innovation in our work concerns the AI in which we employ Machine Learning to classify the sound profiles, analyse the data for abnormal events, and to make decisions for raising alerts with carers. A Neural Network classifier was used to train and identify the sound profiles associated with normal daily routines within a given person’s home, signalling departures from the daily routines that were then used as templates to measure deviations from normality, which were used to make weighted decisions regarding calling for assistance. A practical experimental system was then designed and deployed to evaluate the methods advocated by this research. The methodology involved gathering pre-design and post-design data from both a professionally run residential home and a domestic home. The pre-design data gathered the views on the system design from 11 members of the residential home, using survey questionnaires and focus groups. These data were used to inform the design of the experimental system, which was then deployed in a domestic home setting to gather post-design experimental data. The experimental results revealed that the system was able to detect 84% of abnormal events, and advocated several refinements which would improve the performance of the system. Thus, the research concludes that the system represents an important advancement to the state-of-the-art and, when taken together with the refinements, represents a line of research which has the potential to deliver significant improvements to care provision for the elderly.

1. Introduction

The number of people aged over 60 is expected to quadruple within the next two decades [1], placing existing care provision capabilities under strain and driving the need to research into finding more effective and efficient ways of delivering care. Research [2] has shown the vast majority of elderly people prefer to stay in their own homes whenever possible, rather than living in residential care or nursing homes. This view has further been reinforced with the emergence of the COVID-19 pandemic, which highlighted the vulnerabilities of residential care homes [3]. Thus, the number of people who will need assistance to remain living independently in their own homes is expected to rise sharply in the coming years. It has long been recognised that technology can play a vital part in supporting people’s desire to maintain an independent life as they grow older. For example, sensors and actuators, together with intelligent software, can augment everyday objects to provide additional functionalities to support those in need of care. Products that combine embedded computers, sensors/actuators, and intelligent software are referred to as ‘smart objects’, and form a fundamental building block of the Internet-of-Things (IoT) [4]. Technological advancement in the form of machine learning [5] has enabled these smart objects to work collaboratively, as a whole, creating intelligent environments [6] that can learn how to adapt their collective behaviour to meet the needs of their occupants. As a result, IoT can be used to foster a uniquely supportive environment, which is particularly useful for elderly care settings, such as assisted living systems, where the environment can be aware of and sensitive to the particular needs of its occupants [7,8,9].
While much thought has gone into creating effective financial and care models, a significant ongoing research challenge is how best to lever advances in smart technologies, to develop context-aware systems which promote the well-being of elderly people by improving their day-to-day living experiences. Towards those ends, monitoring a person’s physical state is an effective approach, as it can establish the functional status of an individual which, in turn, can be used to determine the level of assistance required, with minimum intervention [10,11,12]. In systems that monitor and gather behavioural data about people, inference techniques are commonly used [10,13], as no prior knowledge about the individuals is required. While applications designed to help older people to perform their day-to-day tasks have increasingly gained recognition [14,15], the vast majority of these applications rely on “active monitoring” approaches to deliver key information. Generally, monitoring is classified as either active or passive. Active devices require physical user interaction to function, while passive devices monitor the local environment in autonomous fashion. Active devices are regarded as being intrusive, interfering with the user’s normal routine in order to function, whereas passive devices are seen as non-intrusive, since they place no physical or mental load on the user. Moreover, active devices, such as an alert button, have the disadvantage of requiring the user to be physically conscious to operate it, which, depending on the status of the individual at the time when assistance is required, may not be possible. While active participation can offer benefits within a tele-medical monitoring system [16,17], we argue that using passive or ambient environmental monitoring has the potential to develop a comprehensive assisted living scenario, in a less intrusive and less costly way. This can be achieved by combining Artificial Intelligence (AI) with IoT to create an environment that responds to the occupant’s needs without reliance on their manual input, which overcomes issues such as incapacity, by exploiting the capability of passive seniors [7]. Developments in ambient assisted monitoring include fall-detection and ‘lack of mobility’ alerts which, for example, involve using passive infrared (PIR) motion sensors, which work in tandem with accelerometers built into furniture and flooring [13]. Other techniques, such as ubiquitous sensing, which are driven by a specific knowledge-driven framework, have also shown promising results [18]. Thus, it has been recognized that pervasive passive monitoring is a good way forward to enable elderly people to sustain greater independence and enjoy a better quality of life [19,20]. The pervasive nature of noise in the modern domestic environment means that monitoring ambient sound levels offers good potential for the detection of abnormal patterns of behaviour. Furthermore, since this approach needs only a few strategically placed sound sensors around a person’s home, it is easy and cheap to install. However, while this is a most promising approach, there is a sparsity of research on this topic with no off-the-shelf solutions being available. Furthermore, the few existing examples (discussed in Section 2) generally have built-in transceivers, with many being designed to be carried by users, which limit their practicality.
An overview of the most common unobtrusive wellness monitoring methods is given in [21], but to be accurate, they need to be backed by ontological-based algorithms [22], especially when the activities to be recognised are similar [23], issues that are even more critical when designers have access to only a reduced training data set.
Thus, investigating these issues with the aim of overcoming existing barriers (especially the ones given by the lack of training data) is the main focus of the work reported in this paper; more precisely, whether sound (acoustic signals) can be used as the primary ambient source to create a relatively low-cost IoT system for identifying abnormal situations in domestic homes where elderly people may need external assistance.
Clearly, the focus of any research project (in our case, the AI methods) will contain bespoke and novel methods and technologies. However, in research that is aimed at practical deployment or commercial exploitation, for the findings to be convincing to a broader base (e.g., the industry), it is helpful if supporting technologies and standards utilised are ones which are readily available to those that might seek to exploit and deploy the work. Thus, in our work, for such secondary platforms, we sought to build on widely used industry standards. For networking, we used the IEEE 802.15.4 protocol, which is the most popular wireless sensor network standard. Likewise, for Middleware, we chose the popular open (and free) standard LinkSmart, developed by the FP7 EU project EBBITS. In this way, the debate on our results (theoretical and practical contributions) was better able to be appreciated by the wider ambient intelligence community. The system functionality requirements were gathered by consulting residents in a residential home for the elderly operated by a UK local government authority. Initial experiments were conducted in a domestic home to assess the capability of the sound sensors for gathering suitable data (the distribution of sensors and its effect on live/dead spots) in areas of interest (e.g., kitchens) based on typical events (e.g., walking, boiling a kettle) that relate to care of elderly people. In this work, we employed a Neural Network classifier combined with a context reasoning system to analyse and classify the acoustic data. In addition, we employed a goal-based approach to determine whether an alert should be triggered when an abnormal event was detected.
The remainder of this paper is structured as follows: Section 2—Low-Cost Wireless Sensor IoT Network Design; Section 3—Methodology, Design and Implementation; Section 4—System Prototyping and Discussions; Section 5—Experimental Setup and Results Analysis; Section 6—Conclusions.

2. Low-Cost Wireless Sensor IoT Network Design

The Internet-of-Things (IoT) comprises a network of embedded networked computers (implanted into everyday ‘things’) that are able to sense the environment, as well as communicate and interact with people and each other [24]. The falling costs of embedded devices and advanced networking has led to a proliferation of IoT applications in industries, cities, homes, and the natural world, all of which enhance people’s lives [25]. Consequently, the ‘assisted living’ community are investing much ongoing research and development into the use of IoT. For example, an important area of IoT research needed to realise the ‘assisted living’ vision is inter-device communication, especially that relating to the development of the wireless sensor network (WSN). WSNs are networks that consist of multiple sensor nodes, with a gateway and a wireless transmission path. Generally, node communication is supported by technologies such as Wi-Fi, Low Energy Bluetooth, ZigBee, 6LoWPAN, and 4/5G.
Table 1 presents a comparison of four main wireless technologies (Wi-Fi, Bluetooth, ZigBee, Zwave) typically used to create WSN-based smart environments, a function of six characteristics: (1) standards, (2) indoor/smart home market share, (3) data throughput, (4) range, (5) reliability, and (6) ease of use. It can be noticed that the Wi-Fi protocol offers the highest data throughput for the transmission range (∼10 m) required in this study, and therefore, the lowest communication latency and highest.
In addition to the connectivity requirements, another key IoT element that requires special attention when designing assisted living systems is the environmental monitoring module [26,27].
To date, most of the approaches taken have been fairly conventional, relying on sensor networks to measure parameters such as temperature or humidity [28,29,30,31] to predict and maintain occupant comfort [26]. The use of IoT ambient acoustic systems for human activity prediction and recognition is somewhat rare in environmental monitoring. For example, in 1995, Reynolds [32] investigated the Gaussian Mixture Models (GMM) technique to predict users’ activities or events. Environmental sound generated from users’ activities/events has been explored by a number of scholars, including Virone and Istrate’s work in 2007 [33], which used an environmental sound sensor for monitoring home activities using a simulator. Later in 2009, Shaikh et al. [34] investigated sound virtualization in a virtual world where they tried to capture environmental sound together with interpreting associated activities. While Virone and Istrate’s research demonstrated the validity via simulation, they did not investigate the implementation of IoT. Virone and Istrate’s work in identification of the types of sound was further developed by Segura–Garcia et al. [35] eight years later in 2015, where the new team focused on analysing urban noise levels. Segura–Garcia et al. used Zwicker’s annoyance model to analyse road traffic noise in the City of Barcelona, utilising a WSN similar to the technological set-up of the IoT application we developed in this study [35]. However, our settings and focus differ to this earlier work in that our application has aimed at a domestic home environment occupied by elderly people. Nevertheless, the conclusions drawn by Segura–Garcia et al. [35] added weight to our hypothesis that an acoustic sensing approach had the potential to deliver a low-cost solution to the provision of care within a domestic setting.

3. Methodology, Design and Implementation

The main goal of this study was to identify sequences of events, and their meaning, that occur in domestic environments based on analysing and contextualising ambient acoustic signals. Subsequently, our study is divided into three areas. The first area is related to understanding the needs of the elderly; the second area is related to the technique used to identify and profile the acoustic signals; and the final area is related to the context system for improving overall system performance. The sections below describe these areas in more detail:

3.1. Understanding the Needs of the Elderly

To better understand the needs of the elderly, we utilised a core-task analysis (CTA) approach developed by Norros [28] and Pesonen et al. [36].

3.1.1. The Analysis Process

The methodology involved analysing the views and preferences of the target end-users and mapping this information into usability requirements, which were then embedded into the design process. The process comprised of three phases:
  • Phase 1: Science-based modelling of core tasks. This relied on gathering expert knowledge on elderly care practices within the UK, drawn from scientific and professional literature. The goal was to collect important information regarding the care needs of elderly people.
  • Phase 2: Consultations with residents and staff in a residential home for elderly people in the UK to identify residents who would be suitable and willing to participate in our research.
  • Phase 3: Identify habitual home activities in order to inform/train the analytic system to classify them, and then infer “abnormal” activities at a later deployment stage.
Our study focused on residential homes for elderly people run by county councils in the UK. The elderly homes studied in this research were a small sample of the 462,000 units of accommodation provided by specialist housing for older people in the social sector across England, Scotland and Wales. The residential homes in our study had either one or two occupants. The local council concerned (Cambridgeshire) recognised that the ageing population had become the key players in the housing market and that they needed to provide better and more cost-effective support to this sector via the use of innovative technology. This policy was echoed in an earlier report [37] published by the UK Department for Work and Pensions (DWP) and the Ministry of Housing, Communities and Local Government (MHCLG) in 2016, highlighting the need to deliver housing solutions to meet older people’s needs through, for example, supporting them to live independently in their own homes for as long as possible. The following section describes the process of end-user data-gathering and preliminary results.

3.1.2. Consultation and Focus Group Interviews

Two separate sessions with the inhabitants of the residential homes were undertaken; one being administration of a questionnaire to identify their usual habitual activities, the second being an interview to assess their attitudes towards the acceptance of technology. As the target group for these sessions were elderly people, the interviews were designed to be broad, “light-hearted”, and conducted in a manner that was as open and sensitive as possible. The questionnaires were prepared based on an earlier questionnaire format used in Oxford Time-Use Research (daily activities diary). In this study, household activities were divided into three categories: kitchen, living room, and other. Each category had between 6 to 16 pre-listed common activities in addition to an open-ended question labelled “other” for participants to complete if they could not find a particular activity that was itemised (Table 2). To cater for age-related vision problems, the questionnaire’s font size was set to 20 points. The interview sessions were conducted in the community room within the residential home complex. No recordings were made, apart from note-taking. The procedure was as follows:
  • Residents of the elderly homes were invited to attend the sessions.
  • Residents were briefed about the study and asked to give their consent.
  • Interviews (first session: 1 h)
  • Questionnaires (second session: 1 h)
Eleven elderly residents participated in each of the sessions, with ages ranging from 60 to 95. Residents were asked for their views on assisted technology and if they had any concerns. They were also asked to record how many times during the day they thought they had conducted the activities listed on the questionnaire. Concerning the views on assisted technology, preliminary results showed that among the 11 elderly people, six expressed the view that they were quite happy to accept help from technology when needed; three had reservations; while two totally rejected the idea. Further analysis showed that for those who expressed reservations in accepting assisted technologies, cost was the main issue, followed closely by privacy.
In terms of household activities, preliminary results showed that the most frequent activity done in the kitchen area was making tea/coffee, whereas in the living room area it was watching TV while sitting on a sofa. Because making tea/coffee involves readily identifiable discrete activities, such as filling up the kettle, switching the kettle on, and waiting for the kettle to boil, the kitchen area was chosen as the experimental focus for our study.

3.2. Neural Network Classifier

Detection of anomalies in ambient sound is a special form of the speech recognition problem. A conventional solution is to train a series of classifiers to map some sound features to the events of interest. One way to train the classifiers is by using neural networks (NN). A NN is a computational model inspired by the structure of a biological brain. It consists of a large number of simple neural units. Each neural unit is connected with many others through weighted strengths that can be increased or reduced to affect the computed value of the adjoining neural units.

3.2.1. Neural Network Classifier Description

The chosen neural algorithm, commonly used in a variety of applications, is the Feed-Forward (FF) algorithm, briefly described through Equations (1) and (2). In the inference phase, the FF algorithm starts with the computation of the output value of each of the neurons as a weighted summation of all inputs (x), with their corresponding weights (w), that is, the net value, (1) followed by the firing of the net value with an activation function, such as the sigmoid function, (2).
Output ( net ) = f ( bias + k = 1 N w k x k )
Output ( net ) = 1 1 + e ( bias + k = 1 N w k x k ) .
In order for the classifier to be able to recognize patterns, it has to be trained. There are many training algorithms available for FF NNs, and their selection is done according to the data to be classified. For large-scale data, one of the most appropriate learning algorithms is the scaled conjugate gradient method. The algorithm is based on the standard training process that minimizes the error function E, defined as the squared difference between the actual output value of the output layer neuron for the applied pattern and t, the target output value, (3), by adjusting the neurons’ weights (for this, a search direction, η , and a step size, ε , are set), such as the new error function (calculated based on the new weights) is less than the previous error function (4).
E ( w ) = k ( output k t k ) 2
E ( w k + ε k η k ) < E ( w k ) .
One option is to set the search in the steepest descent direction that corresponds to the negative gradient of the error function in respect to the weights (the gradient descendent algorithm), but in the context of large-scale NNs, the algorithm shows poor convergence and time-consuming line search. However, if the search direction uses information from the second-order approximation of the error function, combining the Levenberg–Marquardt method with the conjugate gradient approach, the training time can be drastically reduced. A detailed explanation of the algorithm is given by its author in [38].
NNs have been used to solve a wide variety of tasks. For example, in 2013, Swarnkar et al. [39] explored this method for classifying sound signals to auto-detect human coughs. The same technique was recently used in a wearable device to detect sound from a heartbeat [40] and to monitor a person’s health. In hardware, Boes et al. [41] investigated NN learning models on low-cost devices to detect volatile events by learning signalling patterns. In their study, the authors discovered that better results could be achieved by providing greater flexibility and accommodation (in the learning model) but without losing prior knowledge. Recently, Ramanujam et al. investigated NN deep-learning models for human activity recognition, with specifically targeted mobile and wearable devices. However, the study was limited to pre-existing datasets instead of using a real-time data acquisition method [42]. The section below describes the NN classifier design and implementation.

3.2.2. Neural Network Classifier Design and Implementation

In our study, we trained a series of NN classifiers to identify the sound of the most significant events that occur in the daily life of older people. We determined these events from our focus group interviews, which revealed that the most important sounds include sudden bangs, footsteps, water from taps, kettles boiling, and toilets flushing.
First, we collected data through making our own recordings of the sounds (in WAV format) from the home of a volunteer (see Section 5.1), as well as the FreeSound website. In this way, we successfully obtained more than 36,000 s of sound for training and validation purposes. The main objective was to develop a proof-of-concept for classifying acoustic scenes; hence, we used standard feed-forward neural networks (FFNN) with mel-frequency cepstral coefficients (MFCC) as audio features. For this, Matlab programs have been developed to calculate the MFCCs used to train, validate, and test the FFNN.
Data sound collected through recordings and downloaded from the FreeSound website had a sampling frequency of 44.1 kHz, 24 bit depth on one or two channels. For calculating the MFCCs, the short-time Fourier transform has been initially applied on a window size of 2048 samples (46 ms) with a shift of 1024 samples (23 ms). The mel-spectrum was obtained by applying a mel filter bank of 128 bandpass filters in the range of 300 Hz to 22 kHz. The cepstral analysis was later performed, and 13 cepstral coefficients (including the 0th order) were obtained for each window analysed. The audio features were then stored in a data matrix of 130 rows, (which corresponded to 13 MFCC features/time bin multiplied by 10-time bins) and N columns, where N is the corresponding number of frames (windows) of the sound analysed. Each matrix value had been pre-processed by subtracting the minimum value of its corresponding column and normalizing it by dividing it by its column standard deviation. Consequently, the FFNN design consisted of two layers: an input layer with 130 neurons, a hidden layer with 10 neurons, and an output layer with two neurons with the sequence of ‘10’ for true (the sound belongs to the classifier) and ‘01’ for false (the sound does not belong to the classifier). The training algorithm applied was the scaled conjugate gradient backpropagation with a sigmoid activation function on the hidden layer, and a linear one on the output layer. On occasions when more than one classifier gave a positive detection, the one with the highest confidence value was chosen. For cases where all classifiers gave a negative detection (‘01’), the sound was considered to be ambient noise.

3.3. Context Reasoning System and Implementation

In addition to the NN classifier, we also developed a context reasoning system to improve the monitoring performance. In many previous studies of homes monitoring [9,28,43], events were considered as being independent from each other, and therefore an abnormal event was defined as a single event, such as a sudden bang due to falling down or breaking glass. However, in reality, we argue that events are related to each other. For example, before the kettle boils, we can expect for there to be some sound from the tap (to fill the kettle with water); before the sound of the tap, we can expect some footstep sounds (indicating someone entering the kitchen or the area where the kettle is located). This chain of events provides a contextual clue as to what will happen next, and more certainty concerning any particular event. Therefore, an abnormal event is not a single event, but can be viewed as a “surprise” out of an otherwise normal routine of events. This concept is very important for improving the performance of home monitoring. Previously, Chik et al. [44] used this contextual reasoning system to predict prospective risks at home, in a robotic observation system with positive results. Others, in order to overcome the limitations imposed by the large data set requirement to train the supervised classifiers, have used unsupervised approaches in conjunction with ontological (translated in a Markov Logic Network) methods, as reported in [22]. In the current study, we employed the context reasoning system applied on NN models trained with a low data set to achieve activity-increased recognition system reliability.
The data structure for the context reasoning system was developed as shown in Table 3. The first column shows different events that have been recognised by the NN classifier described in Section 3.2.2.
As explained above, we defined the events manually, based on the interviews with occupants of a residential home for the elderly. For the purposes of this paper, our system recognised a fixed number of habitual events, but our plan for the future is to develop a novel detector that will learn and create dynamic events automatically. The second column in the table shows the maximum duration of the event. This was learnt from the training data gathered from the home of a volunteer (Section 3.1.2). The third, fourth, and fifth columns show the top three next most likely events (a pragmatic choice). For this, we counted the frequency of occurrence between subsequent events, and obtained the top three events with the highest likelihood to happen next (A, B, C). Thus, for example, Table 3 shows that, if the current event is tap water, then the next event is likely to be the kettle boiling. Another possibility is, after the sound of tap water, the next event might be a footstep. A total of 1000 s of audio data were collected from the participant’s home. The data contained sequences of events. The sequences were split into individual clips and fed into a NN classifier. The results were then used to construct the context data and train the system. After training, we deployed the system to assess its ability to generate alerts for potential accidents. We derived the alarm criteria as:
  • If the next event is not as expected, then send an alarm;
  • If the duration of the current event is longer than 1.5 times the maximum duration in the context table, then send an alarm.
The alarm is delivered in two ways. First, the sensor nodes will beep to alert the resident. Second, an email will be delivered to the assigned carers. We recognise this is a somewhat simplistic interpretation of alarm detection, but its purpose is simply to illustrate the principle, which would be adapted to reflect more complex conditions which can be dynamically determined and adjusted in our future work.

4. System Prototyping and Discussion

The main concern gathered from interviewing elderly people (see Section 3.1.2 was the cost, which is understandable as they were all retired and living on small pensions.
Our investigation into the development of a low-cost unobtrusive monitoring system that can passively monitor domestic environments via processing its ambient acoustic information could help with this concern. As explained earlier, we adopted an IoT-enabled LinkSmart semantic middleware (Figure 1), an open source technology developed by the FP7 EU project EBBITS, which provides open interoperable capabilities suitable for applications using distributed physical devices with communication. In addition, the LinkSmart middleware has a modular framework, making it readily customisable. Furthermore, its ontological approach, to provide semantic representation of devices in the form of services, provides great flexibility and robustness, making it suitable for low-cost implementation.
As explained in Section 2, a Wi-Fi-based WSN was chosen for our prototype implementation. In our prototype, the LinkSmart Virtual Entity module was used to act as the main semantic virtualization mechanism for representing and handling all the data on the WSN; while the coordinator node provided the data via IoT Service (Figure 1).
Nodes were distributed and discoverable via LinkSmart’s device discovery feature and distributed architecture. LinkSmart is a popular standard, adopted by a number of other projects, including ELLIOT to create smart offices [45] for business use. Our study is the first to explore its use in a domestic assisted-living environment.
The development of a WSN for monitoring construction noise was investigated previously by Hughes et al. [46].
Their work compared different transmission modes, examining the effectiveness of a Bluetooth Low-Energy (BLE) network with a ZigBee network, specifically for an outdoor construction site setting. In their study, Beaglebone, Arduino, Raspberry Pi and Pi Zero computers were assessed for cost and availability. The node-to-gateway data transmission route choice was ZigBee, although BLE was considered. Other examples of low-cost IoT WSNs for environmental monitoring include the work by Ferdoush and Li [47] on analysing the combination of the Raspberry Pi and ZigBee components in a multi-node network. In their study [47], they drew the conclusion that a WSN with a Raspberry Pi base enabled the prototyping of a low-cost flexible WSN. In addition, Ahmed et al. [48] investigated various sensor types using Raspberry Pi and ZigBee.
Their work emphasized the flexibility of the Raspberry Pi, particularly as the foundation of a WSN, and they concluded that it was a robust platform suitable for a wide variety of WSN locations. However, ZigBee has slightly higher power consumption and low data rate capability, and it is not suitable for supporting tasks requiring high bandwidth, such as gathering ambient acoustic signals.
In addition to the cost concerns raised by the participants, a literature review showed the adoption of smart environment technology had many other limiting factors, such as ease of implementation and perceived usefulness [49]. These factors have been taken into consideration in the design of our prototype, which involved the use of low-cost devices: Two Raspberry Pi 3 were selected to act as WSN nodes, one equipped with a USB mini-microphone and the other with a speaker. Pi3 had a built-in Wi-Fi module and data were transmitted between the nodes and the gateway over Wi-Fi (Figure 2).

IoT Hardware and Sensor Nodes Implementation

The prototype development was based on the principles employed in our previous study [28], the key aspects of which can be summarised as:
  • The design of the Wireless Sensor Network (WSN) should be based around affordable devices with high-capacity processing [25].
  • In line with IoT principles, a distributed network of devices should be employed to collect data independently and autonomously of other devices on the network [50].
  • Raw data should be collated by a coordinator (control) node.
  • On-node data processing should be restricted and the data transmission rate balanced against power consumption to work within the limits of the IoT devices.
The overall system architecture is shown in Figure 2. It uses a modular framework comprising hardware/device-level workflow, IoT middleware communication, NN learning modules, and GUIs (as explained in Section 3.2). All sensors/nodes communicated directly with the coordinator (control) node via Wi-Fi. The coordinator (control) node communicated via LinkSmart middleware. As mentioned earlier, the main benefit from adopting the LinkSmart was device interoperability, while all nodes were implemented using a low-cost Raspberry Pi. A wireless LAN and an antenna were built into the board on the Pi-3, reducing the number of USB ports that were required and allowing for a smaller enclosure. All nodes were powered by a mains electricity connection to ensure a continuous and reliable data feed. SD cards of 8 GB were configured with the Linux-based Raspbian (Jessie), a Debian distribution, to provide an operating system for the Raspberry PIs. In addition, for the gateway node, an Apache server was used to present the data in graphical form. The prototype WSN nodes used low-cost, off-the-shelf microphones from Kinobo, which were easily obtained through on-line retailers.
As mentioned, our aims were to use cheap components; hence, the range was limited. However, sound levels recorded by the sensor node were sufficient for prototyping purposes. Consistency in the recordings from the microphones was ensured through comparative testing. Ambient acoustic data were gathered by a node placed in the kitchen, before being passed to the central controller node via Wi-Fi for analysis.

5. Experimental Setup and Results

For this study, we followed an experimental setup similar to the one used in our previous work [28] but with two main differences: the WSN was based on Wi-Fi, and the setup consisted of two nodes. The experiment was conducted in a domestic home, with one bedroom, kitchen, living/dining room, and shower room, measuring approximately 50 sq. meter in total (Figure 3).
One node (acting as a gateway node with an attached speaker) was placed in the living room and another node (a sensor node with an attached microphone) was placed in the kitchen area (Figure 4). The data that were collected informed the size of the different networks, as the optimal number of nodes for different locations varied due to the location/room size and its ambient noise level. The gateway node received the transmissions from the sensor nodes and unpacked the data. A Python-based NN library was installed on the gateway node, which was used to test the data and generate predictions.

5.1. Experimental Procedure

The experiment involved two phases. The first phase was to gather data for training purposes. In this phase, the data were gathered in one of the participant’s home. A total of three sequences of activities, each repeated 10 times, were gathered. The three sequences of activities were:
  • Walking, grasping the kettle, going to the tap, filling the kettle with water, switching the kettle on, the kettle boiling;
  • Walking, grasping water filtered jug, (1a) going to the tap, (1b) filling the jug with water (2) pouring the water into the kettle, switching the kettle on, the kettle boiling;
  • Walking, opening the fridge, closing the fridge, walking.
The data gathered were processed offline and used to train the NN model and context reasoning system using a Dell Optiplex computer, described in Section 3.2 and Section 3.3. Once the models were fully trained, they were then deployed to the gateway node, ready for the second phase of the experiment. The preliminary data of the second phase were then collected and subsequently analysed. We used the methodology developed in our previous work [28].
This involved using the same data collection and transmission script on the node running automatically, from boot, on the sensor node (the recording, analysis of the recording, and transmission of analysis ran in an infinite loop). A new recording and analysis started every second. The on-node analysis of recordings was done using the SoX, a cross-platform audio-processing command line utility. The audio data were then sent via Wi-Fi to the gateway node for further processing and testing. The gateway node stored the data in a local database which were then visualized through real-time rendering on a graphical interface. This allowed quick visual comparisons of consecutive time periods. The participants were first asked to repeat the same sequence of activities as normal, and then ‘walking three steps and stopping’ (remaining quiet). Figure 5 shows an example of the audio data gathered during the first phase of the experiment. The red circle indicates a ’kettle boiling’ event as recognised by the NN classifier. The green circle indicates a ‘footstep’ event.

5.2. Results and Analysis

The data collected from the Phase 1 experiment were used to train a NN classifier and context reasoning system. For comparison purposes, the graphic representation of the sound waveforms, the logarithm of the filter bank energies, and the mel-frequency cepstrum for the kettle and footstep sounds are displayed in Figure 6 and Figure 7, respectively.
The accuracy of the NN classifier was investigated. In the experiment, the aim was to identify the sound of the kettle boiling, as distinct from other distracting noises such as walking, opening/closing the fridge, filling the kettle with water, switching the kettle on, the kettle boiling, and staying quiet. Initially, we used only the neural network classifier without the context reasoning. After training with 20 positive and 20 negative examples, we tested the system with 20 positive and 55 negative events based on the data gathered from one individual’s home. Next, we included the context reasoning. We trained another two neural network classifiers to identify the water tap sound and footstep sound (again using 20 positive and 20 negative examples for each case). The logic of the context reasoning was, if the water tap sound and footstep sound were not detected before a probable detection of the kettle boiling sound, then the probability (or confidence) of kettle boiling detection will be reduced.
The test results are shown in the corresponding confusion matrix shown in Table 4 for the NN without context reasoning, and in Table 5 for the NN with context reasoning. The table shows the performance of the NN trained to recognise the Boiling Kettle sound identified as “True Positive” (TP) when the NN predicted the Boiling Kettle when the Boiling Kettle sound was presented, “True Negative” (TN) when the NN predicted a non-Boiling Kettle when a non-Boiling Kettle sound was presented, “False Positive” (FP) when the NN predicted kettle boiling when a non-Boiling Kettle sound was presented and “False Negative” (FN) when NN predicted a non-Boiling Kettle when a Boiling Kettle sound was presented. The confusion matrix helped calculate the sensitivity (TP/(TP + FN)), specificity (TN/(TN + FP)), precision (TP/(TP + FP), and accuracy ((TP + TN)/(TP + FP + TN + FN)).
Table 6 shows the performances of the NN without and with a context reasoning system. Analysing the statistical data indicates that performance improvement comes from a reduction in false positives by taking the context into account. We expect that the accuracy could be improved further by designing more complex context reasoning logic.
However, putting these prediction percentages in the context of the Bayesian statistics, where the sensitivity and specificity are the conditional probabilities, the prevalence is the prior probability, and the positive/negative predicted values are the posterior probabilities, the model prediction is usually involved in unconditional events, and therefore the positive predictive value (PPV), defined in (5) can better characterise the overall performance of the classifier when the prevalence is not 50%.
P P V = s e n s i t i v i t y p r e v a l e n c e s e n s i t i v i t y p r e v a l e n c e + ( 1 s p e c i f i c i t y ) ( 1 p r e v a l e n c e ) .
The calculated PPV values are shown in Table 7. As expected, the value increased considerably, showing the importance of using a context reasoning system, especially on NNs trained with a low data set size.
Collating the data from all our test experiments and applying the context reasoning system on our NNs trained to recognize Kettle Boiling, Tap Water, and Foot Step, the calculated accuracy and PPv values are displayed in Table 8.
Applied in the Phase 2 experiment, where the participant was asked to repeat the same habitual sequence of activities, the system correctly did not trigger any alarm (as the sequence was in the predefined sequences of activities), but when the participant walked three steps and stopped (not in the predefined list, and therefore treated as “abnormal” sequencing), an alarm was triggered. These demonstrate that the system design is capable of recognising the sounds produced and classifying it as normal or abnormal.

6. Conclusions

The summary of the research, that is, the goals achieved and how they were attained are the following:
(i) The main goal of this study was to investigate a holistic approach using a combination of acoustic sensing, artificial intelligence, and the Internet-of-Things as a means of providing a cost-effective approach to alerting care providers or relatives when an abnormal event is detected. (ii) We investigated the use of low-cost IoT devices and NN methods to unobtrusively monitor an ambient acoustic domestic environment. A prototype was successfully developed for this purpose. In developing the prototype, we investigated the performance of Wi-Fi WSN and the semantic LinkSmart IoT middleware as a platform to support on-node real-time audio sample processing, and distributed logic, in IoT systems, proving its architectural viability for such purposes. (iii) We investigated NN techniques as part of our acoustic processing work and combined the context reasoning system with the NN classifier trained on a low-training data set, and good overall accuracy was obtained. Moreover, our work has shown that by contextualising events within a stream of activities, the overall accuracy, and therefore the reliability of the alerting system can be improved (i.e., reduction of the number of false alarms). Although we have left a more focused and detailed study of the sequential logic and how it might be refined for a follow-on project, the findings from this research confirm that is possible to create a stable, low-cost wireless sensor network for monitoring and modelling ambient acoustic levels using an NN learning model on a distributed Raspberry Pi network over Wi-Fi and LinkSmart middleware.
As we explained earlier in this paper, our main goal was to explore the potential for a combination of low-cost IoT devices, NN models (trained using a low data set), and a context reasoning system to unobtrusively monitor ambient acoustic domestic environments for the purposes of providing care to elderly people. At the core of that work was establishing a set of hardware and software architectural principles (e.g., the nature of, and the distribution of processing arrangements and AI models) capable of ensuring a deployable system in home care, which we believe this work has achieved.
Clearly, there are many directions that this work could follow in the future, and below we are suggesting nine of them (numbered as i-ix): (i) One direction could include the investigation of more complex acoustic profiles (e.g., by including prior knowledge and context). (ii) A wider range of sound measurements could be captured to allow for a more detailed analysis of different types of sound and a more refined alert system (e.g., associating footsteps to particular people). (iii) To improve the context reasoning system, the sequential logic might be refined to involve a larger-scale study covering many more events. Likewise, it would be beneficial to (iv) further refine the NN classifier to optimise the network structure, the number of hidden units and the training features, or make it more sustainable when scaling up the system, (v) to replace the FF-BF NN with a CNN, which is a more accurate NN when recognising picture-type patterns (as the mel-frequency cepstrum provides). (vi) Another useful line of research would be to address the cocktail party problem (several sounds mixing with each other) by, say, using blind source separation or other techniques. (vii) HCI is another aspect worthy of further work.
Although the project has focused on how the system could be of benefit in AAL settings, the demonstrated WSN could be easily adapted for different environments and applications that would benefit from remote ambient sound monitoring and alerts. For example, beyond the care of elderly people, the system might be applied to (viii), baby/child monitoring, monitoring children with ADHD (as proposed in [51], where accelerometers and proximity sensors were used), security systems, and even pest control.
Finally, (ix) concerning commercial deployment routes, given the ever-increasing popularity of smart-home voice control systems such as Amazon’s Echo, it may be possible to integrate the techniques presented in this paper into such systems, thereby endowing them with a broader sound analysis capability, enabling care services to benefit from the ’economics of scale’ associated with commercial smart-home technologies.
Thus, we are confident that our results will encourage further research on applications using systems trained with a low data size and support the increasing interest in applying sound analysis to a range of emerging IoT and smart home applications, which will benefit us all and especially older people who, to date, have not been the main beneficiaries of the technology revolution that is changing our world at an unprecedented rate, which in our opinion, has had a knock-off effect in gathering reliable behavioural data. Ultimately, such work would benefit all of us, since most of us are destined to be elderly later in our lives [51].

Author Contributions

Conceptualization, J.C. and A.T.; methodology, J.C. and A.T.; software, J.C., A.T. and D.C.; validation, J.C., A.T. and D.C.; formal analysis, J.C., A.T. and V.C.; writing—original draft preparation, J.C., A.T. and V.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bloom, D.E.; Canning, D. Global Demographic Change: Dimensions and Economic Significance; Working Paper 10817; National Bureau of Economic Research: Cambridge, MA, USA, 2004. [Google Scholar]
  2. Mack, R.; Salmoni, A.; Viverais-Dressler, G.; Porter, E.; Garg, R. Perceived Risks to Independent Living: The Views of Older, Community-Dwelling Adults. Gerontologist 1997, 37, 729–736. [Google Scholar] [CrossRef] [PubMed]
  3. Savage, M. More Than Half of England’s Coronavirus-Related Deaths Will Be People from Care Homes. Available online: https://www.theguardian.com/society/2020/jun/07/more-than-half-of-englands-coronavirus-related-deaths-will-be-people-from-care-homes (accessed on 13 June 2021).
  4. Chin, J.; Callaghan, V.; Allouch, S.B. The Internet-of-Things: Reflections on the past, present and future from a user-centered and smart environment perspective. J. Ambient Intell. Smart Environ. 2019, 11, 45–69. [Google Scholar] [CrossRef] [Green Version]
  5. Chin, J.; Callaghan, V.; Lam, I. Understanding and personalising smart city services using machine learning, The Internet-of-Things and Big Data. In Proceedings of the 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), Edinburgh, UK, 19–21 June 2017; pp. 2050–2055. [Google Scholar]
  6. Callaghan, V.; Clarke, G.; Colley, M.; Hagras, H.; Chin, J.S.Y.; Doctor, F. Inhabited Intelligent Environments. BT Technol. J. 2004, 22, 233–247. [Google Scholar] [CrossRef] [Green Version]
  7. Patel, P.A.; Gunnarsson, C. A Passive Monitoring System in Assisted Living Facilities: 12-Month Comparative Study. Phys. Occup. Ther. Geriatr. 2012, 30, 45–52. [Google Scholar] [CrossRef] [PubMed]
  8. Feki, M.A.; Kawsar, F.; Boussard, M.; Trappeniers, L. The Internet of Things: The Next Technological Revolution. Computer 2013, 46, 24–25. [Google Scholar] [CrossRef]
  9. Rivera-illingworth, F.; Callaghan, V.; Hagras, H. Detection of Normal and Novel Behaviours in Ubiquitous Domestic Environments. Comput. J. 2007, 53, 142–151. [Google Scholar] [CrossRef]
  10. Kaye, J.A.; Maxwell, S.A.; Mattek, N.; Hayes, T.L.; Dodge, H.; Pavel, M.; Jimison, H.B.; Wild, K.; Boise, L.; Zitzelberger, T.A. Intelligent Systems for Assessing Aging Changes: Home-Based, Unobtrusive, and Continuous Assessment of Aging. J. Gerontol. Ser. 2011, 66B, i180–i190. [Google Scholar] [CrossRef] [PubMed]
  11. Mokhtari, M.; Aloulou, H.; Tiberghien, T.; Biswas, J.; Racoceanu, D.; Yap, P. New Trends to Support Independence in Persons with Mild Dementia—A Mini-Review. Gerontology 2012, 58. [Google Scholar] [CrossRef]
  12. Hossain, M.A. Perspectives of human factors in designing elderly monitoring system. Comput. Hum. Behav. 2014, 33, 63–68. [Google Scholar] [CrossRef]
  13. Arcelus, A.; Jones, M.H.; Goubran, R.; Knoefel, F. Integration of Smart Home Technologies in a Health Monitoring System for the Elderly. In Proceedings of the 21st International Conference on Advanced Information Networking and Applications Workshops (AINAW’07), Niagara Falls, ON, Canada, 21–23 May 2007; Volume 2, pp. 820–825. [Google Scholar] [CrossRef]
  14. Lindenberger, U.; Lövdén, M.; Schellenbach, M.; Li, S.C.; Krüger, A. Psychological Principles of Successful Aging Technologies: A Mini-Review. Gerontology 2008, 54, 59–68. [Google Scholar] [CrossRef] [Green Version]
  15. Reijnders, J.; van Heugten, C.; van Boxtel, M. Cognitive interventions in healthy older adults and people with mild cognitive impairment: A systematic review. Ageing Res. Rev. 2013, 12, 263–275. [Google Scholar] [CrossRef] [PubMed]
  16. Bao, W.; Chen, Y.; Zeng, X. Wearable Wireless system with embedded real-time fall detection logic for elderly assisted living applications. In Proceedings of the 2012 IEEE 11th International Conference on Solid-State and Integrated Circuit Technology, Xi’an, China, 29 October–1 November 2012; pp. 1–3. [Google Scholar] [CrossRef]
  17. Stroiescu, F.; Daly, K.; Kuris, B. Event detection in an assisted living environment. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 7581–7584. [Google Scholar] [CrossRef]
  18. Caroux, L.; Consel, C.; Dupuy, L.; Sauzéon, H. Verification of Daily Activities of Older Adults: A Simple, Non-Intrusive, Low-Cost Approach; Association for Computing Machinery: New York, NY, USA, 2014; pp. 43–50. [Google Scholar] [CrossRef] [Green Version]
  19. Townsend, D.; Knoefel, F.; Goubran, R. Privacy versus autonomy: A tradeoff model for smart home monitoring technologies. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 4749–4752. [Google Scholar] [CrossRef]
  20. Callaghan, V.; Clarke, G.; Chin, J. Some Socio-Technical Aspects Of Intelligent Buildings and Pervasive Computing Research. Earthscan J. 2009, 1. [Google Scholar] [CrossRef]
  21. Ni, Q.; García Hernando, A.B.; De la Cruz, I.P. The Elderly’s Independent Living in Smart Homes: A Characterization of Activities and Sensing Infrastructure Survey to Facilitate Services Development. Sensors 2015, 15, 11312–11362. [Google Scholar] [CrossRef] [PubMed]
  22. Civitarese, G.; Sztyler, T.; Riboni, D.; Bettini, C.; Stuckenschmidt, H. POLARIS: Probabilistic and Ontological Activity Recognition in Smart-Homes. IEEE Trans. Knowl. Data Eng. 2021, 33, 209–223. [Google Scholar] [CrossRef]
  23. Bettini, C.; Civitarese, G.; Giancane, D.; Presotto, R. ProCAVIAR: Hybrid Data-Driven and Probabilistic Knowledge-Based Activity Recognition. IEEE Access 2020, 8, 146876–146886. [Google Scholar] [CrossRef]
  24. Pletikosa, I.; Michahelles, F. The Toolkit Approach for End-user Participation in the Internet of Things. Archit. Internet Things 2011, 65–96. [Google Scholar] [CrossRef]
  25. Vujović, V.; Maksimović, M. Raspberry Pi as a Sensor Web node for home automation. Comput. Electr. Eng. 2015, 44, 153–171. [Google Scholar] [CrossRef]
  26. Pitt, L.; Green, P.R.; Lennox, B. A sensor network for predicting and maintaining occupant comfort. In Proceedings of the 2013 IEEE Workshop on Environmental Energy and Structural Monitoring Systems, Trento, Italy, 11–12 September 2013; pp. 1–6. [Google Scholar] [CrossRef]
  27. Choudhury, S.; Kuchhal, P.; Singh, R.; Anita. ZigBee and Bluetooth Network based Sensory Data Acquisition System. Procedia Comput. Sci. 2015, 48, 367–372. [Google Scholar] [CrossRef] [Green Version]
  28. Norros, L. Acting under Uncertainty: The Core-Task Analysis in Ecological Study of Work; Number 546 in VTT Publications; VTT Technical Research Centre of Finland: Espoo, Finland, 2004. [Google Scholar]
  29. Eriksson, M.; Niitamo, V.P.; Kulkki, S. State-of-the-Art in Utilizing Living Labs Approach to User-Centric ICT Innovation—A European Approach. Available online: http://84.88.32.6/openlivinglabs/documents/SOA_LivingLabs.pdf (accessed on 15 December 2015).
  30. Juan, C.; Yelmo, J.M.Á.; Trapero, R.; Martín, Y.S. A user-centric approach to service creation and delivery over next generation networks. Comput. Commun. 2011, 34, 209–222. [Google Scholar]
  31. Akao, Y.; Mazur, G.H. The leading edge in QFD: Past, present and future. Int. J. Qual. Reliab. Manag. 2003, 20, 20–35. [Google Scholar] [CrossRef]
  32. Reynolds, D.A. Speaker identification and verification using Gaussian mixture speaker models. Speech Commun. 1995, 17, 91–108. [Google Scholar] [CrossRef]
  33. Virone, G.; Istrate, D. Integration of an Environmental Sound Module to an Existing In-Home Activity Simulator. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2007, 2007, 3810–3813. [Google Scholar] [CrossRef]
  34. Shaikh, M.A.M.; Rebordao, A.R.F.; Nakasone, A.; Helmut, P.; Hirose, K. An automatic approach to virtual living based on environmental sound cues. In Proceedings of the 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, Amsterdam, The Netherlands, 10–12 September 2009; pp. 1–6. [Google Scholar] [CrossRef]
  35. Segura-Garcia, J.; Felici-Castell, S.; Perez-Solano, J.J.; Cobos, M.; Navarro, J.M. Low-Cost Alternatives for Urban Noise Nuisance Monitoring Using Wireless Sensor Networks. IEEE Sens. J. 2015, 15, 836–844. [Google Scholar] [CrossRef]
  36. Pesonen, L.; Koskinen, H.; Rydberg, A. InfoXT—User-Centric Mobile Information Management in Automated Plant Production; Nordic Council of Ministers: Copenhagen, Denmark, 2008; Volume 05103. [Google Scholar]
  37. LGA Report 2017. Available online: https://www.local.gov.uk/sites/default/files/documents/5.17%20-%20Housing%20our%20ageing%20population_07_0.pdf (accessed on 13 June 2021).
  38. Møller, M.F. A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw. 1993, 6, 525–533. [Google Scholar] [CrossRef]
  39. Swarnkar, V.; Abeyratne, U.R.; Amrulloh, Y.; Hukins, C.; Triasih, R.; Setyati, A. Neural network based algorithm for automatic identification of cough sounds. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 1764–1767. [Google Scholar] [CrossRef]
  40. Shi, W.Y.; Chiao, J. Neural network based real-time heart sound monitor using a wireless wearable wrist sensor. In Proceedings of the 2016 IEEE Dallas Circuits and Systems Conference (DCAS), Arlington, TX, USA, 10 October 2016; pp. 1–4. [Google Scholar] [CrossRef]
  41. Boes, M.; Oldoni, D.; De Coensel, B.; Botteldooren, D. Long-term learning behavior in a recurrent neural network for sound recognition. In Proceedings of the 2014 International Joint Conference on Neural Networks (IJCNN), Beijing, China, 6–11 July 2014; pp. 3116–3123. [Google Scholar] [CrossRef] [Green Version]
  42. Ramanujam, E.; Perumal, T.; Padmavathi, S. Human Activity Recognition with Smartphone and Wearable Sensors using Deep Learning Techniques: A Review. IEEE Sens. J. 2021, 1. [Google Scholar] [CrossRef]
  43. Griffiths, N.; Chin, J. Towards unobtrusive ambient sound monitoring for smart and assisted environments. In Proceedings of the 2016 8th Computer Science and Electronic Engineering (CEEC), Colchester, UK, 28–30 September 2016; pp. 18–23. [Google Scholar] [CrossRef]
  44. Chik, D.; Tripathi, G.; Wagatsuma, H. A Method to Deal with Prospective Risks at Home in Robotic Observations by Using a Brain-Inspired Model. In Proceedings of the International Conference on Neural Information Processing, Daegu, Korea, 3–7 November 2013; Volume 8228. [Google Scholar] [CrossRef]
  45. Furdík, K.; Lukac, G.; Sabol, T.; Kostelnik, P. The Network Architecture Designed for an Adaptable IoT-based Smart Office Solution. Int. J. Comput. Netw. Commun. Secur. 2013, 1, 216–224. [Google Scholar]
  46. Hughes, J.; Yan, J.; Soga, K. Development of wireless sensor network using Bluetooth Low Energy (BLE) for construction noise monitoring. Int. J. Smart Sens. Intell. Syst. 2015, 8, 1379–1405. [Google Scholar] [CrossRef] [Green Version]
  47. Ferdoush, S.; Li, X. Wireless Sensor Network System Design Using Raspberry Pi and Arduino for Environmental Monitoring Applications. Procedia Comput. Sci. 2014, 34, 103–110. [Google Scholar] [CrossRef] [Green Version]
  48. Ahmed, N.; Hussain, S.; Siddiqui, A. Raspberry Pi based online parameters monitoring and control system implemented using four sensor nodes. In Proceedings of the International Conference on Modern Communication and Computing Technologies, Nawabshah, Pakistan, 26–28 February 2014; pp. 49–56. [Google Scholar]
  49. Wilson, C.; Hargreaves, T.; Hauxwell-Baldwin, R. Smart homes and their users: A systematic analysis and key challenges. Pers. Ubiquitous Comput. 2014, 19, 463–476. [Google Scholar] [CrossRef] [Green Version]
  50. Ashton, K. That ‘Internet of Things’ Thing. Available online: http://www.itrco.jp/libraries/RFIDjournal-That%20Internet%20of%20Things%20Thing.pdf (accessed on 13 June 2021).
  51. Berrezueta-Guzman, J.; Pau, I.; Martín-Ruiz, M.L.; Máximo-Bocanegra, N. Smart-Home Environment to Support Homework Activities for Children. IEEE Access 2020, 8, 160251–160267. [Google Scholar] [CrossRef]

Short Biography of Authors

Electronics 10 01433 i001Jeannette Chin Jeannette Chin is an academic at the University of East Anglia, UK. Her expertise lies in ubiquitous computing and development of real-time End User Programming systems for smart environments and the Internet of Things (IoT). Jeannette’s contributions and research interests include secure distributed systems, HCI, personalised services, assisted living/healthcare leveraging everyday artefacts and connectivity. More recently, her work involves applying ML on big data to solve problems. Jeannette is a member of the editorial team for the IET Smart Cities Journal, Journal of Ambient Intelligence and Smart Environments, and Ambient Intelligence and Smart Environments. She is a Senior Fellow Higher Education Academy (SFHEA) and a Fellow British Computer Society (FBCS). Jeannette has led three funded projects, as the principal investigator, into successful completion, attracting over £200,000 in funding. She receives the Chancellor Innovation Award on “Outstanding Achievement” (2021) for her work to help control the spread of COVID-19 in Norwich.
Electronics 10 01433 i002Alin Tisan Dr. Alin Tisan received a BEng and a MSc in Physics from Babeş-Bolyai University and a Ph.D. in neuromorphic systems for olfaction applications from Technical University of Cluj Napoca, RO. He is currently lecturer in the Electronic Engineering Department at Royal Holloway University of London, UK. His research interests include Pattern recognition on System on Chip (neuromorphic software/hardware systems), Machine learning for neuronal data, Data processing in olfaction, Signal processing in gas sensors, IoT with applications in Pervasive healthcare and smart homes. He published over 60 papers and has been co-chairing Electronic Systems-on-Chip and Embedded Systems related Technical Tracks at several conferences, such as ICIT, OPTIM, ISIE, INDIN or IECON. Alin is also an editor for the MDPI Sensors Journal.
Electronics 10 01433 i003Victor Callaghan Victor Callaghan is an emeritus professor of computer science at the University of Essex (UK), visiting professor at Canterbury Christ Church University (UK), current president of the Association for the Advancement of Intelligent Environments and CEO of the Creative Science Foundation. He holds a BEng in Electronic Engineering and a PhD in Computing from the University of Sheffield. His research has covered areas which include IoT, Artificial Intelligence, Affective Computing, Quantum Computing and Mixed Reality. He attracted over 10 million pounds sterling (UK) in funding and authored over 300 papers. He is a founder of the ‘Intelligent Environments’ Conference and past program chair for various earlier conferences such as ‘Active Media technology’, ‘Pervasive Computing & Applications’ and ‘Ubiquitous Intelligence & Computing’. He is an editorial board member of Springer’s ‘Reliable Intelligent Environments’ journal, IOS’s ‘Ambient Intelligence and Smart Environments’ journal and Springer’s book series “Science & Fiction”.
Electronics 10 01433 i004David Chik David Chik is a Lead Engineer in Hong Kong Applied Science and Technology Research Institute. He received his PhD degree in Physics from The University of Hong Kong. His research interest includes spiking neural networks and deep learning. He is leading a team of engineers to develop innovative products in manufacturing and healthcare areas. He also works closely with local companies and investors on the commercialization of research outputs.
Figure 1. Modular architecture of the LinkSmart IoT middleware in the EBBITS project. Image from Furdik et al. [45].
Figure 1. Modular architecture of the LinkSmart IoT middleware in the EBBITS project. Image from Furdik et al. [45].
Electronics 10 01433 g001
Figure 2. System architecture.
Figure 2. System architecture.
Electronics 10 01433 g002
Figure 3. Floor plan of the experiment setup with two nodes in a domestic environment.
Figure 3. Floor plan of the experiment setup with two nodes in a domestic environment.
Electronics 10 01433 g003
Figure 4. Sensor nodes with the USB microphone and speaker attached.
Figure 4. Sensor nodes with the USB microphone and speaker attached.
Electronics 10 01433 g004
Figure 5. Audio signal recorded by Pi 3 and transmitted to the computer for analysis.
Figure 5. Audio signal recorded by Pi 3 and transmitted to the computer for analysis.
Electronics 10 01433 g005
Figure 6. The sound waveforms, the logarithm of the filterbank energies, and the mel-frequency cepstrum graphic representation for the kettle sound.
Figure 6. The sound waveforms, the logarithm of the filterbank energies, and the mel-frequency cepstrum graphic representation for the kettle sound.
Electronics 10 01433 g006
Figure 7. The sound waveforms, the logarithm of the filterbank energies, and the mel-frequency cepstrum graphic representation for the sound of the footsteps.
Figure 7. The sound waveforms, the logarithm of the filterbank energies, and the mel-frequency cepstrum graphic representation for the sound of the footsteps.
Electronics 10 01433 g007
Table 1. Key characteristics of wireless protocols.
Table 1. Key characteristics of wireless protocols.
CharacteristicWi-FiBluetoothZigbeeZwave
StandardYesYesYesNo
Supports HANYesYesYesYes
Data throughput *72 Mbps2 Mbps250 kbps40 kbps
Range *100 m750 m10 m30 m
TopologyStarP-2-P/MeshStar/MeshMesh
IP enabledYesNoNoNo
PC/Mobile supportYesYesNoNo
Widely deployedYesYesNoNo
* Maximum.
Table 2. Household activities questionnaire.
Table 2. Household activities questionnaire.
Kitchenmaking tea/coffee
open/close fridge/freezer (put/take food, milk etc.)
making toast
preparing food in the sink
preparing food (on the chopping board)
making soup
cooking (on the stove)
baking (food in the oven)
microwaving
listening to radio
open living-room-kitchen door
open kitchen-garden door
open window
washing dishes
washing clothes
drying clothes
other:
Living Roomsitting on the sofa
watching TV
reading books/papers/magazines
knitting
talking with someone on the phone
resting
listening to music (on CD/DVD/MP3 player)
watching a movie (on CD/DVD/MP3 player)
other:
Otherpreparing medication
taking medication
monitoring blood pressure/sugar level/home diagnostic tests
taking a shower
taking a bath
playing games
other:
Table 3. Context data structure.
Table 3. Context data structure.
Current EventMax DurationNext Event ANext Event BNext Event C
Foot step100 sTap waterQuiet
Tap water100 sKettle boilingFoot stepQuiet
Kettle boiling120 sFoot stepQuiet
Fridge opening30 sFridge closingFoot stepQuiet
Table 4. Neural network confusion matrix without context reasoning.
Table 4. Neural network confusion matrix without context reasoning.
Predicted:Predicted:
Boiling KettleNon-Boiling Kettle
Actual Boiling KettleTrue Positives: 12False Negatives: 820
Actual non-Boiling KettleFalse Positives: 9True Negatives: 4655
2154
Table 5. Neural network confusion matrix with context reasoning.
Table 5. Neural network confusion matrix with context reasoning.
Predicted:Predicted:
Boiling KettleNon-Boiling Kettle
Actual Boiling KettleTrue Positives: 12False Negatives: 820
Actual non-Boiling KettleFalse Positives: 4True Negatives: 5155
1659
Table 6. NN performance.
Table 6. NN performance.
NN without Context ReasoningNN with Context Reasoning
prevalence26.6%26.6%
sensitivity60%60%
specificity83.6%92.7%
precision57.1%75%
accuracy77.3%84%
Table 7. NN performance.
Table 7. NN performance.
NN without Context ReasoningNN with Context Reasoning
PPV57%74.8%
Table 8. The accuracy and PPV values of the NN classifiers with context reasoning.
Table 8. The accuracy and PPV values of the NN classifiers with context reasoning.
EventAccuracy of NNPPV of NN
Kettle boiling84%74.8%
Tap water68%58.6%
Footstep72%62.8%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chin, J.; Tisan, A.; Callaghan, V.; Chik, D. Smart-Object-Based Reasoning System for Indoor Acoustic Profiling of Elderly Inhabitants. Electronics 2021, 10, 1433. https://doi.org/10.3390/electronics10121433

AMA Style

Chin J, Tisan A, Callaghan V, Chik D. Smart-Object-Based Reasoning System for Indoor Acoustic Profiling of Elderly Inhabitants. Electronics. 2021; 10(12):1433. https://doi.org/10.3390/electronics10121433

Chicago/Turabian Style

Chin, Jeannette, Alin Tisan, Victor Callaghan, and David Chik. 2021. "Smart-Object-Based Reasoning System for Indoor Acoustic Profiling of Elderly Inhabitants" Electronics 10, no. 12: 1433. https://doi.org/10.3390/electronics10121433

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop