Edge AI for Real-Time Anomaly Detection in Smart Homes
Abstract
:1. Introduction
1.1. Background and Motivation
1.2. Research Problem and Objectives
- Design an efficient anomaly detection architecture that leverages Edge AI to process data locally, reducing latency and preserving privacy.
- Develop and implement lightweight machine learning models suitable for deployment on resource-constrained devices commonly found in smart homes.
- Evaluate the proposed framework using simulated IoT data representative of typical smart home environments to assess its effectiveness in detecting anomalies.
1.3. Contributions
- Proposing a novel edge-based architecture for real-time anomaly detection tailored to the constraints and requirements of smart home environments.
- Developing lightweight machine learning models that balance detection accuracy with computational efficiency, facilitating deployment on typical smart home devices.
- Conducting comprehensive evaluations using simulated IoT data to validate the proposed framework’s performance and provide insights into its practical applicability.
2. Related Work
2.1. IoT Anomaly Detection
2.2. Edge Computing and Edge AI
- Reduced latency: Local data processing enables immediate detection of anomalies, which is crucial for timely responses in smart home applications.
- Enhanced privacy: Processing data on-device minimizes the exposure of sensitive information, addressing privacy concerns associated with cloud computing.
- Lower bandwidth usage: By analyzing data locally, the need for constant data transmission to the cloud is reduced, conserving network bandwidth.
2.3. Research Gap
3. Proposed System Architecture
- IoT Network Anomaly Detection in Smart Homes Using Machine Learning: This study presents a machine learning-based approach for detecting anomalies in smart home IoT networks, utilizing Random Forest and decision tree classifiers. The methodology and results informed our selection of efficient models suitable for edge deployment [5].
- Enhancing Smart Home Security: Anomaly Detection and Face Recognition: This research study explores the integration of deep learning models for anomaly detection and face recognition within IoT devices, highlighting the importance of real-time, on-device processing to enhance security and privacy in smart homes [17].
- Edge AI for Smart Home Applications: This article discusses the application of Edge AI in smart homes, emphasizing its role in enabling functionalities like motion detection and anomaly detection without relying on cloud services, thereby ensuring data privacy and reducing latency [3].
3.1. Overview of the Edge AI-Based Framework
3.2. Anomaly Detection Approach
3.2.1. Advantages of Edge AI over Cloud-Based Detection
- Lower latency: On-device processing eliminates the delay associated with cloud-based computation, enabling on-device anomaly monitoring and immediate response [20]. Edge AI-based detection can reduce latency from several hundred ms (typical for cloud processing) to under 50ms, making it suitable for time-sensitive applications.
- Enhanced privacy: Keeping data local reduces the risk of exposing sensitive user information by limiting transmission to external servers. This approach ensures compliance with data protection regulations and enhances user trust.
- Reduced bandwidth usage: Processing sensor data at the edge minimizes network congestion and reliance on cloud infrastructure, reducing overall bandwidth requirements by up to 80% compared to continuous cloud-based processing.
3.2.2. Model Selection Rationale
3.2.3. Future Work: Exploring Lightweight Transformer Variants
- MobileViT—A mobile-friendly vision transformer that reduces computational complexity while preserving accuracy [22].
- TinyBERT—A distillation-based transformer designed for low-power AI applications.
3.2.4. Mathematical Formulation of Anomaly Scores
3.3. Edge Deployment Considerations
4. Experimental Setup
4.1. Data Simulation
- Temperature (°C): Simulated using a normal distribution around typical indoor temperature.
- Motion detection (binary: 0/1): Randomized based on expected human activity patterns.
- Energy usage (wattage): Simulated with baseline values and occasional spikes to mimic high-power consumption events.
- Feature scaling: Min-Max normalization was applied to standardize sensor readings.
- Noise reduction: A rolling average filter was used to smooth transient fluctuations in sensor data before feeding it into the anomaly detection models.
4.2. Model Implementation
- Isolation Forest (unsupervised anomaly detection): Identifies outliers based on data point separation.
- LSTM Autoencoder (deep learning for sequential anomaly detection): Detects long-term deviations in sensor data patterns.
- UNSW BoT-IoT Dataset (captures real-world IoT traffic anomalies)
- CASAS Smart Home Dataset (records user activity and environmental sensor data)
- Raspberry Pi 4 (Edge AI baseline device)
- NVIDIA Jetson Nano (optimized for deep learning inference)
- ESP32 microcontroller (low-power anomaly detection)
4.3. Evaluation Metrics
- Accuracy, precision, recall, and F1-score: Used to assess detection performance.
- Inference time: Measured the time taken for on-device anomaly monitoring on the Raspberry Pi 4 compared to a cloud-based model execution.
- Memory usage: Evaluated the model footprint on edge devices, ensuring feasibility for deployment on resource-limited hardware.
- Energy consumption: Measured using power monitoring tools to compare the runtime efficiency of Isolation Forest and LSTM Autoencoder.
5. Results and Discussion
5.1. Anomaly Detection Performance
5.2. Computational Efficiency
5.3. Key Findings and Discussion
5.3.1. Justification for the Hybrid Model Approach
- Step 1: Isolation Forest is used as a first-stage anomaly filter, quickly flagging potentially anomalous data points with minimal computational cost.
- Step 2: LSTM Autoencoder is applied only to the flagged data, performing sequential validation to refine anomaly detection and reduce false positives.
5.3.2. Comparative Analysis of Detection Approaches
5.4. Generalization on Real-World Smart Home Data (CASAS Dataset)
6. Conclusions and Future Work
6.1. Conclusions
6.2. Future Work
- Custom firmware adaptations for compatibility with smart home hubs (e.g., OpenHAB, Home Assistant).
- Edge-cloud hybrid models, where initial anomaly detection occurs on the device, and high-confidence anomalies are verified via cloud APIs.
- Real-time API connectivity for direct integration with IoT security services, enabling automated responses to detected anomalies (e.g., activating security cameras or sending alerts).
- Ultra-low-power AI hardware (e.g., TensorFlow Lite, TinyML, NVIDIA Jetson Nano optimizations).
- Event-triggered inference, where the anomaly detection model only activates when sensor deviations surpass predefined thresholds, reducing unnecessary energy consumption.
- Implement continuous model monitoring to assess real-world false positive and false negative rates over time.
- Investigate federated learning strategies, where multiple IoT devices collaborate to update models without sharing raw sensor data, enhancing privacy while improving detection performance across different smart homes.
- Time Series Transformer (TST)—A variant designed for sequential anomaly detection, emphasizing capturing long-range dependencies.
- Temporal Fusion Transformer (TFT)—A deep learning model that combines attention mechanisms with interpretable AI for time-series forecasting.
- MobileViT and TinyBERT—Transformer-based architectures optimized for low-power devices.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A
Appendix A.1. Isolation Forest Implementation (Code 1)
import numpy as np import pandas as pd from sklearn.ensemble import IsolationForest from sklearn.preprocessing import MinMaxScaler # Generate synthetic IoT sensor data np.random.seed(42) time_steps = 1000 temperature = np.random.normal(loc = 22, scale = 2, size = time_steps) motion = np.random.choice([0, 1], size = time_steps, p = [0.95, 0.05]) energy_usage = np.random.normal(loc = 500, scale = 50, size = time_steps) # Inject anomalies anomaly_indices = np.random.choice(time_steps, size = 30, replace = False) temperature[anomaly_indices] += np.random.uniform(5, 10, size = 30) motion[anomaly_indices] = 1 energy_usage[anomaly_indices] += np.random.uniform(200, 400, size = 30) # Create a DataFrame data = pd.DataFrame({“Temperature”: temperature, “Motion”: motion, “Ener-gy_Usage”: energy_usage}) # Normalize data scaler = MinMaxScaler() data_scaled = scaler.fit_transform(data) # Apply Isolation Forest iso_forest = IsolationForest(contamination = 0.03, random_state = 42) data[“Anomaly_Score_IF”] = iso_forest.fit_predict(data_scaled) data[“Anomaly_IF”] = data[“Anomaly_Score_IF”].apply(lambda x: 1 if x = −1 else 0) # Display anomaly detection results print(data.head()) |
Appendix A.2. LSTM Autoencoder Implementation (Code 2)
import numpy as np import pandas as pd import tensorflow as tf from tensorflow.keras.models import Sequential from tensorflow.keras.layers import LSTM, Dense, Dropout, RepeatVector, TimeDistributed from sklearn.preprocessing import MinMaxScaler # Generate synthetic IoT data np.random.seed(42) time_steps = 1000 temperature = np.random.normal(loc = 22, scale = 2, size = time_steps) motion = np.random.choice([0, 1], size = time_steps, p = [0.95, 0.05]) energy_usage = np.random.normal(loc = 500, scale = 50, size = time_steps) # Inject anomalies anomaly_indices = np.random.choice(time_steps, size = 30, replace = False) temperature[anomaly_indices] += np.random.uniform(5, 10, size = 30) motion[anomaly_indices] = 1 energy_usage[anomaly_indices] += np.random.uniform(200, 400, size = 30) # Create a DataFrame data = pd.DataFrame({“Temperature”: temperature, “Motion”: motion, “Ener-gy_Usage”: energy_usage}) # Normalize data scaler = MinMaxScaler() data_scaled = scaler.fit_transform(data) X_train = data_scaled.reshape((data_scaled.shape [0], 1, data_scaled.shape [1])) # Reshape for LSTM # Define LSTM Autoencoder model = Sequential([ LSTM(32, activation = ’relu’, return_sequences = True, input_shape = (1, X_train.shape [2])), Dropout(0.2), LSTM(16, activation = ’relu’, return_sequences = False), RepeatVector(1), LSTM(16, activation = ’relu’, return_sequences = True), Dropout(0.2), LSTM(32, activation = ’relu’, return_sequences = True), TimeDistributed(Dense(X_train.shape [2])) ]) model.compile(optimizer = ’adam’, loss = ’mse’) # Train model model.fit(X_train, X_train, epochs = 10, batch_size = 32, validation_split = 0.1, ver-bose = 1) # Make predictions X_pred = model.predict(X_train) mse = np.mean(np.abs(X_pred − X_train), axis = (1, 2)) # Identify anomalies threshold = np.percentile(mse, 97) # Top 3% as anomalies data[“Anomaly_Score_LSTM”] = mse data[“Anomaly_LSTM”] = (mse > threshold).astype(int) # Display results print(data.head()) |
References
- Advantech Co., Ltd. Edge AI Explained: Uses, How it Works & More. Available online: https://www.advantech.com/en-us/resources/industry-focus/edge-ai (accessed on 13 February 2025).
- Meidan, Y.; Avraham, D.; Libhaber, H.; Shabtai, A. CADeSH: Collaborative Anomaly Detection for Smart Homes. IEEE Int. Things J. 2022, 10, 8514–8532. [Google Scholar] [CrossRef]
- Gill, N.S. Edge AI for Smart Home Applications. Available online: https://www.xenonstack.com/use-cases/edge-ai-for-home-applications (accessed on 13 February 2025).
- Xiao, J.; Xu, Z.; Zou, Q.; Li, Q.; Zhao, D.; Fang, D.; Li, R.; Tang, W.; Li, K.; Zuo, X.; et al. Make Your Home Safe: Time-aware Unsupervised User Behavior Anomaly Detection in Smart Homes via Loss-guided Mask. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Barcelona, Spain, 25–29 August 2024; pp. 3551–3562. [Google Scholar] [CrossRef]
- Sarwar, N.; Bajwa, I.S.; Hussain, M.Z.; Ibrahim, M.; Saleem, K. IoT Network Anomaly Detection in Smart Homes Using Machine Learning. IEEE Access 2023, 11, 119462–119480. [Google Scholar] [CrossRef]
- Iturbe-Araya, J.I.; Rifà-Pous, H. Enhancing unsupervised anomaly-based cyberattacks detection in smart homes through hyperparameter optimization. Int. J. Inf. Secur. 2024, 24, 45. [Google Scholar] [CrossRef]
- Holdren, W. Deep Learning Anomaly Detection Using Edge AI. Master’s Thesis, University of Central Florida, Orlando, FL, USA, January 2022. Available online: https://stars.library.ucf.edu/etd2020/1026 (accessed on 14 March 2025).
- Arm Ltd. Edge AI, Arm|The Architecture for the Digital World. Available online: https://www.arm.com/markets/iot/edge-ai (accessed on 15 February 2025).
- Kotevska, O. Privacy by Design in Distributed Edge Systems: Innovating Secure Workflows for Smart Cities. Newsletter. Available online: https://smartcities.ieee.org/newsletter/october-november-2024/privacy-by-design-in-distributed-edge-systems-innovating-secure-workflows-for-smart-cities (accessed on 15 February 2025).
- Zhao, Y.; Zhao, J.; Jiang, L.; Tan, R.; Niyato, D.; Li, Z.; Lyu, L.; Liu, Y. Privacy-Preserving Blockchain-Based Federated Learning for IoT Devices. IEEE Int. Things J. 2021, 8, 1817–1829. [Google Scholar] [CrossRef]
- Li, H.; Ge, L.; Tian, L. Survey: Federated learning data security and privacy-preserving in edge-Internet of Things. Artif. Intell. Rev. 2024, 57, 130. [Google Scholar] [CrossRef]
- Ragab, M.; Ashary, E.B.; Alghamdi, B.M.; Aboalela, R.; Alsaadi, N.; Maghrabi, L.A.; Allehaibi, K.H. Advanced artificial intelligence with federated learning framework for privacy-preserving cyberthreat detection in IoT-assisted sustainable smart cities. Sci. Rep. 2025, 15, 4470. [Google Scholar] [CrossRef] [PubMed]
- Chen, I.; Yan, H.; Liu, Z.; Zhang, M.; Xiong, H.; Yu, S. When Federated Learning Meets Privacy-Preserving Computation. ACM Comput. Surveys 2024, 56, 1–36. Available online: https://dl.acm.org/doi/full/10.1145/3679013 (accessed on 19 February 2025). [CrossRef]
- Shimillas, C.; Malialis, K.; Fokianos, K.; Polycarpou, M.M. Transformer-Based Multivariate Time Series Anomaly Localization. arXiv 2025, arXiv:2501.08628. [Google Scholar]
- Chen, Z.; Chen, D.; Zhang, X.; Yuan, Z.; Cheng, X. Learning Graph Structures With Transformer for Multivariate Time-Series Anomaly Detection in IoT. IEEE Int. Things J. 2022, 9, 9179–9189. [Google Scholar] [CrossRef]
- Ruff, L.; Vandermeulen, R.; Goernitz, N.; Deecke, L.; Siddiqui, S.A.; Binder, A.; Müller, E.; Kloft, M. Deep One-Class Classification. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Volume 80, pp. 4393–4402. Available online: https://proceedings.mlr.press/v80/ruff18a.html (accessed on 14 March 2025).
- Rahim, A.; Zhong, Y.; Ahmad, T.; Ahmad, S.; Pławiak, P.; Hammad, M. Enhancing Smart Home Security: Anomaly Detection and Face Recognition in Smart Home IoT Devices Using Logit-Boosted CNN Models. Sensors 2023, 23, 6979. [Google Scholar] [CrossRef] [PubMed]
- Liu, F.T.; Ting, K.M.; Zhou, Z.-H. Isolation Forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar] [CrossRef]
- Moll, P.J.W.; Potter, A.C.; Nair, N.L.; Ramshaw, B.J.; Modic, K.A.; Riggs, S.; Zeng, B.; Ghimire, N.J.; Bauer, E.D.; Kealhofer, R.; et al. Magnetic torque anomaly in the quantum limit of the Weyl semi-metal NbAs. Nat. Commun. 2016, 7, 12492. [Google Scholar] [CrossRef] [PubMed]
- Kuzdeba, S.; Carmack, J.; Robinson, J. RF Fingerprinting with Dilated Causal Convolutions–An Inherently Explainable Architecture. In Proceedings of the 2021 55th Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 31 October–3 November 2021; pp. 292–299. [Google Scholar] [CrossRef]
- Lim, B.; Arik, S.O.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for Interpretable Multi-horizon Time Series Forecasting. Int. J. Forecast. 2020, 37, 1748–1764. [Google Scholar] [CrossRef]
- Mehta, S.; Rastegari, M. MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer. arXiv 2022, arXiv:2110.02178. [Google Scholar] [CrossRef]
- Wardana, I.N.K.; Gardner, J.W.; Fahmy, S.A. Optimising Deep Learning at the Edge for Accurate Hourly Air Quality Prediction. Sensors 2021, 21, 1064. [Google Scholar] [CrossRef] [PubMed]
- Mahmud, H.; Kang, P.; Desai, K.; Lama, P.; Prasad, S. A Converting Autoencoder Toward Low-latency and Energy-efficient DNN Inference at the Edge. arXiv 2024, arXiv:2403.07036. [Google Scholar] [CrossRef]
- Power Consumption Benchmarks|Raspberry Pi Dramble. Available online: https://www.pidramble.com/wiki/benchmarks/power-consumption?utm_source=chatgpt.com (accessed on 17 February 2025).
- Eames, A. How Much Power Does the Pi4B Use? Power Measurements. RasPi.TV. Available online: https://raspi.tv/2019/how-much-power-does-the-pi4b-use-power-measurements (accessed on 17 February 2025).
- Jaramillo-Alcazar, A.; Govea, J.; Villegas-Ch, W. Anomaly Detection in a Smart Industrial Machinery Plant Using IoT and Machine Learning. Sensors 2023, 23, 8286. [Google Scholar] [CrossRef] [PubMed]
- Bohutska, J. Anomaly Detection—How to Tell Good Performance from Bad. Towards Data Science. Available online: https://towardsdatascience.com/anomaly-detection-how-to-tell-good-performance-from-bad-b57116d71a10/ (accessed on 17 February 2025).
Layer | Function | Key Components |
---|---|---|
IoT Sensor Layer | Collects temperature, motion, and energy usage data | DHT22, PIR motion sensor, smart plugs |
Edge Processing Layer | Detects anomalies in real time at the edge | Isolation Forest, LSTM Autoencoder |
Cloud Layer (Optional) | Stores long-term data, retrains models if needed | Cloud ML APIs, federated learning |
User and Alerting System | Notifies users and enables real-time monitoring | Mobile app, web dashboard |
Feature | Cloud-Based Detection | Edge AI-Based Detection |
---|---|---|
Latency | High (data transmission delays) | Low (on-device real-time processing) |
Privacy | Data stored in the cloud (risk of exposure) | Processed locally (minimized risk) |
Bandwidth Usage | Requires frequent data uploads | Minimal data transmission |
Scalability | Scalable but network-dependent | Scalable but hardware-limited |
Power Consumption | Low (processing offloaded to cloud) | Moderate (depends on device efficiency) |
Aspect | Isolation Forest | LSTM Autoencoder | Hybrid Approach |
---|---|---|---|
Accuracy | Medium | High | High |
False Positive Rate | High | Low | Low |
Energy Consumption | Low | High | Medium |
Retraining Frequency | Rarely Needed | Weekly | Adaptive |
Best Use Case | Real-Time Flagging | Sequential Pattern Analysis | Balanced Approach |
Model | Key Hyperparameters |
---|---|
Isolation Forest | contamination = 0.03, n_estimators = 100 |
LSTM Autoencoder | layers = 3 LSTM layers, units = 32, activation = ReLU, epochs = 50, batch size = 32 |
Model | Total Anomalies Detected | Percentage of Data Anomalous |
---|---|---|
Isolation Forest | 30 | 3.00% |
LSTM Autoencoder | 30 | 3.00% |
Actual/Predicted | Anomaly (1) | Normal (0) |
---|---|---|
Anomaly (1) | 24 (TP) | 6 (FN) |
Normal (0) | 5 (FP) | 65 (TN) |
Actual/Predicted | Anomaly (1) | Normal (0) |
---|---|---|
Anomaly (1) | 27 (TP) | 3 (FN) |
Normal (0) | 3 (FP) | 67 (TN) |
Temperature | Motion | Energy Usage | Anomaly Score IF | Anomaly IF |
---|---|---|---|---|
21.72 | 0.0 | 462.39 | 1.0 | 0.0 |
23.30 | 0.0 | 515.96 | 1.0 | 0.0 |
25.05 | 0.0 | 567.02 | 1.0 | 0.0 |
21.53 | 0.0 | 406.24 | 1.0 | 0.0 |
Temperature (°C) | Motion (0/1) | Energy Usage (W) | Anomaly Score (LSTM) | Anomaly (LSTM) |
---|---|---|---|---|
22.99 | 0 | 484.54 | 0.04372 | 0 |
21.72 | 0 | 462.39 | 0.03132 | 0 |
23.29 | 0 | 515.96 | 0.05711 | 0 |
25.04 | 0 | 567.02 | 0.11847 | 0 |
21.53 | 0 | 406.24 | 0.05983 | 0 |
Actual/Predicted | Anomaly (1) | Normal (0) |
---|---|---|
Anomaly (1) | 28 (TP) | 2 (FN) |
Normal (0) | 2 (FP) | 66 (TN) |
Metric | Isolation Forest | LSTM Autoencoder |
---|---|---|
Inference Time (ms) | 22 | 35 |
Memory Usage (MB) | 5 | 50 |
Power Consumption (W) | 2.8 | 4.2 |
Energy Savings from Quantization (%) | N/A | 35% |
Feature | Isolation Forest (IF) | LSTM Autoencoder (LSTM-AE) | Hybrid Approach (IF + LSTM-AE) |
---|---|---|---|
Anomaly Type Detection | Point anomalies (e.g., sudden energy spikes) | Temporal anomalies (e.g., gradual changes over time) | Captures both types of anomalies |
Accuracy | Moderate (higher false positives) | High (better precision and recall) | Balanced (reduces false positives and improves recall) |
Computational Efficiency | Very high (low complexity, fast inference) | High computational cost (deep learning-based) | Medium (uses IF for quick filtering and LSTM-AE for refinement) |
Energy Consumption | Low (suitable for low-power devices) | High (requires more processing power) | Moderate (optimized through event-triggered execution) |
Latency | Very low (real-time inference possible) | Higher latency due to model complexity | Lower latency than LSTM-AE alone but higher than IF |
Scalability | Highly scalable for large datasets | Requires significant memory and processing | Moderate (trade-off between efficiency and detection quality) |
Suitability for Edge Devices | Ideal for lightweight devices | Best suited for high-performance edge hardware (e.g., Jetson Nano) | Optimized for mixed hardware constraints |
Feature | Cloud-Based Detection | Edge AI Detection (IF and LSTM-AE) |
---|---|---|
Latency | High latency due to data transmission to cloud servers before analysis | Low latency, real-time anomaly detection at the device level |
Privacy | Data are transmitted over the internet, increasing privacy risks | Local processing minimizes data exposure, ensuring better user privacy |
Bandwidth Usage | High bandwidth required for continuous data uploads | Minimal bandwidth use; only necessary metadata are transmitted (if needed) |
Scalability | Scalable but dependent on network availability and cloud infrastructure | Scalable for local devices but limited by computational capacity |
Power Consumption | Relies on remote computation (low device power usage but high cloud energy demand) | Consumes local processing power, but optimizations (e.g., quantization, event-triggered execution) can reduce energy usage |
Model | Precision | Recall | F1-Score | Avg. Inference Time (ms) |
---|---|---|---|---|
LSTM-AE (CASAS) | 86.1% | 83.7% | 84.9% | 48.5 |
Isolation Forest (CASAS) | 79.2% | 69.4% | 74.0% | 17.9 |
Hybrid (CASAS) | 88.2% | 87.3% | 87.7% | 33.2 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Reis, M.J.C.S.; Serôdio, C. Edge AI for Real-Time Anomaly Detection in Smart Homes. Future Internet 2025, 17, 179. https://doi.org/10.3390/fi17040179
Reis MJCS, Serôdio C. Edge AI for Real-Time Anomaly Detection in Smart Homes. Future Internet. 2025; 17(4):179. https://doi.org/10.3390/fi17040179
Chicago/Turabian StyleReis, Manuel J. C. S., and Carlos Serôdio. 2025. "Edge AI for Real-Time Anomaly Detection in Smart Homes" Future Internet 17, no. 4: 179. https://doi.org/10.3390/fi17040179
APA StyleReis, M. J. C. S., & Serôdio, C. (2025). Edge AI for Real-Time Anomaly Detection in Smart Homes. Future Internet, 17(4), 179. https://doi.org/10.3390/fi17040179