1. Introduction
Buildings remain the largest end-use energy consumer, responsible for roughly one-third of global final demand and about one-quarter of anthropogenic CO
2 emissions [
1,
2,
3]. Because people spend up to 90% of their time indoors, IEQ directly affects physiology, cognition, and economic output. Indoor environmental quality (IEQ)—the combined thermal, acoustic, visual, and air quality conditions—has measurable effects: CO
2 increases from 600 to 1 000 ppm slow decision-making by ~15%, and ±2 °C thermal deviations correlate with higher complaints and absenteeism [
4,
5]. Balancing energy demand with IEQ is therefore a multi-objective optimization challenge.
The engineering toolkit for tackling that challenge has progressed through several eras. The 1970s introduced centralized building automation systems (BAS) that applied closed-loop control and delivered tangible energy savings [
4]. Scientometric analyses show that the interest in smart building research has accelerated markedly in the past decade [
6], and systematic reviews underline the persistent integration challenges and opportunities that accompany this growth [
7]. From the early 2020s, low-power wireless protocols such as Zigbee, Thread, and LoRaWAN reshaped BAS into IoT-centric platforms capable of interconnecting thousands of sensing and actuation points [
8].
A fourth, emerging era layers artificial intelligence (AI) onto this IoT fabric. Reviews trace the evolution of AI techniques aimed at boosting building energy efficiency [
9] and show how machine learning models coupled with IoT data streams advance that goal [
10], while field implementations demonstrate intelligent control that reduces energy use without compromising comfort [
11]. When multimodal occupancy sensing—passive infrared, ultrasound, Wi-Fi analytics, and computer vision—is added to the mix [
12], sequence models forecast presence patterns with mean absolute errors below 10% [
13], providing a foundation for model-predictive and reinforcement learning controllers that modulate HVAC equipment proactively.
These capabilities typically reside in a three-layer architecture. The perception layer hosts modular nodes for temperature, humidity, illuminance, CO
2, acoustics, and motion; the cognitive layer embeds gradient boosting, convolutional, and graph neural networks, refining digital twin simulations with sub-room granularity; and the control layer executes multi-objective optimization constrained by predicted occupancy and weather. Hybrid physical–statistical baselines reduce CV-RMSE by up to 8% relative to classical regression [
14], and urban-scale extensions that account for morphology, albedo, and micro-climate improve heat flow assessment across districts [
15]. Parallel work on occupant-centric metrics links control decisions to subjective comfort and disease prevention [
16], and the publication of large, open datasets is accelerating the replication and benchmarking of these approaches [
17].
To ground the discussion, narrative literature mapping was performed across the Web of Science, Scopus, and IEEE Xplore databases. The search used combinations of keywords including “smart building”, “indoor environmental quality”, “energy efficiency”, “machine learning”, “IoT”, “comfort”, “health”, and “HVAC control”. We retained peer-reviewed articles published from 2015 to 2025 that explicitly link building energy consumption to IEQ, comfort, or health, while excluding gray literature, duplicates, and studies with non-transparent methodologies. The resulting corpus of 180 publications provides the evidentiary foundation for the analysis that follows. The context of climate change further highlights the urgency of implementing such integrated solutions [
18].
This review offers an integrative synthesis across building sensing, digital twins, AI-based HVAC control, and emerging LLMs for building management systems (BMS), consolidating segmented surveys rather than claiming algorithmic novelty. Specifically, we frame the literature along an end-to-end sequence—sensors → digital twin → AI control → LLMs in BMS. By mapping interfaces, data handoffs, and evaluation metrics across these layers, this review bridges strands that are often treated in isolation and surfaces system-level gaps (occupancy inference-to-control feedback, twin calibration-to-policy transfer, and LLM grounding in BMS ontologies). In contrast to specialized studies, our paper traces the full evolution—from sensors to generative AI—within a single, human-centric architecture. We introduce the integrated concept of digital twins of the building and the human (DT-B + DT-H) and analyze how advanced AI models, including LLMs, enable the shift from reactive control to a proactive and hyper-personalized environment aimed at maintaining health and comfort.
The next parts present recent statistics on energy use alongside findings that link IEQ to health and productivity. The section titled “Ecosystem of Sensors, Data, and Predictive Intelligence” then examines the shift from legacy building automation systems (BAS) to IoT-enabled buildings and details the architecture and algorithms that deliver proactive comfort and health management. A dedicated methodology section elaborates on the narrative mapping process, followed by a discussion of outstanding challenges and practical recommendations. The conclusion synthesizes the findings and outlines avenues for future work. The remainder of the paper follows this pipeline to maintain conceptual continuity and highlight cross-layer dependencies.
2. Methodology
To frame the evidence synthesis and ensure transparent, reproducible reporting, the study selection process followed the PRISMA-ScR guideline and is summarized in
Figure 1. Of the 300 records initially identified, 21 duplicates were removed, leaving 279 publications for title/abstract screening; at this step, 11 records were excluded, and full texts were sought for 268 articles. However, 12 full texts could not be retrieved, resulting in 256 reports assessed for eligibility. Following full text assessment, a further 27 publications were excluded for different reasons (out of scope—10; insufficient data—7; non-ENG/RUS—10), and 229 studies were included in the analysis. Screening was performed by two independent reviewers, with disagreements resolved by a third expert; the search window was 2009–2025; sources: Scopus, Web of Science, Nature, IEEE Xplore, ACM Digital Library, PubMed, and Google Scholar. Thus, the sequence of stages from “Identification” to “Studies included” and the counts at each step are fully consistent, as shown in
Figure 1.
The distribution of included sources aligns with the review structure.
Section 1 (Introduction) comprises 18 papers covering sectoral statistics, the direct link between IEQ and health/well-being/productivity, the significant energy footprint of buildings, and the transition from traditional buildings to BAS and then to IoT-enabled smart buildings; it also includes overview pieces on AI for smart buildings (architecture “building + comfort + health + AI”) and methodological notes (databases, keywords, and publication window).
Section 3 (Ecosystem of Sensors, Data, and Predictive Intelligence) aggregates 75 studies on how the architectural layers work together (IoT layer → communication → processing and control → actuation), on multimodal sensing (incl. smartwatches, fitness trackers, and smartphones), and on the personal comfort and health digital twin concept.
Section 4 (Artificial Intelligence for Prediction, Control, and Interaction) includes 107 studies spanning foundational ML (supervised/unsupervised), advanced AI for adaptive control, reinforcement learning for HVAC control, generative AI and LLMs for comfort/health management, and personalization via fine-tuning.
Section 5 (Discussion) synthesizes 29 publications focusing on data quality and scarcity, the robustness and generalization of fine-tuned models, cybersecurity risks, the paradox of control and loss of agency, privacy, unintended health consequences of over-optimization, ethical and societal implications, and future trajectories/solutions.
Table 1 consolidates the essential design and outcome information across the reviewed studies, covering building and climatic context, sensing and control stacks, analytical and optimization methods, data provenance, baseline configurations, metric definitions and validation procedures, the experimental scenario (simulation, field, or testbed), and a concise appraisal of costs and benefits. Where studies report comparable quantities, ranges or absolute changes (e.g., energy use, reliability, latency) are included. Missing details are marked “not reported.” Bracketed numerals map to the reference list.
3. Ecosystem of Sensors, Data, and Predictive Intelligence
Smart buildings have evolved from simple collections of stand-alone, hard-wired controls into richly instrumented, data-driven ecosystems that sense, think, and act in real time [
21,
26]. At the foundation of this ecosystem sits a dense fabric of environmental sensors—temperature, humidity, CO
2, and illuminance—and occupancy detectors ranging from passive infrared (PIR) and image-based sensors to Bluetooth beacons [
27]. These devices stream time-stamped telemetry into an Internet-of-Things (IoT) network that blends legacy building automation protocols with low-power wireless standards like Zigbee and LoRaWAN, guaranteeing a secure communication between devices and controllers [
26,
28].
The first computational stop for these data is the edge layer, where embedded gateways perform noise filtering and low-latency inference [
20]. Edge-resident machine learning models can predict imminent threshold violations—say, a rapid rise of CO
2 in a crowded meeting room—and trigger timely corrective actions while circumventing round-trip delays to the cloud [
20].
Cleaned and summarized datasets flow onward to cloud platforms that offer virtually unlimited storage and processing power [
29,
30]. Here, historical telemetry is fused with external variables like weather forecasts and dynamic electricity tariffs [
31]. The same cloud tier hosts holistic digital twins: data-driven replicas that couple Building Information Model (BIM) geometry with live sensor feeds [
32,
33]. These twins enable what-if experiments, fault diagnostics, and model-predictive control routines that optimize comfort, air quality, and energy consumption simultaneously [
32,
34].
At the application layer, operators and occupants interact with the system through dashboards, mobile apps [
35], and, increasingly, conversational agents underpinned by large language models (LLMs) [
36,
37]. A facilities manager can visualize thermal maps and equipment health indices, while an employee can nudge personal set-points or receive notifications.
The closed-loop interaction unfolds as follows. Sensors perceive the indoor environment and occupant presence [
22]; edge analytics detect patterns [
23]; predictive models in the cloud forecast future states [
26]; optimization engines compute control trajectories that balance comfort, health, and energy costs [
30]; and actuators in HVAC and lighting systems execute the commands [
24]. Throughout, cybersecurity and privacy measures—device authentication, encrypted channels, and privacy-preserving federated learning—safeguard the integrity of both data and people [
23,
33].
Several concrete scenarios illustrate the value of this architecture. Demand-controlled ventilation systems use CO
2 predictions to increase airflow minutes before concentrations reach cognitive performance limits [
23]. Behavior-aware HVAC scheduling mines occupancy data to pre-condition spaces only when needed, trimming peak loads [
34,
35]. Multi-objective reinforcement learning agents, trained inside digital twins, can orchestrate environmental controls to cut electricity bills significantly [
35,
36]. In safety-critical events such as fires, the same sensor and control backbone can automatically shut down air handling units and unlock egress routes [
28].
By intertwining pervasive sensing, reliable data pipelines, and predictive intelligence across edge and cloud domains [
25], modern smart building ecosystems transcend reactive automation and deliver proactive, human-centric environments [
9]. Occupants benefit from healthier air and greater comfort, while owners realize significant gains in energy efficiency, maintenance planning, and asset resilience [
9,
10].
The realization of proactive, human-centric control is underpinned by a multi-layered cyber–physical architecture that orchestrates the cyclical flow of data from the physical environment to cloud-based intelligence and back to physical actuators (illustrated in
Figure 2). This integrated framework is designed to sense, reason, and act in real-time, transforming raw telemetry into intelligent control actions that continuously optimize for occupant comfort, health, and energy efficiency.
At its foundation lies the perception and actuation layer, which constitutes the direct interface with the building and its occupants. A dense fabric of environmental sensors (for temperature, humidity, and CO
2) and multimodal occupancy detectors—ranging from passive infrared (PIR) to advanced millimeter wave (mmWave) radar—generates a continuous stream of high-granularity telemetry [
21,
22,
31]. Conversely, actuators within HVAC systems, lighting, and automated shading execute the control commands that physically alter the indoor environment.
This raw data is transmitted upwards through the communication and IoT layer. This layer leverages a hybrid of legacy wired protocols like BACnet and modern wireless standards such as Zigbee and LoRaWAN to ensure reliable and secure connectivity between the physical devices and the computational layers [
20,
22,
33]. The first computational stop for this data stream is the edge computing and local layer. Here, local gateways and embedded machine learning (ML) modules perform critical low-latency functions, including data filtering, aggregation, and immediate inference for time-sensitive responses. This stage is crucial for immediate actions and for reducing the data load on the cloud infrastructure [
24,
25].
Cleaned and pre-processed data, often fused with external data from sources like weather APIs and dynamic energy tariffs [
26], proceeds to the cloud analytics and digital layer. This central intelligence hub hosts scalable data lakes for long-term storage and automated machine learning (AutoML) platforms for developing predictive models [
25]. The core of this layer is the digital twin—a dynamic, physics-informed virtual replica of the building continuously synchronized with real-world sensor feeds. This twin serves as an invaluable sandbox for simulating complex scenarios, performing fault diagnostics, and training reinforcement learning agents for multi-objective optimization [
27,
29].
Finally, the applications and UI layer manages the interaction with human stakeholders. Through intuitive interfaces like dashboards, mobile applications, and conversational agents, facility managers and occupants receive actionable insights derived from the cloud analytics [
19,
32]. Crucially, this layer also allows users to provide direct feedback and explicit preferences, creating a human-in-the-loop control system that is vital for personalization and closing the reinforcement cycle [
19].
The entire architecture operates as a continuous, closed loop. Data flows upward from sensors to the cloud, where it is transformed into intelligent decisions. These decisions propagate downwards as control signals to actuators. The system’s adaptive capability is perpetually reinforced by feedback mechanisms, including model updates sent from the cloud back to the edge layer, ensuring it learns and evolves over time. This synergistic integration of sensing, connectivity, and predictive intelligence is what creates a truly proactive, resilient, and human-centric building ecosystem.
3.1. Environmental Sensing Technologies
The proactive management of comfort and health in smart buildings relies on a multi-layered technological architecture. This framework provides the essential structure for sensing the environment, communicating data, processing information, and enacting physical changes. This section deconstructs this architecture by outlining its operational workflow, presenting a visual model, and analyzing its role in current research and future development.
The effectiveness of smart building systems hinges on the comprehensive and accurate data gathered by the sensing layer, which comprises a variety of IoT environmental sensors designed to capture real-time information about the indoor environment. Studies on indoor air quality (IAQ) in sensitive environments, such as daycare centers, emphasize the important role of IoT sensors in monitoring parameters like temperature, humidity, and CO
2 levels to ensure safe and healthy conditions for occupants [
37]. Beyond environmental parameters, the sensing layer also extends to energy consumption monitoring, with industrial IoT (IIoT)-based submetering solutions deploying IoT-enabled submeters to provide real-time energy consumption data from critical equipment, enabling optimized energy management and waste reduction in manufacturing facilities [
38]. Similarly, for building equipment energy saving optimization, online monitoring systems leverage IoT sensors to collect data that, when integrated with Building Information Modeling (BIM), allows for the intelligent control of systems like air conditioners [
39]. The optimal design of communication topology for wireless sensor networks (WSNs) is also crucial for efficient data collection, considering factors like network energy consumption and stability to implement fully distributed optimal control approaches in IoT-enabled smart buildings [
40].
The communication layer is responsible for the reliable and efficient transmission of data from sensors to processing units and control signals to actuators, employing various wireless and wired technologies. The communication layer is responsible for the reliable and efficient transmission of data from sensors to processing units, as well as control signals to actuators, using various wired and wireless technologies. LPWAN (Low-Power Wide-Area Network) technologies are gaining increasing popularity due to their ability to transmit data over long distances with low energy consumption, making them ideal for many IoT applications in smart buildings [
41]. Furthermore, the integration of advanced networking paradigms like Named Data Networking (NDN) with edge computing can significantly enhance IoT performance in wireless and mobile networks by optimizing data retrieval and caching, thereby reducing latency and improving reliability [
42]. This focus on efficient communication protocols is vital for supporting the real-time demands of smart building operations.
The processing layer is where raw sensor data is transformed into actionable insights through advanced computational techniques, often involving a combination of edge computing and cloud-based analytics. Edge computing allows for localized data processing, reducing latency and bandwidth requirements, which is particularly beneficial for real-time IoT applications [
42]. For more complex analyses and long-term data storage, cloud integration remains essential. The development of decentralized machine learning frameworks for IoT is also enhancing security, privacy, and efficiency in these cloud-integrated environments [
43]. The architectural flexibility of microservices-based IoT platforms is a key enabler in this layer, allowing for scalable, interoperable, and dynamic ecosystems that can efficiently handle the distributed nature of IoT devices [
44]. Moreover, machine learning techniques are increasingly employed within this layer for critical tasks such as attack detection in IoT networks, bolstering the cybersecurity posture of smart building systems [
45]. The final layer, actuation and control, translates the intelligent decisions made by the processing layer into physical changes within the building environment. This involves sending commands to various devices, such as HVAC systems, lighting controls, and air purifiers, to adjust conditions according to desired comfort, health, and energy efficiency parameters. The continuous feedback loop from the sensing layer allows the system to monitor the effects of these actions and refine its control strategies, leading to adaptive and optimized building performance.
The architectural structure of smart buildings is not isolated but operates within a broader Internet of Things (IoT) ecosystem. This wider context includes initiatives aimed at creating open IoT innovation ecosystems for smart cities, which focus on ensuring interoperability through open communication and data standards [
46,
47]. These principles also extend to the industrial sector, where IoT smart factory ecosystems are being developed based on Software-Defined Networking (SDN) to enhance communication and increase efficiency in industrial processes [
48].
A key aspect across all levels of this architecture is security. As the number of IoT devices grows, so does the number of potential vulnerabilities, prompting ongoing research focused on developing countermeasures and building robust security systems [
49]. Additionally, the sustainability of widespread IoT deployment is becoming an increasingly important issue, driving interest in energy supply solutions. Thermoelectric Generators (TEGs) are one promising option for powering autonomous sensors, converting thermal gradients into electrical energy and thereby supporting their self-sufficiency while reducing the environmental footprint of smart building components [
50]. The ongoing evolution of this multi-layered architecture, coupled with advancements in AI, communication technologies, and sustainable power solutions, promises to deliver increasingly intelligent, responsive, and resilient smart buildings that proactively manage occupant comfort and health.
Table 2 provides an overview of the sensor and IoT technologies used in smart buildings, including measured parameters, areas of application, and advantages. The first technology is indoor air quality sensors (IAQ sensors), which measure temperature, humidity, and CO
2 levels. They are used for monitoring air quality in sensitive areas such as medical facilities or laboratories and provide data on air conditions to maintain safe environments. The second technology is industrial IoT, designed for measuring and optimizing energy consumption in industrial buildings. It enables real-time monitoring, energy resource management, and a reduction in energy losses. The third technology is IoT sensors integrated with Building Information Modeling (BIM). These sensors collect data aimed at improving equipment energy efficiency and are used for the online monitoring and automatic control of building systems. The fourth technology is wireless sensor networks (WSNs), which collect data for implementing distributed control. They are used for data acquisition in smart buildings and support the development of communication topologies that consider energy consumption and system stability. All listed technologies are supported by references to scientific sources.
3.2. Personal and Occupancy Sensing Technologies
The cornerstone of any system designed to proactively manage comfort and health is the ability to collect accurate, non-stop data about the people in a building. Choosing the right sensing technology, however, is a major challenge, forcing a trade-off between how detailed the data is, how much it costs to implement, and, most critically, how much it invades occupant privacy. This section offers a deep dive into the key sensing technologies, assessing how they fit into the main goal of creating personalized, healthy, and comfortable spaces.
The first layer of data collection comes from ambient (or occupational) sensors placed within a space. The most basic of these, passive infrared (PIR) detectors, are effective for simply detecting presence, but their binary “present/not present” logic cannot assess the number of people or their metabolic activity, which limits their use in advanced HVAC systems [
51]. Computer vision (CV)-based systems offer the highest level of detail: they can accurately count people, identify their postures, and determine activity levels, which are all direct inputs for thermal comfort (PMV) calculation models. However, their use is associated with severe privacy violation concerns, making them unacceptable for most residential and office environments [
52].
To solve this dilemma, privacy-preserving technologies have been developed. Wi-Fi Channel State Information (CSI) analysis uses existing infrastructure to estimate occupancy with reasonable accuracy [
53], but its performance can be unstable due to changes in the environment [
54]. A more robust alternative is millimeter wave (mmWave) radar. These devices do not create images but can determine the number and location of people with high precision, as well as detect micro-motions like respiration rate, which is directly relevant for assessing air quality and health status [
55,
56].
However, to transition from group-level to personalized management, wearable devices are essential, as they provide physiological data that is inaccessible to ambient sensors. These devices track key parameters such as Heart Rate (HR) and Heart Rate Variability (HRV), which are reliable proxies for metabolic rate (heat production) and stress levels, enabling control systems to deliver personalized cooling or adapt lighting to reduce stress [
57]. In addition, skin temperature offers a direct indicator of a person’s thermal balance, allowing the system to react more quickly to individual discomfort. Activity levels, captured via an accelerometer, are also a critical parameter for dynamically calculating metabolic rate in thermal comfort models like PMV [
58].
A comparison of wearable devices reveals that while budget-friendly trackers (Xiaomi Mi Band) are suitable for general activity assessment, more expensive devices (Oura Ring, Apple Watch) provide more accurate temperature and HRV data, which is preferable for sophisticated health and stress models [
59].
As it turns out, no single technology is a perfect solution by itself. Therefore, the most sensible approach is sensor fusion—combining data from multiple sources. For example, a system might use an mmWave radar to accurately count people in a zone, while data from their wearables determines the average metabolic load for that zone. This allows the HVAC system to be both energy-efficient and highly personalized to individual needs.
Table 3 summarizes this comparative analysis, contrasting the primary technologies based on their key characteristics and relevance for comfort and health management systems. The data acquired from this multi-layered sensor network serves as the foundation for the digital twin models discussed in the next section.
In this review, thermal comfort is operationalized using the PMV/PPD and/or adaptive model, with compliance expressed as the proportion of occupied hours within the ASHRAE 55 acceptable zone (approximately PMV −0.5…+0.5, i.e., ≤10% PPD) [
60,
61]. Indoor air quality (IAQ) adequacy is assessed according to the Ventilation Rate Procedure (VRP) specified in ASHRAE 62.1; no fixed indoor CO
2 limit is imposed, with CO
2 instead applied as a proxy indicator for ventilation performance [
62,
63,
64]. In steady-state conditions, indoor CO
2 levels approximately 600–800 ppm above the outdoor baseline are used as a practical indicator of insufficient ventilation, guiding demand-controlled ventilation (DCV) settings in conjunction with space-specific CO
2 metrics [
62,
63]. These anchors enable the direct linkage of sensor data streams to measurable comfort and IAQ targets, which are then applied in the subsequent control and evaluation stages [
60,
62,
65].
In
Table 4 we demonstrated the detailed mapping of key sensor modalities to their primary measurable metrics, associated standard anchors, representative multimodal fusion strategies, and trade-offs across cost, energy consumption, privacy, and bias dimensions. The “Primary metric(s) mapped” column specifies the measurable outputs linked to comfort (e.g., operative temperature, and PMV/PPD), IAQ (e.g., CO
2 as a ventilation adequacy proxy), or ergonomics (e.g., acoustic noise levels). “Standard anchor” identifies the normative framework or procedural logic—such as ASHRAE Standard 55 for thermal comfort or ASHRAE Standard 62.1 [
66] VRP for ventilation—used to interpret each metric in a building performance context. “Typical fusion” lists common multimodal combinations (e.g., CO
2 + PIR/mmWave for robust demand-controlled ventilation, Temp + RH + air speed for PMV computation, and mmWave + wearables for metabolic load estimation) that enhance the accuracy and resilience to single sensor limitations. The “Key trade-offs” column summarizes practical considerations that influence deployment decisions, including initial and operational costs, energy overhead, data privacy implications, and inherent biases or sensing limitations. This synthesis supports the review’s objective to link sensing technologies explicitly to measurable targets, applicable standards, and integrative control strategies for occupant-centric and energy-efficient building management.
3.3. Digital Twin Technology
The data gathered from this sensor network serves as the foundation for the next leap in building management: digital twin (DT) technology. A DT is a living virtual replica of a physical asset, synchronized with it in real time. In this context, as outlined in foundational reviews [
68,
69], the concept is split into two interconnected entities: the digital twin of the building (DT-B) and the digital twin of the human (DT-H).
A DT-B is a dynamic, physics-based model of the building. For over a decade, the consensus has been that “gray-box” models, which blend physical principles with data-driven AI, are optimal [
70]. The historical challenge, however, has been making them fast enough for live control. Recent research has tackled this problem from different angles, addressing key bottlenecks that previously hindered real-time operation. For instance, to solve the problem of calibration speed, a 2024 framework called GenPhysiCal can perform a full calibration cycle in just 0.04 s [
71]. In parallel, to overcome the challenge of simulation complexity, a 2025 study uses an AI-based “surrogate model” to predict complex airflow in milliseconds instead of minutes, maintaining a high accuracy [
72]. Together, these advances make real-time DT-B operation feasible.
A fast, accurate DT-B provides a ‘virtual sandbox’ for learning control policies. Comparing the evolution of this approach from classical MPC [
70] to modern AI shows rapid progress. A foundational 2019 review [
73] and subsequent application papers [
74] showed that deep reinforcement learning (DRL) could achieve significant energy savings of 15–40%. More recently, a 2024 study using Bayesian optimization demonstrated a more holistic benefit: it improved occupant comfort by 38% while simultaneously reducing energy consumption [
75]. This highlights a critical shift in research goals, moving from a pure energy focus to a dual-objective approach that balances efficiency with human well-being.
Parallel to the DT-B, the concept of the human digital twin (DT-H) is evolving. While the DT-H is already a major topic in personalized medicine for testing treatments in silico, as shown in recent 2024 reviews [
76,
77], its role in smart buildings is to model an individual’s comfort. This is achieved by feeding biometric signals from wearables into sophisticated thermoregulation models. For example, the Fiala model can be parameterized in real time with wearable data to predict personal thermal sensation, a method demonstrated by Al-Khafaji et al. [
78].
The culmination of this technology is the hybrid “building–human” twin, which links the DT-B and the DT-H, as shown in
Figure 3. This architecture materializes the “human-in-the-loop” control concept reviewed by Papantoniou et al. [
79], where a person’s physiological state (from the DT-H) directly informs the building’s control systems (the DT-B). This is no longer just a theoretical concept; a 2024 pilot study on a Korean smart campus provided hard evidence, demonstrating that this hybrid system increased the time occupants spent in their individual comfort zone from 62% to an impressive 85% [
80].
Despite this significant potential, widespread implementation faces hurdles. As a major 2024 systematic review by El-Amroussi et al. confirms, challenges like the interoperability between different models and the standardization of data exchange remain critical barriers [
81]. Nevertheless, the digital twin architecture provides the necessary infrastructure to treat thermal comfort, air quality, and health not as separate challenges, but as a single, dynamically optimized objective function.
Interoperability and semantic standards such as brick ontology and IFC (ISO 16739) [
82] have emerged as critical enablers for seamless digital twin integration in the built environment. A recent analysis of IFC-based workflows for embedding Environmental Product Declaration (EPD) data illustrates that, despite its comprehensive schema, semantic alignment challenges persist when integrating Life Cycle Assessment (LCA) information into Building Information Modeling (BIM) and digital twins [
83]. Addressing these gaps requires not only standardized data models but also structured ontologies capable of supporting cross-domain integration. In this context, multi-domain ontologies anchored in IFC have been proposed as a basis for incremental digital twin conceptualizations [
84]. Beyond structural information, semantic web technologies have been applied to domains such as indoor environmental quality, where the IFC ontology structure is complemented by frameworks like the Smart Applications REFerence ontology (SAREF) to enhance semantic interoperability and data reuse in building performance monitoring [
85]. At the asset end-of-life stage, reviews of BIM-based digital deconstruction approaches reveal that ontologies accepting IFC inputs can streamline demolition planning, material recovery, and reuse processes, supporting sustainability objectives within digital twin environments [
86]. This interoperability imperative also extends to infrastructure, where strategies for reinforced concrete bridge management in compliance with Italian regulations emphasize open formats such as IFC to maintain compatibility between inspection data, 3D modeling, and maintenance systems [
87].
Cultural heritage contexts further highlight the role of semantic standards. Heritage BIM (HBIM) approaches for twentieth-century concrete structures leverage IFC as a foundation for integrating geometric documentation, historical metadata, and sensor data into coherent digital twins [
88]. In a similar way, HBIM workflows for built heritage utilize IFC to support interoperable virtual and augmented reality applications, fostering wider access and collaboration among stakeholders [
89]. Interoperability is equally relevant in operational management, where integrating BIM, Internet of Things (IoT), and facility management systems through semantic construction digital twins—structured around IFC—addresses the challenges of linking real-time sensor data with as-built models [
90]. Systematic reviews of BIM-based structural health monitoring confirm that ISO 16739-compliant IFC schemas facilitate sensor data integration, enabling the continuous monitoring of assets such as historical churches and masonry bridges [
91]. The role of IFC in enhancing stakeholder collaboration is underscored in studies on BIM-driven sustainable heritage tourism, where embedding semantic information into IFC models supports richer, more accessible cultural heritage experiences [
92]. These works demonstrate that adherence to semantic standards such as IFC—complemented where appropriate by domain-specific ontologies—remains central to achieving the full interoperability potential of digital twins across building, infrastructure, and heritage domains.
5. Discussion
An analysis of studies reveals a steady shift from rigid rule-based systems to adaptive environments that account for occupant comfort, health, and behavior. AI algorithms are the main driver: applying deep reinforcement learning lowers energy consumption by 10–35% without sacrificing comfort [
199]. These gains are enabled by a growing array of sensors, connected devices, analytics tools, and user interfaces that form the technological foundation of smart buildings.
The transition from pilot projects to large-scale deployment is hampered by data, integration, and ethical issues. AI models require large, high-quality datasets, yet standard public datasets are scarce, complicating the objective comparison of solutions [
199]. The spread of audio and video sensors increases privacy risks [
200].
Data quality is closely tied to broader ethical challenges, including the reproducibility crisis. Only a small fraction of key findings can be confirmed by independent studies, casting doubt on the reliability of the underlying datasets [
201]. The lack of harmonized data management practices across research and clinical centers exacerbates scarcity: quality control methods vary, making datasets hard to interpret and compare [
202,
203]. Data sharing is hindered by concerns over privacy, intellectual property, and the reputational risks associated with exposing data quality shortcomings [
201,
204].
Recent studies highlight the necessity of aligning smart building systems with frameworks like the General Data Protection Regulation (GDPR). GDPR’s “privacy by design” principle mandates data anonymization and user consent in occupant monitoring systems, emphasizing the need for granular access controls to mitigate re-identification risks in HVAC and occupancy datasets [
202]. Federated learning (FL) has been proposed as a privacy-preserving approach, allowing decentralized model training without sharing raw data and thereby maintaining occupant anonymity [
203]. Cross-border data transfers introduce additional complexities, with EU–US Privacy Shield alternatives evaluated post invalidation. Hybrid cloud–edge architectures are recommended to localize data processing and avoid jurisdictional conflicts [
197]. Ethical AI deployment also requires explainability and bias mitigation. Explainable AI (XAI) tools like SHAP and LIME can audit black box models in building automation, ensuring transparency for stakeholders [
205]. Integrating XAI with BIM allows occupants to query and understand AI-driven HVAC adjustments, providing transparency in automated decision-making [
206]. Synthetic data generation, for instance via GANs, addresses data scarcity while adhering to privacy laws. Synthetic occupancy patterns can retain statistical fidelity without exposing real identities [
207].
The following approaches mitigate these risks. Federated learning trains models on distributed devices without transferring sensitive information [
207]. Synthetic datasets generated with GANs and variational auto-encoders enlarge training corpora without using real occupant records [
207]. Blockchain provides secure, auditable data exchange between stakeholders [
208]. Unified data models and ontologies enhance system interoperability and simplify information interpretation across platforms [
209].
One of the primary cybersecurity risks linked to fine-tuning is the heightened susceptibility to adversarial attacks. Because fine-tuning can erode the robustness of the original pre-trained model, attackers may exploit these weaknesses by injecting minor perturbations into inputs that force the model to misclassify or generate harmful outputs. Adversaries can engineer inputs that cause a machine learning-based intrusion detection system to ignore malicious activity or, conversely, to flood it with false alerts that overwhelm security operations [
210]. Deploying fine-tuned models in mission-critical cybersecurity applications without strong adversarial defenses can lead to major breaches, data leaks, and operational disruptions [
210,
211].
One of the key technical challenges lies in integrating modern AI systems with the existing infrastructure in buildings. Most current automation systems were not designed with openness or compatibility in mind, which makes integration costly and technically demanding [
206]. Digital twins (DTs) have become a promising solution. By creating virtual replicas of real-world buildings, DTs allow researchers and engineers to simulate AI-driven management strategies safely before deploying them in actual environments [
212]. These digital models can merge data from multiple sources—such as Building Automation Systems (BAS), IoT devices, and weather forecasts—into one framework. However, building and maintaining these models is resource-intensive and becomes more difficult when scaling across many buildings [
213]. Research now points to the value of standard frameworks and open-source platforms, which could simplify development and reduce compatibility barriers. Using widely accepted communication protocols like BACnet/IP, Modbus TCP, or newer ones like Matter also helps unify devices from different vendors [
214]. Middleware tools and API gateways are essential for bridging the gap between legacy systems and modern AI-driven platforms.
As AI tools grow more complex, understanding how they make decisions becomes more difficult, especially with methods like DRL and large-scale language models. This lack of transparency, sometimes called the “black box” issue, makes it harder for stakeholders to trust the system [
208]. An AI system acting autonomously could make decisions that are not immediately understandable, raising concerns about comfort, safety, or operational reliability.
To counter this, explainable AI (XAI) methods are gaining traction. Tools like SHAP and LIME help break down model predictions and offer insights into why a particular decision was made [
209]. Incorporating human-in-the-loop (HITL) approaches is also important—these combine AI automation with oversight from human operators who can intervene or validate outcomes when needed [
214]. Frameworks for testing and verifying AI systems are essential to ensure they meet safety and performance requirements before being deployed [
215]. On a broader scale, legal and ethical guidelines tailored to AI use in buildings are needed to ensure responsible deployment and increase public acceptance [
216].
Future developments must blend engineering, policy, and user-focused design. Successful smart building systems should be open, secure, and easy to use, while always centering the needs of the people living and working inside them. Future research should explore scalable architectures, fair data governance policies, and technologies that are both adaptable and explainable. Continued collaboration between researchers, developers, policymakers, and occupants will be essential to ensure that AI in smart buildings improves quality of life without introducing new risks.
A forward-looking roadmap is required to steer innovation in smart building AI while maintaining ethical responsibility and practical feasibility. Recent studies converge on staged frameworks that coordinate research, development, and policy over clearly defined short- and mid-term windows. These frameworks extend beyond technical targets to foreground transparency, trust, security, and sustainability.
Figure 11 distills the priority areas and concrete action steps for the next five years, outlining a clear path toward responsible and scalable smart building systems.
Large-scale smart buildings deploy extensive networks of sensing devices to monitor environmental and operational parameters, yet the size and complexity of these networks create significant maintenance and optimization challenges. Thermoelectric Generators (TEGs) have shown significant potential for powering autonomous sensors by converting thermal gradients into electricity, which reduces maintenance requirements and enhances sustainability [
50]. Similarly, optimal wireless sensor network (WSN) topologies can minimize energy consumption and improve stability through distributed control, addressing scalability challenges in IoT-enabled buildings [
40]. Accurate device localization and optimization in large buildings remain complex due to heterogeneous sensor densities and dynamic environmental conditions. Integrating edge computing with Named Data Networking (NDN) can reduce latency and improve reliability in data retrieval, enabling real-time sensor coordination [
43]. Decentralized machine learning frameworks further enhance privacy and efficiency in cloud–edge architectures, which is crucial for managing distributed sensor networks [
42]. Autonomous optimization is another critical focus. Continuous calibration frameworks automate the synchronization of a building’s digital twin with its physical counterpart by dynamically adjusting model parameters in real time, ensuring a high accuracy with minimal human intervention. The system continuously ingests sensor data and uses a pre-trained model to estimate unobservable variables like occupant count or equipment heat, updating the digital twin instantly. A real-time validation loop compares the model with live data to maintain precision, while the automated denoising and handling of missing data enhance resilience. This enables the digital twin to self-correct, adapt to changing conditions, and support proactive decisions [
71]. AI-driven surrogate models can predict airflow patterns, optimizing HVAC sensor placement and reducing computational overhead [
72]. Energy harvesting solutions integrated with Building Information Models (BIMs) support intelligent energy management, reducing the dependency on external power sources [
39].
Privacy-preserving techniques such as federated learning and blockchain help mitigate security risks while ensuring scalable sensor networks [
207]. Hybrid cloud–edge architectures automate the distribution of computational tasks, with edge devices handling lightweight processing and cloud servers managing intensive computations. This setup enables real-time autonomous optimization, where local models are continuously trained and updates aggregated to refine the global AI model. Automated feedback loops allow the system to adapt quickly to changing conditions without human intervention. Privacy and data integrity are maintained by keeping raw data on local devices and transmitting only anonymized model updates. This approach ensures efficient, secure, and intelligent operation across the network [
197]. These advancements prioritize sustainability, scalability, and autonomy in large-scale smart building systems.
6. Conclusions
This review summarizes the current state of research on the application of artificial intelligence technologies for proactive comfort and health management in smart buildings. The analysis shows that the integration of IoT-based sensor networks, advanced communication protocols, and AI-driven data processing forms the foundation of a new generation of buildings capable of continuously adapting to changing environmental conditions and individual user needs. Through a multi-layered architecture—including sensing, edge and cloud computing, actuation systems, and user interfaces—modern buildings are shifting from reactive automation to proactive, human-centered ecosystems.
The review demonstrates that AI methods, particularly supervised and unsupervised machine learning, reinforcement learning, and their deep and multi-objective variations, significantly enhance the efficiency and adaptability of control strategies. Applications include occupancy and thermal comfort prediction, as well as the implementation of energy saving policies without compromising user well-being. Digital twin technology amplifies these capabilities by enabling safe training, optimization in simulation environments, and real-time closed-loop control. Studies report energy savings in the range of 15–35% while also improving indoor environmental parameters such as CO2 concentration, temperature stability, and air circulation.
Despite this progress, several challenges remain. Data silos, the lack of standardized evaluation methods, and privacy concerns related to personal sensing technologies hinder widespread adoption. Moreover, the complexity and resource intensity of deploying advanced AI models—especially those involving deep learning and hybrid digital twins—require substantial investment and interdisciplinary collaboration. The need for explainable AI, privacy-preserving algorithms, and ethical principles is growing as buildings evolve into cyber–physical systems that handle increasingly sensitive user data. Future research should focus on the development of modular and interoperable platforms that allow integration across different building systems and equipment vendors. Equally important is the publication of open datasets and the creation of simulation environments that ensure the reproducibility and comparability of results. Personalized control strategies, comfort modeling using wearable devices, and self-adaptive intelligent agents are key directions to bridge the gap between technical feasibility and practical usability.