Next Article in Journal
Edge Intelligence: Edge Computing for 5G and the Internet of Things
Previous Article in Journal
IoV and Blockchain for Traffic Optimization in Ro-Ro Terminals: A Case Study in the Spanish Port System
Previous Article in Special Issue
Are Strong Baselines Enough? False News Detection with Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Applying Large Language Model Analysis and Backend Web Services in Regulatory Technologies for Continuous Compliance Checks

by
Jinying Li
1 and
Ananda Maiti
2,*
1
Australian Maritime College, University of Tasmania, Newnham, TAS 7248, Australia
2
School of IT, Deakin University, Waurn Ponds, VIC 3216, Australia
*
Author to whom correspondence should be addressed.
Future Internet 2025, 17(3), 100; https://doi.org/10.3390/fi17030100
Submission received: 10 January 2025 / Revised: 13 February 2025 / Accepted: 17 February 2025 / Published: 22 February 2025

Abstract

:
Regulatory technologies (RegTechs) are a set of electronic and digital technologies applied to check compliance in industrial processes. Such applications also aim to simplify the process of data collection and exchange according to the expected format over the cloud or the internet. Industrial processes are required to meet basic regulatory requirements according to law and follow a set of industry practices. Industry practices must be compliant with the basic regulatory requirements. Such applications also need a high level of privacy to protect the individual participant’s data from competitors but are revealed to the relevant regulatory agencies. However, there cannot be a standard data procurement method, as the industrial processes are different for individual businesses and often involve various stages of data collection with different aims. Also, the regulatory requirements may be changed over time. These challenges can be addressed over an online system that uses large language models (LLM) to perform continuous compliance checks. With LLMs, RegTech can be easily scaled up to meet new requirements. It can also help with data analysis and reformatting for different stakeholders in RegTech, such as producers, supply chains, regulators, and financial institutions. It can check for acceptable values with regards to RegTech through either numeric comparisons or enumerations matching. In this paper, we propose a comprehensive RegTech framework backed by LLM and web services. We propose a method to measure the accuracy of LLM in returning appropriate responses for RegTech queries and herein analyze several LLMs to conclude that they are satisfactory for basic tasks, but a dedicated LLM is needed for RegTech. Furthermore, we test the LLM’s tool-calling capabilities to identify and use dedicated functions in the form of web services to enhance the analytical accuracy and consistency of RegTech-related prompts.

1. Introduction

Primary industries, such as agriculture, construction, fisheries, and forestry, are vital to global economies [1,2,3]. However, these industries are subject to a complex web of regulations [4,5,6,7], often involving multiple jurisdictions and agencies [8]. Compliance with these regulations is crucial for ensuring product safety, environmental sustainability, and fair-trade practices [9,10,11]. Traditional compliance methods can be time-consuming, costly, and prone to human error [8,12,13,14]. This is where regulatory technology (RegTech) can streamline compliance checks and documentation with an online process enhanced by modern computing and Internet of Things (IoT) tools [15]. RegTech uses such technology to improve regulatory efficiency, reduce compliance costs, and enhance regulatory oversight [16,17,18]. The primary objective of applying RegTech solutions in primary industries is to release regulatory burdens for all stakeholders by streamlining compliance processes, reducing administrative burdens, and improving decision making.
In this paper, we consider the agriculture sector, which is a major primary industry, as a case study. It involves the typical stakeholders of RegTech: assessors, including regulators; assessees, including individual farmers; and artificial intelligence (AI) tools like large language models (LLM) and deep learning services. The main aim of this system is compliance checks.
Despite its potential, implementing RegTech in agriculture faces significant obstacles. Technically, limited connectivity and access to necessary technologies and digital skills [19], coupled with the ever-changing landscape of agricultural regulations [15], pose substantial challenges. Integrating disparate agricultural systems and addressing cybersecurity and data privacy concerns [20,21] further complicate matters. Economically, the high cost of developing and deploying RegTech solutions [22], combined with a lack of awareness and understanding among farmers [21], creates significant barriers to adoption [23]. Socially, resistance to change from those accustomed to traditional farming methods can also hinder progress [24]. Finally, the political landscape, with its complex global supply chains, multiple jurisdictions, and diverse stakeholders, makes standardizing regulatory compliance a particularly difficult undertaking. LLMs can mitigate many of these problems by providing a continuous, easy-to-use interface to technology.

1.1. Compliance Check and Its Challenges

Compliance checks comprise the following processes:
(1)
Recording information during production in industries;
(2)
Auditing and checking the correctness of the information and its adherence to the rules;
(3)
Considering exceptional situations and making appropriate decisions;
(4)
Certification of products as suitable;
(5)
Making the outcomes available to the public in an immutable manner.
A compliance check is mainly a one-to-one process between a producer, such as a manufacturer or farmer, and the corresponding industry regulator. However, there are many such producers and many levels of compliance regulators, e.g., governments and assessment bodies, as shown in Figure 1. The critical challenges of compliance on a large industrial scale are as follows:
  • It is impossible for a human to remember all the rules;
  • Although the majority of the compliance regulations are fixed, some critical unexpected issues can appear in a lengthy process, such as in agriculture, leading to exceptional assessments;
  • Conditions external to the industrial process may change, leading to minor temporary changes in the compliance expectations;
  • Data are distributed among many individual sources, especially considering that data are generated by both humans and smart IoT devices. Currently, many data are paper-based, and there is a lack of digital synchronization;
  • There are privacy concerns about the sharing of data among peer producers and the regulator.
The agricultural sector confronts a multitude of compliance challenges, primarily stemming from a complex and ever-evolving regulatory landscape [25,26,27]. Frequent regulatory updates necessitate constant adaptation and vigilance from regulated entities, especially small farms, which often struggle with limited resources to prioritize and meet compliance requirements [4,9,28,29]. While digital technologies offer potential solutions, they also introduce new challenges, such as the digital divide [30,31,32], effective data management [33], non-invasive sensors [34], and cybersecurity risks [33,35,36,37]. Regulators also face obstacles such as geographic diversity of the fragmented sectors [38] and resource constraints such as EU cross-compliance, which highly relies on extensive monitoring, ref. [39] hindering effective oversight and enforcement. Challenges also exist in the auditor’s technical competence and field knowledge, and the potential negative impact of the common interests between certifying companies and their clients on audit quality may lead to the issue of certification reliability [40].

1.2. RegTech LLM–Web Service: Contributions and Limitations

This paper explores the potential of large language models (LLMs) to address real-time issues in compliance checks for industries, specifically for agriculture. We herein investigate a scenario involving regulators and farmers who must collaborate effectively. By employing an LLM as an impartial intermediary, real-time challenges can be swiftly resolved. This paper utilizes data from a cherry orchard and honeybee apiary to analyze and illustrate the practical application of LLMs in agricultural regulatory contexts. We delve into the advantages, limitations, and capabilities of LLMs in this specific use case. Also, here, we consider a continuous compliance check process that keeps on going for a lengthy period of time for the industrial process and is repeated in cycles. The key contributions of this paper are as follows:
  • Basic framework/set of functionalities of different automation needed for compliance checks in a RegTech system, including the following:
    o
    A new multi-level prompt categorization with respect to difficulty in processing time series data;
    o
    Measuring accuracy with respect to consistency as needed by RegTech.
  • Analysis of LLM characteristics for RegTech and its performance with regards to these basic compliance checks automation.
  • A web service (WS)-based improvement on the performance of LLM.
The remainder of the paper is structured as follows: Section 2 discusses the current state of LLM, automated compliance checks, and regulatory technologies. This provides the context and motivation for the proposed LLM–web service–RegTech system. Section 3 presents the critical process of compliance check, using agriculture as an example, and the LLM–web services-based regulatory technology architecture. Section 4 presents the LLM performances, and Section 5 presents the observations and future work.

2. Related Work

This section discusses the current status of developments in LLMs, along with the compliance check process.

2.1. Role of Large Language Models vs. Traditional AI/ML

Traditional AI systems, such as rule-based systems and expert systems, are often rigid and inflexible, and they struggle to adapt to changing environments and unforeseen circumstances [41]. These systems excel at specific tasks but cannot generalize knowledge and apply it to new situations. Traditional AI systems rely heavily on structured data and predefined rules [42], limiting their ability to learn from unstructured or noisy data. Some classic AI techniques, such as expert systems, can be computationally intensive, especially for complex problems [43]. To address these limitations, modern AI approaches like natural language processing have emerged. These techniques enable AI systems to learn from data, make intelligent decisions, and adapt to changing conditions. In addition, the explainable AI (XAI) system is emphasized by many researchers, as it can enhance trust, transparency, and accountability in the decision-making process [44,45,46].
LLMs are excellent at summarizing text [47,48,49,50]. They are good at logical reasoning in the short term and when the context is clear [51]. However, they are limited in analytical capabilities if asked to perform a calculation in real time. They can understand the requirements and develop the procedure to obtain the results, but they are poor at executing the procedure to obtain the results, particularly if the data are great in quantity. In order to solve this, tool-calling abilities are being developed for LLM [52,53]. These enable the LLM to call external functions for performing a dedicated task. The LLM must create the proper input for these functions, and the function should return a proper output for LLM consumption.
In this paper, we analyze the current general purpose of LLM’s capabilities for compliance-related functionalities. We term LLM’s tool calling as “back-end web services (WS)”. Web services are well-established technologies that execute codes with inputs via an HTTP request and return a response in the form of text or other media [54]. They are suitable in terms of security and scalability when deployed over the cloud.

2.2. Compliance Checks

Compliance checks are typically time-limited manual activities between a regulator and producers that happen occasionally. In this paper, we propose a web-based compliance check mechanism that is not time-limited but a continuous automated process between the producers and online LLM and WS as well as the regulators, when needed.

2.2.1. Source of Data

Data about production can come from various sources, including human and IIoT devices. Human data can be recorded through mobile apps and desktop applications. While there are still producers relying on manual paper-based data recording, even that data can be converted to digital information. However, the data are mainly structured, i.e., specific data are recorded for specific purposes. The following are some categories of data in the case of agriculture:
  • Sensors collect real-time numeric data on soil moisture, temperature, plant health, and other factors [55,56]. This is a well-established technology;
  • Pest control measures with traps are still mostly manually recorded, although IoT devices are being developed for this as well [57,58]. On-site detection of pesticide residues, such as through biosensors and colorimetric sensors, is also mostly recorded manually [59,60];
  • Drones and satellites capture high-resolution images of fields, allowing farmers to assess crop health, identify pest or disease outbreaks, and monitor irrigation effectiveness [61,62,63];
  • QR code, RFID, and NFC offer a secure and easy way for consumers to access product information with their smartphones [64,65,66,67];
  • Several advanced technologies that can directly generate digital data in the form of images or text are being developed to uniquely identify farm products, such as DNA barcodes and unique chemical makeup [68,69,70,71];
  • Compliance checks in the industry can involve a variety of methods or work, including document review, field inspections, and laboratory testing, if necessary [72]. These will ultimately create documents that are usually in a structured or semi-structured format.

2.2.2. Industrial Process and General Compliance

The industrial process is a series of activities that create a product. The product must be worthy of human consumption. The process must follow the best practices. Often, there are industrial practices that are above the bare minimal requirement for safe production. The industrial process will involve human activities as well as autonomous equipment, thus involving a degree of uncertainty. Another key feature of an industrial process is that it will be finite, i.e., as a duration called a process cycle. The process cycle would be repeated again and again to produce the same product, possibly with minor variations with time. From a regulatory point of view, the aim is to ensure that the process complies with the safety, environmental, and human resource requirements. Regulatory procedures ordinarily do not judge the quality of the product or process as long as it is above the minimum requirements.
In the case of agriculture, the industrial process is the crop cycle, which takes a few months. The end product is food, which should be safe in terms of content and pests. If exported, it needs to meet biosecurity standards. It also involves both manual labor and IIoT devices to generate localized and continuous data. There are compliance requirements throughout the whole process, from seeding to packaging. Regardless of the length of the process, it is often irreversible. For example, if the agriculture process is corrupted, it impacts food safety and quality. Also, certain events may need to be recorded at a particular time, and failing to record them leaves unfillable gaps in the data for regulatory compliance checks.
Compliance efforts are hindered by three key challenges: accurate regulatory interpretation [73], seamless data collaboration or transformation [74], and efficient compliance verification. Accurate interpretation of complex regulations is essential to establish clear compliance standards. Seamless data sharing and analysis across systems and organizations is vital for effective compliance monitoring.
Traditional methods, such as third-party audits and certifications, can be costly and time-consuming, often involving on-site visits [75]. Additionally, communication barriers between producers and regulators [76], inconsistent regulatory application [25], and high turnover of supervisors or auditors [77] can further impede and complicate compliance. These third-party entities can provide valuable expertise and assistance, but they can also increase costs and introduce additional layers of complexity to the communication process [78]. Inconsistencies in the application of regulations by different regulators, leading to varying interpretations of rules, inconsistent enforcement actions, and increased compliance costs, can create uncertainty and unfairness for regulated parties. Changes in actual human staff within supervisory units can introduce variability in regulatory oversight. When experienced auditors leave, their knowledge and expertise can be lost, making it difficult to maintain consistent standards.
Automated and digital solutions may help to address these challenges by streamlining processes, improving accuracy, and enhancing transparency. By leveraging digital technologies, auditors can reduce the frequency and duration of on-site visits, leading to significant cost savings. Measures such as remote monitoring, data analytics, automated document review, and virtual audits can all contribute to a more efficient and cost-effective compliance process.

2.2.3. Automated Compliance Check

Automating compliance checks and verification processes can significantly enhance efficiency and accuracy [79]. This empowers human experts to concentrate on higher-value endeavors, such as strategic decision making and intricate problem solving. Early approaches to automation, such as “Black Box” and “Gray Box” methods [80], relied heavily on hardcoded rules and manual interpretation [79]. These methods proved to be costly to maintain and challenging to adapt, and they lacked a generalized framework for modeling rules and regulations [79]. Semi-automated methods, exemplified by the RASE (requirement (R), applicabilities (A), selection (S), and exceptions (E)) methodology, aimed to translate regulatory texts into machine-processable formats [81]. While this approach improved efficiency, it still necessitated substantial manual intervention [79]. More advanced methods leverage natural language processing (NLP) techniques [82]. Rule-based NLP methods, although effective, demand significant manual effort in data training [79]. Domain-specific regulatory texts are generally more amenable to NLP analysis than general texts [83]. Semantic and syntactic information extraction methods, frequently combined with ontology and machine learning, offer a more automated approach [84].
In the general compliance automation field, some researchers focus on provision interpretation or requirement analysis [85,86,87,88,89,90,91], while others also cover compliance checking [92,93,94]. The technologies involved in these above works include AI, NLP, ML, etc. However, these methods encounter challenges in scalability and adaptability. NLP techniques, such as the rule-based approach, often necessitate extensive manual feature engineering [95] or manual verification for information [82], while deep learning models typically require large, annotated datasets for optimal performance [96,97,98]. Large language models (LLMs) such as GPT-4 [99], PaLM [100], Galactica [101], and Llama [102] offer promising solutions to these challenges. These models excel in few-shot learning and can understand complex language structures with minimal labeled data [103,104]. They have the potential to adapt to evolving regulations and extract structured information from regulatory texts, making them useful tools for automated compliance checking [79,105,106,107,108].
Ref. [109] delved into the potential of large language models (LLMs) to enhance the explainability of AI systems. Ref. [110] explored the potential of using large language models (LLMs) like ChatGPT in the field of diagnostic medicine, particularly in digital pathology. Despite some challenges, LLMs offer significant potential to revolutionize diagnostic medicine by improving accuracy, efficiency, and accessibility [110]. In the architecture, engineering, construction, and operations (AECO) industry, the current methods for compliance evaluation are time-consuming and prone to errors [111]. Ref. [112] developed an LLM-based automatic compliance check (ACC) artifact to evaluate its performance in classifying regulations, identifying rule dependencies, and extracting information from BIM (building information modeling) models. The results demonstrate the potential of LLMs to significantly improve efficiency, transparency, and trust in building permitting.
Ref. [113] provided a comprehensive overview of the potential applications of ChatGPT and other large language models (LLMs) in the fields of accounting and finance. The review highlighted the potential benefits of LLMs in various tasks, including financial analysis, auditing, tax compliance, financial reporting, and educational purposes. Ref. [114] also discussed LLM’s transformative potential in financial analysis, risk management, fraud detection, customer service, and regulatory compliance. Ref. [115] confirmed these opportunities but also outlined the challenges, such as data quality, model bias, and ethical considerations. Ref. [105] shared the view that large language models (LLMs) have the potential to revolutionize the financial auditing industry by automating the process of regulatory compliance verification. However, they face challenges, particularly with non-English languages. The study compared the performance of different LLMs and suggested that LLMs can be a valuable tool for financial auditors, even without relying on expensive proprietary software. However, the paper focused on a pre-defined compliance check. The data used seem pre-determined, with all the levels and parameters already established. While this approach is viable, a conventional deep-learning method could achieve higher accuracy. Thus, the potential advantages of LLMs were not fully realized in this specific application. This approach may not be applicable in a dynamic environment like agriculture. The reliance on custom datasets would not translate well to the complexities and real-time needs of agricultural compliance.
In the food industry, ref. [116] demonstrated the significant potential of LLMs to improve efficiency and accuracy in legal compliance and regulatory analysis by using BERT and GPT models to classify legal provisions and automate compliance checks. The study also highlighted challenges like handling complex legal language and ensuring the reliability of LLM-generated outputs [116]. Other researchers have focused on its potential in agriculture [117,118,119,120]. Ref. [118] conducted a comprehensive review of the potential of multi-modal large language models (MM-LLMs) to revolutionize agricultural practices, including applications in crop management, soil health, livestock management, and precision agriculture. Ref. [121] highlighted the potential of tools like ChatGPT to address various challenges faced by farmers by providing accurate and timely information to enhance farm management and simplify regulatory processes in agriculture. LLMs are considered to be valuable tools to assist in understanding, interpreting, and applying agricultural safety and health regulations [122]. They can provide summaries, explanations, and practical guidelines. However, it is crucial to remember that LLMs are not a substitute for a thorough understanding of the regulations and industry’s best practices [122]. Any information or advice provided by an LLM must be reviewed and verified, especially when it is specific and detailed [122]. ChatGPT can analyze large datasets to generate valuable insights for digital and precision agriculture as well as nanotechnology applications [121]. It can also simplify complex regulatory documents, making them more accessible to farmers and stakeholders [121]. These studies provide evidence that LLM has the potential to revolutionize agriculture.

2.2.4. Compliance Cost

Every industry, such as the agricultural sector, is subject to a wide range of regulations, including those related to the environment, water resources, biotechnology, transport, and work health and safety. The sheer volume and diversity of these regulations impose a significant time burden on producers, e.g., farmers or manufacturers, who must diligently track and comply with them. This burden is further compounded by the need to report to multiple agencies at various jurisdictional levels and the frequent changes to these regulations.
The financial impact of regulatory compliance on agribusinesses is substantial. Costs can be categorized into one-off and ongoing expenses [76]. One-off costs are incurred when new regulations are introduced, while ongoing costs vary with business operations. These costs encompass direct payments to the government, expenses associated with meeting regulatory requirements, time and resources spent on compliance activities, and losses resulting from delays in applications and approvals [123]. It was found that environmental regulatory requirements such as acquiring licenses and permits were the costliest area, requiring significant external assistance and posing a significant burden on these industries [123].

2.2.5. Law Is Code and Code Is Law

“Law is Code” is a revolutionary approach that involves translating legislation, regulations, and policies into machine-readable code [124]. By encoding rules, ambiguity is reduced, interpretation becomes more manageable, and compliance becomes more straightforward. Such transformation could facilitate the development of digital services that can automatically apply and enforce rules, streamlining processes and improving efficiency. Furthermore, it enhances transparency by making rules accessible and understandable to a broader audience. “Code is Law” [125], a related concept suggesting that rules can be directly encoded into software, making code itself a form of law. This approach can further enhance the efficiency and precision of rule enforcement. However, it also raises important ethical and legal questions regarding accountability, bias, and the potential for unintended consequences.
When combined with the power of large language models (LLMs), “Law is Code” becomes even more powerful. LLMs can effectively generate, refine, and understand rules, reducing the technical burden and making the process accessible to a wider range of users. This synergy between “Law is Code” and LLMs empowers organizations to create and manage complex rule sets with greater ease and accuracy. However, the “Code is Law” principle is not easily applicable with generative AI systems due to their inherent opacity. The black-box nature of current LLM systems requires new regulatory approaches, as traditional methods based on explicit code control are no longer sufficient [126].

2.2.6. Key Issues with Compliance Checks

Expert knowledge is crucial in several key areas of traditional compliance procedures. First, understanding and interpreting complex regulations, standards, and guidelines is essential. Experts must identify critical compliance points and define specific entities (objects) and their relevant characteristics (attributes) to be checked. Second, compliance checking involves validating data against regulatory requirements, detecting anomalies, and interpreting data within the broader industry and regulatory context. Finally, handling arbitrary events requires expert decision making, clear explanations, and adaptability to changing circumstances.
However, manual compliance checks that rely on experts in primary industries can be time-consuming, costly, and prone to human error. As regulatory landscapes become increasingly complex, the need for efficient and accurate compliance solutions becomes even more critical. Additionally, manual checks can be inconsistent, as different inspectors may interpret regulations differently. Scaling manual checks to accommodate growth and increased regulatory requirements can be difficult. Furthermore, periodic inspections may not be sufficient to identify and address real-time issues.
To overcome these limitations, digital solutions offer a promising avenue for improving compliance efficiency and accuracy. For instance, IoT sensors and data analytics enable real-time monitoring of environmental conditions, livestock health, and crop growth [62,63,127,128,129]. Blockchain technology ensures product traceability and verifies ethical and sustainable practices [130,131,132]. AI and machine learning algorithms can identify patterns, predict risks, and automate routine tasks [133,134,135]. Remote sensing and satellite imagery facilitate land use monitoring and environmental compliance [136,137,138]. Drone technology offers efficient inspection, monitoring, and mapping capabilities [139,140].
These digital tools can automate routine tasks, analyze vast amounts of data, and identify potential compliance risks. Businesses and regulators can streamline compliance processes, reduce costs, and improve accuracy. For example, a cherry farm can use IoT sensors and blockchain technology to automate the tracking of pesticide use and environmental conditions, ensuring compliance with minimal manual intervention. The data collection at the producers’ end about the products and their production process is streamlined and very well developed. This can be used for continuous monitoring by AI tools in real time. However, human users are unable to ask analytical or pre-emptive questions about these processes unless they are experts in interpreting and analyzing the collected data. Often, this is due to the complex nature of processes and a single human’s inability to understand such complexities.
The current research in automated legal text processing faces significant limitations. Firstly, it heavily relies on sentence-level analysis, which can be insufficient for capturing complex legal concepts [107]. Secondly, automation strategies often lack clear justifications for their decisions and either adopt a coarse-grained approach or require substantial manual effort to build [107]. These works also focus on a one-time compliance check. However, for continuous compliance checks, a lack of clear justification in compliance automation strategies within primary industries may lead to several significant problems. It can undermine trust in the automation system, especially when decisions have far-reaching consequences for producers and consumers. It can potentially hinder maintenance efforts. In contrast to one-time compliance checks [112], we consider a continuous compliance check in this paper. It goes on for a time period during which the compliance rules may be altered due to external factors, and the events are expected in a specific order. In the single compliance check, all data products are gathered before being checked against a fixed set of rules. In the continuous process, the data may sometimes be old or experienced, and there are more chances of data corruption. This is the reason we designed our LLM-based RegTech compliance check around a time series dataset, which reflects the changing data and the relevant conditions over time.
A comprehensive and accurate compliance report generated by LLM can enhance transparency and accountability. RegTech LLM can develop mechanisms to explain the system’s outputs to its users, particularly in cases of unexpected or controversial decisions. RegTech LLM can enable human users to intervene immediately when necessary, especially for complex or high-stakes decisions, and provide a safety net.

3. Digitalization of Compliance Checks with RegTech LLM

In this section, we present the regulatory system requirements, the proposed regulatory systems supported by LLM—RegTech LLM—their key features, and expected functionalities.

3.1. Design Requirements of the Digital Regulatory Systems

The design of a regulatory system should consider the expected inputs from various sources, capabilities, and industry expectations.

3.1.1. Changes in Regulations

Most industries, like the agri-food industry, are subject to frequent regulatory changes, such as new food safety standards, labeling requirements, or traceability protocols. These changes can be complex and time-consuming to implement, especially for large-scale operations. To address this challenge, AI agents can be employed to monitor regulatory changes, interpret their impact, and automatically update systems and processes, such as in the medical device area [141]. As illustrated in Figure 2, AI agents continuously monitor regulatory databases and news feeds, utilizing NLP to extract key information from regulatory documents. They analyze these changes, assess their impact on existing traceability procedures, and update systems accordingly. Additionally, AI agents can monitor product movements in real time, flagging deviations from regulatory requirements and generating alerts for potential issues.

3.1.2. Key Regular Capabilities of RegTech LLM

With regard to the regular use of RegTech LLM, there are four key capabilities that an LLM should possess to assist in compliance checks effectively:
  • Flagging: This is a process for setting alerts on variables to check if they cross a threshold. Flagging can be set with conventional software, but with LLM, it can be customized with multiple conditions based on human input;
  • Analysis: The LLM must possess numerical analysis capabilities to answer user questions regarding regular and unexpected situations. This is further outlined in the next section.
  • More advanced features include the following:
  • Reasoning: This is an advanced feature from the farmer’s perspective, allowing the creation of conclusions based on the analysis. Reasoning requires the LLM to be trained in past decisions made by human users and the rules interpreted from the guidelines;
  • Suggestions: Once the reasoning is completed, LLM can potentially provide suggestions on the next steps to follow.
In this paper, we do not delve into all the reasoning and suggestions and limit ourselves to flagging and analysis only.

3.1.3. Industry Expectations

In many compliance scenarios, businesses need to make quick decisions. In today’s fast-paced business environment, the ability to quickly adapt to changing regulations and respond to compliance challenges can be a significant competitive advantage. Delays in compliance can lead to costly fines, reputational damage, and operational disruptions. Responsive RegTech solutions can help businesses minimize these risks.
In terms of human factors, LLM can resolve the two biggest problems with respect to compliance. A key issue with compliance is having the knowledge of the requirements at the right time. While it is possible for expert workers to remember the most common rules and record the data accordingly in a timely manner, it is still difficult to remember everything. This problem becomes more complex if the compliance is checked by a non-expert or if the rules are relaxed/changed for a particular process cycle. The other issue is on-demand analysis. Often, the producers may want to try new things by analyzing their own past performance and potentially stimulating new deals to see if they would still meet the competition. However, if they do not have an expert on-site, such analysis would require additional costs. LLM can help with this by providing a comprehensive guide to all the rules, both base and temporary. The two main features are as follows:
  • Responsiveness: RegTech LLM should be ready to answer in a real-time chat, such as in ChatGPT. It can respond in seconds with a human-friendly output, thus allowing users to receive fast advice. The producers do not need to wait for expensive expert advice;
  • On-demand analysis: Existing compliance software can provide fixed compliance check features. LLM can give flexibility in the questions users can ask. This means an analysis can be fine-tuned to specific user situations and requirements. LLM has to be capable of the following:
    Step 1.
    Understanding the question.
    Step 2.
    Determining the procedure to obtain the results.
    Step 3.
    Performing the analysis according to step 2.
    Step 4.
    Producing a human-friendly output.
In this context, the evaluation of the RegTech LLM is different from that of conventional LLM applications.
(a)
The inputs, i.e., the prompts, need to be succinct and short, as human users are not expected to explain too much. This can be addressed by considering only 0-shot prompts;
(b)
The outputs must satisfy the question’s requirements. A majority of RegTech questions and their corresponding data will be discrete. This means the output will contain a countable number of distinct elements ( ε ), and thus, if any of these elements (m) are missed by LLM or if new extra elements are added to the output (e), it will result in an incorrect response from the LLM. As such, we define the output accuracy ( λ ) of a RegTech LLM response as follows:
λ = ε m   ε × ε ε + e = ε m ε + e  
where ε can be lists of strings, sets of rows from a time series dataset, sets of rules, sets of web service functions for selection, or sets of format specifications. Variations in the outputs do not matter for RegTech LLM. If the same information is represented in multiple formats, it does not create any problems for the target users.
Ultimately, with respect to RegTech, from an industry perspective, RegTech LLM is expected to achieve the following targets:
  • Product safety: This is related to the safety of using or consuming the industrial process’s product. For the agriculture sector, this means adhering to stringent food safety regulations, including Hazard Analysis Critical Control Point (HACCP) principles, to prevent foodborne illnesses and protect public health;
  • Environmental protection: This is about quantifying and reducing the environmental impact of the industrial process. In agriculture, this is complying with environmental regulations to minimize the industry’s impact on natural resources, e.g., responsible pesticide use, water conservation, and soil conservation practices;
  • Human resource management: This concerns the human resources in the industrial process and how their time and efforts are managed. It is important to relate human performance to the product, e.g., timeliness issues; if someone has to check and enter data manually, that should be recorded;
  • Traceability: An industrial product is expected to change ownership multiple times. Thus, it is important to implement a robust traceability system to identify the source of any potential issues, such as product recalls or quality concerns.

3.2. Key Components of LLM Compliance Checks—Automation Functionalities Needed by LLM

For streamlining the compliance check process, several types of functionalities must be provided by AI and/or LLM, with or without the use of external dedicated services:
  • Numeric checks: These checks are arithmetic and logical checks to determine if the recorded numeric values are within acceptable thresholds. In the context of agriculture, environmental and production data are often recorded as time series. Typically, if IIoT is implemented, the time series is generated by devices and thus records time. If IIoT is not implemented, and the data are recorded manually, the time series may be more incomplete or less orderly. However, the aim would be to continuously monitor the recorded numeric value to check if it is within the acceptable range;
  • Milestone checks: This is a special case of numeric checks that check against a specific date. These checks are meant to confirm that the expert event has happened within the documented timeline. The events are not expected to be continuous but to be within time limits. Milestone checks are when we expect the event to happen within time, and compliance fails if the event does not occur by that time;
  • List checks: A key aspect of compliance checks is labeling. This is obviously in the form of strings, and list checks would involve logically comparing the data entered with the expected values from the list. It is essential for users to record their data using pre-approved terms to ensure verifiability;
  • Flag/binary/fault checks: Sometimes, unexpected events occur and must be flagged immediately. Unforeseen events do not inherently mean a compliance failure, as they can be remedied within a time limit. These checks are typically binary, i.e., whether there is an error or not, for example, whether there was a pest on the farm.
Apart from the above four, there is a more complex form of compliance checks when the compliance data involve more than numeric lists and date/time data.
5.
Document check: This is checking the format, relevance, and correctness of data entered in each section and pocket of data. It is invariably composed of list checks and could also have flag checks;
6.
Process checks: These higher-level checks ensure the correctness of the whole industrial process by analyzing smaller chunks of data. This would involve checking for events that should occur in an expected sequence or for missing data;
7.
Image and video checks: LLMs have multimodal capabilities to interpret and describe visual data by analyzing a given image (and possibly videos). These models, when integrated with vision modules, can process and analyze images to identify objects, recognize patterns, detect anomalies, and provide descriptive captions or insights. For RegTech LLM, computer vision algorithms are already used for visual inspection of several production-related activities, e.g., seed quality checks.
Figure 3 shows a hierarchy of different checks. A process check would comprise document and numeric checks, which have to be in the correct order with respect to time. Document checks can be compared to list checks, milestone checks, and binary checks based on writing documents (digital or scanned from paper) as well as document completeness. Numeric checks are composed of milestone checks and binary checks as well as basic arithmetic and logical checks.
In agriculture, effective compliance is crucial to ensure adherence to regulations, standards, and best practices. A robust compliance framework should incorporate a multi-faceted approach, encompassing process checks, document and numeric checks, and milestone and list checks. Process checks verify the sequence and timing of agricultural processes, such as planting, harvesting, and pesticide application, and they identify any gaps or missing data. Document and numeric checks assess the accuracy, completeness, and format of documents, such as farm records, certification documents, and regulatory reports. They also monitor continuous data, such as temperature, humidity, and soil moisture, to identify anomalies or trends that may signal non-compliance. Milestone and list checks track key events in the agricultural process and verify the accuracy and consistency of lists, such as approved pesticides, fertilizers, or seed varieties.

3.3. Architecture of the LLM–Web Services for RegTech

Figure 4 shows the RegTech LLM–web service architecture. There are multiple sources of data from producers such as farmers and shipping companies. Data can be generated by sensors as well. The source of these data is the assessee, who regularly checks for compliance. Human users would ask LLM questions under the following conditions:
(a)
Regular checks where the user asks the same question but expects updated results;
(b)
Some adverse situations have happened, and they require advice. In this situation, the question would be very specific and more difficult for the LLM to answer. As such, the prompt questions need to be detailed to provide accurate answers and specify the exact situation in which the user is asking the question;
(c)
Ask questions with relatively relaxed rules. Human users may ask about a hypothetical situation to see if making any changes in the industrial process could still satisfy compliance, even if less comfortably;
(d)
Comparative questions where human users would want to compare their own performance with historical data or, in the case of the regulator (or assessor), compare between the producers concurrently.
The assessor provides the following guidelines:
(i)
Base guidelines, which currently are in the form of text and diagram documents. These are created by governments or regulatory bodies;
(ii)
Additional temporary guidelines are in response to changes in the situation during a process cycle. These guidelines would not be part of the base guideline and are temporarily implemented in the specific cycle;
(iii)
There may be additional operational guidelines from the manufacturers of components used in subsequent productions by other manufacturers. In these cases, compliance also includes adhering to the manufacturer’s guidelines along with the regulatory guidelines.
In this paper, we only consider (i), as agriculture typically does not involve (ii) or (iii).
The LLM acts as an interface to human users, acting upon all the data sources, providing compliance features, and performing any formatting if needed. We define this as a virtual AI agent. It takes all the raw data from the assessee or data generators; i.e., the producer takes all the real-time guidelines/rules from assessors to perform the compliance checks. The output of the compliance checks can be put into any format corresponding to the assessor’s requirements from different jurisdictions. Ultimately, the output of the compliance is made available to the end consumers who only view the compliance checks.
We tested LLM’s functionalities and, unfortunately, found them inadequate for RegTech on its own. It is either too slow for human interaction or cannot handle large quantities of data. So, it needs to be backed by web services that perform specific functionalities. Such web services can be called by the LLM as part of the tool-calling features of LLM and the output used to create the final output. The web services would have limited flexibility in operation and cannot answer any specific question. The LLM, on the other hand, can determine precisely what the user needs and use the web services to calculate the output.

3.4. Key Characteristics of a Virtual AI Regulatory Agent

The virtual AI agent uses LLM to act as an assistant to both parties, facilitating communication and reducing uncertainties. Unlike traditional compliance processes, the LLM promotes a more collaborative approach. Queries in the form of prompts are submitted to the LLM, which analyzes the information and forwards it to the relevant farmer. The farmer provides additional context, and the LLM can iteratively analyze the data to reach a mutual understanding. Such queries will almost always be 0-shot; i.e., the user does not have the ability to give examples for more complex queries.
A virtual agent can monitor data continuously, albeit at intervals, for each user during each iteration of an industrial process. This continuous monitoring does not inherently need LLM. However, the reports or results generated by constant monitoring with deep learning or machines become the source of information for the LLM later when the users actually ask a query. Thus, it is important to design and record the outcomes correctly in a timely manner.
Virtual AI regulatory agents are capable of continuous operation, accumulating knowledge and experience over time. They can gain experience indefinitely. Additionally, they can share their knowledge with new regulatory human officers, ensuring knowledge preservation and training. However, it is essential to recognize that AI agents are not infallible. Human oversight will remain indispensable to ensure accuracy, fairness, and ethical decision making. Virtual AI regulatory agents with LLM can enable large-scale summarization of reported information. They can work 24/7 and consume data continuously from multiple sources, widen monitoring, and present summarized information. These conjectures can then be verified by humans quickly.
The virtual agent’s LLM interpretations must be automated and upgradable. The LLM needs to extract rules from given textual regulatory guideline sources, such as manufacturing or production guidelines, standard operating procedures, and laws. It then preserves these rules to check the data against. The models can also be retrained to remove or modify a rule. Rule identification and upgrade are developer- or regulator-centric activities. This feature of LLM is effectively a form of “Law is Code”. However, as the codes are not explicitly released for regulations, it is up to LLM to create the rules. The key challenges in this are the following:
(a)
The number of variables being tracked (α)
(b)
The number of logical conditions being tracked (ε).

3.5. Traditional Software vs. LLM for Compliance Check

LLM-based compliance checks offer significant advantages over traditional software in various aspects. Traditional software totally relies on rigid, predefined rules and static datasets, struggling to adapt to dynamic regulatory landscapes. In contrast, LLM-based software can learn from new data and evolve to meet changing compliance requirements.
Traditional software requires significant manual effort and code modifications for updates, whereas LLM-based software can be easily expanded through training and fine-tuning processes such as retrieval-augmented generation (RAG). While traditional software may not provide explanations for unexpected outcomes, LLM-based software can offer more nuanced and customized explanations derived from raw data. Traditional software is often limited by its reliance on structured data sources, while LLM-based software can process a broader range of data, including unstructured text and real-time streams. Additionally, LLM-based software leverages advanced AI techniques like NLP and machine learning to perform complex tasks and provide human-friendly interactions. LLMs can identify complex compliance issues and continuously adapt to new regulations. This allows for more accurate, efficient, and adaptable compliance checks, ultimately leading to better decision making and risk mitigation. LLM, deep learning, and traditional software are compared in Figure 5, and the differences are presented in Table 1.
LLMs mainly provide a user-friendly interface. This removes the requirement for users to be experts, integrate knowledge of operational parameters, and have a deeper understanding of what could go wrong. It also eliminates the need for humans to remember all rules in real-time assessments. Current LLM-based software may require more frequent updates and maintenance and have limitations in processing large amounts of real-time data. However, future LLMs may be explicitly created for regulatory systems.

4. Results

This section presents the result of using LLM for RegTech for each role of the virtual agent mentioned above. In most cases, there is a predefined value for ε, and measurement of accuracy was based on the fractions of this, i.e., data or rules that are extracted or checked. It is to be noted that the aim of this work was to determine the suitability of LLMs in general for RegTech, and as such, we did not directly compare and contrast individual LLMs with each other. Instead, in each test, we focused on data about the best-performing LLMs for RegTech, showing that at least one LLM can perform reasonably.

4.1. Experimental Setup

The experiments conducted were comprised of raw data used as data sources and LLMs used for both extracting information/rules and acting as agents for users.

4.1.1. Data Sources

For regulations, we used the Australian Cherry Guidelines guidebook (2011) [142], which is used by farmers to grow cherries. This lays out the industry standards’ expectations as well as the minimum regulatory requirements. It contains 209 pages of text in English, tables, and some figures.
For time series data, we used a dataset of 297 rows, each containing timestamped numeric values collected at an on-field weather station, which recorded basic information such as time, temperature, and absence/presence. The system was designed to record a time series value every 5 min. However, if pests were detected, they were recorded earlier as well. Hence, a linear but variable time gap time series was used. The data were not preprocessed and were fed as raw values as saved in CSV format by the IoT system.
For format checking in Section 4.5, we used the official compliance recording forms used by Honeybee Industries and government regulators [143]. Each form contains a table with columns recording the information. There is also a two-page guideline on how to fill in the forms, i.e., expected values.

4.1.2. LLM Platforms and Models

We used a variety of platforms and models for this work. Ollama [102] is an AI tool that allows users to run and interact with LLMs locally on their own devices, ensuring privacy and security. It offers a simple command-line interface to download, customize, and use models efficiently. With a focus on accessibility and performance, Ollama was designed for developers, researchers, and AI enthusiasts seeking local AI solutions. Ollama allows one to use one of the many available LLM models. It was also used for the tool-calling feature. For hardware, we used a 2.9 GHz CPU with an Intel i5-9400 processor, 32 GB RAM, and a 16 GB GPU card.
We also used a few commercial-grade, publicly available cloud-based LLMs. Copilot is an AI coding assistant that helps developers write and debug code by providing real-time suggestions within their IDEs [144]. ChatGPT (GPT-4o), by OpenAI, is a conversational AI tool that engages users in natural dialogue, assisting with questions, problem-solving, and content creation [145]. We also used Gemini 2.0 and Claude3.

4.2. LLM Agent Monitoring Tests

The LLM agent’s monitoring comprises continuous checks of data inputs in various forms. Continuous checks are meant to provide timely feedback to users, either autonomously or when users ask questions. Such checks are like spot checks, which use limited data that have usually been available since the last check.

4.2.1. List Extraction and Checking

We start with the most straightforward analysis for the compliance check, which is the list extraction and list checking. For list extraction, we used the cherry guidebook and asked the LLMs to extract a variety of certain items like cherries from the guidebook. This is an important enumeration in the compliance check where the item names entered in any form must be from a list of pre-approved items, i.e., the cherry names. For this experiment, ε = 29   when the full document was used, and ε = 23 when only five pages were used; the difference is due to most names/strings appearing as a list on five pages, but some other names are embedded randomly in the document. To test the speed and accuracy, we changed the number of words we submitted to the LLM.
The list extraction only concerns picking the entities from the texts. For RegTech, the order of this list does not matter much, but it indicates a level of expected similarity of the extraction for every user prompt. Another indication of consistency is the item-wise lighting. In the original text, the list is embedded in the text in a specific order (not necessarily in alphabetical order), and the LLM application should ideally extract them in the same order. Assuming p(item, list) is the index of an item in a list, the actual item order in the document (D), and the list obtained after LLM (L), we used two metrics:
  • List order variation rate: This is the likelihood of the list not being in the expected order. It is given by the number of times an LLM can return the perfect order divided by the number of times we ask the prompt;
  • Average item displacement: This is the likelihood of an item from the expected list being out of place, given by the following:
k = 1 ε j = 1 n 1   p ( k ,   D )     p ( k , L j ) 0   p k ,   D   =   p ( k , L j )     /   n / ε
where n = 10 is the number of times we ran the LLM chats independently. Even without a perfect extraction, the LLM performance in this situation was shown to be good enough for RegTech. Also, this order checking is only relevant to list extraction. Other extraction, such as rules, is not expected in any particular order. A lower value for both of these indicates that the LLM is reading the document in a consistent manner. A higher value does not necessarily mean incorrect outputs, as the output list can still contain the desired items in a different order but could be challenging to use in subsequent applications. The following are the results of the list extraction (Table 2):
The results indicate that the LLMs are reasonably good for extraction with a 0-shot prompt from a general body of text. Also, all the results were obtained within a reasonable 11 s. The list variations and displacements are pretty high regardless of whether a full document was given or only the five pages. For further LLM applications, once we can correctly extract the list of strings, e.g., cherry names, the LLM can quickly check if any given cherry is on the list.

4.2.2. Numeric/Milestone/Binary Analysis for RegTech

In order to test the numeric, milestone, and binary (NMB) check, we used the time series. We created 10 prompts, as shown in Appendix A. The prompts were designed with varying levels of difficulty in obtaining the target answers:
  • Level 1 prompts asked for an analysis whose answers were in the dataset itself with a 0-shot prompt, e.g., finding maximum and minimum;
  • Level 2 and successive prompt levels asked for analysis that required performing mathematical operations on the raw data or data from the previous level to generate another set of derived data. These data can be used for the next level.
We prompted LLMs with a 0-shot prompt from Appendix A. Each time, the LLM started creating a Python code to generate the answer. Once the Python code was created, LLM used it to check the IoT data, which were in the form of a time series in CSV format, as provided in the prompt (297 rows of data). For each prompt for each LLM model, we ran the test in a new chat every time to isolate the conversations. We repeated this five times, each time with a new chat for each prompt for each model. Each time, the CSV data were appended to the prompts 2–11, itemized in Appendix A, when given to the LLM. Every time, the LLMs generated the code again and executed it or tried to execute it with the given data. The prompts were from three different levels:
  • Level 1 (L1): Finding minimum, maximum, or row(s) greater or lower than a value, e.g., the time when the height of a tree goes above 50. The answers to these questions are not calculated but determined. The answer is already in the data given to the LLM;
  • Level 2 (L2): Finding the average time gap or standard deviation and checking for secondary conditions. Answering these questions requires the LLM to perform one set of operations on the raw data provided, often estimating a new set of values, e.g., a new column based on the original values;
  • Level 3 (L3): Counting the anomalies based on average time and standard deviation. This requires a level 2 operation to be performed before another round of analysis is performed.
Table 3 shows the accuracy results using Equation (1), and Table 4 shows the average execution time. The value of ε is fixed and known for each prompt based on the fixed time series dataset providing the ground truth.
Note that for prompts 2–5, the LLM was able to produce a perfect or high-quality result at least once. However, as the responses must be consistent for RegTech, we focused only on the average accuracy. If an LLM can produce a perfect result, it is possible to update it in the future to reduce randomness and be consistent with a focus on RegTech prompts.
This means that if we ask the LLM about RegTech compliance without any RAG, it knows what to do but is limited in operational accuracy in generating a correct output. Tool calling can improve this, as shown later. The response time is also suitable for RegTech applications. It is not instantaneous and could be further impacted when there are multiple users of the same resources. It is not large enough for users to discontinue the conversation with LLM at the level 1 prompts, but it is quite large for complex prompts. This is because a code was generated most of the time successfully to answer the prompts, but the data were not fed correctly to the code, leading to poor efficiency for L2 and L3.
The LLM can handle time series where the time gap recorded is not consistent. LLMs can understand and interpret the rules but have difficulty checking. The LLM functions as an additional layer, augmenting the capabilities of existing deep learning models. It acts as an assistant, providing support and enhancing the overall performance.

4.3. LLM Agent Rule Identification and Upgradeability

From a RegTech developer’s perspective, if the LLM can convert the written text to code and apply it to a dataset, it can apply a set of rules and conclude if the regulations were met. Also, the LLM agent must be able to alter the execution of the rules according to any expected situation, i.e., understand anomalies and apply the rules accordingly. This requires the rules to be updated for a particular compliance check for a user for a specific process.
We prompted LLMs to create a set of rules for each of these tables with a 0-shot prompt. Once the rules were created, we asked via another prompt to create a Python code to check the rules, assuming the IoT data were in the form of a time series in CSV format.
Automatic rule identification is another prospective application of LLM in regulatory technology. Rule identification and extraction are complex processes depending on the source of the rules—such as whether it is from a prepared table or a text paragraph. Here, we tested the LLM’s capacity to interpret rules from a given table. In order to do this, we considered tables that contained the desired rules in the form of ranges. LLM was asked to provide a pseudo code for the rule and then convert it to a Python code. From the given guideline document, we identified four tables and asked to convert the tables to code. The tables defined thresholds for environmental parameters and/or their conversion rules for regulatory recordkeeping and further analysis. Tables had either a single variable or two variables (α = 1, 2). There was also a fixed set of rules, i.e., a set of if/else conditional statements (ε = 7, 23, 18, 16). For example, in Table 1, in [142], there is 1 variable—“temperature”, which can be between six different numeric ranges, resulting in a corresponding “chilling unit” value.
We provided Ollama models llama3.2 and llama3.1 and Copilot with the relevant pages from the guidebook, which is in PDF format. We recorded the time required and accuracy for converting to a code with varying levels of complexity 10 times for each table in isolated chats. We only considered a single 0-shot prompt for this rather than a set of prompts, as the users may not have the expertise to use multiple prompts to attain a good code. Also, LLMs are known to forget or hallucinate; hence, if the LLM is not able to obtain the perfect output in a single prompt, it may not be helpful for RegTech.
Table 5 shows the results of the rule extraction and rule-to-code conversions. There is a difference in the accuracy of identifying the total number of rules ε correctly between 1- and 2-variable rule sets. For single variables, the numeric rule checks were very accurate; for string checks, where it had to map a set of strings with another set of strings, it was slightly inaccurate in maintaining consistency with strings that had spaces. With the 2-variable numeric check, it was pretty inaccurate with Ollama but good with CoPilot, with the second table returning a high accuracy due to one of the variables having only two possible values, which were put as an if/else conditional check every time.
In all cases, the time consumed was less than 10 s, which is suitable for RegTech applications. If the tables are fed to the LLM in stages, a perfect set of rules and codes may be possible with α > 1. However, this will increase the time consumed and also require considerable software development expertise.
A critical aspect of rule identification and execution is the ability to (a) upgrade the rules or (b) ignore some rules on demand. Eventually, after additional prompts, it will be possible to add more of the conditions to the existing rules set and code. This would follow typical practices of software coding with LLM [146,147]. However, this would require expertise from the users and is not suitable for all RegTech users.

4.4. Web Services Selection and Execution

As expected, the current LLM performance, as shown above, does not inherently meet the standards required for tasks above a certain complexity, i.e., L2 prompts and α > 1, in regulatory technologies. They can solve problems but are too inconsistent for complex tasks. However, automatic rule identification is partially successful in extracting executable code from a set of rules written in regulatory documents. These codes can be perfected and preserved as web services. The web services can be made available through cloud resources and used inside the LLM for data extraction. One of the critical aspects of the LLM for this purpose is to select the correct web service based on the user’s current prompts.
To test this, we used the same prompts from before but this time with tool calling enabled for LLM. We used Ollama models llama3.1 and llama3.2 for this. We defined four parameters to measure the effectiveness of web service selection:
  • Function selection: This refers to accuracy in selecting the right function, method, or web service for the given prompt;
  • Function calling: This is accuracy in correctly formatting the inputs to the selected function, method, or web service for the given prompt. The expected inputs are adequately defined as part of the web services;
  • Function output correctness: This is the accuracy of the output of the selected function, method, or web service after correctly calling it. It is expected to be quite accurate, as the functions are properly hard-coded. The output of the function call is equivalent to the results described in Table 2 and Table 3 with direct analytical prompts. The web service function outputs are structured properly for human interpretation. However, this may not be easy enough, and we asked LLM to present a sentence as an explanation for the output of the web service functions;
  • Function output presentation: The final criterion is the accuracy of properly presenting the information to human users in a readable sentence containing the exact outputs of the function. This is expected to be the weakest, as it depends heavily on LLM’s ability to understand the description of the functions/web services.
While LLM’s performance in selecting, preparing outputs, and executing the web service functions was very accurate, it was not good in explaining the results in this fourth aspect.
When we used raw prompts, the success rate was not very good, as seen in Table 2 and Table 3. The LLM could determine the procedure but made a substantial error in getting the inputs correct. Furthermore, the contained output of the LLM was extremely corrupted, giving unstable outputs. LLM has a default nature of writing code by itself to obtain the desired data, which is irrelevant to this RegTech application.
So, we modified the prompt to append the string “Use one of the given tools and …” at the beginning to ask the LLM to use the prompts explicitly. This increases the accuracy of the service/function selection, correctly deciding the inputs to it, executing the code, and correctly forming the human-friendly output. The experiment was run 50 times separately on two Ollama models.
The results are presented in Figure 6. For level 1 and 2 prompts, the function selection, calling, and output were near 100% for both Ollama models llama3.2:3b and llama3.1:8b. For the level 3 prompt, the accuracy in selecting the functions was not good every time. However, this was still a significant improvement from raw prompts (Table 3), as when the web service functions were correct, the output was perfect.
The function output representation was inadequate in all cases. For this, we only counted the function outputs that were correct in answering the prompt with a human-friendly sentence based on the function’s output, which is in JSON format. The time consumed for this was always less than 5 s, with an average of 3.2 s.

4.5. Formatting of Compliance Checks

The final component in the RegTech system is format checking. Often, the data are written in forms that are structured as tables and graphs. Such forms can be both digital and paper-based. However, in either case, the form can be available in PDF format, which could be fed directly to LLMs. The main parameters to determine the accuracy of LLM’s interpretation of data in the table are as follows:
o
Relevance of data entry for particular sections in the form. This is the LLM’s ability to correctly identify whether an entry in a particular section of a form is relevant to the section’s requirements;
o
The LLM can also estimate the correctness of the entries. Some of the entries can be written in paragraphs. As such, the LLM can identify whether the written text contains all required details as pre-determined by the form designer.
For the data entry experiment, we used the honeybee compliance forms [143], which are in PDF or docx format. We provided the LLM with a filled (or semi-filled) form in docx format and asked it to verify the entries. Each test was repeated five times in separate isolated chats. The results of the tests—average accuracy and average time consumed—are presented in Table 6. The values of ε are also known based on the forms used for this test. The accuracy of LLM format checking was determined with Equation (1). The time was recorded in seconds.
LLMs are quite successful in identifying relevant information in a given paragraph with respect to a given raw guidelines/section in the form. The are able to complete the checking mostly within 10 s. These results were obtained with 0-shot prompts only.
To experiment with the data entry, we used the GACC and honeybee forms [143], which are in PDF or docx format. These forms act as the base guidelines on what information is required from the users in a written paragraph. We first provided an empty form for the LLM. Then, we provided a paragraph containing the required information (in part) for the corresponding form and asked LLM to verify whether the paragraph contained all the relevant data with respect to that form. Each test was repeated five times in separate isolated chats, starting with feeding the empty form and then providing the paragraph. The results of the tests—average accuracy and average time consumed—are presented in Table 7. The values of ε are also known based on the forms used for this test. The accuracy of LLM format checking was determined with Equation (1). The time was recorded in seconds.
LLMs are remarkably successful in identifying incomplete information in a given paragraph with respect to a given raw guideline. They are also able to do this mostly within 10 s.

5. Discussions and Future Works

5.1. RegTech LLM Performance

5.1.1. Industry Expectations

The experiment carried out in this paper shows that RegTech can be implemented and enhanced by using a combination of LLM and traditional software. LLM has the capabilities to identify rules, keep track of sensor and human data, and check for compliance with time series information. LLM can also perform quickly enough to continue a conversation. All the experiments were completed with 0-shot prompts, which is the likely nature of prompts from a novice user. The current LLM has the capability to understand the problem but lacks analytical ability in a large dataset. It was also unable to solve complex prompts with level 2 or higher analysis. But still, it performed with high accuracy for simple checking. This shows that the proposed RegTech virtual agent can successfully deploy LLM as an interface to the regulators and producers.

5.1.2. LLM–Web Services

To overcome the lack of analytical capabilities, back-end web services were used, and each time, LLM was successful in identifying the correct web service functions for a given prompt, creating the inputs, and obtaining the correct analysis. It was less successful in explaining the results. However, given that the results were obtained through a correct analysis, producing the results “as is” is sufficient for RegTech applications. Obviously, further work could improve the final statement produced by the LLM when presenting the rest of the web services.

5.2. Future Work

In this paper, we defined and demonstrated that LLM and web services are promising for RegTech applications. However, there is great potential for improving performance to realize actual RegTech applications with LLM.

5.2.1. k-Shot Prompts and Questions

Throughout this paper, we used only a 0-shot prompt. It is obvious that most RegTech users will use 0-shot prompt questions. However, it is possible to train such users to use k-shot prompting as well. Such prompts can increase the reliability and performance accuracy of LLMs. Future work can lay out the k-shot prompt design techniques for RegTech. The following are example scenarios on how and where few-shot prompts can be used in RegTech for agriculture.
i. 
Scenario 1: Streamlining Pesticide Usage Reporting with Few-Shot Learning
A farmer is required to submit monthly pesticide usage reports to the local agricultural regulatory authority. To simplify this process, a few-shot learning approach can be employed using a large language model (LLM). The farmer provides the LLM with a few examples of reports from previous months (the “few shots”). These examples demonstrate the required format, including specific data fields and their organization. The farmer then inputs the current month’s pesticide usage data, such as the types of pesticides used, quantities applied, and application dates. Leveraging the few-shot examples, the LLM automatically generates the new monthly report, populating it with the current data and ensuring adherence to the regulatory authority’s prescribed format.
ii. 
Scenario 2: Worker Safety Training Records
Documenting worker safety training is essential for compliance. The LLM can be trained with examples of compliant training records, including training topics, dates, and attendees. The farmer inputs the details of recent training sessions, and the LLM generates the necessary documentation.

5.2.2. Multi-Level Checking and Rule Relaxation

LLMs can enable multi-level compliance checking. It would involve a tiered approach for compliance verification, with non-linear dependencies of guidelines issued by different authorities. LLMs can keep track of these changes in rules in real time and alert relevant stakeholders if they are impacted.
Another feature of LLM is severity detection of non-compliance. If the producer meets the bare minimum of the regulatory requirements, they pass the compliance checks. However, when they fail to meet the compliance, it would be interesting for regulators to know how much they failed to comply. On the other hand, once they meet the minimum requirements, the regulator does not need to be concerned about the quality above the minimum standards. Severity detection is also helpful for early warning systems.
There may be opportunities for rule relaxation in certain circumstances. For example, less stringent regulations could be applied to small-scale, family-owned farms that pose minimal environmental or food safety risks. In emergencies, such as natural disasters or disease outbreaks, temporary regulation relaxation may be necessary to address urgent issues. With this, regulators can strike a balance between ensuring compliance and promoting agricultural innovation and sustainability.

5.2.3. Proof of Compliance

As the LLM virtual agent continuously chats with the human users in RegTech, it can keep a record of all the requests and responses. Traditional software can enable such record keeping, but LLM can also monitor for any malicious intent from the users. Also, as LLM can create a range of formatted outputs, it can create a digital certificate by combining all the recording information. This can be proof of compliance for regulators.
LLM can also act as an auditor for compliance checks, although significant improvement in reasoning skills is needed. However, as LLM can constantly interact with the data, it can build up reports, including any reasons and conditions over time.

5.2.4. LLM–Web Service Balancing

While LLMs have the potential to convert existing guidebooks into rule sets, their current capabilities are limited by factors like processing speed and data processing capacity. Hence, they are best suited for normal human conversations and basic analysis but are incapable of handling composite or derived computational tasks yet.
To overcome these limitations, we showed that LLMs can be integrated with backend web services developed by software developers to handle large datasets and complex computations. This integration will enhance the overall performance and scalability of LLM-powered applications, making them more practical and effective in real-world scenarios. Instead of having the LLM attempt to write its own code, we can instruct it to identify the necessary task and delegate it to a suitable backend service. To achieve this, the backend services should provide the LLM with properly formatted output, enabling the LLM to interpret and utilize the results effectively.
LLMs, while powerful, cannot solve all problems independently. By using a collaborative approach, where LLMs work in conjunction with specialized services, we can overcome these limitations and create more robust and effective AI solutions. In this work, we tested with only one level of web service functions, but future work can consider a chain of function selection or nested functions. In such cases, the LLM must also decide the sequence of the web service functions to be called and how to transfer the output of one to another as needed.

6. Conclusions

LLMs excel at tasks involving language and information retrieval, such as summarizing text, translating languages, and generating creative text formats. However, they may struggle with tasks that require complex reasoning, mathematical calculations, or the execution of code. For RegTech, it is essential to have speed, consistency, and accuracy in results. Thus, it is necessary to develop a new LLM dedicated to RegTech purposes. Such LLM has to be more accurate with a 0-shot prompt, but on the other hand, it does have to answer any random questions from the users, allowing it to be more focused in the response.
In this paper, we show that current LLMs are capable of the basic features needed for an LLM-supported RegTech application. LLMs can process basic analytical prompts. They can also understand how to solve complex analytical prompts and use external web services to solve those problems. LLMs can also automatically handle format checking, enumerations, and numerical checking for compliance. They can do this in a reasonable time for RegTech conversations. LLMs can also identify rules and convert them to code in publicly available human-generated bodies of text that are conventionally used by the relevant industries. Future work can focus on the development of LLM for RegTech, further refining these features.
RegTech is expected to be a fundamental component of Industry 5.0, and LLMs can make them user-friendly and easily deployable.

Author Contributions

Conceptualization, J.L. and A.M.; methodology, A.M. and J.L.; software, A.M.; validation, J.L. and A.M.; formal analysis, A.M. and J.L.; investigation, A.M. and J.L.; resources, J.L. and A.M.; data curation, J.L. and A.M.; writing—original draft preparation, J.L. and A.M.; writing—review and editing, A.M. and J.L.; visualization, A.M.; supervision, A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Department of Agriculture, Fisheries, and Forestry, Australian Government.

Data Availability Statement

Data is contained within the article.

Acknowledgments

We would like to acknowledge Jiangang Fei for his guidance. He played a vital role in obtaining funds for this work and supervised J. Li until 2024 but was unable to contribute further due to medical leave.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; the collection, analysis, or interpretation of data; the writing of the manuscript; or the decision to publish the results.

Appendix A

Sl. NoLevelPromptPurpose
11Consider the following text. Tell me how many types of cherries are mentioned and give me a list.List Extraction
21Consider the time series, which has the first column as time and the fourth column as temperature data. What were the maximum and minimum temperatures?Numeric Analysis
31Please give me the row where the temperature is higher than 17.Numeric Analysis
41Consider the time series data, where column 4 is temperature. Tell me the first record when the temperature exceeds 17.8 before 3 p.m. during a day.Numeric Analysis
52Consider the following time series data and tell me the appropriate chilling units for each row of temperature recorded in the 4th column. The 1st column is the time of recording the temperature. (Note: related to No. 12 below)Numeric Analysis
62Consider the following time series data, which has the first column as time and the fourth column as temperature data. What is the average time gap between the recordings in each row?Numeric Analysis
72Based on the time series data, please compute the standard deviation of the time gaps.Numeric Analysis
82Consider the following time series data, which has the first column as time, and tell me how many rows the time gap over 6 min or less than 5 min was.Numeric Analysis
92Consider the time series data, which has the third column as a security sensor with 0 meaning no problem at that time and 1 meaning someone is detected at that time. How many rows with time series between 6 p.m. and 9 a.m. indicate no problem?Numeric Analysis
103Based on the time series data, if the time gap is more than twice of 5 min, please identify the rows.Numeric Analysis
113Consider the following time series data which has the first column as time, please count the anomalies based on average temperature and standard deviation. An anomaly is a sudden drop in temperature of 3 degrees below the average temperature.Numeric Analysis
12 Define a set of rules for the chilling unit in simple English.
Write a full python code for the rules when taking temperature values from a csv file. Do not explain the code.
Rule Conversion
13 Define a set of rules for possible pollenisers given a particular variety.
Write a full python code for the rules when taking a particular variety. Do not explain the code.
Rule Conversion
14 Define a set of rules for the critical temperature for frost damage where there are 2 conditions of 10% and 90% in simple English.
Write a full python code for the rules when taking temperature values and cherry events from a csv file. Do not explain the code.
Rule Conversion
15 Define a set of rules for expected levels of chemical, e.g., nitrogen, phosphorus, etc. for Budbreak, Postharvest, and total.
Write a full python code for the rules when taking a level of a particular chemical. Do not explain the code.
Rule Conversion
16 Consider this report; please check whether the form is filled out completely.Format Check
17 Consider this report; please check whether all required information is includedFormat Check

References

  1. Wang, J.; Azam, W. Natural resource scarcity, fossil fuel energy consumption, and total greenhouse gas emissions in top emitting countries. Geosci. Front. 2024, 15, 101757. [Google Scholar] [CrossRef]
  2. Rotowa, O.; Adekunle, E.; Adeagbo, A.; Nwanze, O.; Fasiku, O. Economic analysis of agriculture, forestry and fisheries to the economic development of Nigeria. Int. J. Web Eng. Technol. 2019, 6, 15–22. [Google Scholar]
  3. Arce, J.J.C. Forests, inclusive and sustainable economic growth and employment. In Background Study Prepared for the Fourteenth Session of the United Nations Forum on Forest (UNFF), Forests and SDG8; 2019; Available online: https://www.un.org/esa/forests/wp-content/uploads/2019/04/UNFF14-BkgdStudy-SDG8-March2019.pdf (accessed on 22 December 2024).
  4. Muriithi, B.W.; Mburu, J.; Ngigi, M. Constraints and determinants of compliance with EurepGap standards: A case of smallholder french bean exporters in Kirinyaga district, Kenya. Agribusiness 2011, 27, 193–204. [Google Scholar] [CrossRef]
  5. Beach, T.H.; Rezgui, Y.; Li, H.; Kasim, T. A rule-based semantic approach for automated regulatory compliance in the construction sector. Expert Syst. Appl. 2015, 42, 5219–5231. [Google Scholar] [CrossRef]
  6. Neitzel, R.L.; Seixas, N.S.; Ren, K.K. A review of crane safety in the construction industry. Appl. Occup. Environ. Hyg. 2001, 16, 1106–1117. [Google Scholar] [CrossRef]
  7. Gritten, D.; Greijmans, M.; Lewis, S.R.; Sokchea, T.; Atkinson, J.; Nguyen Quang, T.; Poudyal, B.; Chapagain, B.; Sapkota, L.M.; Mohns, B. An uneven playing field: Regulatory barriers to communities making a living from the timber from their forests–examples from Cambodia, Nepal and Vietnam. Forests 2015, 6, 3433–3451. [Google Scholar] [CrossRef]
  8. Emeihe, E.V.; Nwankwo, E.I.; Ajegbile, M.D.; Olaboye, J.A.; Maha, C.C. The impact of artificial intelligence on regulatory compliance in the oil and gas industry. Int. J. Life Sci. Res. Arch. 2024, 7, 28–39. [Google Scholar]
  9. Akiri, M.; Mbugua, F.; Njunge, R.; Agwanda, C.; Gurmessa, N.E.; Phiri, N.A.; Musebe, R.; Kalisa, J.P.; Uzayisenga, B.; Kansiime, M.K. Intervention Options for Enhancing Smallholder Compliance with Regulatory and Market Standards for High-Value Fruits and Vegetables in Rwanda and Zambia. Sustainability 2024, 16, 6243. [Google Scholar] [CrossRef]
  10. Zikankuba, V.L.; Mwanyika, G.; Ntwenya, J.E.; James, A. Pesticide regulations and their malpractice implications on food and environment safety. Cogent Food Agric. 2019, 5, 1601544. [Google Scholar] [CrossRef]
  11. Doyle, E.; McGovern, D.; McCarthy, S.; Perez-Alaniz, M. Compliance-innovation: A quality-based route to sustainability. J. Clean. Prod. 2019, 210, 266–275. [Google Scholar] [CrossRef]
  12. Boda, V.V.R.; Allam, H. Automating Compliance in Healthcare: Tools and Techniques You Need. Innov. Eng. Sci. J. 2021, 1. [Google Scholar]
  13. Xue, X.; Zhang, J. Evaluation of seven part-of-speech taggers in tagging building codes: Identifying the best performing tagger and common sources of errors. In Proceedings of the Construction Research Congress 2020, Tempe, Arizona, 8–10 March 2020; American Society of Civil Engineers: Reston, VA, USA, 2020; pp. 498–507. [Google Scholar]
  14. Coglianese, C. E-Rulemaking: Information technology and the regulatory process. Admin. L. Rev. 2004, 56, 353. [Google Scholar] [CrossRef]
  15. Li, J.; Maiti, A.; Fei, J. Features and Scope of Regulatory Technologies: Challenges and Opportunities with Industrial Internet of Things. Future Internet 2023, 15, 256. [Google Scholar] [CrossRef]
  16. Olawale, O.; Ajayi, F.A.; Udeh, C.A.; Odejide, O.A. RegTech innovations streamlining compliance, reducing costs in the financial sector. GSC Adv. Res. Rev. 2024, 19, 114–131. [Google Scholar] [CrossRef]
  17. Abikoye, B.E.; Umeorah, S.C.; Adelaja, A.O.; Ayodele, O.; Ogunsuji, Y.M. Regulatory compliance and efficiency in financial technologies: Challenges and innovations. World J. Adv. Res. Rev. 2024, 23, 1830–1844. [Google Scholar] [CrossRef]
  18. Grassi, L.; Lanfranchi, D. RegTech in public and private sectors: The nexus between data, technology and regulation. J. Ind. Bus. Econ. 2022, 49, 441–479. [Google Scholar] [CrossRef]
  19. Mhlanga, D.; Ndhlovu, E. Digital technology adoption in the agriculture sector: Challenges and complexities in Africa. Hum. Behav. Emerg. Technol. 2023, 2023, 6951879. [Google Scholar] [CrossRef]
  20. Khanna, M.; Atallah, S.S.; Kar, S.; Sharma, B.; Wu, L.H.; Yu, C.Z.; Chowdhary, G.; Soman, C.; Guan, K.Y. Digital transformation for a sustainable agriculture in the United States: Opportunities and challenges. Agric. Econ. 2022, 53, 924–937. [Google Scholar] [CrossRef]
  21. Aleke, B.; Ojiako, U.; Wainwright, D.W. ICT adoption in developing countries: Perspectives from small-scale agribusinesses. J. Enterp. Inf. Manag. (In English). 2011, 24, 68–84. [Google Scholar] [CrossRef]
  22. Bosompem, M. Potential challenges to precision agriculture technologies development in Ghana: Scientists’ and cocoa extension agents’ perspectives. Precis. Agric. (In English). 2021, 22, 1578–1600. [Google Scholar] [CrossRef]
  23. Nagariya, R.; Mukherjee, S.; Baral, M.M.; Patel, B.S.; Venkataiah, C. The Challenges of Blockchain Technology Adoption in the Agro-based Industries. Int. J. Math. Eng. Manag. Sci. (In English). 2022, 7, 949–963. [Google Scholar] [CrossRef]
  24. Rodriguez, J.M.; Molnar, J.J.; Fazio, R.A.; Sydnor, E.; Lowe, M.J. Barriers to adoption of sustainable agriculture practices: Change agent perspectives. Renew. Agric. Food Syst. 2009, 24, 60–71. [Google Scholar] [CrossRef]
  25. The Productivity Commission. Regulation of Australian Agriculture, Report No. 79; Research Report; The Productivity Commission: Canberra, Australia, 2016.
  26. Adelman, D.E.; Barton, J.H. Environmental regulation for agriculture: Towards a framework to promote sustainable intensive agriculture. Stan. Envtl. LJ 2002, 21, 3. [Google Scholar]
  27. Trivedi, P.; Mattupalli, C.; Eversole, K.; Leach, J.E. Enabling sustainable agriculture through understanding and enhancement of microbiomes. New Phytol. 2021, 230, 2129–2147. [Google Scholar] [CrossRef]
  28. Mercado, G.; Hjortsø, C.N.; Honig, B. Decoupling from international food safety standards: How small-scale indigenous farmers cope with conflicting institutions to ensure market participation. Agric. Hum. Values 2018, 35, 651–669. [Google Scholar] [CrossRef]
  29. Jafri, S.H.; Adnan, K.M.; Baimbill Johnson, S.; Talukder, A.A.; Yu, M.; Osei, E. Challenges and Solutions for Small Dairy Farms in the US: A Review. Agriculture 2024, 14, 2369. [Google Scholar] [CrossRef]
  30. Hackfort, S. Patterns of inequalities in digital agriculture: A systematic literature review. Sustainability 2021, 13, 12345. [Google Scholar] [CrossRef]
  31. Rotz, S.; Gravely, E.; Mosby, I.; Duncan, E.; Finnis, E.; Horgan, M.; LeBlanc, J.; Martin, R.; Neufeld, H.T.; Nixon, A. Automated pastures and the digital divide: How agricultural technologies are shaping labour and rural communities. J. Rural Stud. 2019, 68, 112–122. [Google Scholar] [CrossRef]
  32. Choruma, D.J.; Dirwai, T.L.; Mutenje, M.; Mustafa, M.; Chimonyo, V.G.P.; Jacobs-Mata, I.; Mabhaudhi, T. Digitalisation in agriculture: A scoping review of technologies in practice, challenges, and opportunities for smallholder farmers in sub-Saharan Africa. J. Agric. Food Res. 2024, 18, 101286. [Google Scholar] [CrossRef]
  33. Fuentes-Peñailillo, F.; Gutter, K.; Vega, R.; Silva, G.C. Transformative technologies in digital agriculture: Leveraging Internet of Things, remote sensing, and artificial intelligence for smart crop management. J. Sens. Actuator Netw. 2024, 13, 39. [Google Scholar] [CrossRef]
  34. Saiz-Rubio, V.; Rovira-Más, F. From smart farming towards agriculture 5.0: A review on crop data management. Agronomy 2020, 10, 207. [Google Scholar] [CrossRef]
  35. Salam, A. Internet of things for sustainability: Perspectives in privacy, cybersecurity, and future trends. In Internet of Things for Sustainable Community Development: Wireless Communications, Sensing, and Systems; Springer: Berlin/Heidelberg, Germany, 2024; pp. 299–326. [Google Scholar]
  36. Yazdinejad, A.; Zolfaghari, B.; Azmoodeh, A.; Dehghantanha, A.; Karimipour, H.; Fraser, E.; Green, A.G.; Russell, C.; Duncan, E. A review on security of smart farming and precision agriculture: Security aspects, attacks, threats and countermeasures. Appl. Sci. 2021, 11, 7518. [Google Scholar] [CrossRef]
  37. Adewusi, A.O.; Chikezie, N.; Eyo-Udo, N.L. Cybersecurity in precision agriculture: Protecting data integrity and privacy. Int. J. Appl. Res. Soc. Sci. 2023, 5, 693–708. [Google Scholar] [CrossRef]
  38. Baldock, D.; Hart, K. Risks and Opportunities of a Post-EU Environmental Regulatory Regime for Agriculture in England; Institute for European Environmental Policy: London, UK; Brussels, Belgium, 2020. [Google Scholar]
  39. Rude, J.; Weersink, A.; Jeffrey, S.; Unterschultz, J. Linking Environmental Goals with Business Risk Management Programs in Canadian Agriculture; Alberta Land Institute: Edmonton, AB, Canada, 2015. [Google Scholar]
  40. Mazé, A.; Aït-Aïssa, M.; Mayer, S.; Verjux, N. Third-party certifications and the role of auditing policies in sustainability: The time and space of materiality within combined audits. Organ. Environ. 2016, 29, 308–331. [Google Scholar] [CrossRef]
  41. Florian, R.V. Autonomous Artificial Intelligent Agents; Center for Cognitive and Neural Studies (Coneural): Cluj-Napoca, Romania, 2003. [Google Scholar]
  42. Tsakonas, A.; Dounias, G.; Jantzen, J.; Axer, H.; Bjerregaard, B.; von Keyserlingk, D.G. Evolving rule-based systems in two medical domains using genetic programming. Artif. Intell. Med. 2004, 32, 195–216. [Google Scholar] [CrossRef] [PubMed]
  43. Hayes-Roth, F. The knowledge-based expert system: A tutorial. Computer 1984, 17, 11–28. [Google Scholar] [CrossRef]
  44. Holzinger, A.; Biemann, C.; Pattichis, C.S.; Kell, D.B. What do we need to build explainable AI systems for the medical domain? arXiv 2017, arXiv:1712.09923. [Google Scholar]
  45. Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
  46. Adadi, A.; Berrada, M. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  47. Van Veen, D.; Van Uden, C.; Blankemeier, L.; Delbrouck, J.-B.; Aali, A.; Bluethgen, C.; Pareek, A.; Polacin, M.; Reis, E.P.; Seehofnerova, A. Clinical text summarization: Adapting large language models can outperform human experts. Res. Sq. 2023. [Google Scholar] [CrossRef]
  48. van Schaik, T.A.; Pugh, B. A field guide to automatic evaluation of llm-generated summaries. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, Washington, DC, USA, 14–18 July 2024; pp. 2832–2836. [Google Scholar]
  49. Subbiah, M.; Zhang, S.; Chilton, L.B.; McKeown, K. Reading subtext: Evaluating large language models on short story summarization with writers. Trans. Assoc. Comput. Linguist. 2024, 12, 1290–1310. [Google Scholar] [CrossRef]
  50. Ravaut, M.; Sun, A.; Chen, N.; Joty, S. On context utilization in summarization with large language models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; pp. 2764–2781. [Google Scholar]
  51. Hao, S.; Gu, Y.; Ma, H.; Hong, J.J.; Wang, Z.; Wang, D.Z.; Hu, Z. Reasoning with language model is planning with world model. arXiv 2023, arXiv:2305.14992. [Google Scholar]
  52. Shen, Z. Llm with tools: A survey. arXiv 2024, arXiv:2409.18807. [Google Scholar]
  53. Zhuang, Y.; Yu, Y.; Wang, K.; Sun, H.; Zhang, C. Toolqa: A dataset for llm question answering with external tools. Adv. Neural Inf. Process. Syst. 2023, 36, 50117–50143. [Google Scholar]
  54. Gottschalk, K.; Graham, S.; Kreger, H.; Snell, J. Introduction to Web services architecture. IBM Syst. J. 2002, 41, 170–177. [Google Scholar] [CrossRef]
  55. Dewi, C.; Chen, R.-C. Decision making based on IoT data collection for precision agriculture. In Intelligent Information and Database Systems: Recent Developments 11; Springer: Cham, Switzerland, 2020; pp. 31–42. [Google Scholar]
  56. Muangprathub, J.; Boonnam, N.; Kajornkasirat, S.; Lekbangpong, N.; Wanichsombat, A.; Nillaor, P. IoT and agriculture data analysis for smart farm. Comput. Electron. Agric. 2019, 156, 467–474. [Google Scholar] [CrossRef]
  57. Saranya, K.; Dharini, P.U.; Darshni, P.U.; Monisha, S. IoT based pest controlling system for smart agriculture. In Proceedings of the 2019 International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 17–19 July 2019; pp. 1548–1552. [Google Scholar]
  58. Nayagam, M.G.; Vijayalakshmi, B.; Somasundaram, K.; Mukunthan, M.; Yogaraja, C.; Partheeban, P. Control of pests and diseases in plants using iot technology. Meas. Sens. 2023, 26, 100713. [Google Scholar] [CrossRef]
  59. Umapathi, R.; Ghoreishian, S.M.; Sonwal, S.; Rani, G.M.; Huh, Y.S. Portable electrochemical sensing methodologies for on-site detection of pesticide residues in fruits and vegetables. Coord. Chem. Rev. 2022, 453, 214305. [Google Scholar] [CrossRef]
  60. Issaka, E.; Wariboko, M.A.; Johnson, N.A.N.; Nyame-do Aniagyei, O. Advanced visual sensing techniques for on-site detection of pesticide residue in water environments. Heliyon 2023, 9, e13986. [Google Scholar] [CrossRef]
  61. Abbas, A.; Zhang, Z.; Zheng, H.; Alami, M.M.; Alrefaei, A.F.; Abbas, Q.; Naqvi, S.A.H.; Rao, M.J.; Mosa, W.F.; Abbas, Q. Drones in plant disease assessment, efficient monitoring, and detection: A way forward to smart agriculture. Agronomy 2023, 13, 1524. [Google Scholar] [CrossRef]
  62. Mohanty, A.K.; Rao, T.K.; KS, M.H.; Agme, R.; Gogoi, C.; Velu, C. IoT Applications for Livestock Management and Health Monitoring in Modern Farming. Educ. Adm. Theory Pract. 2024, 30, 2141–2153. [Google Scholar]
  63. Chamara, N.; Islam, M.D.; Bai, G.F.; Shi, Y.; Ge, Y. Ag-IoT for crop and environment monitoring: Past, present, and future. Agric. Syst. 2022, 203, 103497. [Google Scholar] [CrossRef]
  64. Liu, B. Research on NFC Anti-Counterfeit Traceability Based on Signcryption AlgorithmResearch on NFC Anti-Counterfeit Traceability. In Proceedings of the 2023 4th International Conference on Machine Learning and Computer Application, Hangzhou, China, 27–29 October 2023; pp. 941–949. [Google Scholar]
  65. Pigini, D.; Conti, M. NFC-based traceability in the food chain. Sustainability 2017, 9, 1910. [Google Scholar] [CrossRef]
  66. Bernardi, P.; Gandino, F.; Lamberti, F.; Montrucchio, B.; Rebaudengo, M.; Sanchez, E. An anti-counterfeit mechanism for the application layer in low-cost RFID devices. In Proceedings of the 2008 4th European Conference on Circuits and Systems for Communications, Bucharest, Romania, 10–11 July 2008; pp. 227–231. [Google Scholar]
  67. Khalil, G.; Doss, R.; Chowdhury, M. A comparison survey study on RFID based anti-counterfeiting systems. J. Sens. Actuator Netw. 2019, 8, 37. [Google Scholar] [CrossRef]
  68. Galimberti, A.; De Mattia, F.; Losa, A.; Bruni, I.; Federici, S.; Casiraghi, M.; Martellos, S.; Labra, M. DNA barcoding as a new tool for food traceability. Food Res. Int. 2013, 50, 55–63. [Google Scholar] [CrossRef]
  69. Galimberti, A.; Labra, M.; Sandionigi, A.; Bruno, A.; Mezzasalma, V.; De Mattia, F. DNA barcoding for minor crops and food traceability. Adv. Agric. 2014, 2014, 831875. [Google Scholar] [CrossRef]
  70. Siddiqi, K.; Nollet, L.M. Fingerprinting Techniques in Food Authentication and Traceability; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  71. Zhao, J.; Li, A.; Jin, X.; Pan, L. Technologies in individual animal identification and meat products traceability. Biotechnol. Biotechnol. Equip. 2020, 34, 48–57. [Google Scholar] [CrossRef]
  72. Kotsanopoulos, K.V.; Arvanitoyannis, I.S. The role of auditing, food safety, and food quality standards in the food industry: A review. Compr. Rev. Food Sci. Food Saf. 2017, 16, 760–775. [Google Scholar] [CrossRef]
  73. Zhang, Z.; Ma, L.; Broyd, T. Towards fully-automated code compliance checking of building regulations: Challenges for rule interpretation and representation. In Proceedings of the 2022: EC3 (European Conference on Computing in Construction), Rhodes, Greece, 24–26 July 2022. [Google Scholar]
  74. Thumburu, S.K.R. Optimizing Data Transformation in EDI Workflows. Innov. Comput. Sci. J. 2021, 7. Available online: https://www.scribd.com/document/808551426/Optimizing-Data-Transformation-in-EDI-Workflows (accessed on 8 June 2024).
  75. Baur, P.; Getz, C.; Petersen-Rockney, M.; Sowerwine, J. Compliance is far from standard: Relational conditions of access and exclusion in agriculture. J. Rural Stud. 2024, 109, 103335. [Google Scholar] [CrossRef]
  76. The Productivity Commission. Regulator Engagement with Small Business; Research Report; The Productivity Commission: Canberra, Australia, 2013. Available online: https://www.pc.gov.au/inquiries/completed/small-business/report/small-business.pdf (accessed on 8 June 2024).
  77. Brown, V.L.; Gissel, J.L.; Gordon Neely, D. Audit quality indicators: Perceptions of junior-level auditors. Manag. Audit. J. 2016, 31, 949–980. [Google Scholar] [CrossRef]
  78. Vince, J. Third party certification: Implementation challenges in private-social partnerships. Policy Des. Pract. 2018, 1, 323–336. [Google Scholar] [CrossRef]
  79. Chen, N.; Lin, X.; Jiang, H.; An, Y. Automated building information modeling compliance check through a large language model combined with deep learning and ontology. Buildings 2024, 14, 1983. [Google Scholar] [CrossRef]
  80. Nawari, N.O. A generalized adaptive framework (GAF) for automating code compliance checking. Buildings 2019, 9, 86. [Google Scholar] [CrossRef]
  81. Hjelseth, E.; Nisbet, N. Capturing normative constraints by use of the semantic mark-up RASE methodology. In Proceedings of the CIB W78-W102 Conference, Sophia Antipolis, France, 26–28 October 2011; pp. 1–10. [Google Scholar]
  82. Nadkarni, P.M.; Ohno-Machado, L.; Chapman, W.W. Natural language processing: An introduction. J. Am. Med. Inform. Assoc. 2011, 18, 544–551. [Google Scholar] [CrossRef] [PubMed]
  83. Zhang, J.; El-Gohary, N.M. Semantic NLP-based information extraction from construction regulatory documents for automated compliance checking. J. Comput. Civ. Eng. 2016, 30, 04015014. [Google Scholar] [CrossRef]
  84. Zhou, P.; El-Gohary, N. Ontology-based automated information extraction from building energy conservation codes. Autom. Constr. 2017, 74, 103–117. [Google Scholar] [CrossRef]
  85. Breaux, T.D.; Vail, M.W.; Anton, A.I. Towards regulatory compliance: Extracting rights and obligations to align requirements with regulations. In Proceedings of the 14th IEEE International Requirements Engineering Conference (RE’06), St. Paul, MN, USA, 11–15 September 2006; pp. 49–58. [Google Scholar]
  86. Breaux, T.; Antón, A. Analyzing regulatory rules for privacy and security requirements. IEEE Trans. Softw. Eng. 2008, 34, 5–20. [Google Scholar] [CrossRef]
  87. Kiyavitskaya, N.; Zeni, N.; Breaux, T.D.; Antón, A.I.; Cordy, J.R.; Mich, L.; Mylopoulos, J. Automating the extraction of rights and obligations for regulatory compliance. In Proceedings of the Conceptual Modeling-ER 2008: 27th International Conference on Conceptual Modeling, Barcelona, Spain, 20–24 October 2008; Proceedings 27. Springer: Berlin/Heidelberg, Germany, 2008; pp. 154–168. [Google Scholar]
  88. Breaux, T.D. Legal Requirements Acquisition for the Specification of Legally Compliant Information Systems; North Carolina State University: Raleigh, NC, USA, 2009. [Google Scholar]
  89. Zeni, N.; Kiyavitskaya, N.; Mich, L.; Cordy, J.R.; Mylopoulos, J. GaiusT: Supporting the extraction of rights and obligations for regulatory compliance. Requir. Eng. 2015, 20, 1–22. [Google Scholar] [CrossRef]
  90. Sleimi, A.; Sannier, N.; Sabetzadeh, M.; Briand, L.; Dann, J. Automated extraction of semantic legal metadata using natural language processing. In Proceedings of the 2018 IEEE 26th International Requirements Engineering Conference (RE), Banff, AB, Canada, 20–24 August 2018; pp. 124–135. [Google Scholar]
  91. Sleimi, A.; Ceci, M.; Sannier, N.; Sabetzadeh, M.; Briand, L.; Dann, J. A query system for extracting requirements-related information from legal texts. In Proceedings of the 2019 IEEE 27th International Requirements Engineering Conference (RE), Jeju, Republic of Korea, 23–27 September 2019; pp. 319–329. [Google Scholar]
  92. Amaral, O.; Abualhaija, S.; Torre, D.; Sabetzadeh, M.; Briand, L.C. AI-enabled automation for completeness checking of privacy policies. IEEE Trans. Softw. Eng. 2021, 48, 4647–4674. [Google Scholar] [CrossRef]
  93. Amaral, O.; Abualhaija, S.; Briand, L. Ml-based compliance verification of data processing agreements against gdpr. In Proceedings of the 2023 IEEE 31st International Requirements Engineering Conference (RE), Hannover, Germany, 4–8 September 2023; pp. 53–64. [Google Scholar]
  94. Cejas, O.A.; Azeem, M.I.; Abualhaija, S.; Briand, L.C. Nlp-based automated compliance checking of data processing agreements against gdpr. IEEE Trans. Softw. Eng. 2023, 49, 4282–4303. [Google Scholar] [CrossRef]
  95. Yamusa, M.A.; Abdullahi, M.; Ibrahim, Y.M.; Ahmadu, H.; Abubakar, M. Addressing Compliance Checking Matters of Buildings to Green Standards Using Natural Language Processing: A Review. In Proceedings of the WABER & SuDBE Joint Internafional Conference, Johannesburg, South Africa, 30–31 July 2024; pp. 11–19. [Google Scholar]
  96. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep learning to improve breast cancer detection on screening mammography. Sci. Rep. 2019, 9, 12495. [Google Scholar] [CrossRef]
  97. Alhazmi, K.; Alsumari, W.; Seppo, I.; Podkuiko, L.; Simon, M. Effects of annotation quality on model performance. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 13–16 April 2021; pp. 63–67. [Google Scholar]
  98. Emam, Z.; Kondrich, A.; Harrison, S.; Lau, F.; Wang, Y.; Kim, A.; Branson, E. On the state of data in computer vision: Human annotations remain indispensable for developing deep learning models. arXiv 2021, arXiv:2108.00114. [Google Scholar]
  99. Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
  100. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 1–113. [Google Scholar]
  101. Taylor, R.; Kardas, M.; Cucurull, G.; Scialom, T.; Hartshorn, A.; Saravia, E.; Poulton, A.; Kerkez, V.; Stojnic, R. Galactica: A large language model for science. arXiv 2022, arXiv:2211.09085. [Google Scholar]
  102. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F. Llama: Open and efficient foundation language models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
  103. Han, S.; Yoon, J.; Arik, S.O.; Pfister, T. Large Language Models Can Automatically Engineer Features for Few-Shot Tabular Learning. arXiv 2024, arXiv:2404.09491. [Google Scholar]
  104. Hegselmann, S.; Buendia, A.; Lang, H.; Agrawal, M.; Jiang, X.; Sontag, D. Tabllm: Few-shot classification of tabular data with large language models. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 25–27 April 2023; pp. 5549–5581. [Google Scholar]
  105. Berger, A.; Hillebrand, L.; Leonhard, D.; Deußer, T.; De Oliveira, T.B.F.; Dilmaghani, T.; Khaled, M.; Kliem, B.; Loitz, R.; Bauckhage, C. Towards automated regulatory compliance verification in financial auditing with large language models. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; pp. 4626–4635. [Google Scholar]
  106. Ying, H.; Sacks, R. From Automatic to Autonomous: A Large Language Model-driven Approach for Generic Building Compliance Checking. In Proceedings of the CIB W78 Conference 2024, Marrakesh, Morocco, 1–3 October 2024. [Google Scholar]
  107. Hassani, S.; Sabetzadeh, M.; Amyot, D.; Liao, J. Rethinking Legal Compliance Automation: Opportunities with Large Language Models. arXiv 2024, arXiv:2404.14356. [Google Scholar]
  108. Liu, X.; Li, H.; Zhu, X. A GPT-based method of Automated Compliance Checking through prompt engineering. In Proceedings of the 30th European Group for Intelligent Computing in Engineering—‘Towards Sustainable, Smart and Resilient Buildings, Infrastructures and Cities’, London, UK, 4–7 July 2023. [Google Scholar]
  109. Zytek, A.; Pidò, S.; Veeramachaneni, K. LLMs for XAI: Future Directions for Explaining Explanations. arXiv 2024, arXiv:2405.06064. [Google Scholar]
  110. Ullah, E.; Parwani, A.; Baig, M.M.; Singh, R. Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology—A recent scoping review. Diagn. Pathol. 2024, 19, 43. [Google Scholar] [CrossRef]
  111. Ataide, M.; Braholli, O.; Siegele, D. Digital transformation of building permits: Current status, maturity, and future prospects. Buildings 2023, 13, 2554. [Google Scholar] [CrossRef]
  112. Iversen, O. Leveraging Large Language Models for BIM-Based Automated Compliance Checking of Building Regulations. Master’s Thesis, NTNU, Trondheim, Norway, 2024. [Google Scholar]
  113. Dong, M.M.; Stratopoulos, T.C.; Wang, V.X. A scoping review of ChatGPT research in accounting and finance. Int. J. Account. Inf. Syst. 2024, 55, 100715. [Google Scholar] [CrossRef]
  114. Zhao, H.; Liu, Z.; Wu, Z.; Li, Y.; Yang, T.; Shu, P.; Xu, S.; Dai, H.; Zhao, L.; Mai, G. Revolutionizing finance with llms: An overview of applications and insights. arXiv 2024, arXiv:2401.11641. [Google Scholar]
  115. Desai, A.P.; Mallya, G.S.; Luqman, M.; Ravi, T.; Kota, N.; Yadav, P. Opportunities and Challenges of Generative-AI in Finance. arXiv 2024, arXiv:2410.15653. [Google Scholar]
  116. Hassani, S. Enhancing Legal Compliance and Regulation Analysis with Large Language Models. arXiv 2024, arXiv:2404.17522. [Google Scholar]
  117. Potamitis, I. ChatGPT in the context of precision agriculture data analytics. arXiv 2023, arXiv:2311.06390. [Google Scholar]
  118. Sapkota, R.; Qureshi, R.; Hassan, S.Z.; Shutske, J.; Shoman, M.; Sajjad, M.; Dharejo, F.A.; Paudel, A.; Li, J.; Meng, Z. Multi-modal LLMs in agriculture: A comprehensive review. Authorea Prepr. 2024. [Google Scholar] [CrossRef]
  119. Tzachor, A.; Devare, M.; Richards, C.; Pypers, P.; Ghosh, A.; Koo, J.; Johal, S.; King, B. Large language models and agricultural extension services. Nat. Food 2023, 4, 941–948. [Google Scholar] [CrossRef]
  120. Zhao, B.; Jin, W.; Del Ser, J.; Yang, G. ChatAgri: Exploring potentials of ChatGPT on cross-linguistic agricultural text classification. Neurocomputing 2023, 557, 126708. [Google Scholar] [CrossRef]
  121. Ray, P.P. AI-assisted sustainable farming: Harnessing the power of ChatGPT in modern agricultural sciences and technology. ACS Agric. Sci. Technol. 2023, 3, 460–462. [Google Scholar] [CrossRef]
  122. Shutske, J.M. Harnessing the Power of Large Language Models in Agricultural Safety & Health. J. Agric. Saf. Health 2023, 29, 205–224. [Google Scholar]
  123. Red Tape Action Plan. 2013. Available online: https://trove.nla.gov.au/work/236431188 (accessed on 12 June 2024).
  124. De Filippi, P.; Hassan, S. Blockchain technology as a regulatory technology: From code is law to law is code. arXiv 2018, arXiv:1801.02507. [Google Scholar] [CrossRef]
  125. Wu, T. When code isn’t law. Va. L. Rev. 2003, 89, 679. [Google Scholar] [CrossRef]
  126. Judge, B.; Nitzberg, M.; Russell, S. When code isn’t law: Rethinking regulation for artificial intelligence. Policy Soc. 2024, puae020. Available online: https://academic.oup.com/policyandsociety/advance-article/doi/10.1093/polsoc/puae020/7684910 (accessed on 22 December 2024). [CrossRef]
  127. Mishra, S.; Sharma, S.K. Advanced contribution of IoT in agricultural production for the development of smart livestock environments. Internet Things 2023, 22, 100724. [Google Scholar] [CrossRef]
  128. Farooq, M.S.; Sohail, O.O.; Abid, A.; Rasheed, S. A survey on the role of iot in agriculture for the implementation of smart livestock environment. IEEE Access 2022, 10, 9483–9505. [Google Scholar] [CrossRef]
  129. Alonso, R.S.; Sittón-Candanedo, I.; García, Ó.; Prieto, J.; Rodríguez-González, S. An intelligent Edge-IoT platform for monitoring livestock and crops in a dairy farming scenario. Ad Hoc Netw. 2020, 98, 102047. [Google Scholar] [CrossRef]
  130. Biswas, D.; Jalali, H.; Ansaripoor, A.H.; De Giovanni, P. Traceability vs. sustainability in supply chains: The implications of blockchain. Eur. J. Oper. Res. 2023, 305, 128–147. [Google Scholar] [CrossRef]
  131. Kshetri, N. Blockchain and sustainable supply chain management in developing countries. Int. J. Inf. Manag. 2021, 60, 102376. [Google Scholar] [CrossRef]
  132. Nikolakis, W.; John, L.; Krishnan, H. How blockchain can shape sustainable global value chains: An evidence, verifiability, and enforceability (EVE) framework. Sustainability 2018, 10, 3926. [Google Scholar] [CrossRef]
  133. Soori, M.; Arezoo, B.; Dastres, R. Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cogn. Robot. 2023, 3, 54–70. [Google Scholar] [CrossRef]
  134. Shah, V. Machine learning algorithms for cybersecurity: Detecting and preventing threats. Rev. Esp. Doc. Cient. 2021, 15, 42–66. [Google Scholar]
  135. Bharadiya, J.P. Machine learning and AI in business intelligence: Trends and opportunities. Int. J. Comput. (IJC) 2023, 48, 123–134. [Google Scholar]
  136. Lein, J.K. Implementing remote sensing strategies to support environmental compliance assessment: A neural network application. Environ. Sci. Policy 2009, 12, 948–958. [Google Scholar] [CrossRef]
  137. Willis, K.S. Remote sensing change detection for ecological monitoring in United States protected areas. Biol. Conserv. 2015, 182, 233–242. [Google Scholar] [CrossRef]
  138. De Leeuw, J.; Georgiadou, Y.; Kerle, N.; De Gier, A.; Inoue, Y.; Ferwerda, J.; Smies, M.; Narantuya, D. The function of remote sensing in support of environmental policy. Remote Sens. 2010, 2, 1731–1750. [Google Scholar] [CrossRef]
  139. Hafeez, A.; Husain, M.A.; Singh, S.; Chauhan, A.; Khan, M.T.; Kumar, N.; Chauhan, A.; Soni, S. Implementation of drone technology for farm monitoring & pesticide spraying: A review. Inf. Process. Agric. 2023, 10, 192–203. [Google Scholar]
  140. Lawrence, I.D.; Agnishwar, J.; Vijayakumar, R. Drone Technology in Agriculture for Surveillance and Inspection. Eur. Chem. Bull. 2023, 12, 1253–1263. [Google Scholar]
  141. Han, Y.; Guo, Z. Regulator-Manufacturer AI Agents Modeling: Mathematical Feedback-Driven Multi-Agent LLM Framework. arXiv 2024, arXiv:2411.15356. [Google Scholar]
  142. James, P. Australian Cherry Production Guide. Available online: https://figshare.utas.edu.au/articles/report/Australian_Cherry_Production_Guide/23197466 (accessed on 22 December 2024).
  143. BEEAware. B-Qual Approved Supplier Program Manual Record Templates. Available online: https://beeaware.org.au/code-of-practice/record-keeping/ (accessed on 5 December 2024).
  144. Stratton, J. An Introduction to Microsoft Copilot. In Copilot for Microsoft 365: Harness the Power of Generative AI in the Microsoft Apps You Use Every Day; Springer: Berlin/Heidelberg, Germany, 2024; pp. 19–35. [Google Scholar]
  145. Wu, T.; He, S.; Liu, J.; Sun, S.; Liu, K.; Han, Q.L.; Tang, Y. A Brief Overview of ChatGPT: The History, Status Quo and Potential Future Development. IEEE/CAA J. Autom. Sin. 2023, 10, 1122. [Google Scholar] [CrossRef]
  146. Bairi, R.; Sonwane, A.; Kanade, A.; Iyer, A.; Parthasarathy, S.; Rajamani, S.; Ashok, B.; Shet, S. Codeplan: Repository-level coding using llms and planning. Proc. ACM Softw. Eng. 2024, 1, 675–698. [Google Scholar] [CrossRef]
  147. Yaacov, T.; Elyasaf, A.; Weiss, G. Boosting LLM-Based Software Generation by Aligning Code with Requirements. In Proceedings of the 2024 IEEE 32nd International Requirements Engineering Conference Workshops (REW), Reykjavik, Iceland, 24–25 June 2024; pp. 301–305. [Google Scholar] [CrossRef]
Figure 1. Stakeholders and Compliance Process. There can be a many-to-many relationship between producers and regulatory bodies.
Figure 1. Stakeholders and Compliance Process. There can be a many-to-many relationship between producers and regulatory bodies.
Futureinternet 17 00100 g001
Figure 2. AI in RegTech Application.
Figure 2. AI in RegTech Application.
Futureinternet 17 00100 g002
Figure 3. Hierarchy of checks for compliance.
Figure 3. Hierarchy of checks for compliance.
Futureinternet 17 00100 g003
Figure 4. The architecture of Regulatory Technology with Artificial Intelligence and Backend Web services.
Figure 4. The architecture of Regulatory Technology with Artificial Intelligence and Backend Web services.
Futureinternet 17 00100 g004
Figure 5. LLM and conventional deep learning compared to traditional software. Compared to Traditional software design, deep learning can provide more in-depth analysis but requires application-specific pre-trained models for accuracy. In contrast, LLMs cannot offer in-depth analysis, even though they can provide much more flexible user interaction.
Figure 5. LLM and conventional deep learning compared to traditional software. Compared to Traditional software design, deep learning can provide more in-depth analysis but requires application-specific pre-trained models for accuracy. In contrast, LLMs cannot offer in-depth analysis, even though they can provide much more flexible user interaction.
Futureinternet 17 00100 g005
Figure 6. LLM–web service efficiency: (a) llama3.2; (b) llama3.1.
Figure 6. LLM–web service efficiency: (a) llama3.2; (b) llama3.1.
Futureinternet 17 00100 g006
Table 1. Traditional Software vs. LLM for Compliance Check.
Table 1. Traditional Software vs. LLM for Compliance Check.
Traditional SoftwareLLM
Checking CriteriaFixed criteria for checkingVariable criteria for checking
ExpandabilityLimited expansion is possible; only dedicated developers can add featuresDedicated developers are needed to update the LLM model only
ExplainabilityIt is not possible to explain any unexpected situationPossible to re-interpret customized explanations from raw data
Data SourceLimited sources of data can be consumed in a given period of timeUnlimited data can be processed, although checking types is limited
AIDeep learning, but no conversational paradigmsNaturally conversational and human-friendly
UpgradesDedicated developers are needed; it is more time-consumingNo dedicated developers are required, and it is less time-consuming
Table 2. List Extraction and Accuracy (with Prompt 1 from Appendix A).
Table 2. List Extraction and Accuracy (with Prompt 1 from Appendix A).
LLMInput SizeExtraction
Accuracy
Extraction TimeList Order
Variation
Avg. List Item
Displacements
meλ
Co-pilotFull0.00.01.0010.07 s0.00%0.00%
ChatGPTFull3.20.10.8610.17 s90.00%72.17%
ClaudeFull1.10.00.9510.48 s90.00%43.47%
Co-pilot5 pages0.10.00.992.7 s100.00%34.78%
ChatGPT5 pages1.00.10.955.3 s100.00%82.17%
Claude5 pages0.20.00.999.3 s50.00%29.56%
Table 3. The accuracy (λ) of LLMs in answering NMB compliance checks on time series.
Table 3. The accuracy (λ) of LLMs in answering NMB compliance checks on time series.
Prompt →234567891011
Level →1112222233
Copilot1.00.991.000.820.000.000.640.690.560.0
ChatGPT0.80.180.670.240.000.000.220.120.000.0
Gemini0.250.360.830.00.000.000.270.480.300.40
Table 4. The execution time (s) of LLMs in answering NMB compliance checks on time series.
Table 4. The execution time (s) of LLMs in answering NMB compliance checks on time series.
Prompt →234567891011
Level →1112222233
Copilot2.843.53.950.26.440.633.222.77.738.2
ChatGPT25.023.09.136.355.953.725.212.023.745.1
Gemini12.416.42.962.3101.971.029.44.310.96.6
Table 5. The accuracy of rule identification by LLM with a single 0-shot prompt (with prompts 12–15 in Appendix A).
Table 5. The accuracy of rule identification by LLM with a single 0-shot prompt (with prompts 12–15 in Appendix A).
Ollama ModelαεTypeRule
Identification (λ)
Rule-Code
Conversion (λ)
Time (s)
llama3.2:3b17Numeric1.001.002.1
llama3.1:8b17Numeric1.001.003.1
Copilot17Numeric1.001.002.5
llama3.2:3b123String0.990.992.3
llama3.1:8b123String0.980.982.1
Copilot123String1.001.002.3
llama3.2:3b218Numeric0.150.155.1
llama3.1:8b218Numeric0.140.145.4
Copilot218Numeric0.90.95.2
llama3.2:3b216Numeric0.950.953.2
llama3.1:8b216Numeric0.950.953.1
Copilot216Numeric0.970.975.1
Table 6. Format Checking Accuracy with 0-shot prompts—Relevance.
Table 6. Format Checking Accuracy with 0-shot prompts—Relevance.
LLMForm 1Form 2Form 3Form 4Form 5
λTimeλTimeλTimeλTimeλTime
Copilot0.95.590.75.530.85.500.85.480.75.00
ChatGPT0.99.550.88.410.76.300.88.030.815.9
Table 7. Format Checking Accuracy with 0-shot prompts—Completeness.
Table 7. Format Checking Accuracy with 0-shot prompts—Completeness.
LLMParagraph 1Paragraph 2Paragraph 3
λTimeλTimeλTime
Copilot0.847.500.877.120.775.22
ChatGPT0.9416.690.877.310.8913.80
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Maiti, A. Applying Large Language Model Analysis and Backend Web Services in Regulatory Technologies for Continuous Compliance Checks. Future Internet 2025, 17, 100. https://doi.org/10.3390/fi17030100

AMA Style

Li J, Maiti A. Applying Large Language Model Analysis and Backend Web Services in Regulatory Technologies for Continuous Compliance Checks. Future Internet. 2025; 17(3):100. https://doi.org/10.3390/fi17030100

Chicago/Turabian Style

Li, Jinying, and Ananda Maiti. 2025. "Applying Large Language Model Analysis and Backend Web Services in Regulatory Technologies for Continuous Compliance Checks" Future Internet 17, no. 3: 100. https://doi.org/10.3390/fi17030100

APA Style

Li, J., & Maiti, A. (2025). Applying Large Language Model Analysis and Backend Web Services in Regulatory Technologies for Continuous Compliance Checks. Future Internet, 17(3), 100. https://doi.org/10.3390/fi17030100

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop