Next Article in Journal
High-Precision Forward Modeling of Controlled Source Electromagnetic Method Based on Weighted Average Extrapolation Method
Previous Article in Journal
Unified Aviation Maintenance Ecosystem on the Basis of 6G Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Tsetlin Machine with Enhanced Generalization

by
Usman Anjum
* and
Justin Zhan
Department of Computer Science, University of Cincinnati, Cincinnati, OH 45221, USA
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(19), 3825; https://doi.org/10.3390/electronics13193825
Submission received: 16 August 2024 / Revised: 25 September 2024 / Accepted: 26 September 2024 / Published: 27 September 2024

Abstract

:
The Tsetlin Machine (TM) is a novel machine learning approach that implements propositional logic to perform various tasks such as classification and regression. The TM not only achieves competitive accuracy in these tasks but also provides results that are explainable and easy to implement using simple hardware. The TM learns using clauses based on the features of the data, and final classification is done using a combination of these clauses. In this paper, we propose the novel idea of adding regularizers to the TM, referred to as Regularized TM (RegTM), to improve generalization. Regularizers have been widely used in machine learning to enhance accuracy. We explore different regularization strategies and their influence on performance. We demonstrate the feasibility of our methodology through various experiments on benchmark datasets.

1. Introduction

The Tsetlin Machine (TM) is a novel machine learning algorithm first proposed in 2018 by Granmo et al. [1]. It is based on the Tsetlin Automaton (TA). The TM is able to find complex patterns using propositional logic. It learns by performing an action that can trigger a reward or penalty. The TM can have applications in not only pattern recognition but also in other domains and tasks, such as regression, healthcare, and natural language processing. TMs have the advantage of low memory usage and increased learning speed while achieving competitive predictive performance in several benchmark datasets [2,3]. The TM can easily be implemented using simple low-energy consumption hardware [4,5].
Using propositional logic in the TM has many added advantages over traditional machine learning algorithms. Propositional logic-based learning is more interpretable, and the results are easy for humans to understand. Hence, the TM has important applications in the field of explainable AI, where interpretability and transparency are important for machine learning algorithms [6,7]. Propositional logic-based algorithms are also much faster in execution time than traditional neural networks that are based on multiple layers using vector multiplication. In fact, some traditional machine learning algorithms have proposed using propositional logic based algorithms for learning [8,9].
In the TM, propositional logic is used to create clauses. The clauses are made from literals which are features of the data. The literals within the clauses are learned using Type I or Type II feedback based on the feedback from the classification result. The predicted class is based on the majority sum from the clauses which is then obtained by a step-function defined by the parameter T. The number of clauses and the parameter T are the main hyperparameters that increase the discriminative power of the TM and enable patterns to be learned more efficiently.
Increasing the number of clauses and having a high value of T, may lead to better predictive power of the TM but it may also lead to overfitting and an increase in resources used and a reduction in training speed.
In this paper, we propose a novel variant of the TM called Regularized TM (RegTM). RegTM is based on the idea of using regularizers and smoothness functions, just like in traditional machine learning algorithms [10]. We use two different variants of regularizers. The first regularizer, called the Moving Average Regularizer (MAR), uses the sum of clauses from previous iterations and samples and incorporates them into the current clause sum calculation. The second type of regularizer, called the Weighted Regularizer (WER), adds the sum of the weights to improve generalization. By using these regularization techniques, the generalization capability of the TM can improve, increasing their predictive power, especially in cases where the data is low in count.
Furthermore, we also propose the use of the Sigmoid function [11,12]. The sigmoid function can make the TM differentiable, and it can be further optimized with standard methods like gradient descent.
In summary, the contributions of our paper are as follows:
  • We propose a novel variant of the Tsetlin Machine called Regularized Tsetlin Machine (RegTM) that implements regularization based on past clause sums and weights to improve the accuracy of the TM by increasing its generalizability, making it highly suitable for few-shot learning and meta-learning optimization tasks.
  • We propose to use the Sigmoid function as an alternative to the unit-step function.
  • We perform different experiments using benchmark image datasets to show the efficiency of RegTM with the different regularizers and sigmoid function.

2. Related Works

Granmo et al. proposed the TM in 2018 as a way to provide solutions to the bandit algorithm [1,13,14]. Since then, there has been a lot of significant work in this domain to improve upon the existing idea.
The TM was recast as a contextual bandit algorithm in [15]. Two contextual bandit learning algorithms were proposed: TM with ϵ -greedy arm selection and TM with Thompson sampling.
TM was extended for image classification and was called the Convolution Tsetlin Machine (CTM) [16,17]. CTM employed clauses as a convolution filter on each image as opposed to one clause for each image.
In addition to image classification, TM has also been used for regression [18]. The Regression TM is able to output in a continuous form rather than as distinct categories. The Regression TM employs a new voting mechanism to analyze complex patterns.
For multi-output problems, a coalesced Tsetlin Machine (CoTM) was proposed that simultaneously learns both the weights and the composition of each clause by employing interacting Stochastic Searching on the Line (SSL) and Tsetlin Automata (TA) [19].
TM has also shown great potential for natural language processing, e.g., text encoding [20], text classification [21,22,23], sentiment analysis [24], question classification [25], and fake news detection [26].
TM has also been implemented in the domain of healthcare. For example, a TM architecture was developed for the identification of premature ventricular contraction (PVC) by analyzing long-term ECG signals [27]. The capability of TM using continuous data was proposed for forecasting disease outbreaks [18]. Intelligent prefiltering of healthcare measurement data using Principal Component Analysis and Partial Least Squares Regression was added to TM to improve performance [28].
A label-critical TM was proposed that used a twin of label-critical Tsetlin Automata (TA). Label-TA learned the correct labels of the individual samples guided by a logical self-corrected TM clause [29].
There has also been substantial work on improving the feasibility of TM. One of the ways of improving the TM was by focusing on the clauses. The accuracy of the TM was improved with the addition of a drop clause probability [30]. The drop clause probability dropped clauses with a specific probability to increase the robustness, accuracy, and training time of TM, similar to a dropout layer in neural networks.
Another improvement to TM was the addition of weights called the Weighted Tsetlin Machine (WTM) [31]. The Weighted Tsetlin Machine (WTM) adds weights to the clauses to reduce the computation time and memory usage.
Focused negative sampling (FNS) was introduced to discriminate between classes that are not easily distinguishable from each other to improve TM precision [32].
A method to more efficiently prune clauses using the Markov boundary was proposed [33]. The new scheme aimed to add an additional feedback scheme to the TM, called the Type III feedback.
Another proposed scheme added the absorbing automata to the Tsetlin Machine [34]. Another variant of TM focusing on improving the efficiency of clauses was called the Clause Size Constrained TM (CSC-TM) [35]. In CSC-TM, a soft constraint on the clause size is placed and as soon as the clause includes more literals, that is, more than the allowed constraint, the literals are removed. Another research work looked at finding the optimal clause count by proposing a run-time pruning system [36].

3. Preliminary: Vanilla Tsetline Macnhine

The TM is able to solve complex pattern recognition problems using proposition-based conjunctive logic statements [1]. A standard Tsetlin Automaton (TA) consists of a collection of TAs [37] to find solutions to the multi-armed bandit problem [38,39] and the Finite State Learning Algorithm (FSLA) [40]. The TM uses conjunctive clauses composed of a collection of TAs.
In this section, we provide an overview of the TM to help us understand the algorithm introduced in subsequent sections. The TM takes a vector X of n propositional variables as input within the domain 0,1 and is defined as:
X = x k 0 , 1 n
For example, if n = 2 , the variables are x 1 and x 2 , then the domain is { ( 0 , 0 ) , ( 0 , 1 ) , ( 1 , 0 ) , ( 1 , 1 ) } .
The TM determines patterns using conjunctive clauses. A conjunctive clause consists of literals and is meant to describe the dataset. A literal consists of the variables and their negations L = l k 0 , 1 2 n where each literal l k is defined as:
l k = x k i f 1 k n , ¬ x k n i f n + 1 k 2 n
The first n literals are the actual variables and the next n variables are the negated variables. Using the example of n = 2 , L = { x 1 , x 2 , ¬ x 1 , ¬ x 2 } .
Typically, the number of clauses is a user set parameter. It is an important parameter that determines how well training is done [36]. Let the number of clauses be defined as m and each clause be denoted as C j where j { 1 , . . . , m } , then the clause is a conjunction of a subset of literals L j L :
C j = l k L j l k = l k L j l k
C j { 0 , 1 } . Let m = 2 and the user defined clauses are C 1 = x 1 ¬ x 2 and C 2 = x 1 x 2 . Then, from the previous example, the literals have the index L 1 = { 1 , 4 } and L 2 = { 1 , 2 } .
Each clause is also assigned a polarity p { 1 , + 1 } . The polarity decides whether the clause output is negative or positive. Positive polarity clauses, denoted by C j + { 0 , 1 } m 2 , are assigned to class y = 1 and negative polarity clauses, denoted by C j { 0 , 1 } m 2 , are assigned to class y = 0 . Typically, half of the clauses are positive and half are negative. The positive polarity clauses vote for classifying the input as the target class and the negative polarity clauses vote against the target class.
The final classification decision is done by summing the clause outputs:
s ( X ) = j = 1 m / 2 C j + j = 1 m / 2 C j y ^ = u ( s ( X ) )
where u ( v ) is the unit step function defined as:
u ( v ) = 1 i f v 0 0 o t h e r w i s e
In the end, the output of the sum of summation decides the class y ^ { 0 , 1 } .
Equation (4) was modified in the by adding weights to the clauses Weighted Tsetlin Machine [31]:
s ( X ) = j = 1 m / 2 w j + C j + j = 1 m / 2 w j C j
For example, Figure 1 shows two positive polarity clauses, C 1 = x 1 ¬ x 2 and C 3 = ¬ x 1 ¬ x 2 (negative polarity clauses are not shown). Both clauses evaluate to zero, and leading to a prediction of y ^ = 1 .
For training, each clause is given feedback with a probability p ^ .
p ^ = T + c l a m p ( s ( X ) , T , T ) 2 T i f y = 0 T c l a m p ( s ( X ) , T , T ) 2 T i f y = 1
In the above equation, s ( X ) is the sum of the clause of Equation (6). T is a user-defined parameter. The c l a m p function scales the s ( X ) between T and T.
The above random clause selection is crucial as it guides the clauses to distribute across the sub-patterns in a more efficient manner. The frequency of each clause becomes proportional to ± T . The amount of feedback depends on this and more feedback is only generated when s ( x ) is far from T. In this way, only a few clauses are activated, thus reducing the overhead during training.
The TA is depicted in the upper part of Figure 1. The TA determines the actions to take during learning. The TA can decide to include or exclude the literal in the clause based on the received feedback. The feedback can be categorized as Type I or Type II. Figure 2 depicts a two-action TA with 2N states. For states 1     N , the TA performs exclude (Action 1) and for states N + 1     2 N , include (Action 2) actions are performed. Based on the feedback, a TA can receive a reward or a penalty. If a TA receives a reward, then the literal moves deeper into being included or excluded. If a TA receives a penalty, then the literal moves towards the middle until it moves from action 1 to action 2 and vice versa.
Type I feedback reduces false negatives and overfitting by frequently recognizing patterns. Type II feedback suppresses false positives by increasing the discrimination power of the learned patterns.
Type I feedback: In Type I feedback, clauses of positive polarity receive a positive feedback when y = 1 and negative feedback for clauses with negative polarity and when y = 0 . There are two types of Type I feedback: Type 1a and Type1b. Type 1a feedback reinforces the TA include actions. Type 1b feedback is activated when either the clause output or the literal is 0. Type 1b causes the TA to exclude the literal.
Type I feedback is defined by a user-defined parameter s where s 1 . Let the TA state be defined by a j , k for action on the k t h literal and the j t h clause. The TA state a j , k receives Type 1a feedback with probability s 1 s and increases a j , k by 1. Consequently, the TA receives Type 1b feedback with probability 1 s and reduces a j , k by 1. In this way, Type 1a pushes to include the literal and Type 1b pushes to exclude the literal.
Type II feedback: Type II feedback is given to clauses with positive polarity when y = 0 and to clauses with negative polarity when y = 1 . When the clause output is 0, the TA states are not updated. When the clause output is 1, all 0-valued literals that make the output 0 in TA state a j , k are increased by 1. Thus, this feedback focuses on all literals that differentiate between y = 0 and y = 1 .

4. Methodology

In this section, we introduce our novel TM variant titled the Regularized Tsetlin Machine (RegTM).
RegTM is based on the novel idea of using regularization and smoothing for Tsetline Machines. Regularization has been used extensively in machine learning algorithms. It is a way of infusing knowledge about data and parameters into the learning algorithm and provides a good balance between bias and variance. Regularization adds bias to the learning process and helps reduce the variance.
Regularization can reduce overfitting and improve the generalizability of the model [41].
One of the most common methods of regularization is to constrain the capacity of a model by adding some penalty function to the original objective function and making a new objective function. In TM we consider the objective function as the weighted clause sum equation represented in Equation (6). Hence, the new clause summation equation in RegTM can be written as:
s R e g T M ( X ) = j = 1 m / 2 ( w j + C j + + b j + ) j = 1 m / 2 ( w j C j b j )
where b j + and b j are the two regularization terms for the positive and negative classes.
Next we will describe the two regularization methods we propose for implementation in RegTM.

4.1. Moving Average Regularizer (MAR)

The first regularization method we explore is inspired by the moving average term used in time series analysis [42]. In this case we assume that the the short-term fluctuations between clause sums can cause the model to be less generalizable. To remove the fluctuation, we propose a bias function that is dependent on the past values of the clause sums. We call this Moving Average Regularizer (MAR).
Consequently, b j + and b j from Equation (8) can be written as:
b j + = p ( 1 k n = 1 k C j n + )
b j = p ( 1 k n = 1 k C j n )
where p is a weighting parameter to define how much weight we want to put on the regularizer terms, k is a parameter that determines how far into the past we want to consider the clause sums, and C j k + and C j k are the positive and negative clause sums of the sample ( j k ) t h :
C j k + = j = 1 m / 2 w j + C j +
C j k = j = 1 m / 2 w j C j

4.2. Weighted Regularizer (WER)

The second method of regularization is based on weight decay [41], which we call Weighted Regularizer (WER). For WER, the bias terms from Equation (8) are expressed as:
b j + = p j = 1 m / 2 ( w j + ) k
b j = p j = 1 m / 2 ( w j ) k
where p is a parameter to define how much weight we want to put on the regularizer, k is the power of the weights, and ( w j + ) k and ( w j ) k are weights from Equation (8) for the k t h sample.

4.3. Sigmoid

In this paper, we also introduce an alternative to Equation (7) for the feedback mechanism of the classes. One of the drawbacks of Equation (7) is that it is non-differentiable and cannot be optimized via standard methods such as gradient descent. This might be problematic if the TM is to be integrated with traditional neural networks for hyperparameter optimization and for providing explainability to neural networks.
Hence, we propose the use of the Sigmoid function. The Sigmoid function is a popular mathematical function that has an S-shaped curve and works in a similar way to the clamp function [9]. The Sigmoid function maps values between 0 , 1 . The Sigmoid function is defined as:
σ ( s ( X ) ) = 1 1 + e x p s ( X )

5. Experiments

In this section, we introduce the experimental parameters, data used for the experiments, and the experimental results.

5.1. Experimental Design

All of our experiments are performed using the Github repository of the TM (https://github.com/cair/tmu/tree/main accessed on 1 September 2024).
We use the two standard benchmark datasets (MNIST and CIFAR10) for vision classification. The MNIST database contains 60,000 training images and 10,000 testing images. The dataset is a set of 28 × 28 grayscale images of 10 digits. The CIFAR10 dataset consists of 60,000 32 × 32 color images in 10 classes, with 6000 images per class. There are 50,000 training images and 10,000 test images.
The benchmark methods for comparison are the weighted Tsetlin Machine with literal and drop clauses [30,31]. The RegTM implementation includes using MAR and WER with and without the sigmoid function.

5.2. Experiment Results

The accuracy of the test dataset after 60 epochs is summarized in Table 1 and Table 2. The column TM reports the results with vanilla weighted TM. The column with Drop Probability reports the results with drop clause and drop literal probabilities included in the algorithm. The results are reported for different values of clause counts (C).
The drop probabilities are obtained by using the literal and clause drop probability that results in the best accuracy. For MNIST, the best accuracy is obtained when both clause drop and literal drop probabilities are 0.1. For CIFAR10, the best accuracy is obtained when the clause drop probability and the literal drop probability are 0.4 and 0.0 , respectively.
Similarly, we also use the values for k and p that give the highest accuracy. For the MNIST dataset, when considering MAR, the best accuracy is obtained when k = 0 and p = 0.03 . For MAR with sigmoid, the best accuracy is obtained when k = 1 and p = 0.1 . Similarly, for WER the best accuracy is obtained when k = 2 and p = 0.4 and for WER with sigmoid, by setting k = 3 and p = 0.01 gives the best accuracy.
For the CIFAR10 dataset, the best accuracy are given by k = 0 ,   p = 0.8 for MAR, k = 3 ,   p = 0.5 for MAR and sigmoid, k = 3 ,   p = 0.1 for WER, and k = 2 ,   p = 0.7 for WER and sigmoid.
In general, using RegTM regularizers improves the accuracy of the TM. In addition, the improvement in accuracy by using RegTM is better when compared to using drop literals and drop clauses. In fact, for the MNIST dataset, adding drop probability appears to have reduced the accuracy. Furthermore, combining RegTM with drop clauses improves accuracy even further, especially for the CIFAR10 dataset where the accuracy is already low.
For RegTM with only the sigmoid function (denoted by sigmoid), the accuracy is very close to the vanilla TM for both the MNIST and CIFAR10 datasets. This means that the sigmoid function can be used as a viable replacement for Equation (7), especially when there is a need to develop a differentiable model or integrate TM with neural networks.
When comparing different datasets, the improvement in accuracy for MNIST is much lower compared to CIFAR10. This is because MNIST already has a high accuracy, so improvements in its accuracy are very small.
The results show that RegTM with WER has higher accuracy than MAR, and when it does have lower accuracy (e.g., in the case of the MNIST dataset), the difference is very low. Adding a sigmoid function to MAR and WER also appears to not affect the performance of MAR and WER and in some cases, improves the accuracy.

5.3. Ablation Studies

Our ablation studies include observing the effect of the parameter k and p on the accuracy and comparing the runtime of RegTM with TM and TM with drop probabilities.
The effect of the parameters k and p on the accuracy for both the MNIST and CIFAR10 datasets is shown in Figure 3, Figure 4, Figure 5 and Figure 6. It is observed that for a specific k value and any value of p, the accuracy drops sharply. In the case of the MAR algorithm, significantly lower accuracy is obtained when k = 2 , Similarly, for the WER algorithm when k = 0 much lower accuracy is observed. We will study this in more detail in future work. For the other k and p values, there is very little variation in the accuracy values. For MNIST the accuracy is almost the same. For CIFAR10, the variation is slightly more which may be due to the complexity of the dataset.
We also compare the runtime over 60 epochs for the different regularizers. The results are reported in Table 3. The run-time of RegTM with MAR and WER is comparable to that of the vanilla TM and better than the use of dropping clauses and literals. Adding the sigmoid function does increase the run-time compared to vanilla TM and TM with drop clauses, but the increase is not so significant.

6. Conclusions

In this paper, we introduce a novel variant of the Tsetlin Machine called the Regularized Tsetlin Machine (RegTM) by proposing the idea of using regularizers to improve the performance and generalization of the TM. Even though the improvement in the results is marginal, it opens up new potential for future research, where generalization is important, especially in the field of few-shot learning and meta-optimization [43,44].
Our work shows promise and improves upon the performance of TM. However, there are some limitations, epsecally in case of TM. In some cases, the explanatory advantages that TM provides are not intuitive. Understanding and making the interpretations provided by TMs are important. But this is an on-going research. In our paper, we do not specifically focus on the interpretability of TM but rather on increasing its generalizability through regularizers.
Another limitation of TM is its inefficiency in handling large-scale data. This is because of the increase in the Tsetlin Automata, which can cause memory usage problems and reduce computational efficiency. One way of dealing with this issue is by using Sparse Tsetlin Machine [45] and parallelization [1]. However, it is an active area of research that requires further attention. In our paper, we do not focus on how to handle large-scale data. However, we believe regularizers could be used in combination with Sparse Tsetlin Machines to improve performance.
Another concern regarding TM is how to select the different hyperparameters. Currently, the way to select hyperparameters is using grid search, random search, cross-validation or using Bayesian optimization [1,46]. We have shown that using the sigmoid function does not reduce the performance of TM. This means that gradient descent could be used in combination with TM, allowing us to apply the same optimization techniques as those from traditional neural networks. Hence, we can implement other hyperparameter optimization techniques in TM that are also used in neural networks, which would be an interesting area for future research.
Another limitation of TM is that it can only handle binary features. This means that continuous features can only be handled by discretization, which may lead to information loss. This has been discussed in [47] but it still requires further research and investigation of techniques that can reduce information loss.
The main focus of TM has been in the field of classification and regression tasks, but its application in other tasks (clustering, segmentation, and sequence modeling) is still limited. Research on the application of TM in other tasks is ongoing. Our work on regularizers aims to be a step towards the application of TM in other domains and tasks, more specifically in few-shot learning. The applicability of TM for other tasks would be another potential future research area.
Another interesting area of study can involve investigating more complex regularizers, such as tangentprop [48]. However, implementing such regularizers and even the sigmoid function might be challenging considering that TMs are mainly designed for low-energy hardware. This is something that could also be further studied.

Author Contributions

Conceptualization, U.A. and J.Z.; Methodology, U.A.; Validation, J.Z.; Writing–original draft, U.A.; Writing–review & editing, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

We gratefully acknowledge the support National Science Foundation (#2322109, #2346609, #2333726), the Department of Defense Office of Naval Research (#N00014-23-1-2396), and Cincinnati Children’s Hospital Medical Center for their support and finance for this project.

Data Availability Statement

The data is publicly available at https://keras.io/api/datasets/, accessaed on 1 September 2024.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Granmo, O.C. The Tsetlin Machine—A Game Theoretic Bandit Driven Approach to Optimal Pattern Recognition with Propositional Logic. arXiv 2018, arXiv:1804.01508. [Google Scholar]
  2. Abeyrathna, K.D.; Bhattarai, B.; Goodwin, M.; Gorji, S.R.; Granmo, O.C.; Jiao, L.; Saha, R.; Yadav, R.K. Massively parallel and asynchronous tsetlin machine architecture supporting almost constant-time scaling. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 10–20. [Google Scholar]
  3. Lei, J.; Wheeldon, A.; Shafik, R.; Yakovlev, A.; Granmo, O.C. From arithmetic to logic based ai: A comparative analysis of neural networks and tsetlin machine. In Proceedings of the 2020 27th IEEE International Conference on Electronics, Circuits and Systems (ICECS), Glasgow, UK, 23–25 November 2020; pp. 1–4. [Google Scholar]
  4. Cheng, R.; Vasudevan, D.; Kirst, C. Super-Tsetlin: Superconducting Tsetlin Machines. IEEE Trans. Appl. Supercond. 2024, 34, 1301012. [Google Scholar] [CrossRef]
  5. Morris, J.; Rafiev, A.; Xia, F.; Shafik, R.; Yakovlev, A.; Brown, A. An alternate feedback mechanism for tsetlin machines on parallel architectures. In Proceedings of the 2022 International Symposium on the Tsetlin Machine (ISTM), Grimstad, Norway, 20–21 June 2022; pp. 53–56. [Google Scholar]
  6. Xu, F.; Uszkoreit, H.; Du, Y.; Fan, W.; Zhao, D.; Zhu, J. Explainable AI: A brief survey on history, research areas, approaches and challenges. In Proceedings of the Natural Language Processing and Chinese Computing: 8th CCF International Conference (NLPCC 2019), Dunhuang, China, 9–14 October 2019; Proceedings, Part II 8. Springer: Cham, Switzerland, 2019; pp. 563–574. [Google Scholar]
  7. Došilović, F.K.; Brčić, M.; Hlupić, N. Explainable artificial intelligence: A survey. In Proceedings of the 2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO), Opatija, Croatia, 21–25 May 2018; pp. 0210–0215. [Google Scholar]
  8. Anjum, U.; Zadorozhny, V.; Krishnamurthy, P. Localization of Unidentified Events with Raw Microblogging Data. Online Soc. Netw. Media 2022, 29, 100209. [Google Scholar] [CrossRef]
  9. Petersen, F.; Borgelt, C.; Kuehne, H.; Deussen, O. Deep differentiable logic gate networks. Adv. Neural Inf. Process. Syst. 2022, 35, 2006–2018. [Google Scholar]
  10. Tian, Y.; Zhang, Y. A comprehensive survey on regularization strategies in machine learning. Inf. Fusion 2022, 80, 146–166. [Google Scholar] [CrossRef]
  11. Dombi, J.; Jónás, T. The generalized sigmoid function and its connection with logical operators. Int. J. Approx. Reason. 2022, 143, 121–138. [Google Scholar] [CrossRef]
  12. Dubey, S.R.; Singh, S.K.; Chaudhuri, B.B. Activation functions in deep learning: A comprehensive survey and benchmark. Neurocomputing 2022, 503, 92–108. [Google Scholar] [CrossRef]
  13. Bouneffouf, D. Multi-Armed Bandit Problem and Application. Independently Published. 2023. Available online: https://isbnsearch.org/isbn/9798375018096 (accessed on 25 September 2024).
  14. Bouneffouf, D.; Rish, I.; Aggarwal, C. Survey on applications of multi-armed and contextual bandits. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 3 September 2020; pp. 1–8. [Google Scholar]
  15. Seraj, R.; Sharma, J.; Granmo, O.C. Tsetlin Machine for Solving Contextual Bandit Problems. Adv. Neural Inf. Process. Syst. 2022, 35, 30194–30205. [Google Scholar]
  16. Tunheim, S.A.; Jiao, L.; Shafik, R.; Yakovlev, A.; Granmo, O.C. Convolutional Tsetlin Machine-based Training and Inference Accelerator for 2-D Pattern Classification. Microprocess. Microsystems 2023, 103, 104949. [Google Scholar] [CrossRef]
  17. Granmo, O.C.; Glimsdal, S.; Jiao, L.; Goodwin, M.; Omlin, C.W.; Berge, G.T. The convolutional Tsetlin machine. arXiv 2019, arXiv:1905.09688. [Google Scholar]
  18. Darshana Abeyrathna, K.; Granmo, O.C.; Zhang, X.; Jiao, L.; Goodwin, M. The regression Tsetlin machine: A novel approach to interpretable nonlinear regression. Philos. Trans. R. Soc. A 2020, 378, 20190165. [Google Scholar] [CrossRef] [PubMed]
  19. Glimsdal, S.; Granmo, O.C. Coalesced multi-output tsetlin machines with clause sharing. arXiv 2021, arXiv:2108.07594. [Google Scholar]
  20. Bhattarai, B.; Granmo, O.C.; Jiao, L.; Yadav, R.; Sharma, J. Tsetlin Machine Embedding: Representing Words Using Logical Expressions. arXiv 2023, arXiv:2301.00709. [Google Scholar]
  21. Saha, R.; Granmo, O.C.; Zadorozhny, V.I.; Goodwin, M. A relational tsetlin machine with applications to natural language understanding. J. Intell. Inf. Syst. 2022, 1–28. [Google Scholar] [CrossRef]
  22. Saha, R.; Granmo, O.C.; Goodwin, M. Using Tsetlin machine to discover interpretable rules in natural language processing applications. Expert Syst. 2023, 40, e12873. [Google Scholar] [CrossRef]
  23. Berge, G.T.; Granmo, O.C.; Tveit, T.O.; Goodwin, M.; Jiao, L.; Matheussen, B.V. Using the Tsetlin machine to learn human-interpretable rules for high-accuracy text categorization with medical applications. IEEE Access 2019, 7, 115134–115146. [Google Scholar] [CrossRef]
  24. Yadav, R.K.; Jiao, L.; Granmo, O.C.; Goodwin, M. Human-Level Interpretable Learning for Aspect-Based Sentiment Analysis. In Proceedings of the The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21), Virtually, 2–9 February 2021. [Google Scholar]
  25. Nicolae, D.C. Question Classification using Interpretable Tsetlin Machine. In Proceedings of the 1st International Workshop on Machine Reasoning (MRC 2021), Virtual, 12 March 2021. [Google Scholar]
  26. Bhattarai, B.; Granmo, O.C.; Jiao, L. Explainable Tsetlin Machine Framework for Fake News Detection with Credibility Score Assessment. In Proceedings of the Language Resources and Evaluation Conference, Marseille, France, 20–25 June 2022; pp. 4894–4903. [Google Scholar]
  27. Zhang, J.; Zhang, X.; Jiao, L.; Granmo, O.C.; Qian, Y.; Pan, F. Interpretable Tsetlin Machine-based Premature Ventricular Contraction Identification. arXiv 2023, arXiv:2301.10181. [Google Scholar]
  28. Jenul, A.; Bhattarai, B.; Liland, K.H.; Jiao, L.; Schrunner, S.; Futsaether, C.; Granmo, O.C.; Tomic, O. Component Based Pre-filtering of Noisy Data for Improved Tsetlin Machine Modelling. In Proceedings of the 2022 International Symposium on the Tsetlin Machine (ISTM), Grimstad, Norway, 20–21 June 2022; pp. 57–64. [Google Scholar]
  29. Abouzeid, A.; Granmo, O.C.; Goodwin, M.; Webersik, C. Label-Critic Tsetlin Machine: A Novel Self-supervised Learning Scheme for Interpretable Clustering. In Proceedings of the 2022 International Symposium on the Tsetlin Machine (ISTM), Grimstad, Norway, 20–21 June 2022; pp. 41–48. [Google Scholar]
  30. Sharma, J.; Yadav, R.; Granmo, O.C.; Jiao, L. Drop clause: Enhancing performance, robustness and pattern recognition capabilities of the Tsetlin machine. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 13547–13555. [Google Scholar]
  31. Phoulady, A.; Granmo, O.C.; Gorji, S.R.; Phoulady, H.A. The weighted tsetlin machine: Compressed representations with weighted clauses. arXiv 2019, arXiv:1911.12607. [Google Scholar]
  32. Glimsdal, S.; Saha, R.; Bhattarai, B.; Giri, C.; Sharma, J.; Tunheim, S.A.; Yadav, R.K. Focused Negative Sampling for Increased Discriminative Power in Tsetlin Machines. In Proceedings of the 2022 International Symposium on the Tsetlin Machine (ISTM), Grimstad, Norway, 20–21 June 2022; pp. 73–80. [Google Scholar]
  33. Granmo, O.C.; Andersen, P.A.; Jiao, L.; Zhang, X.; Blakely, C.; Berge, G.T.; Tveit, T. Learning Minimalistic Tsetlin Machine Clauses with Markov Boundary-Guided Pruning. In Proceedings of the 2023 International Symposium on the Tsetlin Machine (ISTM), Newcastle upon Tyne, UK, 29–30 August 2023; pp. 1–8. [Google Scholar]
  34. Bhattarai, B.; Granmo, O.C.; Jiao, L.; Andersen, P.A.; Tunheim, S.A.; Shafik, R.; Yakovlev, A. Contracting Tsetlin Machine with Absorbing Automata. arXiv 2023, arXiv:2310.11481. [Google Scholar]
  35. Abeyrathna, K.D.; Abouzeid, A.A.O.; Bhattarai, B.; Giri, C.; Glimsdal, S.; Granmo, O.C.; Jiao, L.; Saha, R.; Sharma, J.; Tunheim, S.A.; et al. Building concise logical patterns by constraining tsetlin machine clause size. arXiv 2023, arXiv:2301.08190. [Google Scholar]
  36. Rahman, T.; Maheshwari, S.; Shafik, R.; Yakovlev, A.; Das, S. MILEAGE: An Automated Optimal Clause Search Paradigm for Tsetlin Machines. In Proceedings of the 2022 International Symposium on the Tsetlin Machine (ISTM), Grimstad, Norway, 20–21 June 2022; pp. 49–52. [Google Scholar] [CrossRef]
  37. Tsetlin, M.L. On behaviour of finite automata in random medium. Avtomat. I Telemekh 1961, 22, 1345–1354. [Google Scholar]
  38. Robbins, H. Some aspects of the sequential design of experiments. Bull. Am. Math. Soc. 1952, 58, 527–535. [Google Scholar] [CrossRef]
  39. Gittins, J.C. Bandit processes and dynamic allocation indices. J. R. Stat. Soc. Ser. B Stat. Methodol. 1979, 41, 148–164. [Google Scholar] [CrossRef]
  40. Narendra, K.S.; Thathachar, M.A. Learning Automata: An Introduction; Courier Corporation: North Chelmsford, MA, USA, 2012. [Google Scholar]
  41. Moradi, R.; Berangi, R.; Minaei, B. A survey of regularization strategies for deep models. Artif. Intell. Rev. 2020, 53, 3947–3986. [Google Scholar] [CrossRef]
  42. Hamilton, J.D. Time Series Analysis; Princeton University Press: Princeton, NJ, USA, 2020. [Google Scholar]
  43. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. (csur) 2020, 53, 1–34. [Google Scholar] [CrossRef]
  44. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
  45. Østby, S.; Brambo, T.M.; Glimsdal, S. The Sparse Tsetlin Machine: Sparse Representation with Active Literals. arXiv 2024, arXiv:2405.02375. [Google Scholar]
  46. Abeyrathna, K.D.; Granmo, O.C.; Zhang, X.; Goodwin, M. A scheme for continuous input to the Tsetlin machine with applications to forecasting disease outbreaks. In Proceedings of the Advances and Trends in Artificial Intelligence. From Theory to Practice: 32nd International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, IEA/AIE 2019, Graz, Austria, 9–11 July 2019; Proceedings 32. Springer: Cham, Switzerland, 2019; pp. 564–578. [Google Scholar]
  47. Mathisen, E.; Smørvik, H.S. Analysis of Binarization Techniques and Tsetlin Machine Architectures Targeting Image Classification. Master’s Thesis, University of Agder, Grimstad, Norway, 2020. [Google Scholar]
  48. Simard, P.; Victorri, B.; LeCun, Y.; Denker, J. Tangent prop-a formalism for specifying selected invariances in an adaptive network. In Advances in Neural Information Processing Systems 4 (NIPS 1991); Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1991; Volume 4. [Google Scholar]
Figure 1. TM learning dynamics with input ( x 1 = 0 ,   x 2 = 1 ) and output y = 1 .
Figure 1. TM learning dynamics with input ( x 1 = 0 ,   x 2 = 1 ) and output y = 1 .
Electronics 13 03825 g001
Figure 2. A two-action Tsetlin Automaton with 2N states.
Figure 2. A two-action Tsetlin Automaton with 2N states.
Electronics 13 03825 g002
Figure 3. MNIST MAR algorithm (C = 100, T = 5000).
Figure 3. MNIST MAR algorithm (C = 100, T = 5000).
Electronics 13 03825 g003
Figure 4. MNIST WER algorithm (C = 100, T = 5000).
Figure 4. MNIST WER algorithm (C = 100, T = 5000).
Electronics 13 03825 g004
Figure 5. CIFAR10 MAR algorithm (C = 100, T = 5000).
Figure 5. CIFAR10 MAR algorithm (C = 100, T = 5000).
Electronics 13 03825 g005
Figure 6. CIFAR10 WER algorithm (C = 100, T = 5000).
Figure 6. CIFAR10 WER algorithm (C = 100, T = 5000).
Electronics 13 03825 g006
Table 1. Summary of results for MNIST Data (literal drop = 0.1, clause drop = 0.1).
Table 1. Summary of results for MNIST Data (literal drop = 0.1, clause drop = 0.1).
Method ( C = 100 ,   T = 5000 )TMwith Drop Probability
No RegTM96.4396.54
Sigmoid96.4596.15
MAR only ( k = 0 ,   p = 0.03 )96.7595.48
MAR + sigmoid ( k = 1 ,   p = 0.1 )96.6896.42
WER ( k = 2 ,   p = 0.4 )96.5996.23
WER + sigmoid ( k = 3 ,   p = 0.01 )96.6396.23
Method ( C = 200 ,   T = 5000 )TMwith Drop Probability
No RegTM97.2897.08
Sigmoid97.0997.24
MAR only ( k = 0 ,   p = 0.03 )97.1997.06
MAR + sigmoid ( k = 1 ,   p = 0.1 )97.2997.21
WER ( k = 2 ,   p = 0.4 )97.1097.17
WER + sigmoid ( k = 3 ,   p = 0.01 )97.2497.04
Table 2. Summary of results for CIFAR10 Data (literal drop = 0.4, clause drop = 0.0).
Table 2. Summary of results for CIFAR10 Data (literal drop = 0.4, clause drop = 0.0).
Method ( C = 100 ,   T = 5000 )TMwith Drop Probability
No RegTM32.7533.72
Sigmoid32.9233.84
MAR only ( k = 0 ,   p = 0.8 )33.8232.27
MAR + sigmoid ( k = 3 ,   p = 0.5 )34.0233.26
WER ( k = 3 ,   p = 1.0 )33.7533.65
WER + sigmoid ( k = 2 ,   p = 0.7 )33.8633.01
Method ( C = 200 ,   T = 5000 )TMwith Drop Probability
No RegTM34.3735.63
Sigmoid35.8936.31
MAR only ( k = 0 ,   p = 0.8 )34.8436.45
MAR + sigmoid ( k = 3 ,   p = 0.5 )34.7334.15
WER ( k = 3 ,   p = 1.0 )36.2135.57
WER + sigmoid ( k = 2 ,   p = 0.7 )36.5436.57
Table 3. Runtime Comparison (60-epochs, C = 100, T = 5000).
Table 3. Runtime Comparison (60-epochs, C = 100, T = 5000).
MethodMNISTCIFAR10
Vanilla TM00:08:1000:21:10
Vanilla TM with drop clause00:08:3700:22:07
sigmoid only00:08:1800:22:12
MAR only00:08:1400:21:36
MAR + sigmoid00:08:1800:22:10
WER00:08:1000:21:25
WER + sigmoid00:08:2300:23:41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Anjum, U.; Zhan, J. A Novel Tsetlin Machine with Enhanced Generalization. Electronics 2024, 13, 3825. https://doi.org/10.3390/electronics13193825

AMA Style

Anjum U, Zhan J. A Novel Tsetlin Machine with Enhanced Generalization. Electronics. 2024; 13(19):3825. https://doi.org/10.3390/electronics13193825

Chicago/Turabian Style

Anjum, Usman, and Justin Zhan. 2024. "A Novel Tsetlin Machine with Enhanced Generalization" Electronics 13, no. 19: 3825. https://doi.org/10.3390/electronics13193825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop