- Introduction
- Definitions and Basic Properties
- Minimization of Variables: Exact Method
- Minimization of Variables: Heuristic Method
- Two-Class Functions
- Linear Decomposition
- Data Mining and Machine Learning
- Functions with Multi-Valued Inputs
- Easily Reconstructable Functions
- Functions with Continuous Variables
- References
- Conclusions

This book introduces a novel perspective on machine learning, offering distinct advantages over neural network-based techniques. This approach boasts a reduced hardware requirement, lower power consumption, and enhanced interpretability. The applications of this approach encompass high-speed classifications, including packet classification, network intrusion detection, and exotic particle detection in high-energy physics. Moreover, it finds utility in medical diagnosis scenarios characterized by small training sets and imbalanced data. The resulting rule generated by this method can be implemented either in software or hardware. In the case of hardware implementation, circuit design can employ look-up tables (memory), rather than threshold gates. The methodology described in this book involves extracting a set of rules from a training set, composed of categorical variable vectors and their corresponding classes. Unnecessary variables are eliminated, and the rules are simplified before being transformed into a sum-of-products (SOP) form. The resulting SOP exhibits the ability to generalize and predict outputs for new inputs. The effectiveness of this approach is demonstrated through numerous examples and experimental results using the University of California-Irvine (UCI) dataset. This book is primarily intended for graduate students and researchers in the fields of logic synthesis, machine learning, and data mining. It assumes a foundational understanding of logic synthesis, while familiarity with linear algebra and statistics would be beneficial for readers