Symbolic distillation is a machine learning technique that extracts and compresses the knowledge learned by a neural network into a compact, human-interpretable symbolic form, such as a set of logical rules, a decision tree, or a finite-state automaton. This process, also known as rule extraction or model distillation to symbolic form, aims to bridge the gap between the high performance of opaque black-box models and the transparency, verifiability, and efficiency of symbolic AI systems. The distilled symbolic model approximates the behavior of the original neural network while being far more explainable.
