A self-critique loop is an architectural component in which a language model evaluates its own proposed outputs against a predefined set of principles, identifies potential violations, and iteratively revises its response before final generation. This internal feedback mechanism is central to Constitutional AI frameworks, enabling autonomous alignment without constant human oversight. The loop typically involves a critique generation step, where the model analyzes its draft for issues, followed by a revision step to correct them.
