stickler-bot
An introduction to stickler-bot as a practical model for AI-assisted moderation on Reddit.
An Introduction
stickler-bot is an experiment in turning written community standards into durable moderation practice. It begins from a simple premise: on the modern internet, the real character of a community is not determined by its slogans, its branding, or even its ideals in the abstract, but by the boundaries it can enforce consistently under pressure. Rules only become real when they survive repetition, edge cases, bad-faith testing, and moderator exhaustion. The purpose of stickler-bot is to help a subreddit carry that burden with more consistency and less drift, not by replacing human judgment, but by giving it a more stable procedural form.
The project operates specifically on Reddit, where governance is unusually local. Each subreddit is run by volunteer moderators who define their own standards within Reddit's platform-wide rules. They review reported content, remove posts or comments that violate local norms, and communicate those decisions through removal reasons and modmail. This is not a minor implementation detail. It means moderation on Reddit is not best understood as generic platform enforcement, but as a form of community self-government. A useful moderation system in that environment cannot simply apply an external rulebook. It has to work from the community's own language, priorities, and tolerances. stickler-bot is built around that constraint.
What makes stickler-bot distinctive is that it treats removal reasons as operational governance text. It does not begin from the assumption that an AI model should decide what a community ought to value. Instead, it asks the model to classify content against the rules moderators have already written. In practice, that means the tool uses OpenAI as a constrained classifier rather than as an autonomous moderator. It sends rule text together with contribution context, asks for a structured decision, validates the result, and then applies explicit enforcement gates before any action is taken.
Reddit Moderation As Social Infrastructure
Reddit occupies a strange position in the broader internet. It is one of the largest public discussion platforms in the world, yet much of its practical order is produced by small groups of unpaid moderators governing semi-autonomous communities. This creates both an enormous strength and an enormous strain. The strength is that subreddits can develop real cultural specificity. The strain is that this specificity must be defended manually, every day, against spam, harassment, ideological drift, and endless low-effort provocation.
stickler-bot should be understood in that context. It is not merely a convenience tool for clearing queues faster. It is a response to the fact that online governance now happens at a speed and scale that routinely exceeds what volunteer teams can process without automation. The interval between a violation appearing and a moderator acting on it is not politically neutral. It shapes the lived norm of the space. What goes unchallenged, even briefly, begins to feel tolerated. What is enforced clearly and predictably becomes legible as a boundary.
How The System Works
Operationally, stickler-bot is event-driven. It evaluates new posts as they are submitted and reported comments as they are escalated by users or moderators. It does not sweep an entire subreddit indiscriminately. Each contribution is normalized into a moderation payload and screened for obvious exclusions such as already removed items, distinguished moderator content, and content authored by the bot itself.
For posts, the payload centers on the title, body, and URL, with optional visual inputs where usable image assets exist. For reported comments, the system reconstructs thread context by walking the parent chain and pulling top-level post context so that replies can be interpreted as part of a conversation rather than as isolated fragments. Usernames in thread context are mapped into anonymized participant labels before they are sent for inference.
The model response itself is tightly constrained. stickler-bot asks for no rule violation or a single best-matching removal reason, alongside a short justification, a confidence score, and a human-review flag. A contribution is only auto-enforced when there is a concrete rule match, confidence meets a configured threshold, and the model does not request human review. If any of those conditions fail, the bot routes the case to internal modmail triage for moderator review.
The system's boundaries are written directly into code. Reddit-hosted video uploads are skipped unless they contain substantial body text. Candidate image URLs are bounded and retried conservatively when media fetches fail. Duplicate in-flight moderation is suppressed, and triage notifications are rate-limited.
Alpha Pilot Communities
The first alpha pilots are r/antinatalism and r/VeganDating. These are not identical communities, but both are spaces whose identity depends on active boundary maintenance rather than passive audience growth. They are strong test cases for a system whose central claim is that clear rules can be made more durable when translated into disciplined procedure.
For r/antinatalism, the need for disciplined moderation is easy to infer. After the May 17, 2025 Palm Springs fertility-clinic bombing, investigators said they were examining antinatalist views in connection with the suspect's motive, and Reddit separately banned r/Efilism under its self-harm rules in the same period. Those events do not define antinatalism as a philosophy, but they do create a harsh public environment for any large antinatalist forum.
The logic for r/VeganDating is different but equally instructive. Vegan-only online communities depend on boundary integrity. Dating communities sharpen that requirement further, because they are vulnerable to misrepresentation, harassment, and low-effort derailment. For a niche subreddit like r/VeganDating, the primary risk is often the speed with which trust erodes when rule enforcement feels inconsistent.
Boundaries, Potential, And Responsibility
For all its usefulness, stickler-bot has hard limits. It can classify text and selected media against predefined rules, but it cannot fully understand lived history, local subcultural memory, or the moral texture of a situation in the way experienced moderators can. It is strong when the rule surface is clear and much less trustworthy when the rule itself is vague, overlapping, or internally conflicted.
That is why stickler-bot is best understood not as an attempt to automate authority, but as an attempt to discipline it. It asks whether a community can write rules clear enough to be applied consistently, whether a machine can help carry that burden without pretending to wisdom it does not possess, and whether the internet can sustain spaces whose standards are not dissolved by speed.
Links