What We Do

Learn and explore how Guardiobot protects Discord and Roblox communities.

Our Mission

Guardiobot believes everyone deserves a safe online environment. Our mission is to build security infrastructure for the platforms where younger audiences spend the most time — and where they are most vulnerable.

We focus on Discord and Roblox because the combination of their scale and younger user base creates a disproportionate risk. Platforms struggle with maintaining consistent and effective moderation at scale.

How We Protect

Protection runs in three layers: automated rule-based detection that acts in under 50ms, adaptive learning that tailors itself to each community's normal behaviour, and cross-server enforcement that ensures a ban in one community follows a bad actor across the entire network.

Rule-Based Detection

Every message passes through a multi-method content detection engine — no external AI dependency, no per-message API cost, no single point of failure.

  • Keyword matching — case-sensitive or insensitive, whole-word or substring
  • Regex pattern matching against configured rulesets
  • Content fingerprinting of known harmful material
  • URL pattern matching against domain and path blacklists
  • Each rule carries a severity score; the score determines the automatic response

Detection completes in under 50ms from message received to action taken.

Adaptive Learning

Every community has different norms. Guardiobot builds a behavioural profile per server and flags content that meaningfully deviates from it — reducing false positives without sacrificing coverage.

  • Tracks message length, emoji rate, link frequency, caps percentage, and mention patterns
  • Sampling intensity is configurable: passive, active, or aggressive per server
  • Low-severity detections feed directly into the learning system
  • Profiles are updated on a regular schedule — not in real time — to prevent manipulation

Message samples are automatically deleted after 30 days.

Cross-Server Enforcement

A ban issued in one server propagates automatically across every server in the Guardiobot network — bad actors cannot escape by server-hopping.

  • Every member join triggers a live check for active cross-network punishments
  • Bans from report actions and automated detections are both propagated
  • Network participation is opt-in per server, with configurable trust thresholds
  • Low-confidence bans go to a manual review queue before enforcement

Network settings are managed exclusively by Guardiobot staff — server owners cannot modify network participation directly.

What We Detect

Detection rules cover the categories most likely to cause real harm to real people — with specific focus on the threats most common in communities where younger audiences are present.

Illegal Content

  • CSAM indicators, child exploitation attempts, and illegal goods and services distribution
  • Zero tolerance — any detection triggers immediate action
  • Evidence retained for the legally required period for reporting purposes, then automatically purged
  • Action: permanent ban + flagged for authority reporting via NCMEC CyberTipline

NSFW Content

  • Explicit imagery and language, suggestive content, and inappropriate requests
  • Image safety scanning via Azure AI Content Safety
  • NSFW reports are routed exclusively to volunteers with verified age and signed legal clearance
  • Action: message deleted + configurable response per server settings

Violence & Harm

  • Direct threats, self-harm encouragement, violent imagery descriptions, and suicide encouragement
  • Grooming escalation patterns: requests for personal information, isolation attempts, and gradual boundary testing
  • Action: message deleted + warn or ban depending on severity and repeat history

Hate Speech

  • Slurs, discriminatory language, and targeted harassment based on protected characteristics
  • Detection works across character encodings and common obfuscation attempts
  • Per-community tolerance is configurable — gaming servers and other communities with different norms are handled appropriately
  • Action: message deleted + escalating response for repeat offences

Spam & Scams

  • Phishing links checked against URL reputation and safety databases
  • Mass mention spam, cryptocurrency scams, and coordinated raid attempts
  • Rate limiting across joins, messages, and moderation events
  • Action: message deleted + kick or ban based on severity and repeat history

User Risk Factors

  • VPN and proxy detection via multiple third-party providers
  • Alt account detection using account age, join patterns, username similarity, and behaviour signals
  • Roblox user safety score cross-referenced via external API integrations
  • User risk classification checked on every server join — high-risk users can be blocked automatically

Why Trust Us?

Transparent Operations

Clear documentation of how our systems work, what data we collect, and how long we keep it. No vague policies.

Read our Privacy Policy — full details on exactly what is stored and for how long.

GDPR & CCPA Compliant

Privacy compliance is built into the architecture, not added afterwards.

  • Sensitive data is cryptographically hashed at write time — never stored in plaintext
  • Message samples purged after 30 days, audit logs after 90 days
  • Right to erasure and right to data portability both implemented
  • Deleted users retain only the minimum necessary data for ongoing ban enforcement

Appeal System

All bans are appealable by default. Appeals are reviewed by senior volunteers and tracked in the audit log.

  • Every moderation action records whether the action is appealable
  • Users are notified of their appeal status via direct message
  • Senior volunteers can restrict or restore appeal rights on a per-case basis

Submit and track appeals through your dashboard.

Human Review Process

Automation handles the clear-cut cases. Humans handle everything else. All volunteers are verified and credentialled.

  • All volunteers are publicly verifiable at guardiobot.org/verify/{id} — no login required
  • Volunteer credentials use industry-standard password hashing
  • Credentials expire annually and must be actively renewed
  • Entry-level volunteers cannot see usernames or reporter identities — redacted at the API layer, not just the UI
  • NSFW reports are restricted to volunteers with verified age and signed legal documentation

High Reliability

Automated health monitoring, connection pooling, and background maintenance keep protection running without gaps.

  • Health checks monitor all core services continuously
  • Database connection pooling with automatic recovery on failure
  • Background tasks handle housekeeping automatically — expired punishments, stale sessions, rate limit cleanup

View live operational status on our status page.

How It Works

Guardiobot integrates with your Discord server in minutes. Once active, every message passes through the detection pipeline automatically — no configuration required to get started, and deep customisation available when you need it.

1

Easy Integration

Add Guardiobot to your Discord server. Channel assignment, verification settings, and feature toggles are all managed through your server owner dashboard — no code, no config files.

2

Automatic Protection

Every message passes through a multi-stage detection pipeline in under 50ms:

  1. Spam Detection — rapid duplicate message checking with no database dependency. Confirmed spam is removed and the user is temporarily muted.
  2. Content Detection Engine — rule-based matching using keywords, regex patterns, content fingerprints, and URL blocklists against your server's configured ruleset.
  3. Adaptive Learning — a community-specific behavioural profile updated regularly. Learns the norms of your server and flags content that deviates meaningfully from them.
  4. External Risk Checks — cross-references third-party safety APIs for Roblox user risk scores, condo server detection, and VPN/proxy detection.

All detections are logged with context for audit and review purposes.

3

Human Review

Each detection carries a severity score that determines the response:

  • High severity — immediate automatic action. No volunteer required.
  • Medium severity — user warned automatically, detection escalated to the volunteer review queue for a final decision.
  • Low severity — logged for pattern analysis only. No action taken. Data feeds the adaptive learning system.

All decisions are recorded in immutable audit logs including action type, target, reviewer, timestamp, and reason. Logs are retained for a minimum of 90 days.

4

Continuous Improvement

The adaptive learning system refines each community's profile on a regular schedule, tracking message length norms, emoji usage, link frequency, capitalisation patterns, and mention behaviour. Content that deviates significantly from a server's established baseline is flagged for review.

Sampling intensity is configurable per server:

  • Passive — light sampling, suitable for low-traffic or established servers
  • Active — standard sampling, suitable for most servers
  • Aggressive — intensive sampling, suitable for high-risk or newly created servers

Message samples are automatically deleted after 30 days.

The Difference We Make

Most Discord servers rely on a handful of volunteer moderators, manual rule enforcement, and reactive bans. Guardiobot replaces that with a system that never sleeps, never misses a pattern, and escalates the hard cases to humans who can make the right call.

Without Guardiobot

  • Manual moderation requires constant moderator availability
  • Threats go undetected until someone reports or complains
  • Inconsistent enforcement — different mods, different standards
  • Bad actors can server-hop to escape bans
  • No visibility into what threats your community is facing
  • Reactive — you respond after harm has already occurred

With Guardiobot

  • Automated detection handles routine threats in under 50ms
  • Rule-based and adaptive systems catch threats before reports come in
  • Consistent enforcement — same rules applied to every message, every time
  • Cross-server enforcement — bans follow bad actors across the entire network
  • Full audit log and dashboard statistics for every action taken
  • Proactive — threats stopped before they reach your members
  • Moderators focus on complex cases that actually need human judgement

Our Story

"He started playing on Roblox… When he was 12, he was targeted by an adult sex predator… the man encouraged him to move their conversations to Discord… On Discord, the man increasingly demanded explicit photographs and videos and threatened the child…" — NBC New York, September 2025, describing a lawsuit against Roblox and Discord following a teen's suicide. View Full Story

There are many other stories like this. It is heartbreaking to see some of the most popular platforms for children do little to combat this — and worse, that young people are often left without the tools or knowledge to protect themselves.

We decided we couldn't watch this happen. So we built Guardiobot — a security service for both Discord and Roblox designed to detect and prevent abusive behaviour automatically, with a human review layer for everything that needs a second opinion.

This exact scenario shaped specific features in Guardiobot:

  • Cross-server enforcement — when a bad actor is banned in one server, that ban propagates automatically across every server in our network. Joining any protected server triggers an immediate check for active punishments and alternate accounts.
  • Grooming pattern detection — detection rules target escalation tactics: requests to move to DMs or private channels, requests for personal information, isolation attempts, and gradual boundary testing.
  • Platform bridging — Guardiobot monitors both Discord and Roblox. Predators frequently use Roblox to identify targets before moving to Discord for exploitation — exactly the pattern described in the NBC News case.
  • Our goal is to reach communities and developers globally. Every person online has the right to a safe experience, and we are building the infrastructure to make that possible.

The Guardiobot Team

Ready to Protect Your Community?

Join Guardiobot and help us build a future with safer platforms for everyone.