Preface
Recently, numerous methods, techniques, and suggestions have been proposed to combat cheaters, providing short-term positive results. However, in the long run, cheaters become smarter and inevitably find ways to circumvent protections and continue their deceptive activities on the platform.
Why Common Methods Fail in the Long Term
Anti-Plagiarism and Timings
Code rapidly generated by Large Language Models (LLMs) often appears human-like due to standardized coding styles, predefined function templates, abbreviated variable names, etc. Additionally, appropriate delays between submissions can be easily simulated by monitoring the timing of other participants.
SMS Verification
Online services offering disposable phone numbers or additional virtual numbers provided by telecom operators render SMS verification largely ineffective. Anyone active on platforms like CodeForces or having even a minimal interest in bypassing verification knows about inexpensive online services providing temporary SMS activation or voice verification. Such services cost mere pennies, allowing number rentals for specific periods or the issuance of additional numbers linked to a primary SIM card (e.g., as offered by some Russian operators like MegaFon for a negligible daily fee). Thus, telephone-based verification is currently insufficient to deter determined cheaters.
Strict Bans
A user losing rating early on can simply create a new account, resetting the cycle. Those with established high ratings, familiar with key detection metrics, will easily evade bans by staying cautious.
Main Idea – Don't Ban, Conceal!
The idea of introducing a reporting system has been suggested previously, but I want to emphasize it further. Let's implement a delayed reporting system and likes from trusted users (elo rating >2000), influencing a suspect’s trust factor. Rather than banning a user for suspicious activity during a round—thus indicating exactly where and how they slipped—we should let them believe they've successfully evaded detection again. Meanwhile, their account receives a hidden mark, and all subsequent competitions involving that user are segregated from the main participant pool. The cheater continues to see their supposed rating and rank, although their scores are excluded from the main leaderboard calculations affecting honest competitors. Suspect participants are shifted into a "hidden pool," competing solely against each other.
Trusted Reporting System
Who can report?
Participants with a rating above 2000
Users significantly contributing to the platform’s community
Trusted participants and coordinators
Report Form:
- Suspicion of AI usage? [Yes/No]
- Key anomalies (timings, coding style, atypical solutions)
- Brief comment
Weighting of Reports:
Reporters are assigned a "weight" based on their current rating, registration date, community contribution, and past report accuracy. The higher these metrics, the greater the credibility assigned. Consequently, reports against highly trusted users don't significantly impact their reputation unless refuted by equally trusted members.
Likes and Reputation System
- Likes can be given by any participant registered at least seven days with a valid rating after round completion.
- Code evaluation: readability, clarity, presence, and comprehensiveness of comments.
Hidden Pool Mechanics
flowchart LR
A[Participant] --> B[Competition]
B --> C{Data Collection}
C --> D[Trusted Reports]
C --> E[Likes/Dislikes]
C --> F[Telemetry]
D & E & F --> G[Bayesian Model]
G --> H{TF ≥ 0.5?}
H -- Да --> I[Hidden Pool]
H -- Нет --> J[Main Pool]
I --> K[Visualization (visible only to cheater)]
J --> L[Real Results]
- Reports and historical data collected before the contest.
- Telemetry gathered during the contest: coding time, submission frequency, page navigation.
- Post-contest Bayesian probabilistic model makes the final decision regarding pool assignment.
UI/UX for "Hidden" vs. "Honest" Participants
Honest participants see their usual rankings without awareness of hidden "attributes" participants.
Hidden participants see fake ratings and rankings in the general leaderboard.
Moderators have access to dynamic trust-factor graphs, detailed telemetry logs and suspect profiles.
Automated Trust Factor Calculation
The trust factor increases or decreases after each contest based on reputation over a specified interval or is adjusted following manual moderator review.
What are your thoughts on this idea? Share your comments below.








