Fact-checked by Grok 2 weeks ago

Scunthorpe problem

The Scunthorpe problem denotes the unintended obstruction of benign online communications, such as emails, forum posts, or account registrations, by automated filters employing rudimentary substring matching that flags sequences resembling profanities irrespective of contextual meaning. This computational limitation, often termed the "clbuttic mistake" in reference to erroneous autocorrections like "classic" becoming "cl*ssic," underscores the pitfalls of context-free algorithmic censorship in processing natural language. The eponymous incident occurred in 1996 when America Online's profanity filter barred residents of Scunthorpe, a town in Lincolnshire, England, from creating accounts because its name embeds the substring "cunt." Similar blocks have affected other innocuous terms, including place names like Penistone and words such as "therapist" parsed as "the rapist," illustrating persistent challenges in balancing overzealous filtering against effective moderation. Despite advances in machine learning for contextual analysis, the problem endures in various digital platforms, prompting developers to adopt hybrid approaches combining whitelists, n-gram models, and human oversight to mitigate false positives.

Definition and Technical Basis

Core Mechanism and Causes

The Scunthorpe problem fundamentally stems from filters employing rudimentary substring matching algorithms that detect and block sequences of characters matching predefined profane terms, irrespective of word boundaries or semantic context. These systems scan text inputs—such as usernames, emails, or search queries—for exact or partial string matches against a blacklist of obscenities, triggering automatic rejection or without further analysis. For instance, the town name "" is flagged due to embedding the substring "cunt," a common profane term, even though the full word is innocuous. This mechanism prioritizes pattern recognition over linguistic nuance, leading to false positives where benign content is erroneously censored. The primary causes trace to design trade-offs favoring simplicity, speed, and reliability in detection over precision. Implementing context-aware filtering—such as requiring whitespace delimiters around matches or integrating () for intent evaluation—demands significantly higher computational resources and , which can introduce delays in high-volume systems like web registrations or . In the mid-1990s, when the issue gained prominence, hardware limitations and nascent software capabilities made advanced parsing impractical for widespread deployment, rendering basic regex-like substring searches the default choice for efficiency. Developers often calibrate filters conservatively to err on the side of blocking potential violations, minimizing false negatives (missed ) at the expense of overreach, as unfiltered explicit content carries higher perceived risks for platforms. Variations in profanity lists exacerbate the issue, as filters must account for obfuscations like leetspeak or misspellings (e.g., "" versus "d1ck"), prompting broader rules that inadvertently capture unrelated terms. While modern advancements in offer potential mitigations through probabilistic context models, many legacy and cost-sensitive systems persist with methods due to their low overhead and ease of maintenance. This persistence reflects a causal prioritization of scalable, rule-based enforcement over resource-intensive alternatives, perpetuating the problem in domains from services to .

Distinction from Intentional Censorship

The fundamentally differs from intentional in that it stems from the mechanical limitations of automated content filters, which employ crude pattern-matching algorithms to detect prohibited s without regard for linguistic context or semantic meaning. These systems, intended to block explicit or , inadvertently flag benign terms containing offensive sequences—such as "" due to its ""—resulting in false positives that affect legitimate users and content. This accidental overreach arises from the filters' reliance on heuristic rules rather than advanced , prioritizing efficiency over precision in high-volume environments like services or forums. In contrast, intentional involves deliberate, policy-driven suppression of information, often targeting viewpoints, ideologies, or specific demographics through human-curated rules, legal mandates, or platform guidelines aimed at enforcing moral, political, or corporate standards. Examples include blocks on media or corporate removals of "hate speech" based on interpretive judgments, where the goal is proactive control rather than incidental . The Scunthorpe problem lacks this volitional element; it represents a technical artifact of under-engineered safeguards, not a strategic effort to curtail expression, as evidenced by historical incidents where filter designers acknowledged the errors as unintended bugs rather than features. This distinction underscores broader challenges in automated moderation: while intentional censorship invites scrutiny for bias or overreach in human , the Scunthorpe problem highlights the inherent brittleness of substring-based detection, which can propagate errors at scale without nuanced exceptions for proper nouns, compounds, or regional variations. Empirical cases, such as early 1990s AOL filters blocking UK place names like Penistone or Clitheroe, demonstrate how such systems fail systematically on edge cases, prompting iterative fixes like whitelists or regex refinements rather than ideological justifications. Critics of expansive filtering argue that conflating these phenomena risks eroding trust in technical solutions, as false positives erode utility without advancing deliberate protective aims.

Historical Development

Initial Discovery in 1996

The problem gained its name from a widely reported incident in 1996, when America Online (AOL), a leading U.S.-based , deployed an automated that blocked residents of , a town in , , from registering accounts or accessing certain services. The operated on basic matching, flagging any input containing sequences like ""—a profane term embedded within ""—without regard for contextual legitimacy or proper nouns. This resulted in widespread frustration among locals attempting to sign up, as the system rejected usernames, addresses, or messages incorporating the town's name, effectively isolating them from AOL's growing network during the early . The issue extended beyond Scunthorpe to other British locales with similar etymological vulnerabilities, such as Penistone (containing "penis") in South Yorkshire and Middlewich (with "dick" in "Middlewich"), where residents encountered identical blocks when providing their locations during registration or communication. AOL's filter, intended to curb explicit content in chat rooms and emails amid rising concerns over online indecency, relied on rigid keyword lists rather than linguistic analysis, exemplifying the pitfalls of overzealous, context-blind automation in nascent content moderation systems. Complaints from affected users prompted media attention and internal adjustments by AOL, including temporary whitelisting of specific terms, marking one of the earliest documented cases of algorithmic overreach in digital filtering. This 1996 episode highlighted the tension between aggressive detection and practical , as AOL's expansion into international markets exposed the limitations of U.S.-centric word lists applied globally. Reports from the time, echoed in later analyses, noted that the blocks disrupted routine online activities for hundreds of residents, underscoring how simplistic regular-expression-based filters could inadvertently censor innocuous text. The incident spurred initial discussions on the need for more sophisticated approaches, though AOL's exact resolution details—likely involving manual overrides—remained proprietary, with the problem persisting in varying forms across providers.

Prevalent Cases in Early Internet Era

In 1996, America Online () implemented a profanity filter for user registrations upon entering the market, which inadvertently blocked account creation for residents of , , due to the town's name containing the substring "". The filter similarly affected users from nearby locales including , ("penis"), , ("clit"), Lightwater, , and the county of ("sex"), preventing them from signing up as the system flagged these as obscene without contextual analysis. temporarily altered "" to "Sconthope" in its system as a while developing a fix, as confirmed by an spokesperson to the Scunthorpe Evening Telegraph. This incident highlighted early reliance on simplistic substring-matching algorithms in and services, which lacked mechanisms for proper nouns or geographic exceptions. Similar blocks occurred with individual names, such as rejecting "Douglas Kuntz" for containing "kunt", a variant of a , underscoring the filter's overreach on non-contextual matches. By the late , as adoption grew, such filters proliferated in nascent services, leading to widespread user complaints but no standardized mitigations until contextual improvements emerged later. These cases demonstrated the causal limitations of rule-based systems, which prioritized crude detection over linguistic nuance, often exacerbating issues in regions with etymologically unrelated but phonetically sensitive place names.

Manifestations and Examples

Email and Registration Blocks

In 1996, America Online's () profanity filter blocked users from , , from creating email accounts during the registration process, as the town's name contained the substring matching a vulgar term. This incident affected multiple British towns with similar innocuous names, such as and , rendering residents unable to complete sign-ups due to substring-based detection without contextual analysis. Email delivery has also been disrupted by such filters; for instance, in the early 2000s, General Hospital's newly implemented system halted all outgoing messages on its activation day because the location reference triggered blocks on substrings. Similar issues persist with personal addresses incorporating surnames like Cockburn or , where filters either reject entire messages or redact portions containing apparent profanities, leading to garbled or undelivered communications. Registration blocks extend beyond geographic names to affect individuals with non-offensive but substring-vulnerable identifiers, such as those including "ass" in "Glasgow" variants or "tit" in place names like Clitheroe, preventing account creation on platforms reliant on basic keyword scanning. These cases highlight how rigid, non-contextual algorithms prioritize substring matches over semantic intent, often requiring manual overrides or whitelist exceptions to resolve, though such interventions expose users to delays or privacy risks during verification.

Search Engine and Domain Restrictions

Search engines mitigate potentially harmful content through safe search filters and detection, but these mechanisms can inadvertently restrict access to legitimate queries containing substrings that match blocked terms, exemplifying the Scunthorpe problem. Such filters scan for obscene patterns within search terms or result snippets, leading to demotion, blurring, or outright suppression of results for innocuous topics like geographic locations or product names. For instance, heightened post-incident sensitivities have prompted temporary blocks; in February 2018, following the Parkland shooting, Google's shopping search filtered out listings for "glue guns," "" albums, and " due to overbroad "guns"-related restrictions, affecting unrelated commercial intent. Domain restrictions arise during registration when automated systems at registrars or oversight bodies like reject names based on substring matches to lists, preventing legitimate acquisitions. These filters aim to curb overtly offensive registrations but often ensnare harmless combinations embedded with profane segments. A documented case occurred in April 1998, when entrepreneur Jeff Gold sought to register "shitakemushrooms.com" to promote mushrooms, only for 's filter to block it owing to the "shit" substring in "shitake." Similar domain blocks have impacted geographic or descriptive names; for example, registrations for domains referencing UK locales like "cockermouth.co.uk"—named after the town of —have triggered filters detecting "cock," complicating local business or informational sites. These incidents highlight the tension between proactive moderation and usability, as registrars prioritize rapid automated checks over nuanced review, resulting in appeals processes or alternative naming for affected users.

Content Platform Incidents

In April 2016, 's automated filter prevented users from promoting posts containing the word "," due to the matching a vulgar term, thereby blocking advertisements for local events, businesses, and bands associated with the town. Affected parties, including residents and promoters like the band October Drift, could post content organically but encountered restrictions on paid boosts, requiring manual appeals to moderators for resolution, often after significant delays that impacted timely marketing efforts. Similar false positives have occurred on other platforms, where filters flag innocuous references to place names or terms with embedded profane substrings during post submissions or account verifications. For instance, users posting about towns like (containing "penis") or surnames such as "Butts" have reported blocks on content sharing or profile setups, as automated systems prioritize substring matches over contextual intent, leading to temporary suspensions or of legitimate discussions. These incidents highlight the challenges of scaling via keyword-based algorithms on high-volume platforms, where overzealous filtering disrupts without distinguishing benign usage.

Specialized Contexts Like Gaming and Media

In multiplayer online , profanity filters applied to in-game chat systems frequently produce false positives by censoring non-offensive words containing substrings of prohibited terms, hindering communication. For example, the common profane substring "" triggers blocks on legitimate terms such as "," "," and "," requiring developers to maintain exception lists for contextual overrides. Similarly, words like "" may be flagged in phrases such as "hell of a ," despite lacking profane intent, which disrupts natural in virtual environments including . Even sophisticated machine learning-based toxicity detectors deployed in gaming chats exhibit elevated false-positive rates, often exceeding 10-20% for edge cases involving compound words or slang variants, as evaluated in comparative studies of real-time moderation tools. These issues are exacerbated in cross-lingual or global player bases, where filters misinterpret culturally neutral terms, leading to fragmented interactions and player frustration reported across titles like Warframe and Dead by Daylight. In media platforms supporting interactive content, such as streaming services with live chat or user forums for gaming broadcasts, analogous filtering challenges arise during automated moderation of comments and captions. Basic keyword-based systems on these platforms censor substrings in viewer inputs, occasionally blocking references to game mechanics or titles (e.g., "ass" in "assault mode" descriptions), though advanced contextual analysis mitigates some occurrences compared to early 2010s implementations. Traditional broadcast media, reliant on manual review rather than real-time algorithms, encounters the problem less acutely, but digital distribution tools for video-on-demand have adopted similar filters prone to overreach in subtitle or metadata processing.

Mitigation Strategies

Basic Filtering Adjustments

Basic filtering adjustments for the Scunthorpe problem primarily involve rule-based modifications to detection algorithms, focusing on matching limitations without requiring advanced contextual analysis. These methods aim to distinguish profane terms from embedded occurrences in innocuous words by enforcing stricter rules, such as requiring profane strings to align with whole-word boundaries. For instance, regular expressions can incorporate boundary anchors like \b to ensure matches occur only at the start and end of words, delimited by non-alphanumeric characters such as spaces or punctuation, thereby preventing false positives in terms like "" where "" appears as a rather than an isolated word. Another foundational adjustment is the implementation of whitelisting, where a predefined list of verified benign terms containing potential profane substrings—such as place names (e.g., , ) or common words (e.g., "assassin")—is exempted from filtering. This approach, often maintained as a simple database or set checked prior to flagging, allows rapid overrides for known edge cases identified through user reports or testing, reducing overblocking in applications like gateways or posts. Whitelists are dynamically updated based on empirical feedback, with studies showing they can resolve up to 80% of recurring false positives in basic setups when combined with boundary checks. Additional rudimentary tweaks include enabling case-sensitive matching where feasible, to differentiate capitalized proper nouns from lowercase , and incorporating minimum length thresholds for flagged terms to avoid partial matches in longer innocent phrases. These adjustments, while effective for static vocabularies, remain vulnerable to novel combinations or non-English languages, necessitating periodic manual reviews; for example, early filters in were retrofitted with such boundaries post-Scunthorpe complaints, halving erroneous blocks within weeks. However, over-reliance on whitelists can lead to burdens, as exhaustive lists grow unwieldy beyond thousands of entries, highlighting their suitability only for low-variability environments like corporate intranets.

Advanced Contextual and Machine Learning Solutions

Machine learning-based profanity detection systems address the limitations of substring matching by incorporating contextual analysis through techniques, such as word embeddings and neural networks, to evaluate word usage within surrounding text. These models are trained on annotated datasets distinguishing offensive intent from benign occurrences, thereby reducing false positives associated with the Scunthorpe problem. For example, architectures like (LSTM) networks combined with word embeddings capture sequential dependencies and semantic nuances, allowing differentiation between innocuous proper nouns or compounds and deliberate . Transformer-based models, including BERT variants, further enhance context awareness by generating contextualized representations of tokens, enabling classifiers to assess based on sentence-level semantics rather than isolated strings. A November 2024 evaluation demonstrated that large language models (LLMs) integrated into pipelines achieve higher in flagging contextual while minimizing erroneous blocks of non-offensive content, such as geographic names or technical terms. Similarly, libraries like profanity-check employ over character n-grams trained via , which learns probabilistic patterns from data to avoid over-matching substrings, reportedly resolving Scunthorpe-like issues without explicit whitelists. Hybrid approaches combine rule-based heuristics with ML for efficiency, where initial keyword triggers prompt deeper contextual scrutiny via or to confirm offensiveness. Context-aware filters, as implemented in tools like Glin-Profanity, incorporate surrounding discourse and cultural factors to lower false positive rates in , adapting dynamically through retraining on feedback loops. Despite these advances, such systems demand substantial computational resources and diverse training data to mitigate biases, with ongoing refinements focusing on multilingual and domain-specific adaptations for platforms like and .

Implications and Debates

Technical and Usability Challenges

The Scunthorpe problem exemplifies the core technical limitation of substring-matching algorithms in profanity filters, which detect offensive content by scanning for prohibited sequences without regard for linguistic context or semantics, resulting in frequent false positives for innocuous terms like place names or compound words. This approach, common in early filters since the , prioritizes exhaustive pattern coverage over precision, as evidenced by AOL's 1996 blocking of residents' registrations due to the town's name containing a profane . Implementing exceptions via whitelists mitigates some cases but introduces maintenance burdens and potential exploits, such as adversaries embedding whitelisted terms to bypass detection. Advanced models, including those using word embeddings and recurrent neural networks like LSTM, attempt to incorporate context but still struggle with edge cases involving proper nouns, sarcasm, or evolving , often requiring vast training data that may embed cultural biases or outdated lists. Resource-intensive real-time processing in high-volume environments, such as or gaming platforms, exacerbates these issues, where even low false-positive rates (e.g., 1-5% in large datasets) can generate thousands of erroneous blocks daily, straining computational . Trade-offs between further complicate design, as overly permissive filters risk unmoderated abuse, while stringent ones amplify errors in substring-heavy languages like English. From a standpoint, these technical flaws degrade by interrupting legitimate interactions, such as search queries for "Scunthorpe United" yielding no results or systems rejecting surnames like "Cockburn," forcing workarounds like phonetic misspellings that undermine accessibility and trust in automated systems. In contexts, such as healthcare or , blocks on medical terms (e.g., "pubic bone" in paleontology discussions) halt workflows, necessitating manual overrides that are infeasible at scale and increase administrative overhead. End-users, particularly in global applications, encounter inconsistent filtering across dialects or scripts, eroding platform reliability and prompting abandonment of features like or registration, as seen in incidents where filters flagged everyday words like "assassin." Overall, without hybrid human-AI validation, these challenges perpetuate a cycle of over-correction that prioritizes perceived over functional .

Criticisms of Overreach in Moderation Policies

Critics contend that moderation policies employing substring-based profanity filters, as exemplified by the Scunthorpe problem, overreach by indiscriminately suppressing non-offensive content, prioritizing superficial over semantic analysis and thereby eroding platform functionality. This approach has drawn rebuke for fostering arbitrary , where benign terms or proper nouns trigger blocks, compelling users to alter legitimate communications or abandon services altogether. A 2018 analysis in the Georgetown Law Journal describes such practices as rushed and inconsistent, subjecting users to opaque that lacks equivalents found in regulation. The overreach manifests in real-world disruptions, such as the blocking of geographic names containing profanity-like substrings, which has impeded access to essential services for affected communities since the problem's in the mid-1990s. Detractors, including technologists, argue that persisting with these filters despite known limitations reflects a policy toward excessive caution, generating false positives that undermine and encourage circumvention tactics potentially evading genuine safeguards. For example, broad rules can flag embedded substrings in compound words like those containing "," as noted in discussions of custom filter development challenges. Broader debates highlight how this overreach exacerbates tensions between safety imperatives and expressive freedoms, with organizations like the warning that flawed automated , including profanity over-filtering, invites external regulatory pressures that compound errors rather than resolving them. In a 2025 policy commentary, centralized systems were flagged for inherent risks of systematic over-censorship, urging shifts toward context-aware alternatives to avert disproportionate harms to innocent users. Such criticisms underscore demands for frameworks that balance empirical with verifiable precision, avoiding the causal pitfalls of rule-based overgeneralization.

References

  1. [1]
    Scunthorpe Problem – Software Filters Misunderstand Words
    Aug 6, 2010 · The problem originated after AOL's profanity filters prevented residents of Scunthorpe from creating accounts because the town's name contains ...Missing: explanation | Show results with:explanation
  2. [2]
    What is Scunthorpe Problem | Glossary - CyberGhost VPN
    The genesis of the Scunthorpe problem can be traced back to the increasing use of automated content filtering systems, particularly prevalent in online ...Missing: explanation | Show results with:explanation<|separator|>
  3. [3]
    What is the “Scunthorpe problem”? - Engage Web
    May 10, 2021 · The town has now lent its name to the problem of computers treating something perfectly innocent as something offensive.Missing: origin | Show results with:origin
  4. [4]
    What Made Scunthorpe Famous - Tedium
    Jul 26, 2016 · And this, friends, was how the Scunthorpe Problem, the tendency for algorithms to not make room for edge cases with curse words, was discovered.Missing: core mechanism
  5. [5]
    The Scunthorpe Problem - When Profanity Filters Go Too Far - Sporcle
    Apr 3, 2017 · It turns out the profanity filters many websites use can actually cause more problems than one would think. Welcome to the Scunthorpe ...
  6. [6]
    The Scunthorpe Problem - Spam Resource
    Jan 19, 2017 · Have you heard of the Scunthorpe Problem? It highlights the side effects associated with trying to filter out profanity.
  7. [7]
    The Scunthorpe Problem, And Why AI Is Not A Silver Bullet For ...
    Aug 31, 2018 · The Scunthorpe problem is the blocking of e-mails, forum posts or search results by a spam filter or search engine because their text contains ...
  8. [8]
    The Scunthorpe effect rides again - Language Log
    May 28, 2020 · This sort of accidental filtering has gained a name in computer science: the Scunthorpe problem, so-called because of the Lincolnshire town's regular issues ...
  9. [9]
    How spam filters dictated Canadian magazine's fate - BBC News
    Mar 29, 2010 · The blocking of harmless websites or e-mails by trigger-happy filters is nothing new. In 1996, residents in the British town of Scunthorpe were ...
  10. [10]
    Scunthorpe Problem | Encyclopedia MDPI
    Oct 14, 2022 · 1. Origin and History. The problem was named after an incident in 1996 in which AOL's profanity filter prevented residents of the town of ...Missing: explanation | Show results with:explanation
  11. [11]
    The 10 best words the internet has given English - The Guardian
    Apr 17, 2013 · The effect was labelled in honour of the town in 1996, when AOL temporarily prevented any Scunthorpe residents from creating user accounts ...
  12. [12]
    Dr. Libshitz Can't Get an E-Mail Address - ABC News
    Dec 9, 2008 · That's when, legend has it, AOL's automated filtering system kept Scunthorpians from registering for accounts because of a certain sequence of ...
  13. [13]
    The Scunthorpe Problem - John M Jennings
    Sep 24, 2018 · The residents of Scunthorpe have a problem: their town name often gets blocked by profanity filters designed to weed out obscene words.Missing: explanation origin
  14. [14]
    DON'T BE A SCUNTHORPE - Dave Trott's Blog
    Nov 16, 2020 · A few years back, Scunthorpe Hospital installed a new computer system. On the first day it was switched on, the staff began using email as usual ...
  15. [15]
    What are some interesting examples of the Scunthorpe Problem?
    Oct 14, 2018 · Scunthorpe hospital got a new e-mail system with a brand new profanity filter. The morning it was turned on, it didn't take long before people started becoming ...What is Scunthorpe problem? - QuoraDo other languages have Scunthorpe problem? For example, in ...More results from www.quora.com
  16. [16]
    The Scunthorpe Problem- messing with your text since 1996 - Medium
    Sep 28, 2021 · Scunthorpe is a perfectly ordinary industrial town in the North of England, known for that time in 1996 when no-one living there could ...Missing: credible | Show results with:credible
  17. [17]
    Facebook bans users from promoting posts with the word ...
    Apr 4, 2016 · Facebook bans its users from promoting posts with the word Scunthorpe in it. The company's profanity algorithm appears to be banning all posts ...
  18. [18]
    Innocent people with dirty-sounding last names face the ... - Quartz
    The name “Scunthorpe problem,” for example, refers to the issues that residents of the English town have encountered with AOL and Google profanity filters ...
  19. [19]
    Life on the Internet Is Hard When Your Last Name is 'Butts' - VICE
    Aug 29, 2018 · When Blackie contacted AOL, he was informed that his registration had failed because of an automatic filtering system that scanned linguistic ...
  20. [20]
    Chat profanity filters - AccelByte Documentation
    Jun 23, 2025 · False Positive: Add list of words that contains the default profanity word and will not be censored when used (e.g., profanity word ass , false ...
  21. [21]
    The Benefits of Profanity Filtering for Virtual Experiences - PubNub
    Dec 15, 2022 · False positives refer to words that have been censored or filtered ... profanity and disruptive behavior from your in-game chat. When ...
  22. [22]
  23. [23]
    The Profanity Filter is a Bit Overzealous - General - Warframe Forums
    Apr 2, 2019 · Profanity filters are the bane of all online games as there really is no way to make them effective without catching just as many false positives as legitimate ...<|control11|><|separator|>
  24. [24]
    Why does a +18 game have a profanity filter? - BHVR
    Dec 23, 2022 · of course, the filter is too sensitive and censoring very innocent words - even in-game names. That should be fixed. But some kind of filter ...
  25. [25]
    Profanity Filters: Advanced Detection & Filtering Solutions
    Users were blocked from being able to create AOL accounts due to filters identifying the town name as profane.<|separator|>
  26. [26]
    Where The *!$% Are Profanity Filters In Games and Streaming ...
    Jul 22, 2021 · This extension filters profanity words in Netflix. It both censors the subtitles and mutes the audio on the appropriate sentence.
  27. [27]
    stephenhaunts/ProfanityDetector: This is a simple library for ... - GitHub
    ... ContainsProfanity, that will check for a profanity without the Scunthorpe checking. ... This method will attempt to remove any false positives from the string.Missing: matching | Show results with:matching<|control11|><|separator|>
  28. [28]
    Facebook FAIL: How To Properly Filter Scunthorpe - Cleanspeak
    Apr 6, 2016 · ... AOL and Google because the town's name contains an embedded 4-letter swear word. AOL first encountered this problem in 1996. It appears ...<|control11|><|separator|>
  29. [29]
    Method of Profanity Detection Using Word Embedding and LSTM
    Feb 10, 2021 · Therefore, in order to solve this problem, this study proposes a method of discriminating profanity using a deep learning model that can grasp ...Missing: awareness | Show results with:awareness
  30. [30]
    GLINCKER/glin-profanity - GitHub
    Glin-Profanity is a high-performance, cross-platform library built to detect, filter, and sanitize profane or harmful language in user-generated content.
  31. [31]
    Deep learning-based approaches for abusive content detection and ...
    This paper focuses on developing a multi-class classification model to identify different types of offensive language.
  32. [32]
    Scunthorpe Problem - Hacker News
    Apr 13, 2023 · The Scunthorpe problem is when names with hidden "bad" words are censored, like "Ishitani Furniture" or "clbuttic" for "classic" due to ...Missing: blocking | Show results with:blocking<|separator|>
  33. [33]
    The Scunthorpe problem: the limits of automatic categorization ...
    Dec 19, 2023 · The Scunthorpe problem is an example of the complex challenges faced by automated categorization filtering solutions.<|separator|>
  34. [34]
    Method of Profanity Detection Using Word Embedding and LSTM
    Feb 25, 2021 · This study proposes a method of discriminating profanity using a deep learning model that can grasp the meaning and context of words.
  35. [35]
    [PDF] Exploring Explanations For Profanity Detection - DiVA portal
    Sep 19, 2023 · the model classify it as being profanity. This technique works well with texts that include obvious hateful content such as swear words. If ...
  36. [36]
    Please fix the Scunthorpe problem with Bedrock's bad word filter
    Jul 5, 2022 · The Scunthorpe problem for those who don't know is a computer limitation that makes it so swearing and profanity filters false flag regular every day words.Filtering profanity from Online Leaderboard : r/gamedev - RedditPSA: Bitdefender Web Content Filtering False Positives this morningMore results from www.reddit.com
  37. [37]
    [PDF] Regulating Online Content Moderation - Georgetown Law
    Online content is moderated by private corporations and human/algorithmic methods, but practices are rushed and incoherent, and users face private censorship.
  38. [38]
    Bad Content Moderation Is Bad, And Government Interference Can ...
    Mar 31, 2023 · Government officials keep asking online services to remove or edit users' speech, raising the specter of unconstitutional interference with private platforms' ...
  39. [39]