Detector24
Graphic Language Detection
ImageText Analysis, QR Codes and OCR

Graphic Language Detection

Detect offensive text, slurs, and hate speech embedded in images. OCR-powered content moderation for memes, screenshots, and user-generated content.

Accuracy
94%
Avg. Speed
15.0s
Per Request
$0.0150
API Name
graphic-language-detection

Bynn Graphic Language Detection

The Bynn Graphic Language Detection model identifies graphic, offensive, or abusive language visible in images. This model detects profanity, slurs, threats, and other harmful text content embedded in images.

The Challenge

Text-based content moderation catches offensive language in posts and comments, but bad actors have learned to embed hate speech, slurs, and threats within images. A meme, product design, or screenshot can carry the same harmful content while evading text filters entirely.

This visual text bypass has become a preferred tactic for spreading hate speech and harassment. Offensive slogans on t-shirts, threatening messages in image macros, slurs disguised as product names—all invisible to traditional moderation. Platforms need detection that reads and understands text within images, applying the same content policies regardless of how the text is delivered.

Model Overview

When provided with an image, the detector analyzes any visible text to identify graphic or offensive language. This includes profanity, vulgar slurs targeting race, gender, or ethnicity, threatening language, and other abusive content that may violate platform policies.

Achieving 94.0% accuracy, the model uses Bynn's Visual Language Model technology to read and understand text within images across multiple languages and text styles.

How It Works

The model performs comprehensive text analysis within images:

  • Text detection: Identifies all readable text in the image
  • Language analysis: Evaluates text content for offensive or graphic language
  • Context consideration: Analyzes the nature and intent of the language used
  • Multi-style support: Reads various text styles including stylized fonts, handwriting, and artistic text

Response Structure

The API returns a structured JSON response containing:

  • graphic_language: Boolean value (true/false) indicating whether graphic language was detected

Detection Categories

Graphic Language (detected as true)

The model identifies the following types of content:

  • Profanities or vulgar slurs targeting race, gender, or ethnicity
  • Harsh language meant to threaten, startle, or demean others
  • Obscenities and explicit vulgar terms
  • Speech that is abusive, rude, or intended to cause harm

No Graphic Language (detected as false)

Content classified as non-offensive includes:

  • Language that is courteous or non-offensive in tone
  • No use of vulgar or explicit terms
  • No threats, insults, or shocking verbal content
  • Racial, ethnic, or gender-based slurs are completely absent

Performance Metrics

Metric Value
Detection Accuracy 94.0%
Average Response Time 15,000ms
Max File Size 20MB
Supported Formats GIF, JPEG, JPG, PNG, WebP

Use Cases

  • E-commerce Moderation: Detect offensive text on product images, apparel designs, or user-uploaded photos
  • Social Media Content: Filter memes, images, and graphics containing hate speech or profanity
  • Advertising Compliance: Ensure ad creatives don't contain inappropriate language
  • User-Generated Content: Moderate images with embedded text in comments, posts, or uploads
  • Marketplace Listings: Prevent offensive product listings and descriptions
  • Brand Safety: Protect brand reputation by filtering harmful text content

Known Limitations

Important Considerations:

  • Text Legibility: Very small, blurry, or heavily stylized text may not be accurately read
  • Language Coverage: While multi-language capable, accuracy may vary across languages
  • Context Ambiguity: Some words are offensive in one context but acceptable in another
  • Creative Spelling: Intentional misspellings or character substitutions may evade detection
  • Cultural Variations: Offensiveness varies by culture; model uses broadly-recognized standards

Disclaimers

This model provides probability-based detection, not definitive content classification.

  • Screening Tool: Use as part of a broader content moderation strategy
  • Context Matters: Some flagged content may be acceptable in specific contexts (quotations, educational content)
  • Human Review: Borderline cases should be reviewed by human moderators
  • Policy Mapping: Map detection results to platform-specific content policies
  • Combined Analysis: Use alongside other moderation models for comprehensive content review

Best Practice: Combine graphic language detection with other content moderation tools for comprehensive image analysis.

API Reference

Version
2601
Jan 3, 2026
Avg. Processing
15.0s
Per Request
$0.015
Required Plan
trial

Input Parameters

Vision Language Model for image/video understanding with reasoning

media_typestring

Type of media being sent: 'image' or 'video'. Auto-detected if not specified.

Example:
image
image_urlstring

URL of image to analyze

Example:
https://example.com/image.jpg
base64_imagestring

Base64-encoded image data

video_urlstring

URL of video to analyze

Example:
https://example.com/video.mp4
base64_videostring

Base64-encoded video data

Response Fields

Structured Graphic Language Detection response

responseobject

Structured response from the model

Object Properties:
graphic_languageboolean
thinkingstring

Chain-of-thought reasoning from the model (may be empty)

Complete Example

Request

{
  "model": "graphic-language-detection",
  "image_url": "https://example.com/image.jpg"
}

Response

{
  "inference_id": "inf_abc123def456",
  "model_id": "graphic_language_detection",
  "model_name": "Graphic Language Detection",
  "moderation_type": "image",
  "status": "completed",
  "result": {
    "response": {
      "graphic_language": false
    },
    "thinking": ""
  }
}

Additional Information

Rate Limiting
If we throttle your request, you will receive a 429 HTTP error code along with an error message. You should then retry with an exponential back-off strategy, meaning that you should retry after 4 seconds, then 8 seconds, then 16 seconds, etc.
Supported Formats
gif, jpeg, jpg, png, webp
Maximum File Size
20MB
Tags:graphic-languageprofanityoffensivetextvlmai-analysis

Ready to get started?

Integrate Graphic Language Detection into your application today with our easy-to-use API.