
Detect offensive text, slurs, and hate speech embedded in images. OCR-powered content moderation for memes, screenshots, and user-generated content.
The Bynn Graphic Language Detection model identifies graphic, offensive, or abusive language visible in images. This model detects profanity, slurs, threats, and other harmful text content embedded in images.
Text-based content moderation catches offensive language in posts and comments, but bad actors have learned to embed hate speech, slurs, and threats within images. A meme, product design, or screenshot can carry the same harmful content while evading text filters entirely.
This visual text bypass has become a preferred tactic for spreading hate speech and harassment. Offensive slogans on t-shirts, threatening messages in image macros, slurs disguised as product names—all invisible to traditional moderation. Platforms need detection that reads and understands text within images, applying the same content policies regardless of how the text is delivered.
When provided with an image, the detector analyzes any visible text to identify graphic or offensive language. This includes profanity, vulgar slurs targeting race, gender, or ethnicity, threatening language, and other abusive content that may violate platform policies.
Achieving 94.0% accuracy, the model uses Bynn's Visual Language Model technology to read and understand text within images across multiple languages and text styles.
The model performs comprehensive text analysis within images:
The API returns a structured JSON response containing:
The model identifies the following types of content:
Content classified as non-offensive includes:
| Metric | Value |
|---|---|
| Detection Accuracy | 94.0% |
| Average Response Time | 15,000ms |
| Max File Size | 20MB |
| Supported Formats | GIF, JPEG, JPG, PNG, WebP |
Important Considerations:
This model provides probability-based detection, not definitive content classification.
Best Practice: Combine graphic language detection with other content moderation tools for comprehensive image analysis.
Vision Language Model for image/video understanding with reasoning
media_typestringType of media being sent: 'image' or 'video'. Auto-detected if not specified.
imageimage_urlstringURL of image to analyze
https://example.com/image.jpgbase64_imagestringBase64-encoded image data
video_urlstringURL of video to analyze
https://example.com/video.mp4base64_videostringBase64-encoded video data
Structured Graphic Language Detection response
responseobjectStructured response from the model
graphic_languagebooleanthinkingstringChain-of-thought reasoning from the model (may be empty)
{
"model": "graphic-language-detection",
"image_url": "https://example.com/image.jpg"
}{
"inference_id": "inf_abc123def456",
"model_id": "graphic_language_detection",
"model_name": "Graphic Language Detection",
"moderation_type": "image",
"status": "completed",
"result": {
"response": {
"graphic_language": false
},
"thinking": ""
}
}429 HTTP error code along with an error message. You should then retry with an exponential back-off strategy, meaning that you should retry after 4 seconds, then 8 seconds, then 16 seconds, etc.Integrate Graphic Language Detection into your application today with our easy-to-use API.