Blog / Model-Specific

How to Detect Midjourney Images: Telltale Signs & Detection API (2026)

May 15, 202611 min read
On this page
MODEL-SPECIFICDetecting Midjourney v7The aesthetic fingerprints that survived the upgrade→ ~35% of all AI images in the wild

Midjourney has been the most popular consumer-facing AI image generator since mid-2023. By mid-2026, an estimated 35-40% of AI-generated images circulating on social platforms originated from Midjourney — a higher share than DALL-E, Stable Diffusion, and Flux combined for general-audience use. If you're trying to detect AI-generated images in the wild, Midjourney is the model you'll encounter most often.

This guide is for journalists, fact-checkers, content moderators, and anyone trying to identify Midjourney outputs specifically. We'll cover the visual fingerprints that have persisted across Midjourney versions, what's changed in v7 (which is significantly harder to spot than v5), the detection-API approach that gives you confident model attribution, and where pure visual inspection fails.

If you want the broader detection workflow that covers all generators, our pillar guide on detecting AI-generated images is the place to start. This piece is the Midjourney-specific deep dive.

Why Midjourney is harder to detect than other generators

Three reasons Midjourney specifically resists detection more than its peers:

1. No C2PA manifest by default. Unlike OpenAI's DALL-E or Adobe's Firefly, Midjourney does not embed Content Credentials in its outputs as of mid-2026. The company has stated public commitment to add support but has not yet shipped it. So one major detection layer — checking for a signed manifest — is unavailable for Midjourney content.

2. Aesthetic prioritization. Midjourney is explicitly tuned to produce images that look beautiful — composed, stylized, atmospheric. The aesthetic polish that makes Midjourney popular also smooths over many of the artifacts that detectors and human reviewers rely on. A Midjourney image is more likely to look "professional" than an unedited Stable Diffusion or Flux image, which removes the "this looks like AI" instinct from casual viewers.

3. Strong text-to-image alignment. Midjourney's v7 release (early 2026) significantly improved its handling of text rendering, hands, and complex compositions. Many of the visual tells that worked on v5 and v6 are weaker or absent in v7 outputs.

That said, model-specific fingerprints persist. Trained classifiers identify Midjourney with 96-98% accuracy on benchmark sets — better than human reviewers do — because the underlying diffusion model leaves frequency-domain signatures that aren't visible to the eye but are detectable to math.

What to Look ForThe 'Midjourney glow'Punchy mid-tones, no clipped highlights, HDR-ish polishPainterly micro-texturesSoft brush-stroke feel at hair-skin and fur transitionsSymmetric stylizationRule-of-thirds always perfect, subjects centeredAtmospheric depthDistant objects always slightly blue-hazedThe 'Midjourney face'Symmetric features, slightly oversaturated lipsBackground object hazeReflections, text, crowds — still soft at distance

Visual fingerprints that persist across versions

Even in v7, Midjourney outputs share several aesthetic and structural signatures that experienced reviewers learn to spot:

The "Midjourney glow"

Midjourney consistently produces images with a slight HDR-ish tonal compression — mid-tones are punchy, shadows are detailed, highlights have a soft roll-off. Real photography shot in challenging lighting almost always has some clipping or some lost shadow detail. Midjourney rarely does.

In a portrait, this manifests as: even illumination on the face from multiple directions, no harsh shadow under the chin or nose, no blown-out highlights even in bright scenes. In a landscape: detail visible in both deep shadow and bright sky in the same frame, no compromise needed.

A real photographer can produce this look with HDR processing or careful exposure bracketing, but it takes effort. When every image in a portfolio looks like this, that's a tell.

Painterly micro-textures

Even in photo-realistic prompts, Midjourney has a subtle painterly quality at the pixel level. Texture transitions — skin to clothing, leaves to bark, asphalt to grass — have a soft brush-stroke feel. Zoom in to 200%+ on any region and look at how textures meet at boundaries.

This is most noticeable on:

  • Hair-against-skin transitions (especially backlit)
  • Fur texture on animals
  • The edge of out-of-focus areas (the bokeh transition is too smooth)
  • Mid-distance background details that should be sharp but slightly aren't

Symmetric stylization

Midjourney compositions tend toward symmetric or rule-of-thirds-perfect framing more than real photography does. When you see a portrait where the subject is exactly centered, eye-level with the camera, perfectly lit from both sides, with a stylistically appropriate background — that's a Midjourney signature. Real photography is messier; subjects look slightly off-center, lit unevenly, with imperfect backgrounds.

Atmospheric haze and depth gradients

Midjourney loves atmospheric perspective. Even when the prompt doesn't request it, distant objects often have a soft haze that creates depth — slightly blue-tinted, slightly desaturated, with falling-off contrast. Real photography has atmospheric haze too, but it's less consistent and often less aesthetically perfect.

The "Midjourney face"

A specific portrait look common across v6 and persisting in v7:

  • Symmetric features even when the pose suggests asymmetry
  • Eyes that are slightly too sharp and contrasty against the rest of the face
  • A subtle "perfectness" to skin texture (pores rendered, but uniformly so)
  • Slightly oversaturated lip color
  • Hairlines that are clean and well-defined even in casual outdoor settings

A trained reviewer can identify this look in under a second on suspect portraits. Untrained reviewers typically can't, which is why detection APIs are necessary at scale.

Background object handling

Where Midjourney still falls short of photographic realism:

  • Text on signs, books, posters — improved a lot in v7 but still inconsistent. Half-legible nonsense words, inconsistent kerning, occasional letter substitutions.
  • Crowds and groups — the third-row-back faces in a crowd are still smeared or melted, even in v7.
  • Mechanical details — clock faces, gauges, control panels, instruments. Midjourney makes them look mechanical without making the specific markings or numbers consistent.
  • Reflections in glass and water — better in v7 than v6 but still occasionally show scenes that don't match the rest of the frame.

Aspect-ratio defaults

Midjourney users often forget to override the default aspect ratio. Common defaults:

  • v6 and earlier defaulted to 1:1 square
  • v7 defaults to 16:9 or 1:1 depending on the prompt context

If you see a content stream where every image is exactly square or exactly 16:9 with no variety, that's consistent with batch Midjourney output. Real photography or mixed-source human content is usually more varied.

What changed in Midjourney v7

V7 launched in early 2026. Key changes from v6 that affect detection:

  • Text rendering improved dramatically. v7 can render legible English text on signs, book titles, t-shirt labels, etc. with reasonable consistency. Several visual tells from v6 ("the text is gibberish") no longer apply.
  • Hands and feet improved further. v7 hands are anatomically correct most of the time, including fingers gripping objects with realistic contact points. The "extra finger" tell from 2023 is essentially dead.
  • Multi-person scenes improved. v6 often had melted-together faces in second-row crowd members; v7 keeps faces distinct further into the depth field.
  • Compositional stylization is somewhat reduced — v7 produces less aggressively stylized output by default, so the "this looks like a movie still" instinct fires less often.

Things that did NOT improve:

  • Reflections in glass and water remain inconsistent (better than v6 but not realistic)
  • Atmospheric perspective and HDR-ish tonal compression are unchanged
  • The "Midjourney glow" aesthetic persists
  • Background object handling at distance is still soft
  • Frequency-domain fingerprints are still detectable by classifiers — even if not by eye

Net effect: v7 outputs are approximately 8-12 percentage points harder to identify by human review than v6 outputs. Detection-API accuracy on v7 is also lower than on v6 — typically 96% vs 98% on benchmark sets — but still high enough for production use.

How detection APIs identify Midjourney specifically

Modern detection APIs don't just classify "AI vs not AI" — the better ones return model attribution. The technical approach:

Frequency-domain fingerprints. Each diffusion model has a slightly different architecture, training data, and sampling process. These differences leave characteristic patterns in the Fourier transform of the output. Midjourney's signature differs from Flux's signature differs from Stable Diffusion's, and a classifier trained on labeled outputs can distinguish them with high accuracy.

Aesthetic style classifier. A complementary classifier learns the visual style fingerprint — the "Midjourney glow," the typical compositional choices, the color palette tendencies. This signal is weaker than the frequency-domain one (real photography can mimic the aesthetic), but combined with frequency analysis it improves robustness.

Metadata patterns. Midjourney typically delivers images with specific resolution defaults (1024x1024 for v6, varying ratios in v7) and specific JPEG compression characteristics from the Discord-based delivery pipeline. These metadata signatures aren't proof of Midjourney origin but they correlate.

In practice, an API like ours returns a structure like:

{  "verdict": "ai_generated",  "confidence": 0.984,  "model_attribution": {    "midjourney_v7": 0.86,    "midjourney_v6": 0.04,    "flux": 0.05,    "stable_diffusion": 0.03,    "dalle": 0.02  },  "version_estimate": "midjourney_v7",  "version_confidence": 0.86}

You get not just "this is Midjourney" but which version, with calibrated confidence. That matters because v6 and v7 detection workflows differ — v7 cases that confuse a reviewer should escalate to a more sophisticated review pipeline than v6 cases that have obvious tells.

A workflow for verifying suspected Midjourney images

For a single image:

  1. Visual scan. Look at the five fingerprints above. Two or more is a strong tell; one is a yellow flag; zero doesn't rule out Midjourney.
  2. Reverse search. Drag the image into Google Images and TinEye. Has it appeared online before? Is the supposed source plausible?
  3. C2PA check. Drag into contentcredentials.org/verify. For a Midjourney image, expect "no manifest." That's not proof of Midjourney origin, but consistent with it.
  4. Detection API. Run through a model-attribution-capable API. If it reports Midjourney with high confidence, you have your answer. If it reports another generator with high confidence, you also have your answer (and you've ruled out Midjourney). If it's uncertain, the image may be heavily edited or in an adversarial format.

For high-volume verification (content moderation, fact-checking pipelines), pipe every suspected image through the API and route flagged ones to human review.

Where Midjourney detection still fails

Honest acknowledgments:

Heavily edited Midjourney outputs. A Midjourney image run through Photoshop with significant edits — tone curves, color grading, retouching, compositing with real photography — degrades both human-detectable visual tells and frequency-domain fingerprints. Hybrid AI-real images are the hardest case.

Adversarially processed images. Specific recompression, downsampling, or noise-injection sequences can degrade detector accuracy. Sophisticated bad actors aware of this can produce Midjourney content that defeats most public detectors.

Cross-generator content. Some users run Midjourney outputs through Stable Diffusion-based "img2img" pipelines for refinement. The result has signatures from both generators and confuses single-attribution classifiers.

Edge cases in content type. Midjourney is best at portraits, landscapes, and stylized scenes. Detection of Midjourney technical content (diagrams, charts, schematic-style imagery) is less well-studied because the training data is photo-heavy.

Detection Accuracy Across VersionsMidjourney v599%Midjourney v698%Midjourney v797%

Testing this yourself

If you're skeptical about claimed accuracy, run your own test. Generate (or download) a set of recent Midjourney v7 outputs, mix them with real photos from comparable scenes, and run the mix through a detection API. Compare the API's verdicts to ground truth.

A few practical notes:

  • Use unedited Midjourney outputs for the cleanest signal. Heavy editing degrades attribution accuracy for any classifier.
  • Match the resolution and compression of the real-photo control set to the AI-generated set. Different file formats can confound the comparison.
  • Include both v6 and v7 outputs separately to see how the detector handles each generation.

If you want to use our API for the test, the free tier covers 500 scans per month, which is enough for a meaningful evaluation. Our accuracy guide walks through evaluation methodology in detail.

Frequently asked questions

Can I tell a Midjourney v7 image from a real photo just by looking?

Sometimes, but not reliably. Trained reviewers correctly identify v7 outputs about 60-70% of the time on average, depending on image complexity. That's better than chance but worse than what a detection API achieves. Don't rely on visual inspection alone for high-stakes use cases.

Does Midjourney watermark its outputs?

No. Midjourney does not embed visible watermarks or invisible cryptographic watermarks in its standard outputs as of mid-2026. The company has discussed adding C2PA support but hasn't shipped it.

What about Midjourney's "Discord seed" — is that a forensic signal?

Midjourney delivers images through Discord with specific filename patterns and size conventions. These are weak signals because they don't survive any image edit or platform re-upload. Useful as a corroborating clue if you have access to the original delivery; useless once the image has been screenshotted or reposted.

How does Midjourney detection accuracy compare to Stable Diffusion or DALL-E detection?

Across major generators in 2026:

  • DALL-E: highest accuracy (98-99%) because of C2PA manifests in many outputs
  • Midjourney v6: high accuracy (~98%) on clean outputs; lower on heavily edited
  • Midjourney v7: slightly lower (~96%) due to recent release and architecture changes
  • Stable Diffusion: high but variable (95-98%) depending on which fine-tune
  • Flux: similar to Midjourney v7 (~96%)
  • Sora-image: lowest (~95%) as the newest mainstream generator

These numbers are from our internal benchmark; published numbers from other vendors vary.

Will Midjourney get harder to detect over time?

Possibly, but the historical trend is mixed. Newer generators are sometimes harder than older ones (v7 vs v6) but sometimes have new fingerprints that detectors quickly adapt to. The arms race is permanent; expect detection accuracy on any given generator to fluctuate within a few percentage points over time.

Is there a "100% Midjourney detector"?

No. Anyone selling one is overselling. Best-case accuracy on clean Midjourney outputs is around 98%; lower on edited or adversarial inputs. Defense-in-depth (multiple detectors, provenance checks where available, human review on borderline cases) is the practical answer.


Midjourney is the most-encountered AI generator in the wild and one of the harder ones to detect by eye, especially in v7. The fingerprints persist — aesthetic stylization, atmospheric perspective, "Midjourney glow," background-detail softness — but reading them reliably takes either training or a detection API.

If you're verifying suspected Midjourney content at any volume, our AI Image Detector API returns model attribution (including v6 vs v7) alongside the AI-detection score in a single call, with sub-100ms latency. Free tier covers 500 scans per month, no credit card required.

Try the AI Image Detector API

500 free scans per month. No credit card. Sub-100ms detection with model attribution and region heatmaps.

Get an API key →