โ† Back to Help Center

Understanding Confidence Scores

What "87 out of 100 robots would flag this" actually means.

Getting Started Fast Scanning Setup How AI Works Understanding Confidence

What is a Confidence Score?

When SkinVault scans a photo, the AI outputs a confidence score between 0% and 100%. This represents how certain the model is that the photo contains sensitive content.

87%
"87 out of 100 robots would flag this photo"

Think of it this way: if we ran 100 slightly different versions of our AI model on the same photo, 87 of them would flag it as sensitive. The higher the number, the more certain the detection.

๐Ÿค” Why not just "yes" or "no"?

Real-world photos aren't always clear-cut. A photo might be ambiguous - swimwear at a beach, artistic photography, medical images. The confidence score tells you how sure the AI is, so you can make informed decisions.

The False Positive / False Negative Tradeoff

No AI is perfect. There are two types of mistakes:

False Positive

AI flags a safe photo as sensitive

Annoying, but harmless

False Negative

AI misses an actual sensitive photo

Defeats the purpose

There's always a tradeoff between these two. You can't minimize both at the same time:

Why We Err Toward False Positives

SkinVault is designed to err on the side of caution. Here's why:

The cost of missing a sensitive photo is higher than the cost of reviewing a safe one.

Consider the consequences:

This is why our default threshold is relatively sensitive. We'd rather flag a few extra beach photos than miss something actually sensitive.

๐Ÿฅ Like a Medical Screening

Medical tests are designed to catch potential issues even if it means some false alarms. A mammogram might flag something that turns out to be benign - that's better than missing actual cancer. Same principle here.

Play With the Threshold

You can adjust how sensitive the detection is. Lower thresholds catch more (including more false positives). Higher thresholds are stricter (but might miss borderline cases).

Sensitive Strict 40%
Photos Flagged
~85%
False Positives
Higher

In the app, you can review flagged photos and mark safe ones. The more you review, the better you'll understand where your personal comfort level is.

Confidence Score Ranges

๐Ÿ”ด 80-100%: High Confidence

The AI is very sure this photo contains sensitive content. These are rarely false positives.

๐ŸŸ  50-80%: Medium Confidence

Likely sensitive, but could be ambiguous cases like swimwear, artistic shots, or partial views. Worth reviewing.

๐ŸŸก 30-50%: Low Confidence

Borderline cases. The AI isn't sure. These have higher false positive rates but are flagged to be safe.

๐ŸŸข 0-30%: Likely Safe

The AI doesn't think this is sensitive. Not flagged by default, but you can lower your threshold to include these.

The Bottom Line

Confidence scores give you transparency into the AI's thinking. Instead of a mysterious black box that just says "yes" or "no," you see exactly how certain the model is.

We default to catching more rather than missing things, because:

The robots are here to help - and they'd rather be a little overcautious than miss something important.

Remember: You're always in control. The AI flags, you decide. Every flagged photo is a suggestion, not a judgment.