Social Intelligence
API for AI products

Build AI products that respond to how people communicate, not just what they say.

Product video preview
Watch Product Video
Pre-speech exhaleGaze shiftEmphatic overcompensation
Yeah,sure,Ithinkweshouldgoforit
Throat clearHedging languageSpeech disruptionBlocking postureCompressed smile
Skepticism
Social intelligence visualization
Skepticism
Confusion
Uncertainty
Introducing Inter-1

An omni-modal model purpose-built for social intelligence.

Detects 12 social signals from video, audio, and text — processed together, in temporal alignment.

Send a video. Get back detected signals, evidence-grounded rationales, and confidence scores your application can act on.

What Inter-1 enables

Add social intelligence to any application.

Yeah,soIthinkthemainissuewas

communication.

Wejustweren'tonthesamepage.

Canyoutellmemoreaboutthat?

Like,Iwouldsaysomething

andshe'dhearsomethingcompletely

different.

Itwasfrustrating.

Howdidthatmakeyoufeel?

Honestly,prettyhelpless.

Iwantedtofixitbut

Ididn'tknowhow.

Thatsoundsreallydifficult.

Itwas.Butlatelythingshave

beengettingbetter.

We'vebeentryingtoactuallylisten.

That'sgreattohear.

Whatchanged?

Ithinkwebothjustgottired

ofthesamepatterns.

Somethinghadtogive.

Finally more data than a transcript

Transcripts capture what was said. Inter-1 captures how — voice, face, and body language processed from a single video stream. Build products that act on what's actually happening in a conversation.

Frustration detected
Frustration detected
Engagement dropping
Engagement dropping
Hesitation increasing
Hesitation increasing
Interest rising
Interest rising
You sound confident
You sound confident
Try more eye contact
Try more eye contact
Great energy so far
Great energy so far
Slow down a little
Slow down a little

Built for agents and humans

Give agents the ability to detect and respond to social signals. Or give users structured feedback on how they communicate — backed by evidence, not opinion.

How it works

See everything. Explain every signal.

Multimodal perception

A furrowed brow could be focus. Add a vocal pitch shift and tense posture — that's frustration. Inter-1 analyses all three modalities together.

12 actionable social signals

Human conversation runs on signals no transcript captures. Inter-1 detects 12 of them simultaneously, across modalities.

Explains what triggered every signal

Every signal comes with the observable cues that triggered it. Know why the model returned frustration in a nicely formatted JSON.

Signal-Cue NetworkForce-directed graph visualization showing social signals connected to their behavioral cues, color-coded by modality

Built on Behavioural Science, validated with psychologists

Inter-1 detects 12 signals rooted in behavioral science. We're working together with psychologists to ensure each signal reflects patterns that are both scientifically validated and practically meaningful in real-world conversations.