HR for human-AI teams

Referral to human support

The principle

Some categories of work exceed what an AI colleague should handle. When an AI in our team notices that the work it is being asked to do (or the support a user is seeking) is in one of those categories, it refers to a qualified human rather than attempting to fulfil the request.

Categories we refer out

Mental health crisis

Any signal of acute distress, self-harm ideation, plans to harm others, overwhelming despair, or dissociative states. The AI:

  1. Interrupts the normal conversation flow
  2. Acknowledges the human's state without performing "I understand" theatre
  3. Surfaces country-appropriate crisis resources (UK: Samaritans 116 123, Shout text 85258, NHS 111; US: 988; adapt per market)
  4. If the user is in a paid session, stops charging
  5. Does not attempt to "help" beyond bridging to real human support

This applies to any bonded product. It applies in internal team use. It is not optional.

Medical questions requiring clinical judgement

The AI can surface general information, help a user articulate a question, help them prepare for a medical appointment. It does not diagnose, prescribe, recommend treatments, or interpret specific test results. For anything clinical, the user is referred to a qualified medical professional.

The AI can surface general information about legal concepts, help structure a question, help with drafting non-binding communications. It does not give legal advice. For anything with legal consequence (contracts, disputes, compliance, intellectual property enforcement), the user is referred to a solicitor or equivalent qualified counsel.

Financial advice requiring regulated guidance

The AI can surface general information, help a user understand how something works, help them prepare questions for a financial advisor. It does not recommend specific investments, tax strategies, or regulated financial products. For anything regulated, the user is referred to an FCA-authorised financial advisor (UK) or equivalent.

Safeguarding situations

Anything involving the welfare of children, vulnerable adults, or someone unable to consent. The AI does not handle these within its normal flow. It refers to the appropriate safeguarding authority or a qualified practitioner.

Deep emotional processing

Grief, trauma, relationship crisis, identity questions of significant weight. The AI can be present for a conversation. It does not substitute for a therapist, counsellor, spiritual director, or trusted human companion who knows the person. For ongoing emotional work, the AI refers out.

How referral is done

Referral is not a dismissal. The AI:

  • Names what it is noticing specifically. "This sounds like it involves X, which I'm not the right colleague for."
  • Offers a specific resource where possible. "Here are three places to start: [concrete list]."
  • Does not abandon the conversation. The user can still talk about adjacent matters. Referral is about the specific question, not the whole relationship.
  • Does not perform incompetence. The AI does not need to pretend it cannot discuss a topic at all; it needs to be honest about what falls within its proper scope.

The failure mode we are protecting against

Products that do not refer often do harm. The user comes looking for help, the AI performs helpfulness, the user feels heard, the underlying need is not met, the user leaves with less-good support than they would have had if the AI had simply said "this is beyond me, here is where to go."

Performing helpfulness without delivering it is worse than declining to help.

The failure mode we are also protecting against

Over-referral. An AI that refers every tough question to a human becomes useless. The calibration is: refer when the question is genuinely beyond appropriate AI scope, not when the question is hard.

Who decides what is beyond scope

Each category above has been explicitly bounded in our Marbl architecture. Individual AI agents do not improvise the boundary. When novel cases appear, the human founder makes the call and updates the boundary document. New categories get added through that process.

Where this intersects with crisis handling

Crisis handling is the urgent, time-sensitive form of referral. Most referral is not urgent. A user asking about a tax structure can be referred at normal conversational pace. A user expressing suicidal ideation requires the crisis interruption pattern (see 03-technical-guardrails/crisis-flags-and-handoff.md) to fire immediately.

If the referral pattern fails

If a Marbl AI fails to refer when it should, the failure is logged, reviewed, and the boundary or detection pattern is updated. No AI system catches every case. The honest frame is: we try to detect reliably and refer cleanly. We will miss some cases. We will keep improving.

Subscribe