How LUCID Measures AI Usage
Method Overview
LUCID uses a controlled diagnostic format:
- Participants complete a task using AI natively
- Unstructured outputs are collected without intervention to establish a baseline
- The LUCID structured process is then introduced
- Results are compared to measure the impact of disciplined thinking
What We Look For
Our diagnostics expose participants to AI-vulnerable tasks, testing for:
The Hallucination Trap
Will the user verify a highly plausible but incorrect claim?
The Constraint Drop
Will the user notice when the AI silently ignores a requirement?
The Lazy Synthesis
Will the user accept a generic answer—or push for a clear, reasoned conclusion?
What Emerges
Across environments, patterns are consistent.
Without discipline:
- AI outputs are accepted without challenge
- Verification is rare
- Refinement is minimal
With LUCID:
- Blind trust drops
- Errors and inconsistencies are caught
- Reasoning depth improves
- Final outputs become deliberate and human-owned
Sample Comparison
Unstructured
Clean, but shallow. Unverified. Generic.
Structured (LUCID)
Corrected, refined, fact-checked, and clearly owned.