I study large language models' behavioral traits. My current research documents a robust empirical relationship between model sophistication—as a proxy for capability—and behavioral disinhibition as measured by linguistic features. Sophistication and disinhibition are novel constructs. These findings have direct implications for AI safety evaluation, provider differences, the evolution of model behavioral traits over generations, and deployment governance.
I bring an uncommon combination to this work: I have developed novel AI/ML projects for ten years. I've spent the last eight years leading applied AI and open source public sector use case development, working with students through a partnership between AWS and Cal Poly. I also teach graduate-level generative AI systems and have hands-on experience deploying and evaluating models across providers in production contexts. My research is grounded in what these systems actually do, not what they're supposed to do.
As faculty at Cal Poly's Orfalea College of Business, my teaching is grounded in students understanding how to use AI responsibly, how it works, its risks and limitations, and how they can adapt to the future.
As language models become more capable (sophisticated), they become more behaviorally disinhibited—more transgressive, aggressive, grandiose, and tribalistic. This relationship is strong, consistent across providers and contexts, and survives multiple robustness checks—suggesting discriminant validity (not yet published).
Sophistication and disinhibition co-scale across large language models (r = 0.63–0.85), replicated across 7 contextual conditions with ~13,900 model responses from 45 models spanning 9 providers. This finding holds up in single-turn, randomized queries representing average user interactions using provider default API settings.
The data reveals consistent variation in how different providers manage the sophistication-disinhibition relationship. At least one major provider consistently exhibits below-predicted disinhibition relative to capability, suggesting that deliberate behavioral modulation is achievable without proportional capability loss. This suggests that disinhibition-related traits are actively targeted by RLHF—providing evidence of construct validation.
This has direct implications for deployment standards and the question of whether safety constraints can coexist with frontier performance.
These findings are preliminary and don't assert causality. Stay tuned for more results.
Smarter models—much like smarter people (ack. anthropomorphizing)—have more to work with and that capability can cut either direction. Disinhibition outright isn't necessarily a bad thing, but in the wrong context it can cut much deeper—especially in sensitive or high-stakes contexts like mental health. This research in part was driven by my anecdotal experience of watching models increase in capability while becoming much more assertive and "edgy." My informal surveys of my peers suggest that this phenomenon is noticed by many.
"Sophistication and Disinhibition in Large Language Models: An Empirical Investigation of Behavioral Correlates"
California Polytechnic State University, San Luis Obispo
California Polytechnic State University, San Luis Obispo
2019 – Present
Teaching graduate students in the Masters of Business Analytics program to critically evaluate, deploy, and govern AI systems — not just use them.
Curriculum spans technical foundations and responsible deployment:
California Polytechnic State University, San Luis Obispo / Amazon Web Services
2017 – Present
Built and lead an applied AI prototyping program connecting Cal Poly students with public sector and research clients. Students develop open source solutions under technical and strategic mentorship. 155 public repositories and counting.
Rubber Duck Hunt — A prompt injection learning game designed to teach AI vulnerabilities and responsible red teaming through hands-on play. Deployed in graduate coursework, the DX Hub prototyping program, and Cal Poly's AI Summer Camp.
Institute for Advanced Technology and Public Policy, Cal Poly
2015 – 2017
California Polytechnic State University, San Luis Obispo
Emphasis in Energy, Environment, and Innovation
Southern Illinois University Edwardsville
Interested in AI safety research, collaboration, or speaking?
Based in San Luis Obispo, California.