LLMpeople
Home People Organizations Reports Fields Schools
Public Atlas People first, reports as evidence, organizations as context.

Atlas / People / Detail

Samuel Marks

Senior research engineer at Anthropic interested in agent foundations, model organisms of misalignment, and human-computer interaction.

Researcher1 organizations6 reports

Profile status: updated

Samuel Marks portrait
Suggest a correction
Suggest a source

Contributions are treated as untrusted leads. Public changes happen only after review against public sources.

Trust signals

Profile completeness72%
Public sources3
Official sources2
CountryUnknown
Last reviewedMar 13, 2026
Review outcomeUpdated
Official homepage Scholar profile
updated Unknown location 3 public sources

Latest review note

Added Anthropic homepage bio, GitHub profile, Google Scholar profile, and GitHub avatar.

Public links

website Personal homepage github GitHub profile google_scholar Google Scholar profile

Organizations

core Anthropic

Reports

Alignment and Safety Many-shot Jailbreaking Alignment and Safety Alignment faking in large language models Alignment and Safety Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming Alignment and Safety Constitutional Classifiers++: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming Interpretability On the Biology of a Large Language Model Interpretability Tracing the thoughts of a large language model

Official and primary sources

https://samuel-marks.com/ Official source · homepage https://scholar.google.com/citations?hl=en&user=Q-KQk40AAAAJ Official source · scholar

Supporting sources

https://github.com/samuelmarks Supporting source · github

LLMpeople is a public atlas for discovering frontier AI researchers with context, provenance, and respect.