About the job
About Anthropic
At Anthropic, we are dedicated to developing AI systems that are reliable, interpretable, and steerable. Our mission is to ensure that AI serves the needs of users and society safely and effectively. Our rapidly expanding team comprises passionate researchers, engineers, policy experts, and business leaders collaborating to construct beneficial AI systems.
About the Role
The Data Infrastructure team is responsible for designing, operating, and scaling secure systems that uphold user privacy while facilitating data-driven decision-making across Anthropic. Our goal is to deliver data processing, storage, and access solutions that are trustworthy, efficient, and user-friendly.
We are seeking infrastructure engineers who excel at the intersection of data systems, security, and scalability. You will face a variety of challenges, from constructing financial reporting pipelines to architecting access control mechanisms and ensuring the reliability of cloud storage. This position offers the chance to collaborate closely with data scientists, analysts, and business stakeholders while delving into cloud infrastructure fundamentals.
Responsibilities:
As part of the Data Infrastructure team, you may engage with critical business areas, including:
Data Governance & Access Control: Design and implement robust access control systems to ensure that only authorized users can access sensitive data. Develop infrastructure for permission management, audit logging, and compliance. Work with IAM policies, ACLs, and security controls that scale across numerous users and systems.
Financial Data Infrastructure: Build and maintain data pipelines and warehouses that support essential reporting. Ensure data integrity, accuracy, and availability for complex financial systems, including third-party revenue ingestion pipelines; manage external relationships to facilitate upstream dependencies. Take ownership of the reliability of systems processing revenue, usage, and business metrics.
Cloud Storage & Reliability: Design disaster recovery, backup, and replication systems for petabyte-scale data. Ensure high availability and durability of data stored in cloud object storage (GCS, S3). Develop systems that safeguard against data loss and enable rapid recovery.
Data Platform & Tooling: Scale data processing infrastructure using technologies such as BigQuery, BigTable, Airflow, dbt, and Spark. Optimize query performance, manage costs, and enable self-service capabilities for stakeholders.

