About the job
Join Point72's AI/ML – Investment Services Team as a Data Engineer. We are at the forefront of innovating AI solutions that revolutionize our business processes and elevate enterprise intelligence. Our mission is to seamlessly integrate technology with business needs, collaborating with stakeholders across the firm while harnessing our expertise in generative AI, data engineering, and machine learning.
Your Role:
- Architect, develop, and optimize scalable data pipelines and ETL processes utilizing Python and distributed processing frameworks on Databricks and cloud infrastructures.
- Ingest, standardize, and enhance large datasets, both structured and unstructured, including market feeds, vendor datasets, documents, and alternative data sources.
- Work closely with AI engineers, data scientists, and software engineers to establish data schemas, interfaces, and APIs that facilitate model development and production services.
- Implement automated data validation, testing, monitoring, and alerting systems to ensure data quality, freshness, and drift detection for both offline and online workloads.
- Enhance pipeline efficiency and cost-effectiveness through strategic resource management in distributed computing and cloud environments.
- Uphold data governance, access controls, and privacy standards while maintaining metadata and comprehensive data lineage using data cataloging tools.
- Assist in onboarding and integrating external data vendors, managing contract-driven ingestion workflows, metadata capture, and delivery SLAs.
- Create reusable data components, libraries, and deployment automation to enhance the delivery and operational stability of data products.
- Evaluate and conduct proof-of-concepts for emerging data technologies and tools relevant to generative AI and machine learning applications.

