A bit about us:

Our client modernizes legal processes, globally, for many of the world's leading companies. They offer scalable, end to end solutions and connecting you with a large worldwide network of experienced lawyers.


Why join us?

Our client's data platform is nascent, and you are excited by the opportunity that accompanies this. New tooling around automation and data processing will help their ML and analytics teams build out their services.

The role will report directly to and receive mentorship from their Director of Data Infrastructure, and the package includes a competitive base salary, equity, and full benefits. The preferred location for this role is San Francisco.*

*Please only apply if you are coming out of a start-up environment
**Please only apply if you are available to work out of Northern CA, OR, or WA - with quarterly travel to San Francisco.

Job Details

Note - You should be coming out of a start-up environment to be considered for this position!

Responsibilities:
Develop, implement and maintain data pipelines and data processing code to support product features
Create new data sets and tooling to help ML and analytics teams uncover insights and develop NLP-based solutions to legal tech problems
Support usage and maintenance of our data infrastructure (i.e., AWS stack, Snowflake, Airflow)
Champion efforts to enforce data governance, quality and security across the organization

Requirements:
3-5 years of experience in data engineering
Fluency in SQL and python and proven ability to ship flexible, maintainable code
Experience developing and productionizing complex data processing pipelines
Comfort using AWS tooling and experience maintaining modern data infrastructure (warehousing, job scheduling)
Excellent communication and problem solving skills and ability to effectively collaborate with technical and business partners
Familiarity with data governance frameworks and Agile methodology
Previous experience developing ML pipelines is a plus