Our client is a fast-growing AI company specializing in the Pharmaceutical Intelligence space.
Key Responsibilities:
- Design, develop, and maintain complex data pipelines that go through multiple stages and transformations.
- Implement data modeling, normalization, and wrangling of different data formats into a cohesive product.
- Manage and optimize data storage solutions using technologies such as PostgreSQL, S3/GCS.
- Orchestrate workflows using tools like Airflow, Prefect, or Dagster.
- Implement and manage message queues to ensure efficient data processing.
- Collaborate with cross-functional teams to understand data requirements and deliver solutions that meet business needs.
- Contribute to the maturation of data infrastructure in a startup environment, from 0 to 1 and beyond.
Required Qualifications:
- Proficiency in Python and experience with PostgreSQL.
- Hands-on experience with cloud storage solutions such as S3 or GCS.
- Familiarity with workflow orchestration tools like Airflow, Prefect, or Dagster.
- Experience with managing message queues.
- Mid-level or higher experience in managing complex data pipelines.
- Strong understanding of data modeling, normalization, and data wrangling.
Preferred Qualifications:
- Experience in fast-growing AI-native companies.
- Proven track record of maturing data infrastructure in startup environments.
- Ability to handle unstructured and varying data formats effectively.
Why Join Us:
- Be part of an innovative team at the forefront of AI and data engineering.
- Opportunity to work on challenging projects that make a real impact.
- Collaborative and dynamic work environment.
- Competitive salary and benefits package.