Data Pipelines

Data Engineer | AWS, Python & Snowflake | Ridgefield, CT (Hybrid) | $140K–$185K

🧠 Data Engineer

πŸ“ Location: Ridgefield, Connecticut (Hybrid – 2–3 days onsite per week)
πŸ’Ό Openings: 2
🏒 Industry: Information Technology / Life Sciences
πŸŽ“ Education: Bachelor’s degree in Computer Science, MIS, or related field (Master’s preferred)
🚫 Visa Sponsorship: Not available
🚚 Relocation: Available for the ideal candidate
πŸ’° Compensation: $140,000 – $185,000 base salary + full benefits
πŸ•“ Employment Type: Full-Time | Permanent

🌟 The Opportunity

Step into the future with a global leader in healthcare innovation β€” where Data and AI drive transformation and impact millions of lives.

As part of the Enterprise Data, AI & Platforms (EDP) team, you’ll join a high-performing group that’s building scalable, cloud-based data ecosystems and shaping the company’s data-driven future.

This role is ideal for a hands-on Data Engineer who thrives on designing, optimizing, and maintaining robust data pipelines in the cloud, while collaborating closely with architects, scientists, and business stakeholders across the enterprise.

🧭 Key Responsibilities

  • Design, develop, and maintain scalable ETL/ELT data pipelines and integration frameworks to enable advanced analytics and AI use cases.

  • Collaborate with data architects, modelers, and data scientists to evolve the company’s cloud-based data architecture strategy (data lakes, warehouses, streaming analytics).

  • Optimize and manage data storage solutions (e.g., S3, Snowflake, Redshift), ensuring data quality, integrity, and security.

  • Implement data validation, monitoring, and troubleshooting processes to ensure high system reliability.

  • Work cross-functionally with IT and business teams to understand data requirements and translate them into scalable solutions.

  • Document architecture, workflows, and best practices to support transparency and continuous improvement.

  • Stay current with emerging data engineering technologies, tools, and methodologies, contributing to innovation across the organization.

🧠 Core Requirements

Technical Skills

βœ… Hands-on experience with AWS data services such as Glue, Lambda, Athena, Step Functions, and Lake Formation.
βœ… Strong proficiency in Python and SQL for data manipulation and pipeline development.
βœ… Experience in data warehousing and modeling (dimensional modeling, Kimball methodology).
βœ… Familiarity with DevOps and CI/CD practices for data solutions.
βœ… Experience integrating data between applications, data warehouses, and data lakes.
βœ… Understanding of data governance, metadata management, and data quality principles.

Cloud & Platform Experience

  • Expertise in AWS, Azure, or Google Cloud Platform (GCP) – AWS preferred.

  • Knowledge of ETL/ELT tools such as Apache Airflow, dbt, Azure Data Factory, or AWS Glue.

  • Experience with Snowflake, PostgreSQL, MongoDB, or other modern database systems.

Education & Experience

πŸŽ“ Bachelor’s degree in Computer Science, MIS, or related field
πŸ’Ό 5–7 years of professional experience in data engineering or data platform development
⭐ AWS Solutions Architect certification is a plus

πŸš€ Preferred Skills & Attributes

  • Deep knowledge of big data technologies (Spark, Hadoop, Flink) is a strong plus.

  • Proven experience troubleshooting and optimizing complex data pipelines.

  • Strong problem-solving skills and analytical mindset.

  • Excellent communication skills for collaboration across technical and non-technical teams.

  • Passion for continuous learning and data innovation.

πŸ’° Compensation & Benefits

πŸ’΅ Base Salary: $140,000 – $185,000 (commensurate with experience)
🎯 Bonus: Role-based variable incentive
πŸ’Ž Benefits Include:

  • Comprehensive health, dental, and vision coverage

  • Paid vacation and holidays

  • 401(k) retirement plan

  • Wellness and family support programs

  • Flexible hybrid work environment

🧩 Candidate Snapshot

  • Experience: 5–7 years in data engineering or related field

  • Key Skills: AWS Glue | Python | SQL | ETL | CI/CD | Snowflake | Data Modeling | Cloud Architecture

  • Seniority Level: Mid–Senior

  • Work Arrangement: 2–3 days onsite in Ridgefield, CT

  • Travel: Occasional

πŸš€ Ready to power the future of data-driven healthcare?
Join a global data and AI team committed to harnessing the power of cloud and analytics to drive discovery, innovation, and meaningful impact worldwide.

Cloud Data Architect | AWS & Snowflake | Ridgefield, CT (Hybrid) | $170K–$210K

☁️ Cloud Data Architect

πŸ“ Location: Ridgefield, Connecticut (Hybrid – 2–3 days onsite per week)
🏒 Industry: Pharmaceutical / Biotech / Information Technology
πŸŽ“ Education: Bachelor’s degree in Computer Science, Information Technology, or related field (or 10+ years equivalent IT experience)
πŸ’Ό Experience Level: Mid–Senior (7–10 years)
🚫 Visa Sponsorship: Not available
🚚 Relocation: Available for ideal candidate
πŸ’° Compensation: $170,000 – $210,000 base salary + benefits + potential performance bonus
πŸ•“ Employment Type: Full-Time | Permanent

🌟 The Opportunity

Join a global leader in life sciences as a Cloud Data Architect, driving innovation and digital transformation within the Enterprise Data, AI & Platforms organization.

In this role, you’ll design and implement scalable, secure, and intelligent cloud-based data architectures that power analytics, AI, and digital products across the business. You’ll partner with data leaders, business stakeholders, and cross-functional technology teams to shape how data is collected, managed, and transformed into actionable insights.

If you’re passionate about modern data ecosystems, cloud architecture, and leveraging AI-driven solutions to impact patient outcomes β€” this is an opportunity to make a global difference.

🧭 Key Responsibilities

  • Design and implement efficient, scalable cloud-based data architectures that align with business and technology goals.

  • Develop and optimize data models, schemas, and database designs supporting structured and unstructured data.

  • Collaborate with business stakeholders, data domain owners, and data scientists to define data requirements and implement robust solutions.

  • Lead data modernization projects and support cloud migration strategies using AWS technologies.

  • Partner with governance teams to establish and maintain data policies, access frameworks, and sharing standards.

  • Enhance data pipelines for ingestion, transformation, and integration of diverse datasets across domains.

  • Provide technical leadership and mentorship to data engineering and development teams.

  • Stay ahead of emerging data technologies, AI/ML innovations, and architectural best practices to drive continuous improvement.

🧠 Core Qualifications

βœ… 7+ years in data management and architecture, including experience as a Data Architect.
βœ… 5+ years hands-on experience with cloud platforms (AWS required) and enterprise data solutions.
βœ… Proven track record leading cloud modernization or data transformation projects.
βœ… Strong experience with AWS components (S3, Glue, Lambda), Snowflake, Apache Parquet, and SQL/NoSQL databases.
βœ… Proficiency with ETL tools such as dbt, SnapLogic, and middleware integrations.
βœ… Expertise in data modeling, data pipeline optimization, and semantic data structures.
βœ… Experience developing Knowledge Graphs and semantic data models using ontologies and taxonomies.
βœ… Understanding of AI/ML data pipelines, including use cases in NLP, recommendation engines, and predictive analytics.
βœ… Deep knowledge of data governance, quality, and lifecycle management.
βœ… Excellent communication and collaboration skills to effectively engage technical and business teams.

πŸ’‘ Preferred Skills & Certifications

⭐ AWS Solutions Architect certification (Associate or Professional).
⭐ Familiarity with Innovator or similar enterprise architecture tools.
⭐ Experience within pharmaceutical or biotech industries β€” particularly in commercial data domains or GxP environments.
⭐ Strong analytical and problem-solving mindset, with the ability to defend and present technical design decisions.
⭐ Exposure to Agile/Scrum delivery models and distributed global teams.

πŸ’° Compensation & Benefits

πŸ’΅ Base Salary: $170,000 – $210,000 (commensurate with experience)
🎯 Bonus: Role-specific or performance-based
πŸ’Ž Benefits Include:

  • Comprehensive medical, dental, and vision coverage

  • Paid time off, holidays, and flexible hybrid schedule

  • 401(k) retirement plan with company match

  • Wellness, family support, and professional development programs

  • Global collaboration and mobility opportunities

🧩 Candidate Snapshot

  • Experience: 7–10 years in data architecture and cloud systems

  • Specialization: AWS Cloud | Data Modeling | ETL | Data Pipeline Design | Data Governance

  • Certifications: AWS Solutions Architect (preferred)

  • Seniority: Mid–Senior

  • Work Arrangement: 2–3 days onsite in Ridgefield, CT

  • Travel: Occasional

🌍 Why This Role Matters

This position sits at the heart of the company’s global data strategy β€” shaping how data is collected, structured, and leveraged to accelerate drug discovery, digital medicine, and patient outcomes. You’ll be part of a collaborative, innovation-driven environment where your work has a tangible impact on global healthcare advancements.

πŸš€ Ready to architect the future of cloud data innovation?
Join a team that’s transforming how data powers scientific breakthroughs and smarter decision-making β€” for patients, healthcare providers, and communities worldwide.

 

Senior / Lead Machine Learning Engineer | Python, PyTorch, AI | Fully Remote | $180,000–$215,000

Senior / Lead Machine Learning Engineer

🌍 Location: Fully Remote
πŸ’Ό Employment Type: Full-time
πŸ’° Compensation: $180,000 – $215,000 (base salary, depending on experience)
πŸ“Š Benefits: Full package included

About the Role

We’re seeking a Senior/Lead ML Engineer to drive the development of advanced enterprise AI and intelligent data applications. This is a hands-on role that combines machine learning, data engineering, and software development to deliver practical, production-ready solutions with measurable impact.

If you’re excited about tackling complex engineering challenges in high-standard environments, this role offers strong career growth opportunities, including senior technical leadership pathways.

Key Responsibilities

  • Lead platform upgrades to ensure products remain cutting-edge and effective.

  • Design and manage dynamic dashboards using Python SDKs to turn data into actionable insights.

  • Optimize data pipelines and access patterns for performance and scalability.

  • Troubleshoot and resolve runtime and performance challenges.

  • Architect robust, scalable, and user-friendly applications designed for long-term growth.

  • Collaborate closely with Product Managers to improve usability and ensure real-world impact.

What You Won’t Do

❌ Work in silos – this role requires versatility across ML, data systems, and software engineering.
❌ Focus solely on research without real-world implementation.

Tech Stack

  • Languages & Tools: Python (primary), Docker, Git

  • Libraries & Frameworks: pandas, numpy, scikit-learn, PyTorch

  • Systems & Processes: CI/CD pipelines, monitoring tools, testing frameworks

Requirements

βœ… 4+ years of professional Python software engineering with experience in production ML deployment (beyond prototyping).
βœ… Proven experience with the end-to-end ML lifecycle: model development β†’ deployment β†’ monitoring.
βœ… Strong production systems background in rigorous engineering environments (Big Tech or top-tier startups preferred).
βœ… Bachelor’s degree in Computer Science from a top 15 university (Ivy League, Stanford, MIT, CMU, etc.).
βœ… U.S. Citizenship and ability to obtain a government security clearance.

Preferred Qualifications

  • Experience in defense-related applications.

  • Exposure to multiple programming languages and diverse tech stacks.

Soft Skills

  • Strong written and verbal communication.

  • Pragmatic approach with a focus on delivering incremental value.

  • Collaborative, with the ability to mentor and influence peers.

Candidate Profile – Not a Fit If

🚫 Job hopper (<2 years per role).
🚫 Focused mainly on research/data science without production deployment.
🚫 Strong theoretical ML background but lacking hands-on implementation.
🚫 No experience with CI/CD, monitoring, or scalable architecture.
🚫 Consulting/contract-heavy career history.

Compensation & Benefits

πŸ’° Base Salary: $180,000 – $215,000
πŸ“¦ Benefits: Comprehensive full package
πŸ›« Travel: Occasional, interview travel reimbursed
πŸ“ Relocation: Not available

πŸ‘‰ Ready to shape the future of AI-driven enterprise applications? Apply now and step into a role where your engineering expertise drives real-world innovation.

 

Data Engineer | Azure, Databricks, Python, SQL, Spark | Hybrid – Netherlands (€3,500–€5,000/month)

Data Engineer

πŸ“ Location: Eindhoven area or Randstad, Netherlands (Hybrid – 3 office days / 2 home days)
πŸ’Ό Employment Type: Full-time
πŸ’΅ Salary: €3,500 – €5,000 per month (€45,360 – €64,800 annually)
🎯 Experience Level: Mid-level | 2–3 years’ experience

About the Role

Do you love working with data β€” from digging into sources and writing clean ingestion scripts to ensuring a seamless flow into a data lake? As a Data Engineer, you’ll design and optimize data pipelines that transform raw information into reliable, high-quality datasets for enterprise clients.

You’ll work with state-of-the-art technologies in the cloud (Azure, Databricks, Fabric) to build solutions that deliver business-critical value. In this role, data quality, stability, and monitoring are key β€” because the pipelines you create will be used in production environments.

Key Responsibilities

  • Develop data connectors and processing solutions using Python, SQL, and Spark.

  • Define validation tests within pipelines to guarantee data integrity.

  • Implement monitoring and alerting systems for early issue detection.

  • Take the lead in troubleshooting incidents to minimize user impact.

  • Collaborate with end users to validate and continuously improve solutions.

  • Work within an agile DevOps team to build, deploy, and optimize pipelines.

Requirements

  • πŸŽ“ Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field.

  • 2–3 years of relevant experience in data ingestion and processing.

  • Strong knowledge of SQL, Python, and Spark.

  • Familiarity with container environments (e.g., Kubernetes).

  • Experience with Azure Data Factory, Databricks, or Fabric is a strong plus.

  • Experience with data model management and dashboarding (e.g., PowerBI) preferred.

  • Team player with strong communication skills in Dutch and English.

  • Familiarity with enterprise data platforms and data lakes is ideal.

What We Offer

  • πŸ’Ά Salary: €3,500 – €5,000 per month

  • 🌴 26 vacation days

  • πŸš— Lease car or mobility budget (€600)

  • πŸ’» Laptop & mobile phone

  • πŸ’Έ €115 monthly cost allowance

  • 🏦 50% employer contribution for health insurance

  • πŸ“ˆ 60% employer contribution for pension scheme

  • 🎯 Performance-based bonus

  • πŸ“š Training via in-house Academy (hard & soft skills)

  • πŸ‹οΈ Free use of on-site gym

  • 🌍 Hybrid work model (3 days in office, 2 days at home)

  • 🀝 Start with a 12-month contract, with option to move to indefinite after evaluation

Ideal Candidate

You are a hands-on data engineer who enjoys data wrangling and building robust pipelines. You take pride in seeing your code run smoothly in production and know how to troubleshoot quickly when issues arise. With strong technical skills in SQL, Python, and Spark, plus familiarity with cloud platforms like Azure, you’re ready to contribute to impactful enterprise projects.

πŸ‘‰ Ready to make data flow seamlessly and create business value? Apply now to join a passionate, innovation-driven team.