Data Engineer at Darwoft

We are redirecting you to the source. If you are not redirected in 3 seconds, please click here.

Data Engineer at Darwoft. Location Information: LATAM. Location:. LATAM (Remote supporting US-based teams). Job Type:. Contractor (Full-Time, Remote). Project:. Data Platform Modernization Healthcare Sector. Time Zone:. Aligned with GMT-3 (Argentina). English Level:. B2/C1. . Get to Know Us. . At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.. . Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using . DBT, Spark, Argo, and AWS. .. . Were Looking For a Senior Data Engineer (DBT + Spark + Argo). . In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.. . If you're passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.. . What Youll Be Doing. . Translate legacy . T-SQL logic. into modular, scalable DBT models powered by . Spark SQL. Build reusable and performant . data transformation pipelines. Develop testing frameworks to ensure . data accuracy and integrity. in DBT workflows. Design and orchestrate workflows using . Argo Workflows. and CI/CD pipelines with . Argo CD. Manage mock data and . reference datasets. (e.g., ICD-10, CPT), ensuring version control and governance. Implement . efficient storage/query strategies. using Apache . Hudi. , . Parquet. , and . Iceberg. Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying. Collaborate with DevOps teams to optimize . S3 usage. , enforce . data security. , and ensure compliance. Work in Agile squads and participate in planning, estimation, and sprint reviews. . What You Bring. . Strong experience with . DBT. for data modeling, testing, and deployment. Hands-on proficiency in . Spark SQL. , including performance tuning. Solid programming skills in . Python. for automation and data manipulation. Familiarity with . Jinja templating. for building reusable DBT components. Practical experience with . data lake formats. : Apache Hudi, Parquet, Iceberg. Expertise in . Argo Workflows. and . CI/CD integration with Argo CD. Deep understanding of . AWS S3. data storage, performance tuning, and cost optimization. Strong command of . ElasticSearch. for indexing structured/unstructured data. Knowledge of . ICD-10. , . CPT. , and other healthcare data standards. Ability to work cross-functionally in . Agile. environments. . Nice to Have. . Experience with . Docker. , . Kubernetes. , and container orchestration. Familiarity with cloud-native data tools: . AWS Glue. , . Databricks. , . EMR. , or GCP equivalents. Prior work on CI/CD automation for data engineering workflows. Knowledge of . data compliance. standards: HIPAA, SOC2, etc.. Contributions to open-source projects in DBT, Spark, or data engineering frameworks. . . Perks & Benefits. . Contractor agreement with payment in USD. 100% remote work. Argentinas public holidays. English classes. Referral program. Access to learning platforms. . Explore this and other opportunities at:. www.darwoft.com/careers