Lead Data Engineer – Remote Job. at Bertoni Solutions

We are redirecting you to the source. If you are not redirected in 3 seconds, please click here.

Lead Data Engineer – Remote Job. at Bertoni Solutions. Location Information: Remote, Peru. . Company Description. We are a multinational team of individuals who believe that, with the right knowledge and approach, technology is the answer to the challenges businesses face today. Since 2016, we have brought this knowledge and approach to our clients, helping them translate technology into their success.. With Swiss roots and our own development team in Lima and across the region, we offer the best of both cultures: the talent and passion of Latin American professionals combined with the organizational skills and Swiss mindset.. Job Description. We are seeking a highly skilled . Lead Data Engineer. with strong expertise in . PySpark. , SQL, and Python, as well as a solid understanding of . ETL. and data warehousing principles. The ideal candidate will have a proven track record of designing, building, and maintaining scalable data . pipelines. in a collaborative and fast-paced environment.. Key Responsibilities:. Design and develop scalable data pipelines. using PySpark to support analytics and reporting needs.. Write efficient SQL and Python code. to transform, cleanse, and optimize large datasets.. Collaborate with machine learning engineers, product managers, and developers to understand data requirements and deliver solutions.. Implement and maintain robust . ETL processes. to integrate structured and semi-structured data from various sources.. Ensure data quality, integrity, and reliability. across pipelines and systems.. Participate in . code reviews, troubleshooting. , and performance tuning.. Work independently and proactively to identify and resolve data-related issues.. If applicable, contribute to . Azure-based data solutions. , including ADF, Synapse, ADLS, and other services.. Support . cloud migration initiatives. and DevOps practices, if relevant to the role.. Provide guidance on best practices and mentor junior team members when needed.. Qualifications. 8+ years. of overall experience working with cross-functional teams (machine learning engineers, developers, product managers, analytics teams).. 3+ years. of hands-on experience developing and managing . data pipelines using PySpark. .. Strong programming skills in . Python and SQL. .. Deep understanding of . ETL processes. and . data warehousing fundamentals. .. Self-driven, resourceful, and comfortable working in dynamic, fast-paced environments.. Advanced written and spoken Engllish is must have for this position (B2, C1 or C2 only).. Nice to have:. Databricks certification.. Experience with Azure-native services, including: Azure Data Lake Storage (ADLS), Azure Data Factory (ADF), Azure Synapse Analytics / Azure SQL DB / Fabric.. Familiarity with . Event Hub. , . IoT Hub. , . Azure Stream Analytics. , . Azure Analysis Services. , and . Cosmos DB. .. Basic understanding of . SAP HANA. .. Intermediate-level experience with . Power BI. .. Knowledge of . DevOps. , . CI/CD pipelines. , and . cloud migration best practices. .. Additional Information. Please note that we will not be moving forward with any applicants who do not meet the following mandatory requirements:. . 3+ years of experience with PySpark/Python, ETL and datawarehousing processes.. Proven leadership experience in a current project or previous projects/work experiences.. Advanced written and spoken English fluency is a MUST HAVE (from B2 level to C1/C2). MUST BE located in Central or South america, . as this is a nearshore position . (Please note that we are not able to consider candidates requiring relocation or those located offshore).. . More Details:. Contract type:. Independent contractor (This contract does not include PTO, tax deductions, or insurance. It only covers the monthly payment based on hours worked).. Location:. The client is based in the United States; however, the position is 100% remote for nearshore candidates located in Central or South America.. Contract/project duration: . Initially 6 months, with extension possibility based on performance.. Time zone and working hours:. Full-time, Monday to Friday (8 hours per day, 40 hours per week), from 8:00 AM to 5:00 PM PST (U.S. time zone).. Equipment:. Contractors are required to use their own laptop/PC.. Start date expectation:. As soon as possible.. Payment methods:. International bank transfer, PayPal, Wise, Payoneer, etc.. . Bertoni Process Steps:. Requirements verification video interview.. Technical interview. . Partner/Client Process Steps:. CV review.. 1 Technical video interview with our partner.. 1 or 2 video interviews with the end client.. . Why Join Us?. Be part of an innovative team shaping the future of technology.. Work in a collaborative and inclusive environment.. Opportunities for professional development and career growth.. .