
Senior Data Engineer at BHFT. Location Information: Dubai, Dubai, United Arab Emirates. . Company Description. We are a proprietary algorithmic trading firm. Our team manages the entire trading cycle, from software development to creating and coding strategies and algorithms. We have a team of 200+ professionals, with a strong emphasis on technology—70% of our team is made up of technical specialists.. We operate as a fully remote organization, fostering a culture of transparency, clarity, and open communication. We are expanding into new markets and technologies, continuously innovating in the world of algorithmic trading.. Job Description. Key Responsibilities. Ingestion & Pipelines:. Architect batch + stream . pipelines. (Airflow, Kafka, dbt) for diverse structured and unstructured marked data. Provide reusable SDKs in Python and Go for internal data producers.. Storage & Modeling:. Implement and tune S3, column‐oriented and time‐series data storage for petabyte‐scale analytics; own partitioning, compression, TTL, versioning and cost optimisation.. Tooling & Libraries:. Develop internal libraries for schema management, data contracts, validation and lineage; contribute to shared libraries and services for internal data consumers for research, backtesting and real-time trading purposes.. Reliability & Observability. : Embed monitoring, alerting, SLAs, SLOs and CI/CD; champion automated testing, data quality dashboards and incident runbooks.. Collaboration: . Partner with Data Science, Quant Research, Backend and DevOps to translate requirements into platform capabilities and evangelise best practices.. Qualifications. Required Skills & Experience. 7 + years building production‐grade data systems.. Familiarity with market data formats (e.g., MDP, ITCH, FIX, proprietary exchange APIs) and market data providers.. Expert‐level Python (Go and C++ nice to have).. Hands‐on with modern orchestration (Airflow) and event streams (Kafka).. Strong SQL proficiency: aggregations, joins, subqueries, window functions (first, last, candle, histogram), indexes, query planning, and optimization.. Designing high‐throughput APIs (REST/gRPC) and data access libraries.. Strong Linux fundamentals, containers (Docker) and cloud object storage (AWS S3 / GCS).. Proven track record of mentoring, code reviews and driving engineering excellence.. Additional Information. What we offer:. Working in a modern international technology company without bureaucracy, legacy systems, or technical debt.. Excellent opportunities for professional growth and self-realization.. We work remotely from anywhere in the world, with a flexible schedule.. We offer compensation for health insurance, sports activities, and professional training.. .