Data Engineer
Client: NDA (USA)
Location: remote (LATAM & Europe)
Employment: full-time (40 hours per week)
Our client is a well-known US-based rewards marketplace that helps users earn for engaging with digital offers, while delivering strong value to advertisers through performance-driven solutions. We are looking for a skilled and detail-oriented Data Engineer who will play a key role in building and maintaining scalable data infrastructure within the AWS ecosystem.
The ideal candidate has solid hands-on experience with Apache Airflow, Snowflake, and Python, and knows how to design efficient, fault-tolerant pipelines. This role is perfect for someone who enjoys solving complex data challenges, optimizing workflows, and collaborating across product and engineering teams.
Requirements:
- 3–5 years of proven experience as a Data Engineer, with strong expertise in designing and maintaining production-grade data pipelines.
- 2+ years of experience with Apache Airflow (MWAA), including building and debugging complex DAGs, and applying best practices for error handling and idempotency.
- 2+ years of practical experience with Snowflake, including advanced SQL (MERGE, COPY INTO), query tuning, and orchestration tasks.
- 3+ years of experience with Python, including object-oriented programming, Pandas, and concurrent execution (e.g., ThreadPoolExecutor for parallel data processing).
- 3+ years of experience writing complex SQL queries (CTEs, window functions, stored procedures) and optimizing them for large datasets.
- 2+ years of experience working with AWS (S3, IAM, Lambda).
- Experience with Infrastructure as Code tools (Terraform or CloudFormation) is a strong plus.
- Strong understanding of ETL/ELT principles, data modeling (star schema, slowly changing dimensions), and data quality management.
- Experience troubleshooting production data issues, including debugging race conditions, performance bottlenecks, and data inconsistencies.
- Excellent communication skills, attention to detail, and a proactive, team-oriented mindset.
Nice to have:
- Experience with marketing and mobile attribution data sources (e.g., Adjust, Singular, AppsFlyer).
- Familiarity with Looker or other BI visualization tools.
- Experience with ETL/Reverse ETL platforms (Hevo, Hightouch, Census).
- Knowledge of dbt or Great Expectations.
- Certifications in Snowflake, AWS, or Airflow will be a plus.
Responsibilities:
- Design, develop, and maintain scalable ETL/ELT pipelines using Python, Apache Airflow (MWAA), and Snowflake.
- Build data ingestion workflows from multiple sources, including AWS S3, Aurora MySQL, MongoDB, BigQuery, and third-party APIs (Singular, AppsFlyer, Google Ads, Facebook Ads, TikTok Ads).
- Implement processes for data validation, cleansing, and deduplication to ensure high accuracy and reliability.
- Apply a declarative and idempotent approach to pipeline design to support reliable backfills and recovery.
- Set up monitoring, logging, and alerting systems (e.g., Slack notifications) for data operations and pipeline performance tracking.
- Collaborate with Data, Product, and Engineering teams to translate business needs into reliable technical solutions.
- Continuously optimize data workflows and SQL queries for performance and cost efficiency.
- Ensure adherence to data quality and governance standards across all processes.
We offer:
- Remote work in an international company in the USA.
- Competitive salary in the USD.
- Flexible working hours to help you manage your work/life balance.
- Fully remote service/support
- Career and professional growth.
- Warm and friendly attitude to every specialist.