Data Engineer

Client: NDA (USA)
Location: remote (LATAM & Europe)
Employment: full-time (40 hours per week)


Our client is a well-known US-based rewards marketplace that helps users earn for engaging with digital offers, while delivering strong value to advertisers through performance-driven solutions. We are looking for a skilled and detail-oriented Data Engineer who will play a key role in building and maintaining scalable data infrastructure within the AWS ecosystem.

The ideal candidate has solid hands-on experience with Apache Airflow, Snowflake, and Python, and knows how to design efficient, fault-tolerant pipelines. This role is perfect for someone who enjoys solving complex data challenges, optimizing workflows, and collaborating across product and engineering teams.


Requirements:

  • 3–5 years of proven experience as a Data Engineer, with strong expertise in designing and maintaining production-grade data pipelines.
  • 2+ years of experience with Apache Airflow (MWAA), including building and debugging complex DAGs, and applying best practices for error handling and idempotency.
  • 2+ years of practical experience with Snowflake, including advanced SQL (MERGE, COPY INTO), query tuning, and orchestration tasks.
  • 3+ years of experience with Python, including object-oriented programming, Pandas, and concurrent execution (e.g., ThreadPoolExecutor for parallel data processing).
  • 3+ years of experience writing complex SQL queries (CTEs, window functions, stored procedures) and optimizing them for large datasets.
  • 2+ years of experience working with AWS (S3, IAM, Lambda). 
  • Experience with Infrastructure as Code tools (Terraform or CloudFormation) is a strong plus.
  • Strong understanding of ETL/ELT principles, data modeling (star schema, slowly changing dimensions), and data quality management.
  • Experience troubleshooting production data issues, including debugging race conditions, performance bottlenecks, and data inconsistencies.
  • Excellent communication skills, attention to detail, and a proactive, team-oriented mindset.


Nice to have:

  • Experience with marketing and mobile attribution data sources (e.g., Adjust, Singular, AppsFlyer).
  • Familiarity with Looker or other BI visualization tools.
  • Experience with ETL/Reverse ETL platforms (Hevo, Hightouch, Census).
  • Knowledge of dbt or Great Expectations.
  • Certifications in Snowflake, AWS, or Airflow will be a plus.


Responsibilities:

  • Design, develop, and maintain scalable ETL/ELT pipelines using Python, Apache Airflow (MWAA), and Snowflake.
  • Build data ingestion workflows from multiple sources, including AWS S3, Aurora MySQL, MongoDB, BigQuery, and third-party APIs (Singular, AppsFlyer, Google Ads, Facebook Ads, TikTok Ads).
  • Implement processes for data validation, cleansing, and deduplication to ensure high accuracy and reliability.
  • Apply a declarative and idempotent approach to pipeline design to support reliable backfills and recovery.
  • Set up monitoring, logging, and alerting systems (e.g., Slack notifications) for data operations and pipeline performance tracking.
  • Collaborate with Data, Product, and Engineering teams to translate business needs into reliable technical solutions.
  • Continuously optimize data workflows and SQL queries for performance and cost efficiency.
  • Ensure adherence to data quality and governance standards across all processes.


We offer:

  • Remote work in an international company in the USA.
  • Competitive salary in the USD.
  • Flexible working hours to help you manage your work/life balance.
  • Fully remote service/support
  • Career and professional growth.
  • Warm and friendly attitude to every specialist.