Data Engineer

  • Closed
  • US Company | Small (11-50 employees)
  • LATAM (100% remote)
  • 3+ years
  • Long-term (40h)
  • Healthcare
  • Full Remote

Required skills

  • SQL
  • BigQuery
  • Dataflow
  • Python
  • dbt
  • Dataform
  • Google Cloud Platform

Requirements

Must-haves

  • 3-5 years of data engineering experience
  • Experience with GCP Ecosystem (BigQuery, Dataflow, Pub/Sub, Cloud Functions)
  • Proficiency with dbt or Dataform
  • Proficiency with Python for data manipulation and automation
  • Ability to write and optimize advanced SQL queries for large-scale datasets
  • Deep knowledge of Dimensional Modeling (Star/Snowflake schemas)
  • Deep knowledge of modern variations like Data Vault or Wide Tables for ML
  • Strong communication skills in both spoken and written English

Nice-to-haves

  • Startup experience
  • Experience with CI/CD pipelines, Terraform, Infrastructure as Code, Docker, Devcontainers
  • Interest in MLOps and model deployment using Vertex AI or feature stores
  • Ability to contribute using Golang, Java, and Kotlin
  • Experience with OpenLineage, nix/flakes, uv, poetry
  • Bachelor’s Degree in Computer Engineering, Computer Science, or equivalent

What you will work on

  • Lead the evolution and continuous improvement of a cloud-based data platform
  • Architect scalable data solutions on Google Cloud to support machine learning workflows and self-service BI
  • Own and optimize the BigQuery environment with a focus on performance, cost efficiency, and analytics usability
  • Data architecture and pipeline ownership
  • Design, build, and maintain scalable ETL and ELT pipelines using GCP-native services
  • Develop and optimize streaming and batch pipelines with Dataflow (Apache Beam) and Pub/Sub
  • Automate and orchestrate data workflows using Airflow (Cloud Composer), SqlMesh, or Temporal
  • Performance and cost optimization
  • Implement advanced BigQuery strategies including partitioning, clustering, and materialized views
  • Monitor and reduce cloud costs through SQL optimization and use of BigQuery BI Engine
  • Data governance and quality engineering
  • Implement automated data quality checks including schema validation, deduplication, and anomaly detection
  • Design row-level security and column-level encryption to meet HIPAA-level compliance requirements
  • Maintain metadata, lineage, and data catalogs using OpenLineage or Dataplex
  • Strategic collaboration and enablement
  • Partner with data analysts and product teams to ensure clean, analytics-ready data at ingestion
  • Support high-performance reporting for BI tools (Tableau, Looker)
  • Mentor junior team members and contribute to shared best practices for SQL and Python

Other jobs you might like

  • Senior Data Engineer & Analyst

    • LATAM (100% Remote)
    • US Company | 11-50 employees
    • 6+ years of exp.
    • Long-term (40h)
  • Solutions Engineer

    • LATAM (100% Remote)
    • US Company | 51-250 employees
    • 2+ years of exp.
    • Long-term (40h)

Get matched with the best remote opportunities from today's top US companies

Find great opportunities

Earn more compensation for your hard work

Access exclusive benefits like healthcare, English classes, and more

1-1 individualized training to succeed in the international job market

Sign up