Job Description

πŸš€ Data Engineer (Python, SQL, ETL, Airflow, Snowflake, BigQuery)

Full-Time | Remote | U.S. Business Hours

πŸ’‘ About the Role

We’re hiring a highly technical Data Engineer to build and maintain scalable data pipelines, cloud data infrastructure, and analytics-ready datasets that power business decision-making.

This role is focused on:

βœ… ETL/ELT pipeline development

βœ… Data warehouse architecture

βœ… SQL optimization

βœ… Cloud-based data infrastructure

βœ… Pipeline reliability & monitoring

βœ… Scalable analytics systems

You’ll work closely with:

  • Data Analysts
  • Data Scientists
  • Engineering Teams
  • BI & Leadership Teams

to ensure the organization always has accurate, clean, and trustworthy data.

If you:

  • enjoy building robust data systems,
  • love optimizing pipelines and queries,
  • and care deeply about data quality and scalability,

this role is a strong fit.

πŸ”₯ What You’ll Own

ETL / ELT Pipeline Development

  • Build and maintain scalable ETL/ELT pipelines using:

    • Python
    • SQL
    • Scala
  • Ingest data from:

    • APIs
    • SaaS platforms
    • relational databases
    • cloud applications
    • streaming systems
  • Develop reliable workflows for:

    • data extraction
    • transformation
    • loading
    • validation

Workflow Orchestration & Automation

  • Manage orchestration platforms such as:

    • Apache Airflow
    • Prefect
    • Dagster
    • Luigi
  • Monitor:

    • pipeline health
    • failed jobs
    • scheduling reliability
  • Build automated workflows with:

    • retries
    • alerting
    • dependency management

Data Warehousing & Modeling

  • Design and optimize cloud data warehouses using:

    • Snowflake
    • BigQuery
    • Redshift
  • Develop:

    • star schemas
    • snowflake schemas
    • analytics-ready data models
  • Improve:

    • query performance
    • clustering
    • partitioning
    • warehouse efficiency

Data Quality & Governance

  • Implement:

    • validation checks
    • anomaly detection
    • logging systems
    • lineage tracking
  • Use tools such as:

    • dbt
    • Great Expectations
  • Ensure:

    • consistent naming conventions
    • clean transformations
    • audit-ready datasets
  • Support compliance requirements:

    • GDPR
    • HIPAA
    • industry-specific governance standards

Streaming & Real-Time Data

  • Build and maintain streaming pipelines using:

    • Kafka
    • Kinesis
    • Pub/Sub
  • Support:

    • real-time ingestion
    • event-driven processing
    • low-latency analytics workflows

Infrastructure & DevOps

  • Containerize services using:

    • Docker
    • Kubernetes
  • Build CI/CD workflows with:

    • GitHub Actions
    • Jenkins
    • GitLab CI
  • Manage cloud infrastructure using:

    • Terraform
    • CloudFormation
  • Improve scalability, reliability, and deployment automation

Cross-Functional Collaboration

  • Partner with:

    • analysts
    • data scientists
    • BI teams
    • product teams
  • Deliver curated datasets for:

    • dashboards
    • analytics
    • machine learning workflows
  • Support BI tools such as:

    • Tableau
    • Looker
    • Power BI
  • Maintain documentation for:

    • pipelines
    • schemas
    • workflows
    • data definitions

βœ… Required Experience & Skills

  • 3+ years of Data Engineering or backend engineering experience

  • Strong proficiency with:

    • Python
    • SQL
  • Experience with:

    • Snowflake
    • BigQuery
    • Redshift
  • Familiarity with:

    • Airflow
    • Prefect
    • workflow orchestration tools
  • Strong understanding of:

    • ETL pipelines
    • data modeling
    • cloud infrastructure
    • warehouse optimization

⭐ Ideal Experience

  • Experience using:

    • dbt
    • Great Expectations
    • data lineage tools
  • Streaming experience with:

    • Kafka
    • Kinesis
    • Pub/Sub
  • Experience with:

    • AWS Glue
    • GCP Dataflow
    • Azure Data Factory
  • Background in:

    • healthcare
    • fintech
    • regulated environments
  • Experience optimizing large-scale warehouse costs and performance

🧠 What Makes You a Great Fit

  • You care deeply about clean and reliable data
  • You enjoy debugging complex pipeline and infrastructure issues
  • You think about scalability and long-term maintainability
  • You combine engineering rigor with analytical thinking
  • You communicate effectively across technical and non-technical teams

πŸ“… What a Typical Day Looks Like

  • Review Airflow/Prefect pipeline health and resolve failures
  • Build connectors for new APIs or SaaS platforms
  • Optimize SQL queries and warehouse performance
  • Collaborate with analysts and data scientists on datasets
  • Improve validation and monitoring systems
  • Document pipelines and warehouse structures
  • Reduce warehouse costs and improve pipeline reliability

In short:

You build the data infrastructure that powers analytics, reporting, automation, and business intelligence across the organization.

πŸ“Š Key Success Metrics (KPIs)

  • Pipeline uptime β‰₯ 99%
  • Data freshness within SLA
  • Zero critical data quality issues reaching production
  • Query performance & warehouse cost optimization
  • Reliable and scalable pipeline infrastructure
  • Positive feedback from analysts, BI teams, and leadership

🌟 Why This Role Stands Out

  • Work on modern cloud-native data infrastructure

  • Build scalable ETL and analytics systems

  • Exposure to:

    • streaming pipelines
    • cloud data platforms
    • orchestration frameworks
    • warehouse optimization
  • Opportunity to grow into:

    • Senior Data Engineer
    • Analytics Engineering
    • Platform Engineering
    • Data Architecture
  • Fully remote flexibility with collaborative engineering teams

πŸ§ͺ Interview Process

  • Initial Phone Screen
  • Video Interview with Pavago Recruiter
  • Technical Task

(Build a small ETL pipeline or optimize a SQL query)

  • Client Interview with Engineering/Data Team
  • Offer & Background Verification

πŸ‘‰ Apply Now

If you:

  • love building scalable data systems,
  • enjoy solving complex pipeline problems,
  • and want to work with modern data infrastructure,

this role is a strong fit for you.

Share this job:
Please let Pavago know you found this job on Remote First Jobs πŸ™

19309 similar remote jobs

Explore latest remote opportunities and join a team that values work flexibility.

Remote companies like Pavago

Explore remote-first companies similar to Pavago. Discover other top-rated employers that offer flexible schedules and work-from-anywhere options.

Talent Right AUS Logo

Talent Right AUS

End-to-end recruitment and HR services for small and medium-sized Australian businesses.

View company profile β†’
TLNT Logo

TLNT

Connects businesses with remote talent from Mexico and Latin America, specializing in Marketing, Creative, Operations, and Tech roles.

View company profile β†’
Valatam Logo

Valatam

Connects U.S. and Canadian businesses with pre-vetted, full-time remote Virtual Assistants from Latin America.

View company profile β†’
SheWorks! Logo

SheWorks!

A digital platform connecting businesses with certified global remote talent and empowering women with remote work opportunities.

View company profile β†’
BEON.tech Logo

BEON.tech

51-200 beon.tech

Connecting LATAM tech talent

View company profile β†’
VALUE Virtual Assistants Logo

VALUE Virtual Assistants

Tailored virtual assistant solutions connecting businesses with Filipino talent.

View company profile β†’

Project: Career Search

Rev. 2026.5

[ Remote Jobs ]
Direct Access

We source jobs directly from 21,000+ company career pages. No intermediaries.

01

Discover Hidden Jobs

Unique jobs you won't find on other job boards.

02

Advanced Filters

Filter by category, benefits, seniority, and more.

03

Priority Job Alerts

Get timely alerts for new job openings every day.

04

Manage Your Job Hunt

Save jobs you like and keep a simple list of your applications.

21,000+ SOURCES UPDATED 24/7
Apply