Senior Data Engineer Databricks

Job description

Position Overview

At Hypersonix, we are building the leading Generative AI Platform for Commerce. Our Flagship GenAI Product – Competitor + Pricing AI – scrapes the product catalogue for our Enterprise customers and their competitors, and uses RAG to identify the nearest competitive match for each of our customer’s product, facilitating intelligent pricing strategies that were previously impossible to achieve.

We are seeing strong growth in our Enterprise product, and are building out an end-to-end product on Databricks for Shopify Store owners, specializing in Agentic workflows that automate critical business processes (Pricing + Promotion Strategies, Inventory Management, and Competitive Intelligence).

We are seeking an experienced Senior Data Engineer to design, build, and optimize scalable data pipelines and infrastructure. The ideal candidate will have deep expertise in Databricks and modern data engineering practices, with a strong focus on building robust, production-grade data solutions that drive business value while maintaining cost efficiency.

Key Responsibilities:

Data Platform Development

Design and implement enterprise-scale data pipelines using Databricks on AWS, leveraging both cluster-based and serverless compute paradigms

Architect and maintain medallion architecture (Bronze/Silver/Gold) data lakes and lakehouses

Develop and optimize Delta Lake tables for ACID transactions and efficient data management

Build and maintain real-time and batch data processing workflows

Engineering Excellence

Create reusable, modular data transformation logic using DBT to ensure data quality and consistency across the organization

Develop complex Python applications for data ingestion, transformation, and orchestration

Write optimized SQL queries and implement performance tuning strategies for large-scale datasets

Implement comprehensive data quality checks, testing frameworks, and monitoring solutions

Cost Management & Optimization

Monitor and analyze Databricks DBU (Databricks Unit) consumption and cloud infrastructure costs

Implement cost optimization strategies including cluster right-sizing, autoscaling configurations, and spot instance usage

Optimize job scheduling to leverage off-peak pricing and minimize idle cluster time

Establish cost allocation tags and chargeback models for different teams and projects

Conduct regular cost reviews and provide recommendations for efficiency improvements

DevOps & Infrastructure

Design and implement CI/CD pipelines for automated testing, deployment, and rollback of data artifacts

Configure and optimize Databricks clusters, job scheduling, and workspace management

Implement version control best practices using Git and collaborative development workflows

Collaboration & Leadership

Partner with data analysts, data scientists, and business stakeholders to understand requirements and deliver solutions

Mentor junior engineers and promote best practices in data engineering

Document technical designs, data lineage, and operational procedures

Participate in code reviews and contribute to team knowledge sharing

Required Qualifications:

Technical Skills

5+ years of experience in data engineering roles

Expert-level proficiency in Databricks (Unity Catalog, Delta Live Tables, Workflows, SQL Warehouses)

Strong understanding of cluster configuration, optimization, and serverless SQL compute

Advanced SQL skills including query optimization, indexing strategies, and performance tuning

Production experience with DBT (models, tests, documentation, macros, packages)

Proficient in Python for data engineering (PySpark, pandas, data validation libraries)

Hands-on experience with Git workflows (branching strategies, pull requests, code reviews)

Proven track record implementing CI/CD pipelines (Jenkins, GitLab CI)

Working knowledge of Snowflake architecture and migration patterns

Additional Technical Skills:

Experience with Apache Spark and PySpark optimization techniques (caching, partitioning, broadcast joins)

Understanding of data modeling concepts (dimensional modeling, data vault, normalization)

Knowledge of orchestration tools (Airflow, Databricks Workflows)

Familiarity with cloud platforms (AWS, Azure, or GCP) and their data services

Experience with data governance , security frameworks and SOC2

Ability to use Databricks system tables and monitoring tools for cost analysis

Professional Skills:

Strong problem-solving abilities and analytical thinking

Excellent communication skills with technical and non-technical audiences

Ability to work independently and drive projects to completion

Experience with Agile/Scrum methodologies

Cost-conscious mindset with ability to balance performance and budget constraints

Preferred Qualifications:

Databricks certifications (Data Engineer Associate/Professional, Platform Administrator)

Knowledge of data quality frameworks (Great Expectations, Deequ)

Experience with container technologies (Docker, Kubernetes)

Familiarity with Lakehouse architecture patterns and best practices

Experience migrating from traditional data warehouses to modern data platforms

Nice to Have:

Familiarity with e-commerce and retail domain

Experience with reverse ETL and data activation tools

Knowledge of data catalog and metadata management tools

Track record of achieving significant cost reductions (20%+) in cloud data platforms

Experience with data mesh or data fabric architectural patterns

Familiarity with Power BI, Tableau, or other BI tools

Share this job:
Please let Hypersonix know you found this job on Remote First Jobs 🙏

Similar Remote Jobs

Find Remote Jobs

Connect with top companies hiring for remote jobs, work-from-home roles, and 100% online jobs worldwide.

Discover Hidden Jobs

Unique jobs you won't find on other job boards.

Advanced Filters

Filter by category, benefits, seniority, and more.

Priority Job Alerts

Get timely alerts for new job openings every day.

Manage Your Job Hunt

Save jobs you like and keep a simple list of your applications.

Apply