Database Engineer

🇿🇦 South Africa - Remote
📊 Data🔵 Mid-level

Job description

Job Title: Database Engineer

Position Type: Full-Time, Remote

Working Hours: U.S. Business Hours (CST)

About the Role

Our client builds advanced data intelligence systems that map complex ownership, legal, and property relationships using graph technology. Neo4j powers their knowledge graph and enables highly connected data to be represented with accuracy, speed, and traceability.

We are seeking a Neo4j Database Engineer who will own the graph data model, optimize Cypher queries, and build strong data ingestion pipelines. This role is mission-critical in defining how the system structures, connects, and analyzes data across people, properties, documents, transactions, and events.

Key Responsibilities

1. Graph Data Modeling & Architecture

  • Design, refine, and maintain Neo4j data schemas
  • Ensure entity relationships are consistent, normalized, and high-performing
  • Implement schema updates aligned with business and technical needs

2. Ownership & Relationship Computation

  • Implement logic to compute/update ownership changes over time
  • Maintain transitions, transfers, and events through optimized structures

3. Cypher Query Design & Optimization

  • Build, test, and optimize Cypher queries for traversal, lineage, and aggregations
  • Develop Cypher scripts for migrations and large-scale operations
  • Use PROFILE/EXPLAIN and best practices to improve performance

4. Integration & Data Ingestion

  • Build reliable Python-based ingestion pipelines (neo4j-driver, py2neo)
  • Integrate Neo4j with APIs, SQL databases, and analytical systems
  • Collaborate with teams to onboard new data sources

5. Governance & Validation

  • Implement validation rules, constraints, and QA checks
  • Ensure auditability, version control, and schema compliance

Required Skills & Experience

1. Core Expertise

  • 3+ years hands-on with Neo4j (modeling + querying)
  • Strong SQL/NoSQL database design background

2. Querying

  • Expert-level Cypher (path queries, pattern comprehension, aggregations)

3. Data Modeling

  • Graph normalization, temporal modeling, event-sourced architectures

4. Programming

  • Strong Python skills for ingestion, automation, and processing

5. Nice to Have

  • Experience with Dockerized Neo4j, APOC, pipeline automation

What You’ll Do

  • Serve as the primary graph architect
  • Build scalable graph logic to compute complex real-world relationships
  • Contribute to an AI-driven intelligence platform foundation

Key Performance Indicators (KPIs)

1. Query Optimization

  • 20–30% reduction in Cypher execution time
  • Effective profiling and refactoring of traversals and aggregations

2. Data Model Quality

  • Zero critical integrity issues
  • On-time schema updates within sprint cycles
  • Efficient addition of new entity types

3. Pipeline Reliability

  • 99% success rate for Python-based ingestion
  • Minimal downtime or ingestion failures
  • Clean integration with external sources

4. Overall Delivery

  • Meeting milestones (ownership computation, migrations, validations)
  • Strong sprint velocity and low bug rates

5. Data Governance

  • Validation errors under 1%
  • Full compliance with constraints and auditability

Interview Process

  • Initial Phone Screen
  • Technical Interview with Pavago Recruiter
  • Practical Task (schema design or query optimization)
  • Final Client Interview
  • Offer & Background Verification
Share this job:
Please let Pavago know you found this job on Remote First Jobs 🙏

Benefits of using Remote First Jobs

Discover Hidden Jobs

Unique jobs you won't find on other job boards.

Advanced Filters

Filter by category, benefits, seniority, and more.

Priority Job Alerts

Get timely alerts for new job openings every day.

Manage Your Job Hunt

Save jobs you like and keep a simple list of your applications.

Search remote, work from home, 100% online jobs

We help you connect with top remote-first companies.

Search jobs

Hiring remote talent? Post a job

Apply