← All Jobs
Posted May 4, 2026

Data Engineer with Data Build Tool :: W2 Position :: Remote

Apply Now
Job Summary: We are seeking a highly experienced Senior Data Engineer with 12+ years of hands-on experience in data engineering and data pipeline development. The ideal candidate will have strong expertise in Databricks, DBT, AWS, and Apache Spark, and a proven track record of working with modern data platforms in large-scale enterprise environments. This role requires a professional who is adept at building scalable data pipelines, transforming data using DBT, working extensively with AWS cloud infrastructure, and developing in a Spark/Databricks environment. Strong SQL query writing skills are essential. Mandatory Skills & Experience: • 10+ years of experience in data engineering or relevant fields. • Expert-level knowledge of Databricks – including notebooks, jobs, Delta Lake, and workspace management. • Strong hands-on experience with DBT (Data Build Tool) – including writing models, tests, and documentation. • Proficiency in AWS services – including S3, Glue, Redshift, Lambda, IAM, etc. • Strong experience with Apache Spark – both batch and streaming data processing. • Strong experience writing efficient SQL queries for data transformation and analysis. • Proven ability to work independently in a remote environment. • Solid understanding of data modeling, ETL/ELT processes, and modern data architecture. • Experience in version control (Git), CI/CD pipelines, and Agile methodologies. Preferred Qualifications: • AWS Certification (e.g., AWS Data Analytics, AWS Solutions Architect) • Experience with orchestration tools like Airflow or AWS Step Functions. • Familiarity with DevOps practices in a data environment. • Exposure to data quality frameworks and testing strategies. Responsibilities: • Design, build, and optimize robust and scalable data pipelines using Databricks and DBT on AWS. • Develop, deploy, and maintain ELT/ETL processes to transform raw data into actionable insights. • Collaborate with data analysts, data scientists, and business stakeholders to understand requirements and deliver data solutions. • Ensure data integrity, quality, and governance across all data flows. • Continuously monitor and improve performance, reliability, and cost-efficiency of data pipelines. • Write and maintain high-quality documentation. Apply Now Apply Now
Interested in this role?Apply on iHire