Databricks Data Architect || 100% Remote

Remote Full-time
Position :: Databricks Data Architect Location :: 100% Remote Duration :: 12+ months Interview :: Video Job Description: It’s more of a hands-on architect who has 3+ years with Databricks and can still speak to being hands on with Pyspark, Python, SQL etc. Notes: Remote In an Azure infrastructure implementation Databricks data engineering experience and we are now looking and evaluating -we think it’s a great solution and slowly migrate SQL server systems into databricks Job Description: Position Summary This role is ideal for someone who enjoys building well-architected data solutions, thrives in a modern cloud analytics environment, and wants to influence a next-generation Lakehouse platform. The engineer will play a central role in shaping data quality, architecture, and analytic capabilities across multiple domains. We are seeking a skilled Databricks Data Engineer to help build and optimize a modern data platform leveraging the Databricks Lakehouse architecture. This role focuses on designing reliable ETL/ELT pipelines, implementing data governance within Unity Catalog, and enabling high-quality analytics products for internal and client-facing use. Key Responsibilities o Design and develop data pipelines using Databricks (Python, SQL, Delta Live Tables, Workflows) to ingest, transform, and curate data across bronze, silver, and gold layers. o Implement and enforce governance standards within Unity Catalog, including catalog/schema/volume organization, permissions, lineage, and metadata management. o Build and optimize Delta Lake tables, including performance tuning, schema evolution strategies, audit patterns, and CDC approaches (e.g., Change Tracking, timestamp deltas). o Collaborate closely with architects and data teams to translate business requirements into scalable data models and operational pipelines. o Develop reusable frameworks and best practices for ingestion, orchestration, DevOps, monitoring, and quality validation. o Support CI/CD workflows for Databricks assets via Git-based development and automated deployment pipelines. o Ensure reliability and observability, including logging, testing, and performance diagnostics across the platform. o Contribute to cloud architecture decisions, such as storage design, identity management, and compute strategies (jobs, clusters, SQL warehouses). Required Skills & Experience o 3+ years of experience in data engineering, preferably in a cloud environment (Azure, AWS, or Google Cloud Platform). o Hands-on experience with Databricks (Delta Lake, Workflows, DLT, Unity Catalog). o Strong proficiency in SQL and Python for data transformation and pipeline development. o Solid understanding of ETL/ELT patterns, medallion architecture, data modeling, and schema design. o Experience with CI/CD using Git-based workflows. o Working knowledge of cloud storage technologies (ADLS, S3, S). o Familiarity with table-level governance, permissions, and security models in modern data platforms. Preferred Qualifications o Experience integrating SQL Server Change Tracking, timestamp-based incremental ingestion, or similar CDC patterns. o Exposure to Terraform or IaC for Databricks governance and infrastructure configuration. o Understanding of analytics ecosystems such as Power BI, Tableau, or Business Objects. o Background working with domain-driven data models (e.g., CEDS, education data, ERP/operational domains). o Experience building multi-tenant data products or securely partitioned data architectures. Soft Skills o Strong problem-solving mindset and ability to work in fast-moving, collaborative environments. o Clear, proactive communication that supports cross-team alignment. o A commitment to documentation, quality, and repeatable processes. Apply tot his job
Apply Now

Similar Opportunities

Specialist, Health Information Management (DaVita IKC)

Remote

Analytics and Reporting Sr Analytics (IKC)

Remote

Integrated Care Assistant - Davita IKC - Eastern Time Zone Required

Remote

DaVita RN Case Manager - Hybrid/Remote, Pittsburgh, PA in Pittsburgh, Pennsylvania

Remote

Dialysis Technician Float

Remote

Hartford Full-time Role:Entry Level Flight Attendant

Remote

Artificial Intelligence Scientist, II

Remote

Sr. Process Consultant of Artificial Intelligence

Remote

Cloud Machine Learning Engineer - US remote

Remote

Machine Learning Engineer; L4 Remote - India

Remote

Home-Based Data Collection Clerk (Hiring Immediately)

Remote

Virtual Care Client Service Representative

Remote

Senior Application Security Engineer, Corporate Information Security- Remote (Anywhere in the U.S.)

Remote

Data Entry Clerk - Entry Level Opportunity to Kickstart Your Career at blithequark

Remote

Experienced or Entry-Level Remote Data Entry Specialist – Flexible Part-Time or Full-Time Opportunities for Career Growth and Development at arenaflex

Remote

Experienced or Entry-Level Remote Data Entry Specialist – Flexible Work Arrangements and Professional Growth Opportunities at arenaflex

Remote

Experienced Part-Time Data Entry Specialist - Flexible Remote Opportunity with blithequark

Remote

B2B Mid-Market AE (Remote, USA)

Remote

Remote Data Entry Specialist - Work From Home Opportunity with Flexible Hours and Competitive Pay

Remote

Senior Scientific and Regulatory Affairs Manager (Remote) in Boca Raton, FL

Remote
← Back to Home