Skip to Main Content

Job Title


Databricks Data Engineer


Company : Berkley


Location : Manassas, VA


Created : 2026-03-27


Job Type : Full Time


Job Description

We started in early 2019 as a small group of technologists with a passion for making insurance better. Today we are working with a team of industry experts who run five different insurance brands and collectively control $1 billion in annual premiums.We believe in an idea and execution meritocracy. In other words, a place where the best ideas win and the people who deliver the most value get the most opportunities.As we grow our team, we are looking for inquisitive, entrepreneurial people who are excited to reimagine the insurance industry.Insurance is too complex. Help us make it better.This position requires on-site work Monday-Thursday at either our Manassas, VA or Chesterfield, MO location.The Databricks Data Engineer will help design, build, deploy, and maintain scalable and production grade data pipelines in modern cloud environments, enabling analytics, AI, ML, and decision advantage at scale. This role will work with cutting-edge tools like Databricks, Delta Lake, PySpark, and AI/BI genie to transform raw data into actionable insights. As a hands-on Databricks Data Engineer with deep expertise in Azure Databricks and MLOps, this role will have the opportunity to migrate and translate legacy SSIS ETL logic into scalable, cloud-native data pipelines in Databricks. This role will partner with data engineers, data scientists, and product manager to design features, train/evaluate models, and deploy them to production using MLflow, Databricks and Workflows—with rigorous observability, governance (Unity Catalog), and CI/CD automation.Data Pipeline EngineeringDesign, build, and maintain high-performance, scalable ETL/ELT pipelines using Azure Databricks, Delta Lake, and PySpark.Convert and modernize existing SSIS package logic into cloud-native Databricks pipelines using PySpark notebooks, Delta Live Tables (DLT), and Databricks Workflows.Implement reliable batch and streaming pipelines with robust data quality and validation frameworks.Optimize pipeline performance using Photon, efficient file formats, partitioning, Z-ordering, and caching strategies.Lakehouse Platform DevelopmentDevelop and manage datasets within Delta Lake, ensuring ACID reliability, schema evolution, versioning, and time travel.Architect feature-rich data layers including:Bronze (raw ingestion)Silver (validated, conformed)Gold (analytics-ready and ML-ready)Implement data governance using Unity Catalog for fine-grained access control, lineage, auditability, and metadata management.MLOps & ML-Enabled Data PipelinesPartner with data scientists and data engineers to create feature pipelines, model training pipelines, and production scoring pipelines.Deploy and operationalize models using MLflow, Databricks Model Registry, and Databricks Workflows.Use Databricks built-in AI SQL functions such as ai_query, ai_forecast, ai_analyze_sentiment to generate actionable insight from large amount of unstructured or structured raw dataImplement monitoring for:Pipeline failuresData/feature driftModel performance degradationOperational SLAs/SLIs/SLOsBuild automated CI/CD workflows using GitHub Actions or Azure DevOps for notebook deployment, pipeline testing, and environment promotion.Data Platform, Data Security & Data GovernanceCollaborate with data engineers to design reliable data products on Delta Lake; leverage Delta Live Tables (DLT) for declarative pipelines when applicable.Enforce Unity Catalog for lineage, permissions, and audit; manage secrets, tokens, and keys securely (e.g., Databricks secrets, Key Vault/Secrets Manager).Collaboration & LeadershipWork closely with cross-functional teams: data engineering, data scientist, product manager, and business stakeholders.Serve as a Databricks SME—championing best practices, code standards, governance, and reusable frameworks.Document architecture, workflows, data models, runbooks, and operational procedures.Minimum of 3 years of experience in Databricks, PySpark notebooks, Python, DevOps, software development, and data engineering.Certified Databricks Data Engineer Associate or Professional is a plus.Skills & CompetenciesProficient in designing, building, deploying, and maintaining high-performance, scalable ETL/ELT pipelines using Azure Databricks, Delta Lake, and PySpark Notebook.Proficient in building, deploying, and operating production ML models such as supervised, unsupervised, and anomaly detection, including techniques for imbalanced datasetsProficient with ML engineering and MLOps, including model versioning, CI/CD for ML, monitoring, drift detection, and automated retrainingProficiency in Python including Pandas and PySpark DataframesExpert level of SQL skills including Stored Procedure, experience with SSIS, SSRS, Power BI is a plus.Proficient with cloud data engineering platforms, such as Azure, Databricks, Spark, or SQL, and batch and streaming pipelinesFamiliar with Databricks AI Built-In Functions such as AI_Query, AI_Gen, AI_Classify, AI_Forecast, AI_Analyze_Sentiment, able to use them to extract actionable insights from large amount of unstructured or structured raw dataExperience with Python and ML frameworks, such as PyTorch or TensorFlowExperience improving data quality, lineage, and observability in enterprise data environments and operationalizing rules and model-driven scoring for prioritization, routing, or case selectionExperience with predictive analytics, machine learning and artificial intelligence desired.EducationA Bachelor’s degree in Computer Science, Management Information Systems, Engineering, Math, Physics, or a related quantitative field is required (4-year degree). Master’s degree preferredExperience in the commercial insurance industry is a plus.The Company is an equal employment opportunity employer.We do not accept any unsolicited resumes from external recruiting firms.The company offers a competitive compensation plan and robust benefits package for full time regular employees.Base salary & Benefits include Health, dental, vision, life, disability, wellness, paid time off, 401(k) and profit-sharing plans.The actual salary for this position will be determined by a number of factors, including the scope, complexity and location of the role; the skills, education, training, credentials and experience of the candidate; and other conditions of employment.• Ability to travel locally and nationally up to 5% of the timeSponsorship not Offered for this Role