Skip to Main Content

Job Title


Data Engineer


Company : DataKryptonAI


Location : Hosur, Tamil Nadu


Created : 2025-07-26


Job Type : Full Time


Job Description

Job Title:Data Engineer Location:India (Remote) Department:Data & AnalyticsAbout the RoleWe’re seeking a skilled Data Engineer to join our growing Data & Analytics team. In this role, you’ll design, build, and maintain scalable data pipelines and architectures that power critical business insights. You’ll work closely with data scientists, analysts, and business stakeholders to ensure that data is reliable, accessible, and performant.Key Responsibilities •Data Pipeline Development: Design, implement, and optimize ETL/ELT pipelines using Python, SQL, and Databricks to ingest, transform, and load data from multiple sources. •Data Modeling: Develop and maintain dimensional and normalized models (star/snowflake schemas) in a cloud data lake / warehouse environment. •Platform Engineering: Build and manage scalable data processing frameworks (e.g., Spark on Databricks), ensuring high availability, monitoring, and performance tuning. •Data Quality & Governance: Implement data quality checks, validation rules, and monitoring in pipelines; collaborate with Data Governance to enforce standards and best practices. •Collaboration & Support: Partner with data scientists, BI developers, and business analysts to understand requirements, troubleshoot issues, and deliver reliable data solutions. •Documentation & Automation: Create clear technical documentation, maintain code in version control, and automate deployment processes (CI/CD).Required Qualifications: •Programming: 3+ years of professional experience with Python for data engineering (pandas, PySpark, etc.). •SQL Expertise: Strong SQL skills, including complex joins, window functions, query optimization, and performance tuning in a cloud environment. •Databricks: Hands-on experience building and deploying Spark-based data pipelines on Databricks (jobs, notebooks, clusters, Delta Lake). •Cloud Platforms: Familiarity with at least one major cloud provider (Azure, AWS, or GCP) and their data ecosystem (e.g., Azure Data Factory, AWS Glue, GCP Dataflow). •Data Modeling: Solid understanding of data modeling principles and best practices (dimensional modeling, Data Vault, medallion architecture). •Version Control: Proficient with Git (branching strategies, pull requests, code reviews). •Agile Practices: Experience working in Scrum or Kanban teams, comfortable with iterative delivery and backlog grooming. •Communication: Excellent written and verbal communication skills; ability to present technical concepts to non-technical stakeholders.Preferred Qualifications: •Experience with orchestration tools (Airflow, Azure Data Factory, Prefect). •Familiarity with YAML/JSON configuration for pipeline parameterization. •Knowledge of data quality tools (e.g., Soda, Great Expectations, DQX). •Understanding of CI/CD for data infrastructure (Azure DevOps, Jenkins, GitHub Actions). •Exposure to real-time data streaming technologies (Kafka, Event Hubs). •Bachelor’s or Master’s degree in Computer Science, Engineering, or related field.What We Offer •Competitive salary and comprehensive benefits package. •Opportunities for professional development, certification support, and conference attendance. •Collaborative, inclusive culture with focus on innovation and continuous improvement. •Flexible work arrangements and wellness programs.