Skip to Main Content

Job Title


Senior Data Engineer - Scala (Azure Databricks)


Company : Concentrix


Location : Pune, Maharashtra


Created : 2026-04-10


Job Type : Full Time


Job Description

We’re Concentrix. A new breed of tech company — Human-centered. Tech-powered. Intelligence-fueled.We create game-changing solutions across the enterprise, that help brands grow across the world and into the future. We are trusted by clients across all major sectors, from up-and-coming success stories to iconic Fortune Global 500 brands in over 70 countries spanning 6 continents.Our game-changers:* Challenge Conventions* Deliver outcomes unimagined* Create experiences that go beyond WOWIf this is you, we would love to discuss career opportunities with you.In our Information Technology and Global Security team, you will deliver the latest technology infrastructure, transformative software solutions and industry-leading global security for our staff and clients. You will work with the best in the world to design, implement and strategize IT, security, application development, innovation, and solutions in today’s hyperconnected world. You will be part of the technology team that is core to our vision of develop, build and run the future of CX.Concentrix provides eligible employees with an opportunity to enroll in many benefit programs, generally including private medical plans, great compensation package, retirement savings plans, paid learning days, and flexible workplaces. Specific benefits plans will vary by country/region.We’re a remote-first company looking for the absolute best talent in the world. Experience the power of a game-changing career.Sr. Data Engineer – Scala (Azure Databricks) | Streaming (Event Hubs/Kafka) | SQLJob DescriptionThe Data Engineer – Scala (Azure Databricks) is responsible for re-engineering and modernizing an existing Scala-based Databricks (DBX) ecosystem into another Scala-based ecosystem, with a strong focus on performance and cost optimization. The role includes building and optimizing batch and streaming data pipelines on Azure Databricks, leveraging Scala with foundational SQL, and delivering production-grade implementations based on validated solution approaches/POCs. This position will collaborate closely with engineering and platform teams to ensure scalable, reliable, and efficient data processing. Employment Type: Full-Time (CNX)Work Location: Remote (CNX location)Shift: No constraints as of nowEssential Functions/Core Responsibilities• Re-engineer/migrate an existing Scala-based Databricks ecosystem to a new Scala-based ecosystem, ensuring functional parity and improved efficiency.• Develop and maintain Spark pipelines on Azure Databricks using Scala; apply foundational SQL for transformations, validations, and analytical preparation.• Build streaming ingestion and processing pipelines from sources such as Azure Event Hubs and/or Kafka (design, develop, test, and operationalize).• Perform Spark performance optimization using Spark UI: analyze stages/tasks, identify skew/shuffles, optimize joins, caching, partitioning, file sizing, and cluster usage.• Drive cost optimization initiatives by tuning cluster configurations, job orchestration, data layout (e.g., partitioning strategy), and compute patterns.• Troubleshoot production issues (pipeline failures, latency, throughput, data quality) and implement preventive controls/observability in collaboration with platform/SRE.• Follow engineering best practices: modular/reusable code, code reviews, version control, documentation, and CI/CD-aligned development.• Work with stakeholders to refine requirements, estimate effort, and deliver milestones aligned to project plans. Candidate Profile (Required)• 6–8 years of experience in Data Engineering / Big Data / Cloud Data platforms.• Strong hands-on expertise in Scala for Spark/Databricks development (core requirement).• Experience on Azure Databricks (development + job execution patterns).• Foundational to strong SQL skills (complex queries, joins, window functions, optimization basics).• Strong understanding of Spark internals and performance tuning, including hands-on use of Spark UI for bottleneck analysis and remediation.• Experience implementing streaming ingestion/processing using Event Hubs and/or Kafka (Structured Streaming preferred).• Strong problem-solving skills and ability to work effectively in a remote setup. Candidate Profile (Good-to-Have / Preferred)• Azure data services exposure (e.g., ADLS Gen2, ADF, Key Vault) as part of a data platform ecosystem.• Delta Lake / Lakehouse patterns and optimization techniques (e.g., OPTIMIZE/ZORDER concepts where applicable).• CI/CD for Databricks (Azure DevOps/GitHub), repo-based deployments, and automated testing practices.• Java exposure (nice to have; not mandatory). • Experience supporting security/governance integration with platform teams (e.g., access controls, audit logging concepts) where applicable. Education• Bachelor’s degree in Computer Science/Engineering or equivalent practical experience. Career Level Description• Works independently on medium to complex tasks with limited supervision.• Applies judgment in selecting methods/techniques to obtain solutions.• Owns deliverables for assigned components and collaborates across teams to resolve dependencies and risks. Work Environment / Additional Information• Remote role aligned to CNX location requirements.• No shift constraints currently; flexibility may be required based on project needs.