Skip to Main Content

Job Title


Data Engineer


Company : Neuracore


Location : New delhi, Delhi


Created : 2026-03-24


Job Type : Full Time


Job Description

About UsAt Neuracore, we're building the world's first robot learning cloud service (platform eliminates the complexity of traditional robotics development by providing a complete end-to-end solution for data collection, model training, and deployment that works across different robot types and configurations.Our multidisciplinary team is at the forefront of making robot learning accessible to organisations worldwide, from manufacturing and logistics to healthcare and research institutions. We're transforming how robotics teams develop, train, and deploy intelligent systems by providing cloud-native infrastructure that scales from small research projects to enterprise-wide robot fleets.About the RoleWe are seeking a Data Engineer to design and build the data infrastructure that powers our robot learning platform. You'll architect scalable pipelines for ingesting, processing, and serving massive volumes of multi-modal robotics data — from sensor streams and telemetry to video and point clouds. This role offers the opportunity to build the foundational data layer that enables training across diverse robot embodiments and accelerates AI development for the entire robotics industry.Key ResponsibilitiesDesign and build scalable data pipelines for ingesting, transforming, and storing high-volume multi-modal robotics data including sensor streams, video, and telemetryArchitect and maintain data lake and warehouse infrastructure optimised for large-scale ML training workloadsBuild real-time and batch processing systems for robot data collection across distributed fleetsDevelop data quality frameworks including validation, monitoring, and lineage tracking to ensure reliability across the platformOptimise data storage and retrieval for performance and cost efficiency at petabyte scaleCollaborate with ML engineers to ensure training datasets are properly versioned, reproducible, and efficiently served to distributed training jobsRequired SkillsBachelor's degree or higher in Computer Science, Data Engineering, Software Engineering, or related fieldStrong experience with data pipeline orchestration tools such as Apache Airflow, Dagster, or PrefectProficiency in Python and SQL with experience processing large-scale datasets using frameworks like Spark, Dask, or RayCloud platform experience with AWS, GCP, or Azure including services like S3, BigQuery, Redshift, or equivalentExperience with data modelling, schema design, and storage formats optimised for analytical and ML workloads (Parquet, Arrow, Delta Lake)Solid understanding of distributed systems and event-driven architecturesPreferred SkillsExperience with streaming data systems such as Kafka, Kinesis, or PulsarFamiliarity with ML data tooling including feature stores, dataset versioning (DVC, LakeFS), and experiment trackingKnowledge of time-series data and sensor data processingExperience handling multi-modal data types such as video, point clouds, or IMU dataExposure to robotics data formats, ROS bag files, or similarInfrastructure-as-code experience with Terraform, Pulumi, or similar tools