Skip to Main Content

Job Title


Senior Python Developer - Data Pipelines - Full remote - Full remote - contractor in USD


Company : All European Careers


Location : Chennai, Tamil Nadu


Created : 2026-01-29


Job Type : Full Time


Job Description

For an international project in Chennai, we are urgently looking for a Full Remote Senior Python Developer. We seek a Python Developer to build and maintain reliable data pipelines and analytics workflows that ingest data from multiple internal and external sources, transform it into clean, analysis-ready datasets, and validate quality end-to-end. The work supports reporting, monitoring, and advanced analytics.We are looking for a motivated contractor. Candidates need to be fluent in English.Tasks and responsibilities:- Design, build, and maintain data ingestion pipelines to collect information from APIs, databases, cloud storage, web services, and other internal/external data sources; - Develop data transformation processes for cleaning, transforming, aggregating, and harmonizing data into analytics-ready formats; - Implement data validation and quality assurance checks to ensure data accuracy, completeness, and consistency; - Automate data extraction and loading (ETL/ELT) processes using Python and SQL-based frameworks; - Collaborate with analysts, report developers, product owner, and business users to define data requirements and ensure smooth integration with analytics platforms; - Optimize data pipelines for scalability, reliability, and performance. - Maintain data documentation, metadata, and version control for all data assets;Profile:- Bachelor or Master degree; - +6 years of relevant experience as a Python developer. - Strong proficiency in Python, including libraries such as pandas, numpy, sqlalchemy, and requests; - Expertise in data extraction and integration (REST APIs, SQL/NoSQL databases, file systems, cloud data sources); - Solid understanding of data transformation, cleaning, and validation techniques; - Strong hands-on SQL skills, including query optimization and performance tuning; - Experience with ETL orchestration tools (e.g., Airflow, ADF) and workflow automation; - Familiarity with data warehouses (Oracle, Azure Lakehouse, iceberg, PostgreSQL); - Experience implementing data validation and quality checks (e.g., dbt tests); - Experience with version control systems (Git) and CI/CD pipelines; - Knowledge of containerization (Docker) and cloud platforms (Azure); - Knowledge of Dremio and Pyarrow; - Knowledge of ADF and Databrick using pyspark; - Fluent in English;