Data Engineer - Pyspark, Airflow, AWS
NucleusTeq
Posted on: February 27, 2026
Key Responsibilities:
• Develop Apache Airflow DAGs and PySpark ETL pipelines for high volume data processing.
• Write optimized SQL queries for data transformation and aggregation.
• Build data products serving Business Process, Executive KPIs, and Product Analytics.
• Implement data quality and monitoring solutions.
• Optimize pipeline performance and troubleshoot production issues.
• Collaborate with cross-functional teams.
• Production Pipeline Monitoring (KLO).
Required skills:
• 10+ years of data engineering experience. 7+ years dedicated to the Big data stack.
• Expert in Python and PySpark (DataFrame API, Spark SQL).
• Advanced SQL skills (window functions, complex queries).
• Production experience with Apache Airflow.
• Solid background in data warehousing and dimensional modelling.
Preferred skills:
• Experience with SQL, Trino, Apache Iceberg.
• Knowledge of Tableau CRM/CLOUD, Salesforce platforms.
• AWS/cloud data services experience.
Why Join Us?
• Work in an innovative environment with the business that is shaping the future of data migration.
• Be part of a dynamic, high-growth environment at NucleusTeq.
• Competitive salary and comprehensive benefits package.
About Company
NucleusTeq
Telangana ,IN
https://www.nucleusteq.com
Your next job is waiting
Create your profile and start applying in minutes.