Data Engineer - Pyspark, Airflow, AWS

NucleusTeq

Telangana , IN Full–time
Posted on: February 27, 2026
Key Responsibilities: • Develop Apache Airflow DAGs and PySpark ETL pipelines for high volume data processing. • Write optimized SQL queries for data transformation and aggregation. • Build data products serving Business Process, Executive KPIs, and Product Analytics. • Implement data quality and monitoring solutions. • Optimize pipeline performance and troubleshoot production issues. • Collaborate with cross-functional teams. • Production Pipeline Monitoring (KLO). Required skills: • 10+ years of data engineering experience. 7+ years dedicated to the Big data stack. • Expert in Python and PySpark (DataFrame API, Spark SQL). • Advanced SQL skills (window functions, complex queries). • Production experience with Apache Airflow. • Solid background in data warehousing and dimensional modelling. Preferred skills: • Experience with SQL, Trino, Apache Iceberg. • Knowledge of Tableau CRM/CLOUD, Salesforce platforms. • AWS/cloud data services experience. Why Join Us? • Work in an innovative environment with the business that is shaping the future of data migration. • Be part of a dynamic, high-growth environment at NucleusTeq. • Competitive salary and comprehensive benefits package.

About Company

NucleusTeq

Telangana ,IN

https://www.nucleusteq.com

Your next job is waiting

Create your profile and start applying in minutes.