What are the responsibilities and job description for the PySpark Developer position at Jobs via Dice?
Location: Irving, TX
Salary: $120,000.00 USD Annually - $130,000.00 USD Annually
Description:
Job Description - PySpark Developer
Location: Irving, TX
Experience: 5-10 years
Overview
We are seeking an experienced PySpark Developer with a strong background in data engineering to design, build, and optimize scalable data pipelines. The ideal candidate will have deep expertise in Apache Spark, hands-on PySpark development, and experience working with large, complex datasets. You will collaborate closely with business analysts and stakeholders to deliver high-quality data solutions that support analytics and business intelligence.
Responsibilities
Contact:
This job and many more are available through The Judge Group. Please apply with us today!
Salary: $120,000.00 USD Annually - $130,000.00 USD Annually
Description:
Job Description - PySpark Developer
Location: Irving, TX
Experience: 5-10 years
Overview
We are seeking an experienced PySpark Developer with a strong background in data engineering to design, build, and optimize scalable data pipelines. The ideal candidate will have deep expertise in Apache Spark, hands-on PySpark development, and experience working with large, complex datasets. You will collaborate closely with business analysts and stakeholders to deliver high-quality data solutions that support analytics and business intelligence.
Responsibilities
- Develop, optimize, and maintain data pipelines and ETL workflows using Apache Spark and PySpark.
- Apply advanced knowledge of Spark core concepts, performance tuning, and industry best practices.
- Translate business requirements into scalable and reliable data transformations.
- Work with analysts and stakeholders to understand data requirements and deliver innovative solutions.
- Write and maintain unit tests, follow OOP principles, and ensure high code quality.
- Develop complex SQL queries to extract insights from large-scale datasets.
- Schedule, orchestrate, and monitor data transformation jobs based on business needs.
- Perform root-cause analysis on pipeline failures, resolve data quality issues, and implement corrective fixes.
- Ensure data integrity, reliability, and availability across the data ecosystem.
- 5-10 years of hands-on experience in PySpark and Apache Spark development.
- Strong understanding of Spark RDDs, DataFrames, Spark SQL, and performance optimization.
- Expertise in Python with ability to implement complex data transformations.
- Solid SQL skills with experience handling large datasets.
- Knowledge of object-oriented programming, unit testing, and debugging.
- Experience with scheduling tools (Airflow, Cron, or similar).
- Strong analytical and troubleshooting skills for data pipeline issues.
Contact:
This job and many more are available through The Judge Group. Please apply with us today!
Salary : $120,000 - $130,000