What are the responsibilities and job description for the Data Architect - Databricks with PySpark position at Jobs via Dice?
Dice is the leading career destination for tech experts at every stage of their careers. Our client, Crea Services LLC, is seeking the following. Apply via Dice today!
Title: Databricks Architect
Location: Cincinnati, OH
Role Overview:
We are looking for a Databricks Architect to spearhead a strategic migration initiative from Informatica (Or any similar ETL) to Azure Databricks. The ideal candidate will have deep expertise in Databricks integration, PySpark, and Unity Catalog, combined with strong data engineering fundamentals. This role requires working closely with multiple migration pods and bringing best practices for large-scale migration projects, ensuring smooth transformation from legacy ETL to modern cloud-based solutions.
Key Responsibilities:
Title: Databricks Architect
Location: Cincinnati, OH
Role Overview:
We are looking for a Databricks Architect to spearhead a strategic migration initiative from Informatica (Or any similar ETL) to Azure Databricks. The ideal candidate will have deep expertise in Databricks integration, PySpark, and Unity Catalog, combined with strong data engineering fundamentals. This role requires working closely with multiple migration pods and bringing best practices for large-scale migration projects, ensuring smooth transformation from legacy ETL to modern cloud-based solutions.
Key Responsibilities:
- Lead design and development of scalable data pipelines using Databricks and PySpark.
- Integrate Databricks with enterprise systems and diverse data sources.
- Implement Unity Catalog for governance, security, and lineage.
- Drive migration from Informatica to Azure Databricks, ensuring minimal disruption and high data quality.
- Collaborate with multiple migration pods to align technical solutions and timelines.
- Introduce and enforce best practices for migration projects, including performance optimization and compliance.
- Provide technical leadership and mentor team members on modern data engineering practices.
- Hands-on experience in Databricks integration (focus on data engineering workflows, not platform setup).
- Proficiency in PySpark for distributed data processing.
- Familiarity with Unity Catalog for governance and security.
- Strong Data Engineering background (data modeling, ETL/ELT, performance tuning).
- Proven experience in migration from Informatica (or any other ETL) to Azure Databricks.
- Expertise in Azure Data Services (Data Lake, etc.).
- Strong SQL skills and understanding of big data concepts.
- Ability to work with multiple migration pods and manage dependencies effectively.
- Familiarity with Delta Lake and Lakehouse architecture.
- Knowledge of CI/CD pipelines for data workflows.
- Exposure to data security and compliance frameworks.
- Experience with data cataloging and lineage tools beyond Unity Catalog.
- Strong communication and stakeholder management skills.