What are the responsibilities and job description for the Sr. Data Engineer with Unity Catalog position at GlobalPoint Inc?
Seattle Based client
Location – Seattle WA- 4 days office
Duration – 12 months
Core Skills –
Collaborate with cross-functional teams to support data governance using Databricks Unity Catalog
Need to coordinate with Offshore
10-12 years of experience in data engineering or a related field
Expertise with programming languages such as Python/PySpark, SQL, or Scala
Design, build, and deploy data extraction, transformation, and loading processes and pipelines from various sources including databases, APIs, and data files.
Implement and maintain Delta Lake for optimized data storage, ensuring data reliability, performance, and versioning
Automate CI/CD pipelines for data workflows using Azure DevOps
Location – Seattle WA- 4 days office
Duration – 12 months
Core Skills –
Collaborate with cross-functional teams to support data governance using Databricks Unity Catalog
Need to coordinate with Offshore
10-12 years of experience in data engineering or a related field
Expertise with programming languages such as Python/PySpark, SQL, or Scala
- Experience working in a cloud environment (Azure preferred) with strong understanding of cloud data architecture
- Hands-on experience with Databricks Cloud Data Platforms Required.
- Should have experience migrating to Unity Catalog.
- Experience with workflow orchestration (e.g., Databricks Jobs, or Azure Data Factory pipelines) Required
Design, build, and deploy data extraction, transformation, and loading processes and pipelines from various sources including databases, APIs, and data files.
- Develop and support data pipelines within a Cloud Data Platform, such as Databricks
- Monitor and optimize Databricks cluster performance, ensuring cost-effective scaling and resource utilization
- Demonstrates ability to communicate technical concepts to non-technical audiences both in written and verbal form.
Implement and maintain Delta Lake for optimized data storage, ensuring data reliability, performance, and versioning
Automate CI/CD pipelines for data workflows using Azure DevOps