What are the responsibilities and job description for the Senior Data Engineer with GCP, LLM And RAG Experience (only W2 Position – No C2C Accepted) position at Systems Technology Group, Inc. (STG)?
Title: Senior Data Engineer with GCP, LLM And RAG Experience (only W2 Position – No C2C Accepted)
Description: STG is a SEI CMMi Level 5 company with several Fortune 500 and State Government clients. STG has an opening forSenior Data Engineer with GCP, LLM And RAG Experience.
Please note that this project assignment is with our own direct clients. We do not go through any vendors. STG only does business with direct end clients. This is expected to be a long-term position. STG will provide immigration and permanent residency sponsorship assistance to those candidates who need it.
Position Description:
As a Senior Data Engineer, you will architect and scale end-to-end data pipelines on GCP, transforming complex telemetry and enterprise data into high-quality, analytics-ready assets using Medallion architectures. You will lead the implementation of robust CI/CD workflows, rigorous data governance, and security controls while mentoring junior talent and driving engineering best practices. By collaborating with cross-functional stakeholders and optimizing cloud performance, you will ensure the data platform remains secure, cost-effective, and highly available to power critical business insights. • Operational Excellence: Using Terraform, Git, and Airflow to ensure reproducible, secure, and cost-optimized cloud infrastructure. • Governance & Quality: Prioritizing data lineage, PII protection, and observability to maintain high trust in data assets. • Collaboration: Acting as a bridge between technical teams (Data Science, Security) and business stakeholders to deliver self-service analytics. Strong understanding of Generative AI principles and architectures, including Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) systems. * Proven experience in building and deploying RAG systems, including the use of **Vector Databases**. * Proficiency in Python programming. * Solid experience with SQL for data manipulation and querying. * Hands-on experience with Google Cloud Platform (GCP) services relevant to AI/ML. * Basic understanding and practical experience with Machine Learning model fine-tuning. * Familiarity with data engineering concepts and practices. * Expertise in prompt engineering techniques for interacting with LLMs. * Experience with the OpenAI SDK. * Experience developing robust APIs, preferably with **FastAPI**. * Proficiency with **version control systems (e.g., Git) **. * Experience with **containerization technologies (e.g., Docker) **.
- Collaborate with business and technology stakeholders to understand current and future data requirements
- Design, build and maintain reliable, efficient and scalable data infrastructure for data collection, storage, transformation, and analysis
- Plan, design, build and maintain scalable data solutions including data pipelines, data models, and applications for efficient and reliable data workflow
- Design, implement and maintain existing and future data platforms like data warehouses, data lakes, data lakehouse etc. for structured and unstructured data
- Design and develop analytical tools, algorithms, and programs to support data engineering activities like writing scripts and automating tasks
- Ensure optimum performance and identify improvement opportunities
- GCP – Experience deploying and managing services on Google Cloud Platform, including Compute Engine, Cloud Storage, IAM, and Cloud Functions. For example, designing and implementing a cloud-native application architecture using GKE (Google Kubernetes Engine) with Cloud SQL and Pub/Sub.
- Big Data – Experience working with large-scale data processing frameworks such as Apache Spark, Dataflow, or BigQuery. For example, building ETL pipelines that process terabytes of daily event data and transform it for downstream analytics.
- Data Warehousing – Experience designing and maintaining data warehouse solutions (e.g., BigQuery, Snowflake, Redshift). For example, modeling a star schema for a retail analytics platform that supports reporting on sales, inventory, and customer behavior.
- Artificial Intelligence & Expert Systems – Experience developing or integrating AI/ML models and rule-based expert systems. For example, building a classification model using Vertex AI to predict customer churn, or implementing a rule engine that automates underwriting decisions.
- API – Experience designing, building, and consuming RESTful or gRPC APIs. For example, developing a versioned REST API with OAuth 2.0 authentication that serves as the integration layer between a mobile application and backend microservices.
Skills Preferred:
Google Cloud Platform 1. Google Cloud Platform – Familiarity with advanced GCP services beyond core compute and storage, such as Vertex AI, Dataflow, Cloud Composer (Airflow), and BigQuery ML. For example, using Cloud Composer to orchestrate scheduled data pipelines that feed into a BigQuery data warehouse.
Education Required:
- Bachelor's degree or equivalent qualification in computer science, engineering or related disciplines
Senior Data Engineer with GCP, LLM And RAG Experience is based in Dearborn, MI. A great opportunity to experience the corporate environment leading personal career growth.
Resume Submittal Instructions: Interested/qualified candidates should email their word formatted resumes to Vasavi Konda – vasavi.konda(.@)stgit.com and/or contact @(Two-Four-Eight) Seven- One-Two – Six-Seven-Two-Five (@248.712.6725). In the subject line of the email please include: First and Last Name: Senior Data Engineer with GCP, LLM And RAG Experience.
For more information about STG, please visit us at www.stgit.com.
Sincerely,
Vasavi Konda| Recruiting Specialist
“Opportunities don't happen, you create them.”
Systems Technology Group (STG)
3001 W. Big Beaver Road, Suite 500
Troy, Michigan 48084
Phone: @(Two-Four-Eight) Seven- One-Two – Six-Seven-Two-Five: @248.712.6725(O)
Email: vasavi.konda(.@)stgit.com