What are the responsibilities and job description for the Scala Engineers position at Appex Innovation?
Staff Scala Engineer (Onsite)
Location: Chicago, IL (Onsite)
Overview: We are a technology-driven company seeking a highly experienced and technically profound Staff Scala Engineer to anchor our Data Platform team. This role demands leadership in designing, developing, and optimizing high-throughput, fault-tolerant data solutions using a modern stack centered on Scala, Apache Spark, and AWS cloud services.
Core Responsibilities
System Architecture: Lead the architectural definition and implementation of robust, scalable, and efficient data processing systems utilizing Scala and Apache Spark.
Self-Correction Note: Consider the trade-offs between batch and stream processing architectures (e.g., using Spark Streaming or Flink).
Engineering Excellence: Develop high-quality, maintainable, and performant functional code in Scala. Drive performance tuning and optimization of large-scale Spark jobs.
Cloud Infrastructure: Architect and manage the deployment of data pipelines using core AWS services (e.g., S3, EMR, Glue, ECS). Ensure optimal usage of cloud resources for cost and efficiency.
Technical Leadership: Serve as a subject matter expert for the team. Mentor peers, define coding standards, and lead complex technical design reviews.
Collaboration: Partner with cross-functional teams (Product, DevOps, Analytics) to ensure technical solutions meet business requirements and are seamlessly integrated into the ecosystem.
Operational Support: Implement and manage monitoring, logging, and alerting strategies to maintain the health and reliability of production data services.
Required Technical Qualifications
Minimum 7 years of professional experience in software engineering, with significant time spent building distributed data applications.
Expertise in Scala: Deep, demonstrable experience with production-level Scala development, emphasizing Functional Programming paradigms.
Expertise in Apache Spark: Mastery of Apache Spark (Scala API) for complex, large-scale data transformation, ETL/ELT, and performance optimization techniques (shuffling, partitioning).
AWS Cloud Proficiency: Strong, practical experience with primary AWS data and compute services (e.g., S3, EMR, Glue, Step Functions, IAM, CloudFormation/Terraform).
Foundational Knowledge: Solid grasp of distributed systems design, data structures, and algorithms.
Database Experience: Proficiency with various data storage technologies (relational, NoSQL).
DevOps Practices: Working knowledge of CI/CD pipelines and infrastructure as code tools (Terraform, CloudFormation).
Preferred Qualifications
Experience with stream processing (e.g., Kafka, Kinesis, Flink).
Familiarity with container orchestration (Docker and Kubernetes/EKS).
Prior experience in a Staff, Principal, or Lead Engineer role.
Salary : $50