Demo

Senior Software Engineer, Data Pipelines

Perimeter
Boston, MA Full Time
POSTED ON 4/14/2026
AVAILABLE BEFORE 10/7/2026

Senior Software Engineer, Data Pipelines


About Perimeter


Perimeter builds for resilience in an era of accelerating biological risk. Our end-to-end biosecurity infrastructure platform spans detection, characterization, forecasting, and response — powered by frontier AI and deployed across airports, hospitals, wastewater systems, and critical infrastructure worldwide. Perimeter's infrastructure maps the biological landscape, transforming biological data into biointelligence, enabling decisive response as soon as a threat emerges.


https://www.perimeter.bio/


About The Role


On our BUILD team, you are a software engineer focused on building and operating critical biosecurity data systems. You design reliable data pipelines and models, productionize analytics, and ensure data quality across programs spanning PCR, sequencing, wastewater, biosurveillance, and large-scale environmental monitoring.


This role requires strong software engineering fundamentals—including system design, testing, and code quality—applied to data infrastructure challenges. You will work primarily on backend data systems, designing data warehouses, building ETL/ELT pipelines, and managing data architecture. The role combines platform engineering (e.g., orchestration with Airflow, observability, infrastructure-as-code) with analytics engineering (SQL modeling, testing, documentation) to deliver reliable data products that support threat detection, pathogen attribution, and operational decision-making.


Responsibilities


Data Platform Architecture & Engineering

  • Plan, architect, test, and deploy data warehouses, data marts, and ETL/ELT pipelines primarily within AWS and Snowflake environments
  • Build scalable data pipelines capable of handling structured, unstructured, and high-throughput biological data from diverse sources
  • Develop data models using dbt with rigorous testing, documentation, and stakeholder-aligned semantics to ensure analytics-ready datasets


Data Quality & Governance

  • Ensure data integrity, consistency, and accessibility across internal and external biosecurity data products
  • Develop, document, and enforce coding and data modeling standards to improve code quality, maintainability, and system performance
  • Serve as the in-house data expert, making recommendations on data architecture, pipeline improvements, and best practices; define and adapt data engineering processes to deliver reliable answers to critical biosecurity questions


API & Integration Development

  • Build high-performance APIs and microservices in Python that enable seamless integration between the biosecurity data platform and user-facing applications
  • Design backend services that support real-time and batch data access for biosecurity operations
  • Create data products that empower public health officials, analysts, and partners with actionable biosecurity intelligence


AI & Data Democratization

  • Democratize access to complex biosecurity datasets using AI and LLMs, making data more discoverable and usable for stakeholders
  • Apply AI-assisted development tools to accelerate code generation, data modeling, and pipeline development while maintaining high quality standards


Cloud Infrastructure & Performance

  • Build robust, production-ready data workflows using AWS, Kubernetes, Docker, Airflow, and infrastructure-as-code (Terraform/CloudFormation)
  • Diagnose system bottlenecks, optimize for cost and speed, and ensure the reliability and fault tolerance of mission-critical data pipelines
  • Implement observability, monitoring, and alerting to maintain high availability for biosecurity operations


Technical Leadership & Collaboration

  • Lead data projects from scoping through execution, including design, documentation, and stakeholder communication
  • Collaborate with technical leads, product managers, scientists, and data analysts to build robust data products and analytics capabilities


Minimum Qualifications


  • 7 years of professional experience in data or software engineering, with a focus on building production-grade data products and scalable architectures
  • Expert proficiency with SQL for complex transformations, performance tuning, and query optimization
  • Strong Python skills for data engineering workflows, including pipeline development, ETL/ELT processes, and data processing; experience with backend frameworks (FastAPI, Flask) for API development; focus on writing modular, testable, and reusable code
  • Proven experience with dbt for data modeling and transformation, including testing frameworks and documentation practices
  • Hands-on experience with cloud data warehouses (Snowflake, BigQuery, or Redshift), including performance tuning, security hardening, and managing complex schemas
  • Experience with workflow orchestration tools (Airflow, Dagster, or equivalent) for production data pipelines, including DAG development, scheduling, monitoring, and troubleshooting
  • Solid grounding in software engineering fundamentals: system design, version control (Git), CI/CD pipelines, containerization (Docker), and infrastructure-as-code (Terraform, CloudFormation)
  • Hands-on experience managing AWS resources, including S3, IAM roles/policies, API integrations, and security configurations
  • Strong ability to analyze large datasets, identify data quality issues, debug pipeline failures, and propose scalable solutions
  • Excellent communication skills and ability to work cross-functionally with scientists, analysts, and product teams to turn ambiguous requirements into maintainable data products


Preferred Capabilities & Experience


  • Domain familiarity with biological data (PCR, sequencing, wastewater surveillance, TAT metrics) and experience working with lab, bioinformatics, NGS, or epidemiology teams
  • Production ownership of Snowflake environments including RBAC, secure authentication patterns, and cost/performance optimization
  • Experience with observability and monitoring stacks (Grafana, Datadog, or similar) and data quality monitoring (anomaly detection, volume/velocity checks, schema drift detection)
  • Familiarity with container orchestration platforms (Kubernetes) for managing production workloads
  • Experience with data ingestion frameworks (Airbyte, Fivetran) or building custom ingestion solutions for external partner data delivery
  • Familiarity with data cataloging, governance practices, and reference data management to prevent silent data drift
  • Experience designing datasets for visualization tools (Tableau, Looker, Metabase) with strong understanding of dashboard consumption patterns; familiarity with JavaScript for custom visualizations or front-end dashboard development
  • Comfort with AI-assisted development tools (GitHub Copilot, Cursor) to accelerate code generation while maintaining quality standards
  • Startup or fast-paced environment experience with evolving priorities and rapid iteration
  • Scientific or data-intensive domain experience (life sciences, healthcare, materials science)


You Should Apply If You...

  • Are passionate about working on a mission that matters & has real global impact
  • Are a self-starter who thrives in dynamic, fast-moving environments and gets energized by ambiguity rather than slowed down by it
  • Want to be part of building something from the ground up — we're a small, scrappy team, and we're looking for people who are comfortable operating without a full playbook but are equally excited about helping us build the processes and foundations that will carry us forward as we grow

Salary.com Estimation for Senior Software Engineer, Data Pipelines in Boston, MA
$126,559 to $155,760
If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution. Compensation Planning
Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right. Surveys & Data Sets

What is the career path for a Senior Software Engineer, Data Pipelines?

Sign up to receive alerts about other jobs on the Senior Software Engineer, Data Pipelines career path by checking the boxes next to the positions that interest you.
Income Estimation: 
$123,167 - $152,295
Income Estimation: 
$146,673 - $180,130
Income Estimation: 
$122,257 - $154,284
Income Estimation: 
$143,391 - $179,890
Employees: Get a Salary Increase
View Core, Job Family, and Industry Job Skills and Competency Data for more than 15,000 Job Titles Skills Library

Job openings at Perimeter

  • Perimeter Boston, MA
  • Biosecurity Genomics Analyst About Perimeter Perimeter builds for resilience in an era of accelerating biological risk. Our end-to-end biosecurity infrastr... more
  • 4 Days Ago

  • Perimeter Boston, MA
  • Senior Software Engineer, Data Pipelines About Perimeter Perimeter builds for resilience in an era of accelerating biological risk. Our end-to-end biosecur... more
  • 4 Days Ago


Not the job you're looking for? Here are some other Senior Software Engineer, Data Pipelines jobs in the Boston, MA area that may be a better fit.

  • ginkgobioworks Boston, MA
  • Our mission is to make biology easier to engineer. Ginkgo is constructing, editing, and redesigning the living world in order to answer the globe’s growing... more
  • 15 Days Ago

  • ChatGPT Jobs Boston, MA
  • Job Description Senior Data Engineer, Platform & Pipelines Natera Boston, MA Remote $125K - $155K/yr Full-time Medical, Dental, Vision, Life, Retirement Ab... more
  • 2 Days Ago

AI Assistant is available now!

Feel free to start your new journey!