Senior Software Engineer, Data Pipelines
Our mission is to make biology easier to engineer. Ginkgo is constructing, editing, and redesigning the living world in order to answer the globe’s growing challenges in health, energy, food, materials, and more. Our bioengineers make use of an in-house automated foundry for designing and building new organisms.
Ginkgo Biosecurity is building next-generation biosecurity infrastructure to help governments and partners detect, attribute, and deter biological threats. Our mission extends across public health, national security, and global defense, ensuring nations can rapidly identify dangerous pathogens, understand where threats originate, and respond with confidence.
On our Biosecurity team, you are a software engineer focused on building and operating critical biosecurity data systems. You design reliable data pipelines and models, productionize analytics, and ensure data quality across programs spanning PCR, sequencing, wastewater, biosurveillance, and large-scale environmental monitoring.
This role requires strong software engineering fundamentals—including system design, testing, and code quality—applied to data infrastructure challenges. You will work primarily on backend data systems, designing data warehouses, building ETL/ELT pipelines, and managing data architecture. The role combines platform engineering (e.g., orchestration with Airflow, observability, infrastructure-as-code) with analytics engineering (SQL modeling, testing, documentation) to deliver reliable data products that support threat detection, pathogen attribution, and operational decision-making.
Responsibilities
Data Platform Architecture & Engineering
- Plan, architect, test, and deploy data warehouses, data marts, and ETL/ELT pipelines primarily within AWS and Snowflake environments
- Build scalable data pipelines capable of handling structured, unstructured, and high-throughput biological data from diverse sources
- Develop data models using dbt with rigorous testing, documentation, and stakeholder-aligned semantics to ensure analytics-ready datasets
Data Quality & Governance
- Ensure data integrity, consistency, and accessibility across internal and external biosecurity data products
- Develop, document, and enforce coding and data modeling standards to improve code quality, maintainability, and system performance
- Serve as the in-house data expert, making recommendations on data architecture, pipeline improvements, and best practices; define and adapt data engineering processes to deliver reliable answers to critical biosecurity questions
API & Integration Development
- Build high-performance APIs and microservices in Python that enable seamless integration between the biosecurity data platform and user-facing applications
- Design backend services that support real-time and batch data access for biosecurity operations
- Create data products that empower public health officials, analysts, and partners with actionable biosecurity intelligence
AI & Data Democratization
- Democratize access to complex biosecurity datasets using AI and LLMs, making data more discoverable and usable for stakeholders
- Apply AI-assisted development tools to accelerate code generation, data modeling, and pipeline development while maintaining high quality standards
Cloud Infrastructure & Performance
- Build robust, production-ready data workflows using AWS, Kubernetes, Docker, Airflow, and infrastructure-as-code (Terraform/CloudFormation)
- Diagnose system bottlenecks, optimize for cost and speed, and ensure the reliability and fault tolerance of mission-critical data pipelines
- Implement observability, monitoring, and alerting to maintain high availability for biosecurity operations
Technical Leadership & Collaboration
- Lead data projects from scoping through execution, including design, documentation, and stakeholder communication
- Collaborate with technical leads, product managers, scientists, and data analysts to build robust data products and analytics capabilities
Minimum Qualifications
- 7+ years of professional experience in data or software engineering, with a focus on building production-grade data products and scalable architectures
- Expert proficiency with SQL for complex transformations, performance tuning, and query optimization
- Strong Python skills for data engineering workflows, including pipeline development, ETL/ELT processes, and data processing; experience with backend frameworks (FastAPI, Flask) for API development; focus on writing modular, testable, and reusable code
- Proven experience with dbt for data modeling and transformation, including testing frameworks and documentation practices
- Hands-on experience with cloud data warehouses (Snowflake, BigQuery, or Redshift), including performance tuning, security hardening, and managing complex schemas
- Experience with workflow orchestration tools (Airflow, Dagster, or equivalent) for production data pipelines, including DAG development, scheduling, monitoring, and troubleshooting
- Solid grounding in software engineering fundamentals: system design, version control (Git), CI/CD pipelines, containerization (Docker), and infrastructure-as-code (Terraform, CloudFormation)
- Hands-on experience managing AWS resources, including S3, IAM roles/policies, API integrations, and security configurations
- Strong ability to analyze large datasets, identify data quality issues, debug pipeline failures, and propose scalable solutions
- Excellent communication skills and ability to work cross-functionally with scientists, analysts, and product teams to turn ambiguous requirements into maintainable data products
Preferred Capabilities & Experience
- Domain familiarity with biological data (PCR, sequencing, wastewater surveillance, TAT metrics) and experience working with lab, bioinformatics, NGS, or epidemiology teams
- Production ownership of Snowflake environments including RBAC, secure authentication patterns, and cost/performance optimization
- Experience with observability and monitoring stacks (Grafana, Datadog, or similar) and data quality monitoring (anomaly detection, volume/velocity checks, schema drift detection)
- Familiarity with container orchestration platforms (Kubernetes) for managing production workloads
- Experience with data ingestion frameworks (Airbyte, Fivetran) or building custom ingestion solutions for external partner data delivery
- Familiarity with data cataloging, governance practices, and reference data management to prevent silent data drift
- Experience designing datasets for visualization tools (Tableau, Looker, Metabase) with strong understanding of dashboard consumption patterns; familiarity with JavaScript for custom visualizations or front-end dashboard development
- Comfort with AI-assisted development tools (GitHub Copilot, Cursor) to accelerate code generation while maintaining quality standards
- Startup or fast-paced environment experience with evolving priorities and rapid iteration
- Scientific or data-intensive domain experience (life sciences, healthcare, materials science)
The base salary range for this role is $134,300-$189,900. Actual pay within this range will depend on a candidate's skills, expertise, and experience. We also offer company stock awards, a comprehensive benefits package including medical, dental & vision coverage, health spending accounts, voluntary benefits, leave of absence policies, 401(k) program with employer contribution, 8 paid holidays in addition to a full-week winter shutdown and unlimited Paid Time Off policy.
It is the policy of Ginkgo Bioworks to provide equal employment opportunities to all employees, employment applicants, and EOE disability/vet.
Privacy Notice
I understand that I am applying for employment with Ginkgo Bioworks and am being asked to provide information in connection with my application. I further understand that Ginkgo gathers this information through a third-party service provider and that Ginkgo may also use other service providers to assist in the application process. Ginkgo may share my information with such third-party service providers in connection with my application and for the start of employment. Ginkgo will treat my information in accordance with Ginkgo's . By submitting this job application, I am acknowledging that I have reviewed and agree to Ginkgo's as well as the privacy policies of the third-party service providers used by Ginkgo's associated with the application process.
Recommended Jobs
Pre-K Teacher
We’re hiring immediately, with a Signing Bonus for EEC certification! Come enjoy the best work environment of any preschool. We take care of and support all of our teachers. If you're looking for…
Enterprise Sales Executive - US
Location Boston Employment Type Full time Department Sales About Us Perk (formerly TravelPerk) is the intelligent platform for travel and spend management. Built to tackle the ti…
Sales Development Representative (Greater Boston Area)
About demandDrive: demandDrive is an industry-leading integrated sales and marketing agency with over 400 employees, headquartered in the Greater Boston Area. demandDrive provides B2B clients with a…
DelmiaWorks - Industry Process Consultant - East Coast preferred
Job Responsibilities and Role We are seeking a highly motivated and technically statute ERP Sales Solutions Engineer to join our growing team. In this role, you will be a critical partner for our …
Certified Nursing Assistant / Home Health Aide (CNA/HHA)
Job Description Job Description Description: Prime Home Health and Companion Care is seeking compassionate Certified Nursing Assistants (CNAs) and Home Health Aides (HHAs) to support elderly, di…
Part-Time Stylist (Newbury Street Women's)
Job Description Job Description Love helping folks find their perfect look? As a Stylist at Buck Mason, you’ll deliver exceptional customer service, assist customers with their purchases, and kee…
Certified Application Counselor
VITRA Health is hiring, and we believe that our team is the key to our success. That is why we are looking for a Certified Application Counselor to join our Braintree team. If you're a passionate, …
Account Executive (B2B, EU Market, Nebius Academy)
Nebius Academy empowers companies to achieve their business goals by driving AI Adoption through learning. Our transformative approach includes tailored training programs, informed by comprehensive…
Pile Driver (Andover)
The Middlesex Corporation is a nationally recognized and award-winning leader in the heavy civil construction industry. Since 1972, the family business founded by Robert W. Pereira has developed an e…
Data QA Analyst
Kforce has a client in need of a Data QA Analyst in Boston, MA. Responsibilities: * Define and maintain data quality standards, validation rules, and monitoring processes for public health data * Imp…