Hex, a data platform company that provides powerful, collaborative analytics workflows, is seeking an experienced Cloud Security Engineer. You will be responsible for ensuring the security and resilience of the platform’s cloud infrastructure and providing leadership in cloud security practices, especially regarding their RCE-as-a-Service platform.
This is a full-time position available in San Francisco (SF), New York City (NYC), or Remote (USA). The salary range for this role is $180,000 - $220,000.
Core Responsibilities and Container Security Mandate
This role requires close collaboration with infrastructure and engineering teams to embed security controls into cloud-native applications, focusing heavily on securing Kubernetes deployments and maintaining infrastructure-as-code standards.
- Security Design and Isolation: Design, implement, and manage security solutions and controls for AWS environments and Kubernetes clusters, including appropriate isolation/sandboxing methods for Hex’s RCE-as-a-Service platform.
- IaC and Enforcement: Build, deploy, and maintain infrastructure-as-code using Terraform, ensuring robust security standards are enforced throughout.
- Assessment and Audits: Conduct security assessments, threat modeling, and audits on AWS cloud infrastructure and Kubernetes deployments.
- DevSecOps Integration: Collaborate with development and operations teams to embed security best practices into CI/CD pipelines (Secure Software Development Lifecycle).
- Compliance Expertise: Provide expertise in compliance requirements related to cloud security, including SOC 2, ISO 27001, GDPR, HIPAA, and PCI DSS.
- Incident Response: Monitor and respond to cloud security incidents, identifying root causes and recommending remediation actions.
- Mentorship: Mentor engineers and advocate for cloud security across the organization.
Required Experience and Technical Qualifications
The ideal candidate has extensive expertise in AWS and demonstrable proficiency in Kubernetes security, combining infrastructure experience with deep security knowledge.
- Experience (Required):
- 5+ years of experience in cloud security engineering, with extensive expertise in AWS.
- Expert-level knowledge and hands-on experience with Terraform.
- Kubernetes Security Proficiency: Demonstrated proficiency with Kubernetes security, including cluster hardening, Role-Based Access Control (RBAC), network policies, and container vulnerability management.
- Security Tooling Familiarity:
- Familiarity with AWS security services (e.g., IAM, GuardDuty, Security Hub, CloudTrail, WAF).
- Familiarity with CNAPP solutions such as Wiz.
- Familiarity with SIEM solutions such as Panther.
- Methodology: Solid understanding of secure software development lifecycle practices, CI/CD security, and DevSecOps methodologies.
- Preferred Certifications (Highly Desirable):
- AWS Certified Security – Specialty
- Certified Kubernetes Security Specialist (CKS)
- Terraform Associate certification
- Security certifications from SANS or OffSec.
Job Features
| Job Category | Cloud Engineering |
Fullsteam, a leading provider of vertical software and embedded payments technology for small and medium-sized businesses, is seeking a mid-level Site Reliability Engineer (SRE). This role is focused on ensuring the reliability, scalability, and performance of Fullsteam’s infrastructure, with a critical emphasis on Cloudflare implementation and infrastructure automation.
This is a full-time, remote position in the U.S.
Primary Responsibilities and Cloudflare Mandate
As an SRE, you will work closely with development, operations, and Business Unit teams to build and maintain robust systems. A core function of this role is leading the technical implementation and configuration of Cloudflare across multiple business units.
- Cloudflare Leadership: Lead the technical implementation of Cloudflare across multiple business units.
- Design and configure DNS zones and records using Cloudflare DNS.
- Set up and fine-tune Cloudflare WAF rules, firewall policies, and managed rulesets.
- Implement SSL/TLS certificate lifecycle management via Cloudflare and origin servers.
- Automate Cloudflare configurations using Terraform, API, or IaC pipelines.
- Automation and IaC: Write Infrastructure as Code (IaC) and develop CI/CD pipelines using GitHub Actions.
- Operational Support: Support infrastructure and application teams with deployments, troubleshooting, service requests, and change controls.
- Monitoring and Remediation: Monitor logs daily, report findings, and implement remediation strategies.
- Process Improvement: Analyze, develop, and improve operational procedures and actively participate in change review processes.
Required Experience and Technical Competencies
The ideal candidate has strong dual expertise in a major cloud platform (AWS or Azure) and DevOps tooling, combined with practical experience in large-scale database management and automation.
- Minimum Qualifications:
- At least 5 years of experience working with modern DevOps tooling.
- At least 5 years of experience managing database technologies.
- SSA-A or AZ-305 certification.
- Bachelor’s degree in a technical field (or 4+ years of equivalent experience).
- Technical Expertise:
- Expertise in AWS or Azure with strong capabilities in the other platform.
- Experience designing and architecting infrastructure using Terraform and Ansible.
- Proficiency in DevOps tooling, including Jenkins, SVN, GitHub Actions, and Bitbucket.
- Programming/scripting skills in BASH, Shell, PowerShell, or Python.
- Experience with container technologies and orchestration tools such as Docker, ECS, EKS, and AKS.
- Strong understanding of web traffic flows and HTTP/S protocols.
- Database management experience with MSSQL, MySQL, Postgres, MongoDB, and Metabase.
- Preferred Qualifications (Bonus):
- Hands-on experience configuring Cloudflare DNS, WAF, and SSL/TLS.
- Experience with Cloudflare APIs or Terraform provider.
- Knowledge of defense-in-depth, least-privileged, and secure-by-design principles.
- Experience with Blue-Green deployments.
- Understanding and experience in maintaining a PCI DSS environment.
- Experience building monitoring, logging, security, and scalability solutions on Azure or AWS.
Job Features
| Job Category | Software Engineering |
Claritas Rx, a venture-backed digital health startup, is seeking an experienced DevOps Engineer to design, build, and maintain a secure, scalable, and highly available cloud infrastructure. This role is crucial for ensuring reliable delivery pipelines and automated infrastructure management for platforms handling complex specialty biopharmaceutical data.
This is a full-time position based in Washington, D.C., US, but requires candidates to be based in or able to work standard business hours in the Pacific Time zone. The salary range is $125,000 - $140,000 / year.
Role Summary and AWS Automation Mandate
The DevOps Engineer will be the go-to expert for AWS infrastructure, working closely with engineering, analytics, and product teams. This role requires a proactive approach to problem-solving and deep expertise in infrastructure automation to support both backend and user-facing application platforms at scale in a SaaS, healthcare, or life sciences environment.
Key Responsibilities:
- Infrastructure Management: Design, implement, and manage AWS cloud infrastructure, ensuring high availability, performance, and security.
- Automation (IaC): Develop and maintain infrastructure-as-code (IaC) and automation pipelines using tools like Ansible, CDK, and AWS CodePipelines.
- AWS Core Services Oversight: Manage, monitor, and optimize a comprehensive suite of AWS services, including:
- Compute & Networking: EC2, ECS, ECR, VPC, ELB, Route 53, WorkSpaces, VPN.
- Data Services: RDS, DynamoDB, DMS, S3, Glue, Athena.
- Security & Compliance: IAM Identity Center, Secrets Manager, KMS, CloudTrail, Systems Manager.
- Messaging & Eventing: SQS, SNS, API Gateway, Lambda.
- Observability & Incident Response: Ensure proactive monitoring and incident response using tools such as CloudWatch and Splunk.
- CI/CD & Developer Experience: Support CI/CD pipelines (Github, Bitbucket, AWS pipelines) and partner with engineering teams to improve developer experience, reliability, and scalability.
- Data Pipelines: Manage data and analytics pipeline tools like Airflow and Tableau.
Required Experience and Technical Qualifications
The ideal candidate has significant, hands-on experience across the core AWS ecosystem, specializing in automation, security, and containerization.
- Experience (Required):
- 5+ years of experience in cloud infrastructure, DevOps, or systems engineering, supporting production workloads (ideally in SaaS or Healthcare).
- Proven track record of automating infrastructure and deployments using tools such as Ansible, CDK, Terraform/CloudFormation, or CodePipeline/CodeBuild.
- Core Technical Stack:
- Strong hands-on expertise with common AWS services (EC2, RDS, VPC, IAM, ECS/ECR, Lambda, etc.).
- Skilled in containerization and orchestration (Docker, ECS).
- Experienced in monitoring, observability, and incident response with CloudWatch, Splunk, and related tooling.
- Familiar with data and analytics ecosystems (Glue, Athena, Tableau, RDS, DMS).
- Strong background in networking, DNS, and VPNs (including Route 53 management).
- Solid understanding of security, compliance, and identity management (WAF, MFA, IAM Identity Center).
- Soft Skills: Must be a self-starter, adept at working cross-functionally, and possess strong communication skills.
Job Features
| Job Category | DevOps |
A Cloud Security Engineer is sought to play a pivotal role in safeguarding the organization's cloud environments. The primary focus is to design, implement, and manage security solutions across multi-cloud infrastructure to ensure the confidentiality, integrity, and availability of data and services. This role champions a security-by-design approach, integrating robust controls and continuous monitoring.
This is a full-time, remote position. The budgeted salary range is $127,545 – $159,431 USD.
Core Responsibilities and Security Mandate
The Engineer will collaborate closely with compliance, IT, and DevOps teams to align deployments with regulations and best practices, integrating security throughout the software development lifecycle (SDLC) and CI/CD pipelines.
Major Duties:
- Architecture and Design: Design and implement secure cloud architectures and controls for infrastructure, applications, and data across multi-cloud environments (AWS, Azure, OCI).
- Policy and Enforcement: Develop and enforce security policies, standards, and best practices, ensuring alignment with compliance frameworks (HIPAA, SOC2, SOX).
- Security Automation: Work with DevOps teams to incorporate security automation, utilizing Infrastructure as Code (IaC) tools like Terraform/Ansible to enforce configurations.
- Tool Management: Configure, maintain, and review security tools, including Web Application Firewalls (WAFs), encryption, key management, and endpoint security solutions (CloudFlare, Crowdstrike).
- Monitoring and Response: Monitor and respond to security incidents using SIEM solutions (Splunk, Azure Sentinel) and conduct regular security assessments, vulnerability scans, and penetration tests.
- Guidance: Provide expert security guidance during engineering, IT, and application design reviews.
Required Experience and Technical Qualifications
The ideal candidate has significant experience in public cloud security, hands-on experience integrating security into CI/CD, and a strong understanding of compliance and foundational IT principles.
Experience and Technical Requirements:
- 5-7 years of experience in public cloud security, including AWS, Azure, and Oracle Cloud Infrastructure (OCI).
- DevSecOps: Hands-on experience with DevOps practices and security integration into CI/CD pipelines and automated security testing tools.
- IaC Proficiency: Working knowledge with Infrastructure as Code (IaC) tools like Terraform/Ansible.
- Security Tooling: Advanced experience deploying and customizing security tools such as WAFs, encryption, key management, FIM, and DLP solutions in Cloud environments.
- System Foundation: Strong engineering skills with Windows Server, DNS, DHCP, Active Directory, and network switching.
- Compliance: Familiarity implementing security controls to comply with HIPAA, SOC2, and SOX assessments/audits.
- SIEM: Experience configuring, optimizing, and analyzing logs using Splunk and Azure Sentinel.
Preferred Certifications:
- CCSP (Certified Cloud Security Professional)
- AWS Certified Security – Specialty
- Microsoft Certified: Azure Security Engineer Associate
- GIAC Cloud Security Essentials (GCLD)
Job Features
| Job Category | Cloud Engineering |
A cutting-edge sports betting and analytics startup based in San Francisco is hiring a DevOps Engineer. This is a highly technical role focused on scaling infrastructure that supports real-time data pipelines for fantasy sports and betting platforms, working at the intersection of machine learning and predictive analytics.
This is a Direct Hire, Hybrid position with a salary range of $120,000 - $170,000. Crucially, candidates must reside within the Pacific Time Zone (PST) to support real-time collaboration with the team.
Role Summary and MLOps / Containerization Mandate
This DevOps Engineer will play a critical role in managing and scaling the infrastructure that supports high-velocity, real-time data pipelines and ML workloads in AWS. You will build robust CI/CD pipelines, intelligent automation, and optimized cloud systems, working directly with data scientists and engineers.
Technical Focus Areas:
- Kubernetes & Container Orchestration: 40% of focus
- AWS Cloud Infrastructure Management: 30% of focus (e.g., EKS, EC2, RDS)
- CI/CD Pipeline Development & Maintenance: 20% of focus
- ML Infrastructure Support: 10% of focus
Daily Responsibilities:
- 60% Hands-On DevOps Engineering.
- 20% Collaboration with Data Science & Engineering Teams.
- 20% Monitoring, Incident Response & On-Call Support (for 24/7 production environments).
Required Experience and Technical Qualifications
The ideal candidate brings deep experience with AWS and Kubernetes at scale, proficiency with Infrastructure-as-Code (IaC), and a track record of supporting high-availability production systems.
- Experience (Required):
- 5+ years in DevOps, DevSecOps, or SRE roles.
- 5+ years working with Kubernetes, Docker, Helm.
- 2+ years of scripting/programming in Python or Go.
- Previous experience in 24/7 on-call production environments.
- Core Technical Stack:
- Strong experience with AWS (EKS, EC2, RDS, etc.).
- Proficient with Terraform or Terragrunt for Infrastructure-as-Code.
- Deep CI/CD experience with GitHub Actions, ArgoCD, or Jenkins.
- Solid understanding of networking, distributed systems, and cloud security.
- Desired Skills (Bonus):
- Experience supporting machine learning pipelines or MLOps.
- Understanding of advanced statistics or analytics workflows.
- Interest in sports data, betting, or fantasy.
Job Features
| Job Category | Data, DevOps, Technical Services |
Machinify, a leading healthcare intelligence company that leverages AI to maximize financial outcomes for health plans, is seeking a DevOps Engineer. This critical role is focused on automating and maintaining the company's AI/ML cloud technologies, with a primary focus on the migration of VM technologies into Kubernetes environments.
This is a full-time, fully remote position in the United States. The salary range is $150,000.00 - $180,000.00 per year.
Role Summary and Kubernetes Migration Mandate
The DevOps Engineer will provide technical leadership across automation, programming, and system operations. The mandate is to ensure the Machinify Cloud operates at big data scale, high uptime, and with operational excellence. You will be directly responsible and accountable for the design, engineering, and integration within production environments, focusing on future-proof implementations while solving current scaling challenges.
Key Responsibilities:
- Kubernetes Migration: Facilitate the movement of VM technologies into Kubernetes through migration or replacement for all applicable solutions.
- Automation: Automate everything to ensure zero manual intervention, routinely redeploying systems from the ground up to validate automation integrity.
- Architecture: Architect solutions to achieve a high level of performance, reliability, scalability, and security for the distributed compute environment.
- AI/ML Technologies: Create, maintain, and troubleshoot distributed compute AI/ML technologies running in the Cloud.
- Configuration Management: Plan, design, and execute the successful build and deployment of code updates and implement the configuration management of all underlying technologies.
- Compliance & Security: Work in/create compliant environments such as Hi-Trust/SOC2 and apply good Operations security practices.
Required Experience and Technical Qualifications
The ideal candidate is an experienced, logical problem-solver with a strong background in production cloud environments, deep Kubernetes skills, and exposure to distributed AI/ML frameworks.
- Experience (Required):
- 5+ years of production support, preferably in a Cloud Environment (AWS, Azure, or GCP).
- Experience in the migration of VM Technologies into a Kubernetes Environment.
- Core Technical Knowledge:
- Containerization with Kubernetes
- Scripting (Python, shell, etc.)
- Crossplane / Terraform (Infrastructure as Code)
- Linux (CentOS/RHEL)
- Spark / Machine Learning running in the Cloud
- Frameworks for distributed machine-learning / AI, such as Azure OpenAI, AWS Bedrock, TensorFlow, and MxNet.
- Continuous Integration/Continuous Deployment frameworks.
- Compliance & Security:
- Good understanding of Operations security practices.
- Working in / creating compliant environments such as Hi-Trust / SOC2, etc.
- Education & Skills: Degree in Computer Science or equivalent work experience. Must be extremely logical with the ability to solve problems in creative and efficient ways, and demonstrate critical thinking.
Job Features
| Job Category | Cloud Engineering, DevOps |
Woolpert, an award-winning global leader in architecture, engineering, and geospatial services, is seeking a Cloud Infrastructure Engineer to join its dynamic Cloud Solutions and Interdisciplinary Services (CSIS) team. This role is focused on the successful design and delivery of client projects, with a strong emphasis on Google Cloud (GCP) technologies and DevOps practices.
This is a full-time, client-facing, remote position in the United States.
Role Summary and GCP Implementation Mandate
Reporting to the Software Engineering Manager, this engineer will provide technical leadership, collaborate on client discovery, and perform hands-on delivery and testing within an agile/scrum environment. A key function is researching rapidly evolving GCP and DevOps technologies to ensure the delivery of cutting-edge solutions.
Key Responsibilities:
- Client Engagement & Discovery: Perform client and project discovery to understand business and technical requirements, including architecture and security reviews of existing systems.
- Infrastructure Design: Design cloud infrastructure to ensure cost-effective, reliable, and secure cloud solutions, primarily using Google Cloud.
- Infrastructure as Code (IaC): Write code to implement and test infrastructure using IaC for repeatable, cloud-based projects.
- Automation: Develop CI/CD pipelines to automate deploying both software and infrastructure.
- Site Reliability Engineering (SRE): Implement Site Reliability Engineering practices and processes.
- Security & Compliance: Ensure systems are meeting required security guidelines and frameworks.
- Collaboration: Work closely with the Cloud Solutions Architect, Technical Project Manager, and client engineering teams to drive successful project delivery in an agile/scrum environment.
Required Experience and Technical Qualifications
The ideal candidate is a highly experienced cloud or systems engineer with proven proficiency in Google Cloud infrastructure, containerization, and multiple programming/scripting languages.
- Experience (Required): 8+ years of experience in a cloud or systems engineering role.
- Google Cloud Proficiency: Demonstrated proficiency in designing and deploying cloud infrastructure in Google Cloud (compute, storage, networking, security, monitoring/operations).
- Containerization: Demonstrated proficiency in deploying container-based applications (Kubernetes, Cloud Run, Fargate).
- Infrastructure as Code (IaC): Demonstrated proficiency in at least one IaC language (Terraform, Pulumi, AWS CDK).
- Programming/Scripting: Demonstrated proficiency in at least one programming or scripting language (GO, Bash, Python, NodeJS/ES6, etc.).
- Methodologies: Solid understanding of SDLC and Agile software development methodologies and tools (e.g., Git, Jira).
- Soft Skills: Strong communication and collaboration skills are essential for client-facing work.
Preferred Skills (Bonus Points):
- Google Cloud Platform professional certifications.
- Multi-cloud deployment migration experience.
- Professional services experience, designing and delivering large enterprise solutions.
- Cloud-native application development.
Job Features
| Job Category | DevOps, Software Engineering |
Pinterest is hiring a Site Reliability Engineer II (SRE) to join their Engineering team. The SRE organization is crucial for ensuring Pinterest's overall availability and enhancing engineering teams' ability to design, build, and operate robust systems at scale. This role focuses on developing and building systems that assure the reliability of large-scale distributed systems handling billions of page views and petabytes of data.
This is a Regular, full-time position that is Remote. The role requires being in the office for in-person collaboration 1-2 times per half (twice a year), meaning candidates can be situated anywhere in the country.
Role Summary and Reliability Mandate
The SRE II will apply software engineering principles to infrastructure and operations problems, specializing in scaling, optimizing, and automating critical systems. You will gain a deep understanding of complex system behaviors to identify risks and implement long-term solutions that minimize operational overhead.
Key Responsibilities:
- Software Development for Reliability: Develop software solutions to enable the reliability and operability of large-scale distributed systems that handle petabytes of data.
- System Insight and Risk Identification: Build a deep understanding of how Pinterest’s systems behave, scale, interact, and fail, using that insight to identify risks and opportunities for remediation.
- Toil Elimination and Automation: Build tools and automation to eliminate toil and reduce operational overhead. Create frameworks, processes, and best practices for use across Pinterest Engineering.
- SLI/SLO Implementation: Build meaningful, insightful, and actionable SLIs (Service Level Indicators).
- Process Automation: Automate critical portions of Pinterest’s engineering processes to minimize risk and maximize the speed of innovation.
- Capacity Management: Manage capacity and performance to help scale the infrastructure across both public and private clouds around the world.
Required Experience and Technical Qualifications
The ideal candidate is a software-oriented SRE with experience in large-scale distributed systems, a strong background in Linux internals, and proficiency in modern programming and infrastructure technologies.
- Experience (Required):
- 2+ years of experience programming using Python or Go.
- Strong knowledge of Linux/Unix/BSD internals and experience working with open source software (e.g., MySQL, Hadoop, Envoy, HAProxy, Nginx).
- Experience with technologies such as ElasticSearch, ZooKeeper, HBase, Hadoop, Memcache, and Kafka with a focus on reliability, automation, operability, and performance.
- Education: Bachelor’s degree in Computer Science or a related field, or equivalent experience.
- Infrastructure Knowledge (A Plus):
- Infrastructure as Code (IaC) experience is a plus (e.g., Terraform, Puppet, Chef, Ansible, Salt, Fabric, Docker, etc.).
- Experience with deploying web apps to cloud infrastructure (AWS, etc.) and working with distributed, service-oriented architecture.
Job Features
| Job Category | Software Engineering |
We are seeking an AWS CNS Engineer—a specialized cloud network engineer—to design, implement, and maintain secure, scalable, and high-performing network environments within Amazon Web Services (AWS). This role focuses on using deep technical skill and ingenuity to support mission-critical cloud services.
This is a full-time position based in Mapleton, UT, United States, offering remote work (home office flexibility).
Role Summary and Network Automation Mandate
The AWS CNS Engineer will be expected to continuously improve and deliver advanced AWS network systems while relentlessly pursuing automation. You will be instrumental in creating secure and resilient environments using core AWS networking services and security tools, supporting clients in an increasingly connected world.
Key Responsibilities:
- Design & Deployment: Design and deploy advanced AWS network systems, creating blueprints and roadmaps for VPCs, Transit Gateways, Direct Connect, and VPNs.
- Security & Optimization: Secure and optimize cloud environments using native tools like AWS Shield, NACLs, Security Groups, and industry best practices.
- Automation: Automate relentlessly using Python, Terraform, and CloudFormation to streamline deployment and maintenance and eliminate friction.
- Troubleshooting & Support: Provide high-level troubleshooting and support, ensuring reliability, stability, and resilience of the network systems.
- Collaboration: Work across disciplines with architects, developers, and security teams to deliver scalable, secure solutions.
- Documentation: Document and share designs, processes, and lessons learned to strengthen the team.
Required Experience and Qualifications
The ideal candidate has a strong technical background in both traditional networking and cloud-native AWS services, with a demonstrated ability to automate infrastructure deployments.
- Minimum Requirements:
- Bachelor's degree in computer science, engineering, or a related technical field and 3 years of relevant experience.
- Hands-on AWS networking experience with services like VPC, Direct Connect, Route 53, Transit Gateway, and Load Balancing.
- Strong foundation in TCP/IP, DNS, DHCP, routing, and switching.
- Experience with Linux and Windows environments.
- Must be authorized to work in the USA.
- Compensation & Benefits:
- Competitive compensation: $60–99K salary, 401K, bonuses, and stock options.
- Medical and dental benefits.
- Long-term career development with expert mentorship and continuous training opportunities.
Additional Qualifications (Highly Preferred):
- AWS Certified Advanced Networking Specialty or equivalent certification.
- Familiarity with compliance standards (PCI DSS, HIPAA, etc.).
- United States Government contractor experience (preference given to Veterans Affairs).
- Ability to pass a background check for a public trust position.
Job Features
| Job Category | Cloud Engineering |
We are seeking a Site Reliability Engineer (SRE) with dedicated expertise in database management and optimization to join the DevOps team. This mid-level, experienced role is focused on managing and optimizing SQL and NoSQL database infrastructure at scale, ensuring systems are resilient, performant, and secure through code.
This is a Full-Time, Remote position available in the USA & Canada. The compensation range is $120,000 - $150,000 USD.
Role Summary and Database Reliability Mandate
This SRE role sits at the critical intersection of database administration, software engineering, and operations. You will be responsible for applying SRE principles to mission-critical data systems (e.g., Postgres, MongoDB), driving continuous improvement in availability and performance through automation, advanced monitoring, and rigorous incident response.
Key Responsibilities:
- Database Reliability & Performance: Maintain and optimize SQL and NoSQL database systems, focusing on improving availability, latency, and scalability.
- Automation & IaC: Design and implement automation for provisioning, configuration, and maintenance using Python, Bash, and Infrastructure-as-Code (IaC) tools like Terraform or Ansible.
- Observability & Monitoring: Own the setup and refinement of monitoring systems (e.g., Prometheus, Grafana, Datadog) to ensure deep visibility into database health and anomalies.
- Incident Management: Lead or contribute to on-call rotations, triage production issues, and perform thorough Root Cause Analysis (RCA) to drive long-term reliability.
- Performance Tuning: Analyze slow queries, indexing strategies, and schema design to improve database efficiency and throughput.
- Resilience & Security: Implement and validate robust backup, recovery, and disaster readiness strategies. Enforce database security policies, access controls, and compliance best practices.
- Collaboration: Partner with software engineers, data engineers, and DevOps teams to align database architecture with application needs and business continuity goals.
Required Experience and Technical Expertise
The ideal candidate is an experienced SRE or infrastructure engineer with significant, direct experience supporting production database systems in a cloud-native environment.
- Experience (Required):
- 5+ years in SRE, DevOps, or infrastructure engineering roles, with direct experience supporting production database systems.
- Technical Expertise:
- Solid experience with relational and NoSQL databases (e.g., Postgres, SQL Server, MongoDB), including comfort with query optimization, replication, and failover.
- Proficiency in Python, Bash, or similar scripting languages.
- Experience with CI/CD pipelines and Infrastructure-as-Code tools.
- Hands-on experience with cloud platforms (AWS, Azure, GCP) and container orchestration (Docker, Kubernetes).
- Mindset: Demonstrated ability to troubleshoot complex systems independently and drive resolution. Strong written and verbal communication skills to influence and align stakeholders.
- Education: Bachelor’s degree in Computer Science, Engineering, or a related technical field, or equivalent practical experience.
Job Features
| Job Category | Data, DevOps, Software Engineering |
Ditto, a fast-growing, globally distributed startup that provides a peer-to-peer sync engine for resilient, real-time data at the edge, is seeking a Platform Engineer to own and architect their CI/CD infrastructure. This role is highly strategic, giving the engineer end-to-end ownership of the critical deployment platform used by every engineer at Ditto.
This is a full-time, remote position in the United States. The base salary range is $169,176 – $300,000 USD.
Role Summary and Deployment Architecture Mandate
This Platform Engineer will define how Ditto ships software at scale, balancing rapid velocity with safety and reliability. You will build the systems that enforce quality and enable effortless, confident deployments across a microservices architecture. This is a chance to directly multiply the productivity of the entire engineering organization.
Key Responsibilities:
- Platform Ownership & Architecture: Own the architecture and evolution of Ditto's entire CI/CD platform, spanning from code commit to production deployment.
- Quality & Safety Enforcement: Design and build automated quality enforcement systems, including code coverage analysis, security vulnerability scanning, and compliance tracking infrastructure.
- Deployment Safety Mechanisms: Create advanced deployment safety mechanisms, such as automated validation pipelines, progressive rollout frameworks, and intelligent automatic rollback systems.
- Optimization & Standards: Optimize deployment pipeline infrastructure to enable rapid, confident releases. Establish the technical vision and standards for deployment safety, quality automation, and production risk management.
- Developer Experience: Partner with engineering teams to understand pain points and evolve the platform based on real developer needs, focusing on developer experience and pragmatism.
Required Experience and Technical Mindset
The ideal candidate is a seasoned platform engineer who views CI/CD as a cohesive system and possesses strong systems thinking, experience in production environments, and the ability to influence without formal authority.
- Experience (Required):
- Deep experience building and scaling CI/CD infrastructure, deployment pipelines, or developer platform tooling in production environments.
- Strong systems thinking—thinking-the ability to architect for long-term platform evolution, seeing CI/CD as a cohesive platform, not disconnected tools.
- Track record of driving adoption through empathy and pragmatism (developer experience focus).
- Comfort operating under visibility and production responsibility (deployment infrastructure is mission-critical).
- Excellent stakeholder management skills and the ability to influence without authority.
- A proactive attitude about how AI can support automation and thoughtful decision-making in the role.
Great-to-Haves (Bonus):
- Experience implementing progressive delivery patterns (canary deployments, feature flags, traffic shaping) at scale.
- Background in SRE, production operations, or incident management.
- Prior work in high-growth or hypergrowth environments.
- Familiarity with modern deployment ecosystems (Kubernetes, ArgoCD, GitHub Actions, Harness, or similar) and observability platforms.
- Experience establishing quality standards and achieving enthusiastic adoption across engineering organizations.
Job Features
| Job Category | Cloud Engineering |
AIS (Applied Information Sciences), an employee-owned company focused on federal defense and intelligence missions, is seeking a Cloud Engineer. This individual will be a key member of a cross-functional team, responsible for designing, implementing, and maintaining secure, scalable, and resilient cloud-based solutions specifically aligned with Zero Trust, DevSecOps, and DoD Cloud Smart principles.
This is an Onsite, full-time position, and candidates must be based in the DC Metro area (Washington, D.C., Northern Virginia, or Maryland).
Role Summary and Defense Technology Mandate
This role is central to modernizing DoD environments and delivering secure, cost-effective digital engineering platforms for the warfighter. The engineer will leverage deep technical expertise to deploy, integrate, and optimize mission systems across secure federal cloud environments.
Key Responsibilities:
- Cloud Architecture & Design: Develop and document cloud reference architectures aligned to DoD and federal strategies. Design resilient solutions (IaaS, PaaS, SaaS) across environments including Azure Government, AWS GovCloud, DoD IL5/6, and JWCC.
- Containerization: Leverage containerization and orchestration technologies (Kubernetes, AKS, EKS, OpenShift) for secure, portable deployments.
- Engineering & Automation: Build Infrastructure as Code (IaC) solutions using Terraform, Bicep, or CloudFormation. Implement CI/CD pipelines (Azure DevOps, GitHub Actions, Jenkins) for automated testing and deployment.
- AI/ML Integration: Integrate AI/ML and data platform services into cloud architectures to enable advanced analytics and digital engineering.
- Security & Compliance: Apply DoD Zero Trust principles and STIG compliance to cloud workloads. Ensure compliance with FedRAMP, NIST 800-53, and DISA SRG controls, supporting Authority to Operate (ATO) processes through RMF documentation and validation.
- Operations: Monitor and optimize cloud performance, cost, and availability. Troubleshoot complex issues across networking, compute, storage, and security domains.
Required Experience and Technical Qualifications
The successful candidate will be a mission-driven cloud professional with a foundation in engineering, significant experience with IaC, and the necessary DoD clearance and security certifications.
- Experience (Required): 3–7 years in cloud engineering, architecture, or DevSecOps roles.
- Clearance: Active DoD Secret clearance (minimum), with the ability to obtain and maintain Top Secret/SCI if required.
- Certifications: CompTia Security+ certification required.
- Cloud Expertise: Hands-on experience with Azure, AWS, or Google Cloud; familiarity with federal cloud environments is preferred.
- Infrastructure as Code (IaC): Proficiency in tools such as Terraform, Bicep, and CloudFormation.
- Technical Knowledge: Deep understanding of DoD security frameworks, including RMF, DISA SRGs, and Zero Trust principles. Strong foundation in networking, security, identity, and access management.
- Education: Bachelor’s degree in Computer Science, Engineering, or a related technical field, or equivalent professional experience.
Desired Qualifications (Bonus Points):
- Microsoft Azure Administrator (AZ-104) or Azure/AWS Solutions Architect certifications.
- Experience supporting DoD digital engineering initiatives (e.g., Model-Based Systems Engineering).
- Experience with data platforms and analytics services (Databricks, Synapse, Redshift).
- Familiarity with container security and automated compliance frameworks.
Job Features
| Job Category | Cloud Engineering |
CrowdStrike, a global leader in cybersecurity, is hiring a Software Engineer to join its PEAK (Platform Evaluation and Attack Knowledge Center) team. This role is a hybrid of a Full Stack Developer and a DevSecOps Engineer, focusing on building innovative testing platforms and automation frameworks that validate CrowdStrike's security products against sophisticated, real-world threats.
This is a full-time, remote position in the United States.
Role Summary and Threat Emulation Mandate
The mission of this role is to redefine how security products are tested. The Engineer will combine the mindset of a red teamer with the skills of a software engineer to build scalable, AI-driven testing services. The ultimate goal is to directly measure, analyze, and enhance the effectiveness of CrowdStrike's security products against the latest attack techniques.
Key Responsibilities:
- Full-Stack Development & Testing Platforms: Design and develop full-stack applications for automated security testing and threat emulation.
- Automation & CI/CD: Build scalable testing frameworks that integrate seamlessly with CI/CD pipelines. Create and maintain RESTful APIs for test automation services.
- Security Validation: Implement AI-driven validation systems to analyze product performance. Collaborate with red team members to translate manual testing procedures into automated workflows.
- Visualization: Develop dashboards and visualization tools for reporting security testing metrics.
- Infrastructure & Collaboration: Support and enhance existing testing infrastructure. Collaborate across multiple disciplines, including threat researchers, detection engineers, and data scientists.
Required Experience and Technical Qualifications
The ideal candidate possesses deep full-stack development expertise across multiple languages and is highly proficient in modern cloud-native tools, with a crucial understanding of security testing methodologies.
- Experience (Required):
- 5+ years of experience in full-stack development.
- Core Technical Stack:
- Strong proficiency in Python, Golang, JavaScript, and at least one modern frontend framework (React, Angular, or Vue.js).
- Experience with containerization (Docker) and orchestration (Kubernetes).
- Solid understanding of CI/CD principles and tools (Jenkins, GitLab CI, or similar).
- Strong SQL and NoSQL database experience.
- Experience with REST API design and implementation.
- Security Knowledge:
- Knowledge of security testing methodologies and the MITRE ATT&CK framework.
- Soft Skills: Excellent problem-solving and analytical skills, an energetic "self starter" mentality with a focus on root cause analysis, and clear communication skills.
- Education: Bachelor's degree in Computer Science, Software Engineering, or related field.
Bonus Points:
- Experience with red teaming or penetration testing.
- Knowledge of cloud platforms (AWS, Azure, or GCP).
- Experience with machine learning frameworks and AI implementation.
- Security certifications (OSCP, CISSP, or similar).
- Background in automated testing frameworks.
Job Features
| Job Category | DevOps, Software Engineering |
A global leader in Governance, Risk, and Compliance (GRC) solutions is hiring a Software Engineer: DevOps. This key role involves actively engaging with cross-functional teams to build and maintain the cloud infrastructure and tooling that supports their enterprise GRC SaaS product, Risk Cloud®.
This is a full-time, remote position in the United States. The anticipated base salary range is $100,000 - $150,000 per year + variable + equity + benefits. This role is integral to service excellence and includes some responsibilities for 24/7 on-call support.
Role Summary and AWS Infrastructure Mandate
The DevOps Engineer's primary focus will be on AWS infrastructure management, requiring a proven track record of building and sustaining robust environments. The role spans the full spectrum of modern DevOps practices, from Infrastructure as Code (IaC) and CI/CD to internal tooling development and observability.
Key Responsibilities:
- AWS Infrastructure Management: Build new and support existing infrastructure within AWS environments. This expertise is crucial to operations.
- Infrastructure as Code (IaC): Utilize Terraform to automate and manage cloud infrastructure, build custom modules, and employ Ansible for large-scale automation tasks.
- CI/CD Pipeline: Work closely with Engineering and Information Security teams to deploy, maintain, and operate GitLab CI/CD runners and GitLab infrastructure/pipeline code.
- Internal Tooling & Automation: Develop and maintain internal tools using Golang and Python, leveraging a Git-based development workflow to enhance operational capabilities.
- Observability: Implement and manage monitoring and logging frameworks, primarily using Datadog, for optimal performance and quick issue resolution.
- Containers: Collaborate with teams to construct, maintain, and deploy containerized environments using Docker.
- Financial Management: Monitor and manage cloud costs, including AWS Savings Plans and Reserved Instances, to identify potential savings.
Required Experience and Technical Stack
The ideal candidate is a highly experienced DevOps professional with a strong foundation in cloud security and a deep familiarity with the specified technology stack, particularly within a SaaS environment.
- Required Experience:
- 5 years of experience in a DevOps or related role, preferably in a SaaS or Financial Services environment.
- 3+ years in AWS, Terraform, Kubernetes, Docker, GitLab, Go, Datadog, and OpsGenie (Required).
- Core Technical Expertise:
- AWS Infrastructure Management: Proven track record of successfully building and maintaining robust AWS infrastructures.
- Terraform Expertise: A solid grasp of Terraform for infrastructure management and module creation.
- Programming: Practical experience in programming with Golang and/or Python for automation tools.
- Containers: Proficiency with Docker and a strong understanding of containerization technology.
- Observability: Strong background in monitoring, logging, and using tools like Datadog to enhance system observability.
- CI/CD: Hands-on experience with GitLab and GitLab Runners for setting up and managing CI/CD processes.
- Security Knowledge: Knowledgeable in security best practices for infrastructure, including patching, penetration testing, and familiarity with WAR, FTR, and CIS Benchmarks.
Job Features
| Job Category | DevOps, Software Engineering |
A position is available for a DevSecOps Engineer to join a team focused on developing secure CI/CD pipelines and integrating Artificial Intelligence (AI) capabilities. This role requires a strong blend of DevOps practices, security principles, and practical experience with implementing AI/ML technologies.
This is a full-time, Regular position with travel required up to 10% (Local). No minimum security clearance is required to start.
Role Summary and AI/Security Mandate
The ideal candidate will be instrumental in bridging the gap between secure, continuous delivery and the emerging field of machine learning operations (MLOps). You will be responsible for designing and automating secure software development lifecycles while ensuring the seamless integration and secure deployment of AI/ML models.
Key Responsibilities:
- Secure CI/CD Pipelines: Design, implement, and maintain secure CI/CD pipelines using tools like GitLab.
- AI/ML Integration: Integrate AI/ML models and tools into existing DevOps processes, including deployment and configuration.
- Security Automation: Develop and implement security measures throughout the SDLC, including automating security testing and vulnerability assessments (SAST, DAST).
- Infrastructure & Containerization: Implement and manage containerization and orchestration technologies (Docker, Kubernetes), utilizing Infrastructure-as-Code (IaC) tools (Terraform, Ansible).
- Monitoring & Optimization: Monitor and optimize system performance and security, including developing and maintaining documentation for processes.
- Collaboration: Collaborate with development teams to ensure best practices in security and AI integration within an Agile environment.
Required Experience and Technical Qualifications
The successful candidate will have a foundational background in IT or software engineering and specific experience spanning cloud, automation, security, and AI implementation.
- Required Qualifications:
- Bachelor's degree in Computer Science, Software Engineering, or related field.
- 3+ years of experience in DevOps or related roles.
- Strong knowledge of CI/CD tools (e.g., GitLab).
- Proficiency with Linux/Unix admin and scripting languages (e.g., Python, Bash).
- Experience with cloud platforms (e.g., OCI, AWS).
- Experience with containerization technologies (Docker, Kubernetes).
- Knowledge of security best practices and tools (SAST, DAST).
- Experience with infrastructure-as-code tools (e.g., Terraform, Ansible).
- Experience implementing AI technologies.
- Experience with Agile methodologies.
- Desired Skills (Bonus):
- Experience with AI/ML model deployment and integration.
- Knowledge of AI/ML frameworks (TensorFlow, PyTorch, scikit-learn).
- Familiarity with MLOps practices.
- Certifications in cloud platforms, security, or DevOps (e.g., AWS Certified DevOps Engineer, CISSP).
- Experience with monitoring and logging tools (e.g., ELK stack, Prometheus, Grafana, Langfuse).
Job Features
| Job Category | AI (Artificial Intelligence), DevOps, Software Engineering |