Have a question?
Message sent Close
Remote
United States
Posted 1 week ago

Flock Safety, a leading safety technology platform providing cloud software and hardware for crime prevention to communities and law enforcement, is seeking a Technical Support Engineer. This role is responsible for incident management, advanced technical troubleshooting, and collaborating with Product and Engineering teams to drive resolution of escalated issues.

This is a Full-time, Remote (USA) position with a compensation range of $90,000 – $95,000 annually, plus Stock Options (Equity). A condition of employment requires the ability to obtain and maintain Criminal Justice Information Services (CJIS) certification.


The Opportunity and Impact

As a Technical Support Engineer within the Customer Experience organization, you will act as the primary technical point of contact for hardware and software support cases. The role requires advanced troubleshooting skills, the ability to validate issues, and the creation of "engineering-ready" tickets for true defects.

How You’ll Make an Impact:

  • Advanced Troubleshooting: Troubleshoot and resolve escalated issues by analyzing system logs, error codes, and customer incidents, and reproduce software issues in production/dev environments.
  • API & Database Support: Execute and understand basic SQL queries, including database updates, and utilize API tools like Postman to test and validate API-related issues.
  • Authentication Support: Support SSO and non-SSO authentication systems such as Azure, Okta, and Auth0.
  • Incident Management: Manage the incident management process, ensuring visibility and awareness across the business, and triage/escalate critical issues to the Engineering team.
  • Continuous Improvement: Identify trends and perform proactive analysis on support cases, partnering with Product teams to drive continuous improvement and develop customer workarounds.
  • Defect Reporting: Write technical defect reports with clear reproduction steps for Engineering teams.

Required Skillset

The ideal candidate has strong technical troubleshooting and communication skills, with mandatory experience supporting cloud-based SaaS solutions and foundational knowledge in SQL and APIs.

  • Experience (Mandatory):
    • 4+ years of advanced technical support experience.
    • Prior experience supporting cloud-based SaaS solutions (Salesforce Incident Management experience is a plus).
    • Strong technical troubleshooting skills, including reading logs, interpreting error codes, and reproducing issues.
    • Experience writing technical defect reports with clear reproduction steps for Engineering teams.
  • Technical Knowledge:
    • Experience executing and understanding basic SQL queries.
    • Familiarity with API tools like Postman.
    • Experience supporting SSO systems (Azure, Okta, Auth0).
  • Clearance Requirement: Ability to obtain and maintain Criminal Justice Information Services (CJIS) certification.
  • Soft Skills: Exceptional verbal and written communication skills, with the ability to convey complex technical issues clearly and empathetically.

Job Features

Job CategoryInformation Technology, Support Service, Technical Services

Flock Safety, a leading safety technology platform providing cloud software and hardware for crime prevention to communities and law enforcement, is seeking a Technical Support Engineer. This role is ...View more

Imply, the company built by the creators of Apache Druid and focused on big data analytics, and their new observability warehouse, Imply Lumi, is hiring a Senior Technical Support Engineer. This role is central to the Support Team, providing deep technical troubleshooting and partnership to ensure customer success with the Imply Druid platform and the new Lumi offering across multiple cloud and hosted environments.

This is a Full-time, Remote position with a compensation range of $115,000 - $155,000 annually, plus eligibility for the equity plan.


Role Summary and Deep Troubleshooting Mandate

This is a high-impact, customer-centric role requiring a dynamic and technology-savvy individual who can resolve diverse, real-time problems. You will leverage technical and interpersonal skills to partner with Engineering, Customer Success, and Product teams to drive customer issues to resolution on a complex, distributed data platform.

Responsibilities:

  • Deep Troubleshooting: Perform deep troubleshooting and resolve customer issues on the Imply Druid platform within specified Service Level Agreements (SLAs).
  • Platform Coverage: Attend to operational issues on Imply’s SaaS offering (Polaris), Imply Hybrid hosted on AWS, and self-hosted clusters on major Cloud platforms (AWS, GCP, and Azure).
  • Complex Diagnosis: Diagnose, reproduce, and resolve a wide range of issues spanning Imply components, including Pivot (data visualization), Clarity (monitoring), backend DB (Mysql, Postgres), deep storage (HDFS, S3, Azure blob), and ZooKeeper, sometimes involving clusters with hundreds of nodes.
  • Documentation & Enablement: Write technical content in the form of knowledge-based articles dedicated to operational efficiency to empower teammates and customers.
  • Issue Reporting: Assess the business impact of issues/defects/requests and create detailed reports in Jira for the Engineering team.
  • On-Call: Participate in an on-call rotation to cover after-hours and weekend/holiday escalations.

Required Experience and Technical Qualifications

The ideal candidate has extensive experience in technical support for complex, distributed systems, with strong skills in Linux, big data platforms, and SQL proficiency.

  • Experience (Mandatory):
    • 5+ years of experience in technical support or customer-facing roles.
  • Core Technical Skills:
    • Extensive working knowledge of Linux platforms and system administration.
    • Strong troubleshooting skills and technical knowledge to logically approach problems.
    • Good understanding and working knowledge of big data platform ecosystems and distributed systems.
    • Good understanding and working knowledge of relational database management and proficiency in SQL.
  • Soft Skills: Excellent writing and communication skills with attention to detail; team player mentality.
  • Bonus Points (Preferred):
    • Working experience on major Cloud infrastructure platforms (AWS, Microsoft Azure & GCP).
    • Good understanding and working knowledge of Kubernetes.
    • Experience in supporting streaming applications (Kafka, Kinesis) and ETL applications.
    • Experience in supporting big data ecosystems on Cloud infrastructure or cloud SaaS applications.
    • Experience in data modeling & visualization or with Splunk or other observability tools.

Job Features

Job CategoryCustomer Success, Information Technology, Technical Services

Imply, the company built by the creators of Apache Druid and focused on big data analytics, and their new observability warehouse, Imply Lumi, is hiring a Senior Technical Support Engineer. This role ...View more

A home furnishings brand, operating since 1962, is seeking a Senior End User Support Engineer to act as the technical expert and primary engineer for their core End-User Computing (EUC) systems. This role is highly strategic, focused on the full lifecycle of device management, VDI, and asset management, with a strong mandate for automation and security.

This is a Fully Remote, Individual Contributor position. The minimum starting rate is $92,000.00 up to $115,000.00 annually.


Role Summary and EUC Strategy Mandate

As the Senior End-User Computing Engineer, you will own the analysis, design, implementation, and operation of Device Management (UEM/MDM), Technology Asset Management (TAM), and Virtual Desktop Infrastructure (VDI). Your daily focus is on driving automation, optimizing operations, and enhancing the user experience while enforcing security and compliance controls.

A Day in the Life:

  • Engineering & L3 Support: Act as the primary engineer and Level 3 support for MDM, VDI, and TAM tools through ticket resolution, incident response, and alert monitoring.
  • Automation & Scripting: Script and automate administrative tasks using various languages and techniques to integrate products and ease administration.
  • Strategy & Product Selection: Advise as a knowledge matter expert and contribute to ** product selection** activities, recommending products and presenting pros/cons to meet organizational requirements.
  • Optimization: Analyze and optimize existing deployments to minimize costs, reduce overlap, and enhance capabilities (e.g., printer support, vendor comparison).
  • Documentation & Training: Create and maintain knowledge base documentation and provide technical training to Service Desk and Operations teams.
  • Project Leadership: Participate, coordinate, and execute projects, advising on risks and blockers.
  • Off-Hours Support: Provide after-hours, weekend, and holiday support when necessary.

Required Experience and Technical Qualifications

The ideal candidate has significant experience architecting and engineering EUC technologies, mandatory expertise in device management platforms, VDI, and automation scripting.

  • Experience (Preferred): 5 years of experience in architecting or engineering End User Computing Technologies.
  • Core Technical Expertise (Mandatory):
    • Technical expert in End User Computing technologies, including Device Management (e.g., Workspace ONE), Technology Asset Management (e.g., Axonius, Jira Insight), and Virtual Desktop Infrastructure (e.g., Azure Virtual Desktop).
    • Technical expert in device management practices for all applicable platforms (Windows, macOS, iOS, Android).
  • Automation & Integrations:
    • Strong experience working with common Scripting languages (Python, PowerShell, Bash).
    • Experience working with DevOps pipelines for deploying and managing VDI environments (e.g., Terraform and Azure DevOps Pipeline).
    • Understanding of and experience working and interacting with APIs.
  • Identity & Collaboration:
    • Working knowledge of SSO and MFA tools (e.g., Okta, DUO, EntraID).
    • Strong understanding of End User Collaboration tools (e.g., O365, Google Workspace, Slack).
  • Soft Skills: Ability to collaborate with technical peers across a highly dispersed work environment and communicate effectively with all company associates, including executive leadership.

Job Features

Job CategoryInformation Technology, Operations Management

A home furnishings brand, operating since 1962, is seeking a Senior End User Support Engineer to act as the technical expert and primary engineer for their core End-User Computing (EUC) systems. This ...View more

Remote
United States
Posted 1 week ago

ECP, a market-leading SaaS platform for senior living communities, is seeking an AWS Cloud Engineer to build scalable, zero-trust infrastructure and develop innovative solutions. This role will collaborate closely with the AWS Cloud Architect to maintain and enhance cloud architectures and implement Infrastructure as Code (IaC).

This is a Full-time, Fully Remote position open to candidates within the U.S. Legal authorization to work in the U.S. without sponsorship is required.


Role Summary and Well-Architected Mandate

The Engineer will be tasked with continuous innovation, development, and optimization of infrastructure within the AWS Cloud ecosystem. A core focus is engineering solutions aligned with the AWS Well-Architected Framework pillars: operational excellence, security, reliability, performance efficiency, and cost optimization.

Key Responsibilities:

  • IaC & Provisioning: Provision and manage infrastructure using Infrastructure as Code (IaC) tools such as Onica Runway and AWS CloudFormation (and Terraform).
  • Security & Compliance: Implement and enforce security best practices, including least privilege, encryption, and network segmentation to build a zero-trust infrastructure.
  • Resilience: Implement and maintain cloud-native disaster recovery and backup strategies to ensure resilience and business continuity.
  • Architecture & Design: Work closely with the AWS Cloud Architect on the design, implementation, and maintenance of infrastructure workflows that support application delivery, scalability, and feature development.
  • Operations & Troubleshooting: Monitor, troubleshoot, and resolve issues across cloud services, deployments, and operational workflows.
  • Documentation: Document infrastructure designs, workflows, configurations, and Standard Operating Procedures (SOPs).
  • Continuous Learning: Stay current with AWS services, emerging features, and evolving best practices.

Required Experience and Technical Qualifications

The ideal candidate has significant hands-on AWS experience, proficiency in IaC (Terraform/CloudFormation), and a foundational skill in automation scripting with Node.js.

  • Experience (Mandatory):
    • 3+ years of hands-on experience working within the AWS Cloud ecosystem.
    • Proven experience with Infrastructure as Code (IaC) tools such as Terraform and AWS CloudFormation.
  • Technical Proficiency:
    • Proficiency in scripting for automation using Node.js.
    • Strong understanding of core AWS services, cloud networking, and security, including EC2, ECS, S3, VPC, RDS, Lambda, API Gateway, IAM, CloudWatch, and Security Groups.
  • Work Environment: Experience working in an agile development environment.
  • Education: Bachelor’s degree in Computer Science, Engineering, Information Technology, or equivalent work experience.

Job Features

Job CategoryCloud Engineering

ECP, a market-leading SaaS platform for senior living communities, is seeking an AWS Cloud Engineer to build scalable, zero-trust infrastructure and develop innovative solutions. This role will collab...View more

Vimeo, a major video platform serving as part of the internet's infrastructure, is seeking a Site Reliability Engineer III to join its Site Reliability & Infrastructure Engineering team. The role spans platform engineering, database administration, release engineering, and internal tools, focusing on designing, developing, deploying, maintaining, and optimizing the platform that powers Vimeo.

This is a Full-time position, available in New York, NY, OR US - Remote. The base salary range is $130,000 - $178,750 in major metro areas, or $117,000 - $160,875 in all other US cities.


Role Summary and Reliability/Platform Mandate

You will work with cloud infrastructure at massive scale, focusing on optimizing performance, driving down outages, and building robust internal toolkits. The goal is to make Vimeo faster, simpler, more scalable, more reliable, and more efficient to operate.

What You’ll Do:

  • Platform Evolution: Build, secure, and evolve platforms that power Vimeo's applications.
  • Tooling & Automation: Build and maintain tooling that makes manual infrastructure work obsolete and enables self-service for hundreds of engineers.
  • Reliability: Improve observability and reliability of applications to reduce outages to an absolute minimum, while reducing MTTA and MTTR.
  • Internal Platform: Contribute to an internal self-service infrastructure platform used by all engineers for application development and deployment.
  • On-Call: Participate in a weekly on-call rotation shared between US and India offices, responding to production incidents and providing internal support.
  • Documentation: Write and maintain thorough documentation to ensure the global team functions as a cohesive unit.

Required Experience and Technical Qualifications

The ideal candidate possesses deep expertise in distributed systems architecture, expert-level Kubernetes administration, database management (MySQL), and significant experience with major cloud providers.

  • Experience (Mandatory):
    • At least 3 years of professional experience in software development or DevOps.
    • High proficiency in at least one general-purpose programming language (C/C++, Go, Java, Ruby, PHP, Python, etc.).
    • Significant experience with major cloud providers (Google Cloud, AWS).
    • Experience with "Infrastructure as Code" platforms such as Terraform.
    • Experience with observability systems (e.g., Datadog, Grafana, Prometheus).
  • Core Technical Expertise:
    • Deep understanding of the architectural patterns of high-scalability distributed systems.
    • Expert-level proficiency in maintaining, optimizing, and administering Kubernetes deployments.
    • Significant experience with deploying and administering MySQL.
    • Strong knowledge of container orchestration, Linux system internals, networking, and secure computing.
  • Bonus Skills (Nice to Have):
    • Knowledge of ArgoCD, Atlantis, Varnish, Memcached, and/or Chef.
    • Experience with generalized or language-specific build systems (make, bazel, etc.).

Job Features

Job CategoryCloud Engineering

Vimeo, a major video platform serving as part of the internet's infrastructure, is seeking a Site Reliability Engineer III to join its Site Reliability & Infrastructure Engineering team. The role ...View more

Inspira Financial, a company providing health, wealth, retirement, and benefits solutions, is seeking a highly skilled DevOps Engineer II to extend, maintain, and develop pipelines for their Inspira IRA Platform. This is a crucial role focused on ensuring continuous, secure, and compliant software delivery leveraging Azure DevOps and modern automation tools.

This is a Full-time, Remote position available across numerous US states. Employees within a 90-minute radius of the Oak Brook, IL headquarters are required to adhere to an in-office work guideline of 4 days per month minimum. The pay range is $91,000 - $111,000 per year.


Job Summary and Azure DevOps Mandate

You will play a crucial role in the successful implementation of complex technical initiatives, coordinating cross-functional teams, and ensuring alignment with business goals. The primary technical focus is on continuous delivery and monitoring within the Microsoft Azure ecosystem, utilizing Configuration as Code principles.

Duties and Responsibilities:

  • Pipeline Design: Design, implement, maintain, and monitor DevOps pipelines leveraging Cloud engineering automation tools to ensure smooth and efficient operation.
  • Security & Compliance: Ensure software deployment and management are secure, reliable, and compliant with industry standards and best practices.
  • Azure Expertise: Demonstrate expertise in configuring, testing, and delivering core systems on Azure, ensuring they meet performance and reliability standards.
  • Configuration as Code: Deliver configuration as code for reproducibility and scalability.
  • Monitoring & Troubleshooting: Monitor system performance and troubleshoot issues proactively on Azure, ensuring minimal downtime.
  • Continuous Improvement: Identify and implement improvements to enhance the scalability, reliability, and efficiency of Azure DevOps pipelines.
  • Documentation & Collaboration: Document application releases comprehensively and collaborate with clients and delivery teams to influence technical decisions.

Preferred Qualifications and Technical Expertise

The ideal candidate has strong, hands-on experience across the modern Azure DevOps and containerization stack, with a focus on automation and compliance.

  • Experience (Mandatory):
    • 3-5+ years of experience as a DevOps engineer or related role, with a focus on Azure environments.
    • Experience with databases, secrets, and storage services, and other cloud engineering automation tools.
  • Core Technical Stack:
    • Strong experience with Azure cloud platforms, ADO pipelines, Flux CD, Helm, Kubernetes (AKS), Docker, Terraform, and Ansible.
    • Proficiency with containerization and orchestration technologies like Docker and Kubernetes (AKS).
    • Experience with monitoring and logging tools such as Azure Monitor, Log Analytics, Data Dog, and Application Insights.
  • Soft Skills: Strong problem-solving and troubleshooting skills; excellent communication and interpersonal skills, including the ability to drive continuous improvement; and professional, empathetic client interactions.

Job Features

Job CategoryDevOps

Inspira Financial, a company providing health, wealth, retirement, and benefits solutions, is seeking a highly skilled DevOps Engineer II to extend, maintain, and develop pipelines for their Inspira I...View more

Zapier, a company building a platform for automation and AI that helps millions of businesses globally scale, is seeking a Site Reliability Engineer (SRE). This high-impact role is on the Reliability Platform team, which owns observability, incident response, and service ownership, with the mission of strengthening Zapier's reliability posture at scale.

This is a Full-time, Remote position, specifically for the NAMER (West Coast) region. The salary range is $141,000 – $211,700 annually.


Role Summary and Observability Mandate

This SRE role goes beyond typical infrastructure work, focusing heavily on observability, incident response, and coding to build systems that make Zapier more resilient. You're expected to thrive in writing production-grade code and proactively find ways to reduce toil and automate repetitive work.

Things You’ll Do:

  • Platform Tooling: Build and improve platform tooling that helps Zapier engineers observe and operate their services.
  • Observability Evolution: Operate and evolve core observability systems, including logging, metrics, alerting, and dashboards, using tools like Grafana, Datadog, Opensearch, and Prometheus.
  • Incident Response: Participate in the team’s on-call rotation and contribute to the broader incident response program by improving processes, tooling, and practices used to detect, respond, and learn.
  • Automation & Infra: Write code to automate operations, improve developer experience, and contribute to infrastructure reliability using AWS, Kubernetes, and Terraform.
  • Best Practices: Review instrumentation designs, suggest improvements, and advocate for effective alerting to raise the bar on observability and reliability across product teams.
  • AI Exploration: Explore and pilot AI-augmented tools (e.g., debugging agents, alert correlation) to improve reliability workflows.

Required Experience and Technical Qualifications

The ideal candidate is an experienced engineer with a strong coding background, deep familiarity with the cloud-native SRE stack, and a proactive, problem-solving mindset.

  • Experience (Mandatory):
    • 4+ years in systems, infrastructure, or backend software roles (SaaS, cloud-native environments preferred).
    • Hands-on experience with observability (metrics, logging, dashboards, alerts) and the ability to reason about instrumentation and alert design.
    • Comfortable jumping into incidents, diagnosing across telemetry, coordinating, and contributing to postmortems.
  • Core Technical Stack:
    • Thrives writing production-grade code in Go, Python, or equivalent.
    • Experience with Infrastructure-as-Code (Terraform, or equivalent).
    • Experience with cloud (AWS) and container orchestration (Kubernetes).
  • Attitude: Thinks proactively about reducing toil and is comfortable influencing peers by suggesting better practices and driving cross-team improvements. Approaches new tools and ideas (especially AI in reliability) with curiosity and openness.

Job Features

Job CategoryCloud Engineering, Data

Zapier, a company building a platform for automation and AI that helps millions of businesses globally scale, is seeking a Site Reliability Engineer (SRE). This high-impact role is on the Reliability ...View more

Hybrid, Remote
Posted 1 week ago

A dynamic team is seeking a talented DevOps Engineer to focus on automation, CI/CD integrations, and optimizing software development workflows, particularly within the GitHub ecosystem.

This is a Contract, Hybrid position, available in Hanover, Texas, or Remote.


Role Summary and Automation Mandate

This Engineer will play a key role in enforcing code quality, managing configuration standards, and improving developer velocity through automation and standardized workflows, with a strong emphasis on the GitHub Enterprise environment.

Key Responsibilities:

  • GitHub Automation: Develop and implement automation scripts for GitHub repository management and configuration.
  • CI/CD & Testing: Implement CI/CD integrations and automated testing gates to enhance software development workflows.
  • Code Standards: Establish and maintain best practices for code quality and traceability across projects.
  • Workflow Design: Design and document a standardized pull-request workflow to facilitate efficient code reviews.
  • Reporting & Compliance: Produce insightful reporting dashboards that summarize project health and collaborate with IT teams to ensure compliance with security and configuration standards.

Required Experience and Technical Qualifications

The ideal candidate has proven experience with DevOps methodologies, strong knowledge of GitHub Enterprise, and familiarity with core CI/CD tools and programming languages.

  • Experience (Mandatory):
    • Proven experience in DevOps practices and methodologies.
    • Strong knowledge of GitHub Enterprise Server and related tools.
    • Solid understanding of CI/CD processes and automation frameworks.
  • Technical Skills:
    • Experience with programming languages such as C, C++, and React.
    • Ability to work collaboratively in a team-oriented environment.
  • Preferred Skills (Bonus):
    • Familiarity with Confluence, Jira Software, and CodeQL.
    • Experience in containerization and cloud technologies.
    • Knowledge of security standards related to software data management.

Job Features

Job CategoryDevOps

A dynamic team is seeking a talented DevOps Engineer to focus on automation, CI/CD integrations, and optimizing software development workflows, particularly within the GitHub ecosystem. This is a Cont...View more

Remote
United States
Posted 1 week ago

River, a financial institution building trusted products powered by Bitcoin, is seeking a security-minded Infrastructure Engineer to secure, simplify, and scale their systems. This high-impact role focuses on infrastructure and security engineering within a Google Cloud Platform (GCP) environment, including some on-premises systems.

This is a Full-time, Remote position within the United States. The salary range is $150,000 – $220,000/yr plus significant equity stock options. U.S. Citizenship is a mandatory requirement.


Role Summary and Security/Scalability Mandate

The Infrastructure Engineer will be a key leader in the infrastructure organization, ensuring the reliability, availability, and security of systems that handle client funds and services related to buying, selling, and securing Bitcoin. You will be expected to take full ownership of the systems you build.

What You Will Be Doing:

  • System Building: Build the systems that power the deployment, monitoring, and scaling of software services.
  • Security & Reliability: Lead projects to improve security, availability, reliability, and scalability across the entire infrastructure.
  • Disaster Recovery: Create and execute on disaster recovery plans.
  • On-Call & Culture: Participate in on-call rotations and actively shape the culture of the infrastructure organization.
  • Problem Solving: Develop solutions to ambiguous problems while supporting engineering teams collaboratively.

Required Experience and Technical Qualifications

The ideal candidate has deep expertise in security and GCP services, specifically running production Kubernetes, and demonstrated hands-on coding ability.

  • Experience (Mandatory):
    • At least 3 years of experience in a DevOps, Infra, and/or SRE role.
    • Experience running Kubernetes in production.
    • Hands-on coding ability.
  • Technical Expertise:
    • Deep understanding of computer and network security.
    • Highly proficient with Google Cloud Platform services like GKE, CloudSQL, LoadBalancing, CloudArmor, Networking, BigQuery, Compute, and Security services.
    • Strong network perimeter defense and analysis skills.
  • Citizenship (Mandatory): Must be a U.S. Citizen.
  • Nice to Haves (Preferred):
    • Experience working on financial services infrastructure.
    • Experience working at a fast-growing startup.
    • Interest in Bitcoin.

Job Features

Job CategoryCloud Engineering, Data, Fintech, Information Technology

River, a financial institution building trusted products powered by Bitcoin, is seeking a security-minded Infrastructure Engineer to secure, simplify, and scale their systems. This high-impact role fo...View more

Remote
United States
Posted 1 week ago

CARET, a technology firm empowering legal and accounting professionals with practice management and automation solutions, is hiring an Intermediate DevOps Engineer. This role is crucial for designing and implementing DevOps processes, automating infrastructure, and troubleshooting complex systems that deliver their services.

This is a Full-time, Fully Remote position in the U.S. (Nationwide Remote). The salary range is $90,000 - $110,000 annually.


Role Summary and Automation Mandate

The Intermediate DevOps Engineer will work closely with various teams to ensure the reliable delivery of services. The ideal candidate will have a strong grasp of system architecture and the complex interplay between servers, databases, APIs, and networking components, with hands-on experience across the entire modern DevOps toolchain.

Key Responsibilities:

  • Infrastructure Automation (IaC): Automate the deployment of infrastructure and maintenance via Terraform and Ansible.
  • CI/CD & Delivery: Develop and deploy pipelines via Azure DevOps.
  • Container Management: Configure and administer Kubernetes, Helmchart, and Docker images.
  • Cloud Infrastructure: Design, build, and maintain multi-cloud infrastructure on AWS and Azure.
  • Monitoring: Implement and configure Datadog monitoring for visibility into system health and performance.
  • Troubleshooting: Troubleshoot and resolve complex issues across a wide variety of service components, including Windows, IIS, NGNIX, DNS, Kubernetes, SQL, MongoDB, Postgres, and Redis.
  • Operations: Deliver projects through planning, design, execution, and completion, and support on-call responsibilities.

Required Experience and Technical Qualifications

The candidate must have over three years of DevOps experience, strong knowledge of multi-cloud environments, and proficiency in IaC tools, scripting, and CI/CD pipelines.

  • Experience (Mandatory):
    • 3+ years of experience as a DevOps Engineer or in a related field.
    • Experience with building and maintaining Azure DevOps pipelines.
    • Support on-call responsibilities.
  • Technical Proficiency:
    • Strong working knowledge of cloud computing platforms such as AWS and Azure.
    • Proficient in writing scripts in languages such as Python, Bash, or PowerShell.
    • Strong understanding of logging and monitoring (e.g., Datadog) and the relationship to environmental health and performance.
  • Education: Bachelor's degree in Computer Science, a related field, or equivalent experience.

Job Features

Job CategoryDevOps

CARET, a technology firm empowering legal and accounting professionals with practice management and automation solutions, is hiring an Intermediate DevOps Engineer. This role is crucial for designing ...View more

United States
Posted 1 week ago

This organization is seeking a skilled DevOps Engineer to contribute to the development and maintenance of infrastructure and deployment processes. The role requires a strong background in automation, containerization, cloud computing, and Continuous Integration/Continuous Deployment (CI/CD).

This is a Full-time position based in Fairfax, VA. The target pay range is $112,422 - $168,634 annually. Crucially, U.S. Citizenship is required, and candidates must have the ability to obtain a Public Trust Clearance.


Role Summary and Technical Mandate

The ideal candidate is a proactive problem-solver with a strong passion for automation and optimizing delivery pipelines. You will be responsible for implementing Infrastructure-as-Code (IaC), managing container orchestration platforms, and setting up comprehensive monitoring and logging systems.

Key Responsibilities:

  • Automation: Utilize scripting languages (Python, Ruby, Shell) to automate tasks.
  • Infrastructure-as-Code (IaC): Implement IaC tools like Terraform or CloudFormation for automating infrastructure provisioning.
  • Configuration Management: Maintain and configure servers using tools such as Ansible, Puppet, or Chef.
  • Containerization & Orchestration: Deploy and manage containerized applications using Docker and utilize orchestration platforms like Kubernetes.
  • CI/CD: Develop and maintain CI/CD pipelines to automate builds, testing, and deployments, working with tools such as Jenkins, GitLab CI/CD, CircleCI, or GitHub Actions.
  • Cloud Services: Deploy and manage infrastructure on cloud platforms (AWS, Azure, or Google Cloud), configuring VMs, databases, load balancers, and networking components.
  • Monitoring & Logging: Set up and configure monitoring tools like Prometheus, Grafana, or Nagios.
  • Version Control: Utilize Git for version control and collaboration.

Required Experience and Technical Qualifications

The candidate must possess at least 5 years of DevOps experience with proficiency across the entire modern DevOps toolchain, including cloud platforms, IaC, and container technologies.

  • Experience (Mandatory):
    • 5–7 years of experience in a DevOps or related role.
    • Experience with CI/CD pipelines and tools (Jenkins, GitLab CI/CD).
  • Technical Proficiency:
    • Proficiency in scripting languages (Python, Ruby, Shell scripting).
    • Experience with automation tools (Terraform, Ansible, Puppet, Chef).
    • Strong understanding of containerization technologies (Docker, Kubernetes).
    • Familiarity with cloud computing platforms (AWS, Azure, Google Cloud).
  • Clearance & Citizenship (Mandatory):
    • U.S. Citizenship Required.
    • Ability to obtain Public Trust Clearance.
  • Education: Bachelor’s Degree in Computer Science, Information Technology Management, or Engineering.

Job Features

Job CategoryDevOps, Technical Services

This organization is seeking a skilled DevOps Engineer to contribute to the development and maintenance of infrastructure and deployment processes. The role requires a strong background in automation,...View more

SafeRide Health is seeking a Site Reliability Engineer (SRE) to join their IT Infrastructure team. This critical role is responsible for ensuring that user-facing services and production systems remain highly available, reliable, and scalable by developing and implementing new processes that support software delivery excellence and operational discipline.

This is a Full-time, Remote position in the United States.


Core Responsibilities and Operational Discipline Mandate

The SRE will focus on minimizing downtime, automating tasks, and proactively managing system health and capacity. A key component involves defining and monitoring Service Level Objectives (SLOs) and collaborating closely with development teams.

  • Reliability & Availability: Focus on availability, reliability, and scalability to keep systems and services running smoothly with minimal downtime.
  • Incident Management: Define and monitor SLOs, respond to and diagnose system incidents, and conduct post-mortems to prevent future occurrences.
  • Automation: Develop and maintain tools and scripts to automate repetitive tasks such as deployments, configuration management, and monitoring.
  • Monitoring & Alerting: Implement and manage monitoring and alerting systems (Prometheus, DataDog, New Relic, Grafana, Splunk) to provide visibility and quickly detect potential issues.
  • Capacity & Risk Mitigation: Perform capacity planning by monitoring resource usage to forecast future needs, and collaborate with stakeholders to identify and mitigate operational risks.
  • Optimization: Analyze metrics from operating systems and applications to identify areas for performance improvement.

Required Experience and Technical Qualifications

The ideal candidate has progressive experience in technology operations, with hands-on proficiency in production monitoring, containerized cloud infrastructure, and automation scripting.

  • Minimum Experience (Mandatory):
    • Minimum of 5 years progressive experience in an IT, Software Engineering, Technology Operations, or Business Continuity role.
    • Minimum of 2 years of hands-on experience in a Site Reliability, DevOps, or IT Observability role.
  • Technical Proficiency:
    • Basic proficiency in an AWS containerized environment running infrastructure as code.
    • Demonstrated proficiency with production monitoring and alerting tools (DataDog is a major plus!).
  • Key Skills:
    • Cloud Technologies: Expertise in major cloud platforms such as AWS and Azure.
    • Tools & Technologies: Experience with tools for Infrastructure as Code (Terraform) and containerization (Docker).
    • Systems Engineering: Deep knowledge of operating systems, networking, storage, and distributed systems.
    • Programming & Scripting: Proficiency in coding languages like Python, Ruby, and JavaScript for automation and infrastructure management.

Job Features

Job CategoryInformation Technology

SafeRide Health is seeking a Site Reliability Engineer (SRE) to join their IT Infrastructure team. This critical role is responsible for ensuring that user-facing services and production systems remai...View more

Remote
United States
Posted 1 week ago

Jasper, the leading AI marketing platform (recognized as one of Fast Company's Top 15 Most Innovative AI Companies of 2024), is seeking an experienced DevOps Engineer to join its Platform team. This is a highly autonomous, high-impact role focused on infrastructure engineering, reliability, and delivery pipeline optimization for AI-powered products at scale.

This is a Full-time, Fully Remote role open to candidates located anywhere in the continental US. The expected base salary range is $170,000 - $200,000 annually, plus an equity grant.


Role Summary and AI/Kubernetes Mandate

Reporting to the Staff DevOps Engineer, you will drive developer velocity and system reliability. The core responsibilities focus on scaling cloud-native infrastructure, implementing advanced CI/CD techniques, and supporting specialized AI/ML compute requirements.

What You’ll Do:

  • Cloud-Native Infrastructure: Design, implement, and operate cloud-native infrastructure (GCP focus) that scales efficiently, fails gracefully, and optimizes for performance and cost.
  • Infrastructure-as-Code (IaC): Develop IaC solutions using Terraform and Helm to create self-healing, automated, and observable systems.
  • Delivery Pipelines: Build and refine software delivery pipelines to enable safe, fast, and frequent deployments with robust testing, rollback, and progressive release mechanisms.
  • AI/ML Support: Collaborate with ML and product teams to support AI model training and inference through scalable compute and storage infrastructure, including GPU-based compute.
  • Reliability Engineering: Identify and eliminate single points of failure, performance bottlenecks, and scalability limits through proactive monitoring and reliability practices.
  • Security: Implement and enforce security best practices, including secrets management, access control, and compliance across all infrastructure layers.

Required Experience and Technical Qualifications

The ideal candidate possesses expert-level skills in running production Kubernetes clusters, utilizing Terraform for IaC, and implementing robust observability with tools like Datadog.

  • Core Expertise:
    • Deep experience running Kubernetes in production (cluster management, networking, storage, security).
    • Expertise with Terraform, Helm, and configuration management to build reproducible, version-controlled infrastructure.
    • Proven success designing and maintaining CI/CD pipelines (GitHub Actions, Argo CD, Jenkins, etc.) balancing speed and safety.
    • Practical knowledge of Google Cloud Platform (GCP) and cloud-native architectures.
  • Automation & Observability:
    • Strong background in observability (especially Datadog)—skilled at instrumentation, dashboard creation, and intelligent alerting.
    • Solid scripting skills in Python, Go, or Bash, with a focus on automation and operational efficiency.
  • AI/Security Experience:
    • Experience supporting AI/ML workloads, including GPU-based compute and multi-language environments (TypeScript, Python, Go).
    • Familiarity with container security, secrets management, and policy enforcement.
  • (Bonus): History of open source contributions in infrastructure, CI/CD, or observability projects.

Job Features

Job CategoryAI (Artificial Intelligence), DevOps

Jasper, the leading AI marketing platform (recognized as one of Fast Company's Top 15 Most Innovative AI Companies of 2024), is seeking an experienced DevOps Engineer to join its Platform team. This i...View more

Carlsbad, California
Posted 1 week ago

Viasat, a global leader in satellite communications, is seeking a DevOps Engineer to join their engineering team. This role is focused on collaborating with software development and network engineering teams to analyze, automate, and deploy software, operating systems, and security updates to a large-scale, mission-critical satellite communications system.

This is a Full-time, Regular position located in Carlsbad, California, that requires U.S. Citizenship due to its work on government projects. The general salary range is $111,500 - $176,500 annually.


Role Summary and Automation Mandate

You will be responsible for software bug and security vulnerability life-cycle tracking, system configuration management, and deployment automation. The core mandate is the development and maintenance of automated software test, build, and deployment processes, focusing on CI/CD into both lab and cloud environments.

Day-to-Day Responsibilities:

  • IaC & Orchestration: Building, deploying, and testing Helm Charts, ArgoCD applications, and/or Terraform configuration files.
  • Automation Development: Development of procedures, scripts, and playbooks to automate the update of company software and COTS application packages.
  • Linux Systems: Deploying automation for Linux operating system updates and patches.
  • Containerization: Building software packages and deploying/testing them in containerized environments.
  • Cloud Infrastructure: Deployment and Maintenance of AWS Infrastructure.
  • CI/CD Maintenance: Development and maintenance of automated software test, build, and deployment processes, including Continuous Integration and Continuous Deployment (CI/CD).

Required Experience and Technical Qualifications

The ideal candidate has a strong foundation in systems engineering and scripting, coupled with mandatory experience in automation, Linux administration, and fundamental networking concepts.

  • Experience (Mandatory):
    • 3+ years of work experience in systems engineering, software engineering, or development operations.
    • 3+ years of work experience using, testing, and developing high-level scripts, preferably with Python and/or Bash languages.
    • Experience working with deployment automation and configuration management tools such as Ansible, Puppet, Chef, or Terraform.
  • Technical Knowledge:
    • Good working knowledge of Linux operating systems and being able to administer Unix/Linux servers.
    • Good working knowledge of software build processes and automation principles.
    • Familiarity with software development tools such as GitHub, Jenkins, Artifactory, Make, etc.
    • Working knowledge of the TCP/IP protocol and networking stack.
  • Education: Bachelor’s degree in Computer Science or a related technical field.
  • Helpful Experience (Preferred):
    • 3+ years working with CI/CD tools such as Ansible, GitHub, Jenkins, Artifactory, and ArgoCD.
    • Work experience with containerization technologies such as Docker, Kubernetes, and/or OpenShift.
    • Some experience with virtualization technologies (ESXI, OpenStack) or SQL database systems.

Job Features

Job CategoryDevOps

Viasat, a global leader in satellite communications, is seeking a DevOps Engineer to join their engineering team. This role is focused on collaborating with software development and network engineerin...View more

Resultant, a professional services firm focused on complex problem-solving for public and private sector clients, is seeking a Cloud DevOps Engineer. This role is committed to staying on the cutting edge of cloud and virtualization technologies, requiring a broad blend of technical skillsets to build, automate, and secure infrastructure and deployment pipelines across multiple cloud environments.

This is a Full-time, Remote position within the Professional Services Cloud Team.


Role Summary and Multi-Cloud Automation Mandate

You will be dedicated to building and managing infrastructure, deployment pipelines, CI/CD tools, architecture, and security. The role offers the opportunity to forge the use of new technologies in creative ways to shape client cloud infrastructure and workflows across major providers.

Key Responsibilities:

  • Multi-Cloud Design & Build: Design and build components encompassing network, security, identity, governance, development, and operations in Azure, Amazon Web Services (AWS), and Google Cloud Platform (GCP) environments.
  • Operationalization & Automation: Operationalize solutions from proof-of-concept through production using automation for repeatability, maintainability, and standard enforcement.
  • Infrastructure as Code (IaC): Script and deploy infrastructure components based on an infrastructure as code mindset, using emerging tools and methodologies to automate existing processes and design new cloud infrastructures.
  • DevOps & Cloud Insight: Provide insight and knowledge of the latest cloud architectures, technology, solutions, industry trends, and DevOps best practices.
  • Collaboration & Compliance: Work closely with cross-functional teams to understand business requirements, drive analysis, and design technical solutions in compliance with architectural blueprints and business strategies.
  • Education: Educate teams on the implementation of new cloud-based initiatives, providing associated training as required.

Required Experience and Technical Qualifications

The ideal candidate possesses hands-on experience with at least one major cloud platform, expertise in IaC (Terraform preferred), and familiarity with containers and CI/CD pipelines.

  • Experience (Mandatory):
    • 2+ years of experience with at least one public cloud (Azure, GCP, or AWS) and/or virtualization (VMware).
    • Hands-on experience with containers (Kubernetes, Docker).
    • Proficiency with at least one Infrastructure as Code tool (Terraform preferred, Ansible, Chef, or similar).
    • Scripting skills in PowerShell, Bash, and/or Python.
    • Experience with Git and CI/CD pipelines.
  • Experience (Nice to Have):
    • 3+ years of systems administration experience.
    • Familiarity with monitoring, observability, and performance optimization tools.
    • Proven “automation-first” mindset with a focus on efficiency and scalability.
  • Education & Certifications (Desired):
    • Bachelor’s or Master’s degree in Computer Science, IT, or related field.
    • Certifications in system management or cloud platforms (CKA, ITIL Foundation, Azure/AWS/GCP Associate-level certifications).

Job Features

Job CategoryCloud Engineering, DevOps

Resultant, a professional services firm focused on complex problem-solving for public and private sector clients, is seeking a Cloud DevOps Engineer. This role is committed to staying on the cutting e...View more