# Distributed Systems Engineer - Data Platform - Logs and Audit Logs

**Company**: Cloudflare
**Location**: Hybrid
**Work arrangement**: hybrid
**Experience**: senior
**Job type**: full-time
**Salary**: Competitive salary and benefits package
**Category**: Engineering
**Industry**: Technology

**Apply**: https://job-boards.greenhouse.io/cloudflare/jobs/7340469?utm_source=yubhub.co&utm_medium=jobs_feed&utm_campaign=apply
**Canonical**: https://yubhub.co/jobs/job_16a0f69d-8a6

## Description

About Us

At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.

We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.

About Role

We are actively seeking experienced and highly motivated Distributed Systems Engineers to join Cloudflare’s dynamic DATA Organisation. This is a pivotal opportunity to contribute to the future of data at Cloudflare, working on systems that are fundamental to our global operations and customer insights.

Our organisation is responsible for the entire data lifecycle, encompassing everything from initial ingestion and sophisticated processing to robust storage and efficient retrieval. These systems are the backbone that power critical logs and analytics, providing our customers with real-time, actionable visibility into the health, performance, and security of their online properties.

Our overarching mission is to empower customers to leverage their data effectively, enabling them to drive superior outcomes for their businesses. To achieve this, we design, build, and maintain a suite of high-performance, massively scalable distributed systems that are engineered to handle an unprecedented scale – processing well over a billion events per second.

As an engineer within our organisation, you will be presented with unique challenges across various critical parts of our intricate data stack. This role offers the chance to work on cutting-edge technologies and contribute to solutions that operate at the very edge of internet infrastructure.

Responsibilities

This particular role is focused on the Logs and Audit Logs group. As a Software Engineer you will focus on the following areas:

- Design, build, and operate a robust logging platform, ensuring reliable logging, and secure data transfer to a wide array of customer destinations and third-party integrations.

- Develop and maintain high-performance data connectors and integrations for our log-shipping products, focusing on usability, scalability and data integrity.

- Create and manage systems for handling comprehensive audit logs, ensuring they are delivered securely and adhere to strict compliance and performance standards.

- Scale and optimise the data delivery pipeline to handle massive data volumes with low latency, identifying and removing bottlenecks in data processing and routing.

- Work closely with Product and other engineering teams to define requirements for a new logging platform and integrations.

- Maintain the operational health of our log delivery platform through comprehensive monitoring and participation in an on-call rotation (with flexibility for out-of-hours technical issue resolution).

- Collaborate on the architectural evolution of our data egress platform, researching and implementing new technologies to improve efficiency and reliability.

Key Qualifications

- 3+ years of experience working in software development covering distributed systems and data pipelines.

- Strong programming skills (Go is preferable), with a deep understanding of software development best practices for building resilient, high-throughput systems.

- Hands-on experience with modern observability stacks, including Prometheus, Grafana, and a strong understanding of handling high-cardinality metrics at scale.

- Strong knowledge of SQL, including experience with query optimisation.

- A solid foundation in computer science, including algorithms, data structures, distributed systems, and concurrency.

- Strong analytical and problem-solving skills, with a willingness to debug, troubleshoot, and learn about complex problems at high scale.

- Ability to work collaboratively in a team environment and communicate effectively with other teams across Cloudflare.

- Experience with data streaming technologies (e.g., Kafka, Flink) is a strong plus.

- Experience with various logging platforms or SIEMs (e.g., Splunk,

## Skills

### Required
- Go
- Distributed systems
- Data pipelines
- SQL
- Query optimisation
- Prometheus
- Grafana
- High-cardinality metrics
- Algorithms
- Data structures
- Concurrency
- Analytical and problem-solving skills
- Debugging
- Troubleshooting
- Learning about complex problems at high scale
- Collaboration
- Communication
- Data streaming technologies
- Kafka
- Flink
- Logging platforms
- SIEMs

---

Source: [Apply at job-boards.greenhouse.io](https://job-boards.greenhouse.io/cloudflare/jobs/7340469?utm_source=yubhub.co&utm_medium=jobs_feed&utm_campaign=apply)
