What we do
We invite you to be part of our ambitious, close-knit team creating systems for large customers who need to crunch through Tbps of data in real-time. Our competitive advantage lies in performance-oriented software engineering vs. server sprawl in our customers’ data centres.
Our data processing pipeline starts with our proprietary NCORE engine (capable of pattern matching 5M patterns at 100Gb/s per 2RU), and ends with a declarative aggregation system. It empowers our analysts and helps us respond on demand to the most complex customer requirements.
In the process, we leverage industry-standard tools, the full power of Linux, and powerful in-house software to achieve maximum performance, flexibility, and data integrity.
We know the devil is in the details. You will improve performance through better data processing strategies, improved algorithms, and code optimisation. You will help ensure that the ETL pipelines are fully tested and running smoothly across the board.
We work end-to-end. You will cooperate with our traffic processing and live analytics teams to implement data engineering solutions that are both efficient and secure ― handling events from hundreds of millions of users, to offer marketing and network quality insight on demand with no compromise for individual information protection.
We want to show off. To attract the best programmers we plan to showcase our technology. You can be part of our effort to open source interesting pieces of our technology stack.
Your role as Data Engineer
As a Data Engineer, you will work as part of the Analytics team that handles the streaming, filtering, processing, and storage of aggregated and real-time data.
Using Python, SQL, standard UNIX tools, and in-house software, you will help define, run, implement, and monitor large data processing pipelines, analysing data for hundreds of millions of users all around the world. You will also be responsible for managing, profiling, and improving our ETL pipelines.
You will have to think both big, in terms of overall architecture, and small, in terms of low-level optimisations, to deliver solutions that are reliable, reusable, extendable and can leverage the performance of the best hardware.
What we value
- Proficiency in Linux-based systems and development tools
- Proficiency in profiling Linux systems/applications
- Familiarity with Data Processing and ETL pipelines
- Proficiency in shell scripting (bash)
- Proficiency in SQL
- Minimum 3 years experience in Python programming
- Experience with at least one compiled language (C/Java/Go/C++/C#) and/or Postgres is a plus
- Familiarity with Apache Airflow and/or Zabbix, or similar tools, is nice-to-have