About the Client
Our client is a provider of IT solutions, systems, hardware and software services in banking sector. It operates data centres, bank systems, POS terminals and HW (servers and PC) and is responsible for development, implementation, support and servicing of banking software and its operations in Central and Eastern Europe.
About the Role
The Site Reliability Engineer will be responsible for the administration, optimization, and scaling of our streaming data pipeline ensuring an efficient and reliable IT and data operation. This role requires a dynamic individual capable of managing cloud native systems and ensuring the robustness and efficiency of data flows across various platforms including extensive work with modern No-SQL databases and messaging systems.
Responsibilities
- Oversee the continuous operation and reliability of the data pipeline.
- Develop and execute comprehensive test automation to ensure data quality.
- Monitor, optimize, and ensure high availability and performance of the cloud infrastructure (Own-Cloud, Google Cloud) including server instances, storage, and networking components; identify bottlenecks, and quickly resolve issues.
- Administer, optimize, and scale our distributed technology stack consisting of SQL and noSQL databases such as Oracle, Cassandra and MongoDB as well as real-time data processing using Kafka, thereby ensuring data consistency and availability. At least one of those databases with expert level knowledge.
- Enforce security best practices and compliance standards across the platform.
- Troubleshoot and resolve issues related to the applications and data pipelines.
- Document technical procedures and policies for IT system operations.
- Support and maintain disaster recovery and business continuity strategies to enhance data integrity and availability.
Requirements
- Data Observability Tools: Strong experience with data observability tools such as Dynatrace and Grafana to monitor and diagnose data and system health.
- Educational Background: Bachelor’s degree in Computer Science, Information Technology, or related field.
- Professional Experience: Minimum of 3 years in DevOps, SysOps, or DataOps, particularly in cloud environments and real-time data processing systems.
- Cloud Services Proficiency: In-depth experience with AWS, Azure, or Google Cloud services; understanding of server and network architectures.
- Database Expertise: Experience managing SQL and noSQL databases like Oracle, Cassandra and MongoDB; proficiency in Kafka for real-time data processing.
- Automation and Scripting: Skills in CI/CD automation; Important is any OOP language like Java and any Scripting language.
- Security and Networking: Solid understanding of network architecture and data security protocols
- Communication Sills: Fluent in English and strong communication and collaboration skills
Nice to Have Skills
- IT Service Managment Service Now
- Certifications: Cloud architecture certifications (AWS Certified Solutions Architect, Azure Administrator, etc.)
- Disaster Recovery: Experience with data management and analytics platforms.
- Additional Technologies: Familiarity with other real-time processing frameworks and data management systems.
- Language: Basic German language knowledge