About the job
We are seeking a talented Senior Data Engineer with extensive knowledge of real-time data streaming and distributed data processing to architect, develop, and enhance state-of-the-art data platforms. This pivotal role is essential for advancing event-driven architecture and real-time analytics within critical banking systems, particularly in risk and compliance domains.
In this position, you will work synergistically with data architects, platform engineers, and business stakeholders to create low-latency, high-throughput data pipelines that empower sophisticated analytics and informed decision-making.
Key Responsibilities
- Design, develop, and maintain robust real-time streaming pipelines utilizing Apache Kafka, PySpark, and Flink
- Construct scalable and fault-tolerant event-driven data architectures
- Handle high-volume streaming data ensuring low latency and high reliability
- Integrate diverse data sources into centralized data platforms (Data Lake / Lakehouse)
- Enhance data pipelines for performance, scalability, and cost-effectiveness
- Uphold data quality, governance, and compliance in line with banking regulations
- Collaborate with cross-functional teams to convert business needs into technical solutions
- Monitor and debug streaming jobs and production pipelines
Required Skills & Experience
- 5+ years of experience in Data Engineering
- Demonstrated proficiency in:
- PySpark / Spark Streaming
- Apache Kafka (Producers, Consumers, Kafka Streams)
- Apache Flink or other real-time processing frameworks
- Proven experience in building real-time / near real-time data pipelines
- Strong understanding of distributed systems and event-driven architecture
- Proficiency in Python / Java / Scala
- Experience with data lakes, ETL/ELT pipelines, and big data ecosystems
- Familiarity with cloud platforms (AWS / Azure / GCP) is advantageous
- Knowledge of banking, risk, or compliance data systems is highly preferred
Preferred Qualifications
- Experience in the financial services or banking domain
- Exposure to data governance, regulatory reporting, or compliance systems
- Understanding of CI/CD pipelines and DevOps practices for data platforms

