Skip to content

Data Science Engineer, Assistant Senior Manager at Diamond Trust Bank (DTB)

posted 2 hours ago
Job Overview
Employment FullTime
Location Nairobi Kenya
Experience At least 3-7 years
Education Level Bachelor's Degree
svg background up
Opportunities Meet Aspirations

Data Science Engineer, Assistant Senior Manager

Job Purpose:

In this role, you will design, build, and optimize the data engines that power DTB’s intelligence. You will develop robust data pipelines, feature stores, model‑serving systems, and scalable big‑data platforms that enable advanced credit scoring, fraud detection, customer intelligence, and a wide range of machine‑learning applications.

You will be at the heart of transforming DTB into a data‑driven organization—ensuring that teams across the bank can rely on high‑quality, trusted, and scalable data to drive smarter decisions, stronger governance, and innovative digital solutions. This is a high‑impact role for a builder, a problem‑solver, and a visionary ready to shape the future of data and AI at DTB

Key Responsibilities:

 Science & ML

  • Build and maintain ETL/ELT pipelines that feed modelling datasets from multiple banking systems (CBS, LMS, CRM, Cards, Mobile Banking, Bureau, Collections systems).
  • Develop automated data preparation workflows for credit scoring, fraud models, behavioral models, and IFRS9 modelling.
  • Create end-to-end ML pipelines integrating feature engineering, data validation, model deployment, and monitoring.
  • Manage and Build other Enterprise ETL using tools like ODI , informatica etc.

Big Data Platform Engineering

  • Develop scalable data-processing workflows using Spark, Hadoop, Kafka, Airflow, Flink or similar.
  • Optimize large datasets (transactional, bureau, behavioural, logs) for modelling in batch and real-time environments.
  • Manage distributed computation and ensure reliability and fault tolerance.

Feature Store & Data Assets Management

  • Design and maintain a centralized feature store for credit, fraud, marketing, and customer analytics models.
  • Ensure feature consistency between training and serving environments.
  • Implement versioning, lineage, documentation, and metadata management for data features.

Model Deployment & MLOps

  • Collaborate with data scientists to deploy models using MLflow, Docker, Kubernetes, API gateways, CI/CD pipelines.
  • Develop automated monitoring pipelines for model performance, drift detection, data quality, and explainability.
  • Ensure models operate efficiently in real-time decision engines and batch scoring environments.

Data Quality & Governance

  • Implement robust data validation, profiling, anomaly detection, and reconciliation checks.
  • Work with Data Governance teams to ensure compliance with IFRS9, Basel, CBK, GDPR, and internal data standards.
  • Manage data lineage, cataloguing, and documentation to support audits and regulatory reviews.

Collaboration & Stakeholder Support

  • Partner with Data Scientists, Risk, Credit, Fraud, Marketing, and BI teams to align data pipelines with business use cases.
  • Work with IT and Infrastructure teams on cluster performance, security, access controls, and SLA adherence.
  • Participate in sprint planning, architecture reviews, and model implementation committee sessions.

Performance Optimization

  • Improve the efficiency, scalability, and cost of ML workloads.
  • Optimize database queries, Spark jobs, Kafka streams, and storage systems.

Qualifications & Experience:

  • Strong academic foundation with a Bachelor’s or Master’s in Computer Science, Data Engineering, Data Science, Information Technology, or a related quantitative field.
  • 3–7+ years of impactful, hands‑on experience in data engineering, big‑data processing, or building scalable ML infrastructure—ideally within fast‑paced, data‑driven environments.
  • Advanced programming capability, with strong proficiency in Python, SQL, and PySpark; experience with Scala is an added advantage.
  • Demonstrated expertise in modern data and ML platforms, including:
    • Big‑data technologies: Spark, Hadoop, Kafka, Airflow
    • MLOps & containerization: MLflow, Docker, Kubernetes
    • CI/CD pipelines: GitLab, Jenkins, GitHub Actions
    • Cloud platforms: AWS, GCP, or Azure (highly preferred)
  • Experience working with banking systems, risk data, or credit‑modelling datasets—a significant advantage that accelerates success in this role.

Key Competencies

  • Strong understanding of data structures, distributed systems, and ML workflows.
  • Excellent problem-solving, debugging, and optimization skills.
  • Fast learner with ability to adapt to new technologies.
  • High attention to detail, documentation discipline, and data governance awareness.
  • Strong collaboration and communication skills.


Share This Post

Don't miss out on new jobs listing! Follow our channels Today WhatsApp Channel

Disclaimer Opened Career is a free job-posting website that does not charge applicants. We do not support recruitment agents or entities that demand money or favors to expedite the hiring process. Please use our platform responsibly and report any suspicious activity.
Why Opened Career
OUR OBJECTIVES
At Opened Career, we prioritize inclusivity, diversity, and equal opportunities for all individuals, regardless of their backgrounds or experiences. We believe in creating a level playing field where every candidate has the chance to showcase their skills and potential, and every employer has access to a diverse pool of qualified candidates.
CORE VALUES
Innovation
Integrity
Team Work
Excellence
Customer Focus
Professionalism