Who are we?

Careem is the leading technology platform for the greater Middle East. A pioneer of the region’s ride-hailing economy, Careem is expanding services across its platform to include payments, delivery and mass transportation. Careem’s mission is to simplify and improve the lives of people and build a lasting institution that inspires. Established in July 2012, Careem operates in more than 130 cities across 16 countries and has created more than one million job opportunities in the region

About your new team

We are on a mission to extend and build a much larger dream data warehouse team here at Careem and as a team motto, we truly believe in the famous “no brilliant jerks” policy a brief of which is outlined below:
On a dream team, there are no “brilliant jerks.” The cost to teamwork is just too high. Our view is that brilliant people are also capable of decent human interactions, and we insist upon that. When highly capable people work together in a collaborative context, they inspire each other to be more creative, more productive and ultimately more successful as a team than they could be as a collection of individuals.

The Data Warehouse team at Careem builds and supports solutions to organize, process and visualize large amounts of data. You will be working with technologies such as Hive, Spark, Spark streaming, Kafka, Python, Redash, Presto, Tableau and many others to help Careem become a data-informed company. Some of the problems the team is working on are: Customer 360, ELT engineering, reporting infrastructure, data reliability, data discovery and access management.

About your new role

  • Work with engineering and business stakeholders to understand data requirements
  • Help shape the growth of in-house data resources to support business Intelligence, analytics, and data science
  • Define, design, model, implement, and operate large, evolving, structured and unstructured datasets
  • Take action with quality, performance, scalability, and maintainability in mind
  • Interact and integrate with internal and external teams and systems to extract, transform, and load data from a wide variety of sources
  • Implement secure and auditable data infrastructure as required
  • Help execute the roll out of modern business intelligence and analytics tools and visualizations
  • Train users on data capabilities and best practices

You have:

  • 7+ years of experience with designing, building and maintaining scalable ETL pipelines
  • Good understanding of data warehousing concepts and modeling techniques
  • Hands-on experience working with real time data processing using Kafka, Spark Streaming or similar technology
  • An understanding of Spark core/internals to read and understand spark catalyst plan and perform optimizations. Understanding of spark concepts like RDD, spark data frame, spark APIs
  • Knowledge of data modeling and schema design especially for distributed, column-oriented databases
  • Hands-on experience experience with workflow processing engines like Airflow, Luigi
  • Ability to dig deeper into the issues of the ETL pipelines, understand the business logic and provide permanent fixes

Good to have:

  •  Experience with CICD using Jenkins, Terraform or other related technologies
  •  Familiarity with Docker and Kubernetes
  • Experience with apache Cassandra

Benefit Summary:

  • Competitive remuneration 
  • Premium medical insurance (including spouse and children)
  • Unlimited leave* 
  • Discounted Careem rides
  • Entrepreneurial working environment
  • Flexible working arrangements
  • Mentorship and career growth

Apply for this Job

* Required