24815909 Lead Python Data Engineer - VP (Hybrid)
Job Description
We are looking to hire an experienced Data Engineer to join our team responsible for building technology solutions for the Markets Business Controls Technology and for various business risk and controls functions. You will work with the Architects, Business stakeholders, and Data Scientists to ensure we are meeting their requirements to develop and drive automation. This is an exciting opportunity to work on an important project for the group, which will have a huge impact on the business and future design in this area.
Key Responsibilities:
* Develop and automate large scale, high performance data processing systems (batch and real-time) to drive growth and improve product experience.
* Develop and maintain infrastructure tooling for our data systems.
* Collaborate with software teams and business analysts to understand their data requirements and deliver quality fit-for-purpose data solutions.
* Ensure data quality and accuracy by implementing data quality checks, data contracts, and data governance processes.
* Contribute to the ongoing development of our data architecture and data governance capabilities.
* Develop and maintain data models and data dictionaries.
* Serve as advisor and coach to junior members of the team.
* Identify problems, analyze information, and make evaluative judgments to recommend and implement solutions.
* Work collaboratively in teams and develop meaningful relationships to achieve common goals.
Essential Skills & Experience:
* Extensive experience as a Lead Engineer including evaluating and providing technical solutions to solve a variety of complex and interdependent processes.
* Significant professional experience in Data engineering, ETL processing, and Data Warehousing.
* Commercial exposure on development using PySpark/Spark, particularly on data processing and the associated modules.
* Cloud - Exposure to Kubernetes/Red Hat OpenShift.
* Experience working with Open table/storage formats like Delta Lake, Apache Iceberg, or Apache Hudi.
* Have worked on Delta Lake and exposure on consuming real-time messages via Kafka.
* Experience on Unix, HDFS, and supported libraries such as PyArrow. PySpark experience is preferred as the environment uses Hadoop and Spark.
* Professional working knowledge of Databases (SQL/NoSQL).
* Working knowledge of BitBucket or equivalent SCM tool.
* Knowledge of Apache Airflow is desirable.
* Ability to multi-task and deliver to a tight deadline.
* Experience as an SME to senior stakeholders and/or other team members.
* Skilled in evaluating and performing POC’s on new strategic technical products and applications and its applicability to the strategic platform.
This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.
What we’ll provide you:
By joining Citi London, you will not only be part of a business casual workplace with a hybrid working model (up to 2 days working at home per week), but also receive a competitive base salary (which is annually reviewed), and enjoy a whole host of additional benefits such as:
* 27 days annual leave (plus bank holidays).
* A discretional annual performance-related bonus.
* Private Medical Care & Life Insurance.
* Employee Assistance Program.
* Pension Plan.
* Paid Parental Leave.
* Special discounts for employees, family, and friends.
* Access to an array of learning and development resources.
Alongside these benefits, Citi is committed to ensuring our workplace is where everyone feels comfortable coming to work as their whole self, every day. We want the best talent around the world to be energized to join us, motivated to stay, and empowered to thrive.
#J-18808-Ljbffr