Role Overview
As a Data Architect / Data Engineer, Integrations, you will tackle the challenges of building reliable, scalable data pipelines and integrations in complex, enterprise-scale environments. You will design, build, and optimize solutions that handle large-scale, inconsistent, and unstructured data, ensuring clean, structured, and secure data to power intelligent systems. You will also enable cloud-agnostic data storage and processing solutions to support client-specific environments, including multi-cloud, on-premises, or dedicated tenancies. Collaborating closely with Research Engineers and Research Scientists, you will ensure data is modelled effectively to support advanced retrieval, real-time insights, and AI-enabled decision support.
Key Responsibilities
* Architect and Build: Design, implement, and optimize scalable data pipelines and architectures to process high-velocity, large-scale data from diverse sources.
* Integrate Diverse Systems: Develop and maintain integrations across enterprise platforms, project tools, communication systems, and external APIs to enable seamless bi-directional data flow across the organization.
* Enhance Data Quality: Design processes to validate, transform, and structure inconsistent and unstructured data, ensuring it is clean, reliable, and optimized for downstream use.
* Collaborate Across Teams: Work closely with Research Engineers and Research Scientists to ensure data is structured effectively to support advanced retrieval, query performance, and AI-enabled insights.
* Leverage Best-in-Class Tools: Build and optimize streaming data pipelines using Azure tools such as Event Hubs, Stream Analytics, and Data Explorer to deliver real-time data insights, risk detection, and proactive alerting.
* Utilize Cloud Expertise: Leverage Microsoft Azure services for secure, cost-effective, and high-performance data solutions, while enabling cloud-agnostic data processing and storage solutions to meet client-specific requirements, including AWS, GCP, or on-premises environments.
* Ensure Security and Compliance: Implement role-based access controls, encryption, and logging to ensure data security and compliance with standards such as ISO 27001, SOC 2, and GDPR.
* Drive Efficiency: Identify opportunities to improve data flow, processing times, and cost-effectiveness across the data infrastructure.
* Stay Current: Keep up to date with emerging trends in cloud platforms, data engineering, and real-time processing technologies to future-proof the infrastructure.
Expertise and Skills
Core Technical Competencies:
* Cloud Platforms: Expertise in Microsoft Azure (Event Hubs, Stream Analytics, Data Explorer, Synapse) with a strong understanding of AWS, GCP, and the ability to deliver cloud-agnostic and on-premises data solutions.
* Data Engineering Tools: Experience with Databricks, Apache Spark, or similar frameworks for large-scale data processing.
* ETL/ELT Pipelines: Proven ability to design and manage scalable data pipelines using tools like Azure Data Factory, dbt, or Apache Airflow.
* Programming Proficiency: Advanced proficiency in Python and SQL for building and automating data processing workflows.
Data Architecture & Integration:
* Data Modelling: Experience designing relational, non-relational, and graph-based data models with strong permisioning and access control structures.
* System Integrations: Familiarity with integrating enterprise tools (e.g., ERP, project management, document stores) and working with APIs or streaming frameworks.
* Real-Time Processing: Practical experience with streaming technologies like Azure Event Hubs, Stream Analytics, or other real-time tools.
Security & Compliance:
* Best Practices: Experience implementing encryption, logging, and role-based access control to align with standards such as ISO 27001, SOC 2, and GDPR.
* Data Governance: Understanding of data lineage, cataloguing, and governance frameworks.
Mindset & Approach:
* Problem Solver: Thrives on tackling complex data challenges and delivering robust, scalable solutions.
* Collaborative Partner: Works effectively across teams, particularly with Research Engineers and Research Scientists, to align data infrastructure with business and technical needs.
* Detail-Oriented: Committed to ensuring high data quality, security, and performance.
* Continuous Learner: Eager to explore emerging tools and techniques to push the boundaries of data engineering.
What Success Looks Like
Success in this role will be measured by the robustness and efficiency of our data pipelines, the seamless integration of diverse systems, and the ability to deliver clean, secure, and well-structured data. Your solutions will support client-specific needs, enabling processing and storage across cloud-agnostic, multi-cloud, or on-premises environments. Your work will directly power intelligent, AI-enabled insights and decision-making for enterprise-scale programmes.
What We Offer
* Competitive salary
* Bonus scheme
* Wellness allowance
* Fully remote working (with regular company get-togethers)
* Private medical and dental insurance*
* Life assurance, critical illness cover, and income protection*