Job Purpose / Role Overview This role is responsible for the development and operation of the data warehouse stack and data quality, integrating additional data sources and providing processed data for consumption, BI and reporting. A key focus is enabling the business to make data-driven decisions that will support innovation and continuous sustainable growth. Key Responsibilities • Responsible for the development and operation of the data warehouse stack and data quality, integrating additional data sources and providing processed data for consumption, BI and reporting. • Build and maintain scalable data pipelines written in Python and SQL and ran on AWS/Snowflake • Management and maintenance of the data architecture and configuration. Deployment of resources within the AWS estate to ensure optimal performance, scalability, and cost efficiency. • Taking ownership of data quality within projects • Managing and educating a range of stakeholders when gathering requirements and delivering data projects • Building effective and collaborative relationships with technical and non-technical colleagues • Reacting to change within the business in a positive and constructive way • Documenting processes and data pipelines • Working as part of an Agile (Scrum, kanban) team, and collaborating with other development, product, design and test teams to ensure the best implementation and highest quality product Requirements Experience working with complex data sets. • Data pipeline development experience in Python (especially using pandas, parsing API responses, posting data to API’s etc.) • Experience of building data warehouses, enterprise data warehouses or data lakes with ETL pipelines • Experience with AWS tools including Glue, Lambda, S3, RDS, is preferred • An understanding of CICD concepts including having a working knowledge of GitHub Skills and Competencies • A strong data mindset • Ability to thrive in a fast paced and dynamic environment • Exceptional SQL skills including an appreciation for performance tuning on a variety of database engines • Excellent communication, collaboration, and problem-solving skills. • An appreciation for the big picture while delivering short/mid-term solutions • Understanding of different data warehouse architectures is preferred (especially dimensional modelling and data vault 2.0 architectures) • Experience with agile methodologies such as Scrum or Kanban, and tools such as Jira and Git 5. Qualifications • Desirable - Bachelor's degree in computer science, data science, software engineering, information systems or related discipline. • Desirable - Certifications in scripting/programming language, such as python or SQL. • Desirable – Certifications in cloud computing on platforms such as AWS Benefits Hybrid working, with regular collaboration days in our Sheffield HQ Opportunity to work at an established but rapidly growing EdTech scaleup NEST Pensions scheme Buy & Sell Holiday scheme offering an opportunity to boost holiday allowance to up to 38 days annually Access to company Life Assurance scheme SmartHealth - access to 24/7 virtual GP, mental health support, financial advice and more Who are The National College The National College are a fast growing and innovative EdTech business headquartered in Sheffield. We are a market-leading and multi-award-winning provider of professional development and software tools that support over 45,000 schools worldwide to ensure compliance and drive up standards. Our cutting-edge platform has revolutionised online training and now boasts the world's largest professional development library for educators The National College is part of the National Education Group. CancelAdd Benefits