At STERIS, we help our Customers create a healthier and safer world by providing innovative healthcare and life science product and service solutions around the globe.
Position Summary
The Data Architect demonstrates mastery of skills and knowledge and is a mentor, strategist, thought leader, evangelist, champion, Plans and leads data engineering activities. Responsible for overseeing the design, development, deployment, and maintenance of scalable and robust data solutions. Develops and manages data integration pipelines connecting disparate data sources. Works closely with data architects, data scientists, analysts, and other stakeholders to support business needs in analytical and data solutions/projects. Collaborates with Infrastructure and DBA teams to ensure appropriate infrastructure is in place. Optimizes and streamlines data processing efforts to ensure data quality, security, privacy, on time delivery and compliance. Provides technical leadership, mentorship, reviews deliverables and provides feedback to the data engineering team.
Duties
- Data Architecture and Technical Infrastructure (30%): Defines, plans, designs and support implementation of enterprise data architectures and enterprise data platform. Plans and leads data engineering activities for strategic, large, and complex programs. Leads the selection and development of data engineering methods, tools, and techniques.
- SDLC Methodology & Project Management (15%): Plans technical transitions between development, testing, and production phases of solutions' lifecycle, and the facilitation of the change control, problem management, and communication processes.
- Innovation, Continuous Improvement & Optimization (15%): Develops organizational policies, standards, and guidelines for the development and secure operation of data services and products. Ensures adherence to technical strategies and architectures.
- Data Modelling / Designing Datasets (10%): Coordinates the application of analysis, design, and modelling techniques to establish, modify or maintain data structures and their associated components. Manages the iteration, review and maintenance of data requirements and data models.
- Partnership and Community Building (10%): Collaborates with other IT teams, business community, data scientists and other architects to meet business requirements.
- Data Pipeline/ETL (5%): Sets standards for data modelling and design tools and techniques, advises on their application, and ensures compliance. Defines and implements administration and control activities related to data warehouse planning and development and the establishment of policies and procedures pertaining to its management, security, maintenance, and utilization.
- Support & Operations (5%): Manages the investigation of enterprise data requirements based upon a detailed understanding of information requirements.
- Data Governance and Data Quality (5%): Ensures that data is reliable, secure, and timely. Implement Data privacy and best practices. Defines, designs and implements data quality assessment and improvement methodology, processes and solutions.
- End-User Support, Education and Enablement (3%): Plans, designs, develops and facilitates training and Data Literacy initiatives within the team and End user community.
- Metadata Management & Documentation (2%): Ensure standards, and best practices in documentation of Metadata, Data Engineering processes and Architectures.
Required Experience
- Bachelors Degree in Data Science, Analytics, or a related field
- 7 years’ experience in development, maintenance, and enhancement of Data Pipelines (ETL/ELT) and processes with thorough knowledge of star/snowflake schemas
- 7 years’ experience in developing complex SQL queries and SQL optimization
- 7 years’ experience in development experience must be full Life Cycle experience including business requirements gathering, data sourcing, testing/data reconciliation, and deployment within Business Intelligence/Data Warehousing Architecture.
- 5 years’ experience in designing and implementing data security
- 7 years’ experience in monitoring and optimizing data storage and data processing
- 2 years’ experience in delivering Data Solutions using Cloud Technologies
- Advanced SQL skills and experience with relational databases and database design like Oracle and SQLSERVER.
- Significant experience working with cloud Data Warehouse and Data Lake solutions (e.g., Snowflake, Redshift, BigQuery, Azure Data Lake Storage, Amazon S3, etc.)
- Experience working with data ingestion tools such as Fivetran, stitch, or Matillion.
- Working knowledge of Cloud-based solutions (e.g., Azure, AWS and GCP).
- Experience building and deploying machine learning models in production.
- Strong proficiency in object-oriented languages: Python, Java, C++, Scala.
- Strong proficiency in scripting languages like Bash.
- Strong proficiency in data pipeline and workflow management tools (e.g., Airflow, Azkaban).
- Familiarity with big data frameworks such as Apache Hadoop and Apache Spark
- Strong project management and organizational skills.
- Excellent problem-solving, communication, and organizational skills.
- Demonstrated leadership experience and skills
- Ability to communicate effectively and influence technical and business stakeholders at all levels of the organization.