GCP Data Engineer

Company:  HTC Global Services
Location: Troy
Closing Date: 25/10/2024
Hours: Full Time
Type: Permanent
Job Requirements / Description

Job Description:

You will:

  • Work in collaborative environment including pairing and mobbing with other cross-functional engineers.
  • Work on a small agile team to deliver working, tested software.
  • Work effectively with fellow data engineers, product owners, data champions and other technical experts.
  • Demonstrate technical knowledge/leadership skills and advocate for technical excellence.
  • Develop exceptional Analytics data products using streaming, batch ingestion patterns in the Google Cloud Platform with solid Datawarehouse principles.
  • Be the Subject Matter Expert in Data Engineering and GCP tool technologies.


Skills Required:

  • Experience in working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment. Implement methods for automation of all parts of the pipeline to minimize labor in development and production.
  • Experience in analyzing complex data, organizing raw data and integrating massive datasets from multiple data sources to build subject areas and reusable data products.
  • Experience in working with architects to evaluate and productionalize appropriate GCP tools for data ingestion, integration, presentation, and reporting.
  • Experience in working with all stakeholders to formulate business problems as technical data requirement, identify and implement technical solutions while ensuring key business drivers are captured in collaboration with product management This includes designing and deploying a pipeline with automated data lineage. Identify, develop, evaluate and summarize Proof of Concepts to prove out solutions.
  • Test and compare competing solutions and report out a point of view on the best solution.
  • Integration between GCP Data Catalog and Informatica EDC.
  • Design and build production data engineering solutions to deliver pipeline patterns using Google Cloud Platform (GCP) services: BigQuery, DataFlow, Pub/Sub, BigTable, Data Fusion, DataProc, Cloud Composer, Cloud SQL, Compute Engine, Cloud Functions, and App Engine.



Experience Required:

  • I n-depth understanding of Google’s product technology (or other cloud platform) and underlying architectures
  • 5+ years of analytics application development experience required - 5+ years of SQL development experience.
  • 3+ years of Cloud experience (GCP preferred) with solution designed and implemented at production scale.
  • Experience working in GCP based Big Data deployments (Batch/Real-Time) leveraging Terraform, Big Query, Big Table, Google Cloud Storage, PubSub, Data Fusion, Dataflow, Dataproc, Cloud Build, Airflow, Cloud Composer etc.
  • 2 + years professional development experience in Java or Python, and Apache Beam.
  • Experience developing with micro service architecture from container orchestration framework.
  • Extracting, Loading, Transforming, cleaning, and validating data - Designing pipelines and architectures for data processing.
  • 1+ year of designing and building Tekton pipelines


Education Required:

  • Bachelor’s degree in computer science or related scientific field

Apply Now
An error has occurred. This application may no longer respond until reloaded. Reload 🗙