Experience in working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment.
Implement methods for automation of all parts of the pipeline to minimize labor in development and production
Experience in analyzing complex data, organizing raw data, and integrating massive datasets from multiple data sources to build analytical domains and reusable data products
Experience in working with architects to evaluate and productionalize data pipelines for data ingestion, curation, and consumption
Experience in working with stakeholders to formulate business problems as technical data requirements, identify and implement technical solutions while ensuring key business drivers are captured in collaboration with product management
Experience Required :
5+ years of SQL development experience
5+ years of analytics / data product development experience required
3+ years of cloud experience (GCP preferred) with solutions designed and implemented at production scale
Experience working in GCP native (or equivalent) services like Big Query, Google Cloud Storage, PubSub, Dataflow, Dataproc, Cloud Build, etc.
Experience working with Airflow for scheduling and orchestration of data pipelines
Experience working with Terraform to provision Infrastructure as Code
2 + years professional development experience in Java or Python
Experience Preferred :
In-depth understanding of Google’s product technology (or other cloud platform) and underlying architectures
Experience in working with DBT / Dataform
Experience with DataPlex or other data catalogs is preferred
Experience with development eco-system such as Tekton, Git, Jenkins for CI / CD pipelines
Exceptional problem solving and communication skills
Experience in working with Agile and Lean methodologies
Team player and attention to detail
Experience with performance tuning SQL queries
Education Required :
Bachelor’s degree in computer science or related scientific field