- Company Name
- Loblaw Digital
- Job Title
- Staff Data Engineer
- Job Description
-
**Job title:** Staff Data Engineer
**Role Summary:**
Lead architecture and hands‑on development of high‑scale data platforms for Retail Media. Design and optimize batch, streaming, and real‑time pipelines in GCP, using PySpark, Airflow, and BigQuery. Mentor engineers, enforce code quality, and collaborate with product, backend, and AI teams to deliver reliable, AI‑ready data assets.
**Expectations:**
* Senior‑level ownership of production‑critical systems.
* Proven record of designing scalable, high‑performance data infrastructure.
* Ability to translate complex business requirements into robust technical solutions.
* Strong mentorship, code‑review, and continuous‑improvement mindset.
**Key Responsibilities:**
- Architect and evolve scalable, reliable data platforms (batch, streaming, real‑time).
- Drive architectural improvements across pipelines, metadata, APIs, and service dependencies.
- Build, test, and optimize large‑scale data pipelines with PySpark, Dataproc, Airflow, GCS, and BigQuery.
- Design data models, partitioning strategies, and storage formats (Parquet) to support analytics and measurement.
- Implement data quality, validation, observability, and monitoring practices.
- Collaborate on backend service and API design; reduce coupling to improve performance and flexibility.
- Partner with AI teams to provide inference‑ready data pipelines and support AI strategy.
- Lead design and code reviews, set performance, readability, testing, and maintainability standards.
- Mentor and coach junior and mid‑level engineers.
- Champion continuous improvement of pipeline performance, cost efficiency, and operational excellence.
**Required Skills:**
- Proficiency in PySpark and distributed data processing.
- Strong experience with GCP services (Dataproc, GCS, BigQuery).
- Advanced SQL skills for large analytical datasets.
- Experience with NoSQL and analytical stores (e.g., Druid, Bigtable, Elasticsearch).
- Solid programming in Python, Scala, or Java.
- Orchestration with Airflow; production pipeline experience.
- Data modeling, partitioning, and Parquet expertise.
- Agile delivery experience.
- Excellent communication, ability to explain technical concepts to non‑technical stakeholders.
**Required Education & Certifications:**
- BA/BS in Computer Science, Engineering, Mathematics, or related field (advanced degree preferred).