
Data Platform Engineer - Python & Spark
- Toronto, ON
- Permanent
- Full-time
Experience: 5+ YearsRole Summary: NearSource is looking for a Data Platform Engineer with expertise in Python, Spark, and Snowflake to join our engineering team in Toronto. The selected candidate will architect and scale data pipelines, enable analytics, and support critical business insights for a Fortune 500 client.Key Responsibilities
- Design, build, and scale modern data platforms capable of handling both real-time and batch processing at enterprise scale.
- Translate complex business and technical challenges into actionable solutions through detailed documentation and iterative delivery.
- Automate deployment, monitoring, and management of cloud-based data infrastructure for high availability and reliability.
- Implement and maintain CI/CD pipelines with a strong focus on automated testing and continuous improvement.
- Collaborate closely with data engineers, data scientists, and product stakeholders to gather requirements and drive data best practices.
- Develop and optimize SQL queries and data models to enable efficient analytics and reporting.
- Create intuitive dashboards, reports, and analytical tools that provide actionable insights to business leaders.
- Contribute to analytics initiatives such as performance tracking, segmentation, and revenue optimization.
- Communicate findings effectively by presenting data-driven insights in a clear and strategic manner.
- 5+ years of hands-on experience working with large-scale data systems, processing pipelines, and relational databases.
- Strong expertise with Spark (DataFrames, Spark SQL, PySpark).
- Advanced programming skills in Python and SQL, with a focus on clean, modular code.
- Solid foundation in SQL concepts, dimensional data modeling, and data warehouses such as Hive or Snowflake.
- Experience with workflow orchestration tools like Airflow.
- Proficiency with BI tools such as Tableau, Power BI, or Qlik.
- Knowledge of CI/CD practices and version control systems (Git, Jenkins).
- Hands-on experience with interactive notebook environments (Jupyter, EMR, or Zeppelin).
- Familiarity with Presto and AWS services.
- Experience working with Looker for BI development.
- Understanding of distributed system design with a focus on reliability and resilience.
- Exposure to monitoring and observability tools.
We are sorry but this recruiter does not accept applications from abroad.