Hirejobs Canada
Register
Auckland Jobs
Canterbury Jobs
Northland Jobs
Otago Jobs
Southland Jobs
Tasman Jobs
Wellington Jobs
West Coast Jobs
Oil & Gas Jobs
Banking Jobs
Construction Jobs
Top Management Jobs
IT - Software Jobs
Medical Healthcare Jobs
Purchase / Logistics Jobs
Sales
Ajax Jobs
Designing Jobs
ASP .NET Jobs
Java Jobs
MySQL Jobs
Sap hr Jobs
Software Testing Jobs
Html Jobs
IT Jobs
Logistics Jobs
Customer Service Jobs
Airport Jobs
Banking Jobs
Driver Jobs
Part Time Jobs
Civil Engineering Jobs
Accountant Jobs
Safety Officer Jobs
Nursing Jobs
Civil Engineering Jobs
Hospitality Jobs
Part Time Jobs
Security Jobs
Finance Jobs
Marketing Jobs
Shipping Jobs
Real Estate Jobs
Telecom Jobs

Data Engineer - Jobs in Toronto, Ontario

Job LocationToronto, Ontario
EducationNot Mentioned
SalaryNot Mentioned
IndustryNot Mentioned
Functional AreaNot Mentioned
Job TypeFull time

Job Description

The Data Engineer will play a key role in building and maintaining a modern data infrastructure. They will design, develop, and optimize ELT/ETL pipelines and implement efficient data models to support analytics, reporting, and business decision-making. The role involves deep collaboration with cross-functional teams to deliver scalable and high-performing data solutions using tools such as Snowflake, ADF, dbt, and Python. Responsibilities include:

  • Design, build, and maintain scalable and robust data pipelines using ELT/ETL patterns to ingest, transform, and integrate data.
  • Architect and implement efficient data models using Star, Snowflake, and One Wide Table (OWD) design patterns.
  • Maintain and create documentation of data architecture, data pipelines, and processes to ensure transparency and reproducibility.
  • Integrate data from multiple sources including databases, APIs, and third-party platforms using tools like Azure Data Factory (ADF) and dbt.
  • Lead technical discussions, advocate for best practices, and ensure solid data foundations and high standards in data engineering workflows.
  • Optimize data systems for performance and cost efficiency using partitioning, clustering, caching, indexing, and fine-tuning techniques.
  • Perform QA audits, manage data loads, generate memo files, and handle ad hoc data requests to ensure data integrity and reliability.
  • Support analytics and reporting by developing reusable metrics, dashboards, and self-service tools in Power BI and/or Sisense.
  • Enhance SDLC by incorporating CI/CD pipelines, version control (e.g., Git), and continuous improvement practices into data engineering processes.
  • Collaborate with internal and external stakeholders to gather requirements and deliver comprehensive data solutions.
Education:
  • Bachelor’s Degree in Computer Science, Mathematics, Statistics, Finance, Information systems or equivalent related technical field experience
Experience:
  • 5+ years of professional experience in data engineering, data analytics, or a similar technical role.
  • Strong SQL skills with advanced knowledge of Joins, Unions, CTEs, Aggregations, Lag/Lead, and optimization techniques.
  • Proficiency in Python for data manipulation, scripting, and automation.
  • Experience working with Snowflake, dbt, and Azure Data Factory (ADF).
  • Demonstrated experience in data modeling, including dimensional and modern approaches (Star, Snowflake, OWD).
  • Hands-on experience in building and maintaining data pipelines (ETL/ELT).
  • Understanding of cost optimization, caching, partitioning, and indexing strategies for performance tuning.
  • Familiarity with BI tools such as Power BI, Sisense, Looker, Tableau, and Domo.
  • Experience with customer personalization solutions and handling large datasets.
  • Exposure to scripting languages like Python, Perl, or Shell.
Toolsamp; Skills:
  • Deep understanding of complex SQL and Snowflake SQL syntax, including Time Travel, Streams, Cloning, and Role-Based Access.
  • Strong knowledge of Snowflake, Azure Data Factory, and dbt.
  • Experience with version control systems and CI/CD workflows.
  • Knowledge of DataBricks (ADB preferred) and ability to interpret existing solutions.
  • Familiarity with reporting tools, especially Power BI and/or Sisense.
  • Advanced proficiency in Python and Excel for data analysis and transformation.
  • Understanding of data warehousing, proactive data quality monitoring, and structured/unstructured data formats including JSON.
Key Competencies:
  • Proven problem-solving skills and high attention to detail.
  • Ability to partner with business stakeholders to define questions and build data sets to answer them.
  • Capable of navigating ambiguity and balancing multiple priorities in a fast-paced environment.
  • Excellent communication and presentation skills for technical and non-technical audiences.
  • Self-starter with a spirit of innovation and consistent delivery.
  • Demonstrated ability to work collaboratively in multi-disciplinary teams and produce results quickly.
Assets:
  • Experience in Telecom or banking industries, especially related to data collection or retention.
  • Hands-on experience with ADF data transformations for custom reporting models.
  • Experience in scripting and automation using Python, Perl, or Shell.
  • Familiarity with data transformations using tools like dbt.
  • Data analysis, report development, and business analysis
  • Experience with tools like Looker, Excel, Power BI, Tableau, R, SAS
Powered by JazzHR

APPLY NOW

© 2021 HireJobsCanada All Rights Reserved