Senior Data Engineer
at
Themesoft Inc.
Senior Data Engineer
Location
Mississauga, Ontario, Canada
Date Posted
December 26, 2024
**Job Title:** Sr. Data Engineer
**Location:** Mississauga, ON (Locals Only)
**Salary:** CAD 130K/annum + benefits
**Responsibilities:**
- Integral team member of our Data Engineering team responsible for the design and development of Big Data solutions.
- Partner with domain experts, product managers, analysts, and data scientists to develop Big Data pipelines in Hadoop or Snowflake.
- Deliver data as a service framework.
- Move all legacy workloads to cloud platforms.
- Work with data scientists to build client pipelines using heterogeneous sources and provide engineering services for data science applications.
- Ensure automation through CI/CD across platforms, both in cloud and on-premises.
- Research and assess open source technologies and components to recommend and integrate into design and implementation.
- Be the technical expert and mentor other team members on Big Data and Cloud tech stacks.
- Define needs around maintainability, testability, performance, security, quality, and usability for data platforms.
- Drive implementation, consistent patterns, reusable components, and coding standards for data engineering processes.
- Convert SAS-based pipelines into languages like PySpark and Scala to execute on Hadoop and non-Hadoop ecosystems.
- Tune Big Data applications on Hadoop and non-Hadoop platforms for optimal performance.
- Evaluate new IT developments and evolving business requirements, recommending appropriate systems alternatives and enhancements.
- Supervise day-to-day staff management issues, including resource management, work allocation, and mentoring/coaching.
**Qualifications:**
- 8+ years of experience in Hadoop/Big Data technologies.
- 3+ years of experience in Spark.
- 2+ years of experience in Snowflake.
- 2+ years of experience working on Google or AWS Cloud developing data solutions. Certifications preferred.
- Hands-on experience with Python/PySpark/Scala and basic libraries for machine learning.
- Experience with containerization and related technologies (e.g., Docker, Kubernetes).
- Experience with all aspects of DevOps (source control, continuous integration, deployments, etc.).
- 1 year of Hadoop administration experience preferred.
- 1+ year of SAS experience preferred.
- Comprehensive knowledge of software engineering principles and data analytics.
- Advanced knowledge of the Hadoop ecosystem and Big Data technologies, including HDFS, MapReduce, Hive, Pig, Impala, Spark, Kafka, Kudu, and Solr.
- Knowledge of agile (Scrum) development methodology is a plus.
- Strong development/automation skills.
- Proficient in programming in Java or Python, with prior Apache Beam/Spark experience a plus.
- System-level understanding of data structures, algorithms, and distributed storage & compute.
- Can-do attitude on solving complex business problems, along with good interpersonal and teamwork skills.
**Education:**
- Bachelor’s degree or equivalent experience.