Senior Data Engineer

HRBrain

CONTRACT
Canada
Posted 2 years ago

We are looking for a Senior Data Engineer! Reach out if you are interested and feel free to refer friends/colleagues!

 

Type of Employment: Contract
Title: Senior Data Engineer
Term: 12-month –  40 hours
Location: Toronto, Mississauga, Etobicoke and GTA – Hybrid, must be okay with eventually coming into office 1-2/ week
Job ID number: C1027

 

Brief description of duties:

 

  • Some of the challenges the candidate would tackle are:
  • Design, implement and operate the client’s core datasets that have extreme requirements on scalability, reliability, maintainability, flexibility, auditability, and quality
  • Design large scale streaming and batch data pipelines (billions of data points) that involve analyzing transactions over a long period without reprocessing vast amounts of data
  • Perform data profiling and source to target mapping analysis for the best design
  • Collaborate with business solution analysts, other engineers, solution architects, and other team members to innovate and evolve our datasets to data products to create a single coherent platform with sources of truth that serve a plethora of stakeholders
  • Apply your expertise in data and software engineering to design and implement data products that meet extreme requirements on scalability, reliability, maintainability, flexibility, auditability, and quality
  • Be T-Shaped. Your primary area is data engineering, but you are comfortable working in a secondary area of expertise such as data presentation/visualization, backend engineering, or data modelling (SQL, NoSQL, Graph & Time-series)
  • Work closely with cross-functional teams of data, backend and frontend engineers, product owners, technical product owners, and technical support personnel
  • Gaining technical expertise in building a data platform at scale to solve business, product, and technical use cases
  • Getting hands-on experience with technologies such as Elasticsearch, Apache Airflow, Apache Kafka, Apache Beam, Apache Spark, Hive, HDFS, Kubernetes (Openshift)
  • Getting hands-on experience with Google Cloud Platform and technologies such as BigQuery, Cloud Composer, Pub/Sub, Dataflow, Dataproc, GCS, Looker, and other cloud-native offerings in GCP

 

MUST haves:

 

  • 6+ years of professional software engineering and programming experience (Java, Python) with a focus on designing and developing complex data-intensive applications
  • 3+ years of architecture and design (patterns, reliability, scalability, quality) of complex systems
  • Advanced coding skills and practices (concurrency, distributed systems, functional principles, performance optimization)
  • Professional experience working in an agile environment
  • Strong analytical and problem-solving ability
  • Strong written and verbal communication skills
  • Experience in operating and maintaining production-grade software
  • Comfortable with tackling very loosely defined problems and thrive when working on a team which has autonomy in their day to day decisions
  • In-depth knowledge of software and data engineering best practices
  • Experience in mentoring and leading junior engineers
  • Experience in serving as the technical lead for complex software development projects
  • Experience with large scale distributed data technologies and tools
  • Experience pulling data from a variety of data source types including Mainframe (EBCDIC), Fixed Length and delimited files, databases (SQL, NoSQL, Time-series)
  • Strong coding skills for analytics and data engineering (Java, Python, and Scala)
  • Experience performing analysis with large datasets in a cloud-based environment, preferably with an understanding of Google’s Cloud Platform (GCP)
  • Understands how to translate business requirements to technical architectures and designs
  • Comfortable communicating with various stakeholders (technical and non-technical)

 

Nice to haves:

 

  • Exposure to data-science or machine-learning packages (Pandas, Pytorch, Keras, TensorFlow, etc…)
  • Contributions to open-source software (code, docs, or mailing list posts)
  • GCP Professional Data Engineer Certification
  • Confluent Certified Developer for Kafka

Job Features

Job CategoryData Engineer
Architecture and design3+
Software engineering and programming experience (Java, Python) with a focus on designing and developing complex data-intensive applications6+

Apply Online