Data Ops Developer

Geotab
Job LocationUS
Job TagFull Time

Description:

Geotab ® is a global leader in IoT and connected transportation and certified “Great Place to Work™.” We are a company of diverse and talented individuals who work together to help businesses grow and succeed, and increase the safety and sustainability of our communities.

Geotab is advancing security, connecting commercial vehicles to the internet and providing web-based analytics to help customers better manage their fleets. Geotab’s open platform and Geotab Marketplace ®, offering hundreds of third-party solution options, allows both small and large businesses to automate operations by integrating vehicle data with their other data assets. Processing billions of data points a day, Geotab leverages data analytics and machine learning to improve productivity, optimize fleets through the reduction of fuel consumption, enhance driver safety and achieve strong compliance to regulatory changes.

Our team is growing and we’re looking for people who follow their passion, think differently and want to make an impact. Ours is a fast paced, ever changing environment. Geotabbers accept that challenge and are willing to take on new tasks and activities – ones that may not always be described in the initial job description. Join us for a fulfilling career with opportunities to innovate, great benefits, and our fun and inclusive work culture. Reach your full potential with Geotab. To see what it’s like to be a Geotabber, check out our blog and follow us @InsideGeotab on Instagram. Join our talent network to learn more about job opportunities and company news.

Who you are:

We are always looking for amazing talent who can contribute to our growth and deliver results! Geotab is seeking a Data Ops Developer to develop and maintain data pipelines. If you love technology, and are keen to join an industry leader — we would love to hear from you!

What you’ll do:

The Data Ops Developer is the foundation of Geotab’s data engineering capabilities. They are responsible for the development of scalable and reusable pipelines that minimize time from insight to production. The role continuously collaborates with data analysts and data scientists to design innovative pipelines using new tools and frameworks. They also work closely with the Data Strategist and Data Quality experts to ensure all data assets are delivered with the highest quality, with the right schema and into the right location.

How you’ll make an impact

  • Design, optimize and maintain SQL queries used for creating data products.
  • Advance your SQL knowledge to ensure queries leverage most recent advancements.
  • Deploy and maintain ETL/ELT pipelines using SQL, Python and Airflow.
  • Design and publish reusable pipeline templates (i.e. templates for data integration, derived metrics, reporting, custom runs).
  • Collaborate with data analysts and data scientists to develop complex pipelines involving big data tools (i.e. Spark, Apache Beam, GKE, Kafka, Docker).
  • Lead optimization of pipelines based on requirements and pipeline performance.
  • Contribute to development of data integration connectors to extract external data.
  • Manage pipeline releases through Git and CI/CD.
  • Ensure metadata is captured and stored across the pipeline lifecycle (i.e. creation, execution, deprecation/update).
  • Support remediation of issues within production pipelines.
  • Collaborate with data quality analysts and specialists to ensure all pipelines include automatic quality checks.
  • Recommend features and enhancements to infrastructure and pipeline framework.
  • Contribute to the migration of data assets and pipelines from legacy data structures. 
  • Participate in a 24×7 on-call rotating schedule.

What you’ll bring to the role

  • Post-secondary Degree specialization in Computer Science, Software or Computer Engineering or a related field.
  • 3-5 years experience in Data Engineering or a similar role.
  • 3-5 years of experience with SQL.
  • 3-5 years of experience building ETL/ELT production pipelines in Python.
  • Knowledge of data management fundamentals and data modeling principles.
  • Experience with Big Data environments (e.g. Google BigQuery) is an asset.
  • Experience with CI/CD processes and tools, such as Gitlab runners or Jenkins is required.
  • Knowledge of workflow orchestration tools is required (e.g. Apache Airflow).
  • Knowledge of Linux and command line commands is an asset.
  • Experience working in a cloud based infrastructure, especially Google Cloud Platform, is an asset.
  • Previous experience with python package development is highly regarded.
  • Excellent oral and written communication skills.
  • Strong analytical skills with the ability to problem solve well-judged decisions.
  • Highly organized and able to manage multiple tasks and projects simultaneously.
  • Entrepreneurial mindset and comfortable in a flat organization.
  • Must stay relevant to technology and have the flexibility to adapt to the growing technology and market demands.
  • Strong team player with the ability to engage with all levels of the organization.

If you got this far, we hope you’re feeling excited about this role! Even if you don’t feel you meet every single requirement, we still encourage you to apply.


Understanding Data Science Roles: Who Does What?

Looking to leverage your data boot camp knowledge or jumpstart a career in data science? With data becoming increasingly important to companies’ bottom lines, data scientists are some of the most valuable individuals in the professional world today.

Related Industries: Data
An image shows a student working on a graph.

Top 10 Coding Projects for Beginners

The best way to begin to learn to code is to simply start coding! When you gain experience developing challenging projects from inception to completion, you establish a strong work ethic that will benefit you throughout your career. 

Related Industries: Coding, Data
A woman works at a laptop.