Data Engineer architecting and building cloud-based systems for Semios Group, an agricultural technology company. Responsibilities include managing data interfaces, scalable infrastructure, and delivering actionable insights.
Responsibilities
Architect and build Cloud based systems to manage and improve the interface between Semios data and its consumers.
Design, develop and maintain scalable infrastructure to process and store data, integrate data driven models and automate manual processes.
Implement highly scalable big data analytics systems in a cloud environment.
Design and build reliable, monitorable and fault-tolerant data systems & data processes.
Create data tools for analytics and data science team members that assist them in building and optimizing our product into an innovative industry leader.
Assemble large, complex data sets that meet functional / non-functional business requirements.
Continuously identify bottlenecks in the data stack and optimize queries and processes for cost and performance.
Write clear documentation of data processes and products.
Requirements
Advanced skills in SQL; how to write elegant queries; written for humans first, machines second.
The ability to thrive both autonomously and in a team environment.
Hands-on experience with provisioning and developing on cloud platforms (familiarity with GCP is a definite plus).
Experience with at least one Data Warehouse (BigQuery, RedShift, Snowflake, On-Prem).
Excellent verbal & written communication skills: a talent to distill complex ideas to different audiences.
An in-depth experience with Big Data. A proven track-record of effective collection, storage, and access.
Proven experience with workflow and scheduling tools (e.g., like Prefect, Airflow, Dagster, Kubeflow, etc.) and version control (Git).
A fluency in Python, Node or other imperative language or ability to learn quickly and with enthusiasm.
Excellent troubleshooting skills to rapidly identify and resolve issues.
**Nice to have:**
Significant exposure to at least one relational database (Postgres, MySQL).
Real world experience with containers (Docker) & container management systems (Kubernetes).
Experience or Interest in working with IoT Cloud and IoT data.
Familiarity with data transformation tools (dbt, SQLMesh, Dataform) and syncing tools (e.g., dlt, Fivetran, Airbyte).
Senior Software Developer modernizing Data Transfer Platform for Intrahealth, a healthcare EMR provider. Focusing on scalable and configurable backend systems in a complex environment.
Data Engineer Intern gaining hands - on experience in TD's big data platform. Collaborating on software development and system enhancements while learning about analytical tools and technologies.
Senior Data Engineer at Mozilla managing data lifecycle and quality. Building data pipelines and collaborating with product teams for data - driven decisions.
Principal Product Manager leading product strategy for health data platform at PointClickCare. Collaborating across teams to optimize health data for analytics and care delivery.
Data Engineer optimizing and maintaining data pipelines in Blackline Safety's IoT - enabled safety ecosystem. Collaborating with product, engineering, and analytics teams on impactful data - driven initiatives.
Azure Data Engineer contractor for Ontario Crown Corporation. Design/build data pipelines using Azure Data Factory, Databricks, Python, PySpark. 3 days/week onsite in Oshawa.
Hiring Data Engineers with IMS DB, DB2, IBM MDM Server, and Talend ETL experience. Role involves data mapping, API integration, and SQL query development.
Data Engineer working with Google Cloud's technologies, assisting clients in data solutions and pipelines. Collaborating with teams to optimize data infrastructures and promote agile practices.