Databricks Engineer
Cynet Systems Voir toutes les offres
- Toronto, ON
- Permanent
- Temps-plein
- Designs, develops, and maintains scalable ETL/ELT pipelines using Databricks (PySpark, Spark SQL).
- Implements Medallion Architecture (Bronze, Silver, Gold layers) using Delta Lake.
- Builds and optimizes batch and real-time data pipelines using Structured Streaming.
- Develops data ingestion frameworks for structured and semi-structured data sources.
- Optimizes Spark jobs for performance, scalability, and cost efficiency (partitioning, caching, joins, etc.).
- Manages and optimizes Delta Lake tables, including ACID transactions, time travel, and schema evolution.
- Performs debugging, troubleshooting, and performance tuning of data pipelines and jobs.
- Implements data quality checks, validation frameworks, and monitoring solutions.
- Ensures adherence to data governance, security, and access control best practices.
- Applies knowledge of semantic layer concepts to support reporting and analytics use cases.
- Works with cross-functional teams including data analysts, engineers, and business stakeholders.
- Utilizes Databricks workflows and scheduling tools for orchestration.
- Troubleshoots and resolves issues across pipelines, clusters, and data workflows.
- Bachelor's degree in Computer Science, Information Technology, or a related field preferred.
- 4 5 years of experience in Data Engineering or Big Data environments.
- Strong hands-on experience with Databricks platform.
- Expertise in Python/PySpark, Spark SQL, and advanced SQL.
- Strong understanding of Apache Spark internals and distributed data processing.
- Experience with Delta Lake and batch/streaming architectures.
- Experience working with large-scale datasets.
- Strong debugging, troubleshooting, and performance tuning skills.
- Experience with Salesforce data or system integrations.
- Knowledge of semantic layer concepts.
- Experience with cloud platforms (GCP preferred).
- Familiarity with data orchestration tools.
- Experience with data warehousing solutions such as Snowflake.
- Understanding of data modeling concepts.
- Experience with Unity Catalog or similar governance tools.
- Exposure to real-time streaming technologies.
- Strong analytical and problem-solving abilities.
- Ability to work effectively in a fast-paced Agile environment.
- Strong communication and stakeholder collaboration skills.
Nous sommes désolés mais ce recruteur n'accepte pas les candidatures en provenance de l'étranger.