Specialized IT Consultant – Senior

LanceSoft

Job title:

Specialized IT Consultant – Senior

Company

LanceSoft

Job description

Position: Specialized IT Consultant
Location: Toronto, ON (Hybrid)
Duration: 319 Days Contract + Possible ExtensionExperience:

  • Experience of 7+ years of working with SQL Server, T-SQL, Oracle, PL/SQL development or similar relational databases
  • Experience of 2+ years of working with Azure Data Factory, Databricks and Python development
  • Experience building data ingestion and change data capture using Oracle Golden Gate
  • Experience in designing, developing, and implementing ETL pipelines using Databricks and related tools to ingest, transform, and store large-scale datasets
  • Experience in leveraging Databricks, Delta Lake, Delta Live Tables, and Spark to process structured and unstructured data.
  • Experience working with building databases, data warehouses and working with delta and full loads
  • Experience on Data modeling, and tools – e.g. SAP Power Designer, Visio, or similar
  • Experience working with SQL Server SSIS or other ETL tools, solid knowledge and experience with SQL scripting
  • Experience developing in an Agile environment
  • Understanding data warehouse architecture with a delta lake
  • Ability to analyze, design, develop, test and document ETL pipelines from detailed and high-level specifications, and assist in troubleshooting.
  • Ability to utilize SQL to perform DDL tasks and complex queries
  • Good knowledge of database performance optimization techniques
  • Ability to assist in the requirements analysis and subsequent developments
  • Ability to conduct unit testing and assist in test preparations to ensure data integrity
  • Work closely with Designers, Business Analysts and other Developers
  • Liaise with Project Managers, Quality Assurance Analysts and Business Intelligence Consultants
  • Design and implement technical enhancements of Data Warehouse as required.

Technical Skills (70 points):

  • Experience in developing and managing ETL pipelines, jobs, and workflows in Databricks.
  • Deep understanding of Delta Lake for building data lakes and managing ACID transactions, schema evolution, and data versioning.
  • Experience automating ETL pipelines using Delta Live Tables, including handling Change Data Capture (CDC) for incremental data loads.
  • Proficient in structuring data pipelines with the Medallion Architecture to scale data pipelines and ensure data quality.
  • Hands-on experience developing streaming tables in Databricks using Structured Streaming and readStream to handle real-time data.
  • Expertise in integrating CDC tools like GoldenGate or Debezium for processing incremental updates and managing real-time data ingestion.
  • Experience using Unity Catalog to manage data governance, access control, and ensure compliance.
  • Skilled in managing clusters, jobs, autoscaling, monitoring, and performance optimization in Databricks environments.
  • Knowledge of using Databricks Autoloader for efficient batch and real-time data ingestion.
  • Experience with data governance best practices, including implementing security policies, access control, and auditing with Unity Catalog.
  • Proficient in creating and managing Databricks Workflows to orchestrate job dependencies and schedule tasks.
  • Strong knowledge of Python, PySpark, and SQL for data manipulation and transformation.
  • Experience integrating Databricks with cloud storage solutions such as Azure Blob Storage, AWS S3, or Google Cloud Storage.
  • Familiarity with external orchestration tools like Azure Data Factory
  • Implementing logical and physical data models
  • Knowledge of FHIR is an asset

Design Documentation and Analysis Skills (20 points)

  • Demonstrated experience in creating design documentation such as:
  • Schema definitions
  • Error handling and logging
  • ETL Process Documentation
  • Job Scheduling and Dependency Management
  • Data Quality and Validation Checks
  • Performance Optimization and Scalability Plans
  • Troubleshooting Guides
  • Data Lineage
  • Security and Access Control Policies applied within ETL
  • Experience in Fit-Gap analysis, system use case reviews, requirements reviews, coding exercises and reviews.
  • Participate in defect fixing, testing support and development activities for ETL
  • Analyze and document solution complexity and interdependencies including providing support for data validation.
  • Strong analytical skills for troubleshooting, problem-solving, and ensuring data quality.

Communication and Leadership Skills (10 points):

  • Ability to collaborate effectively with cross-functional teams and communicate complex technical concepts to non-technical stakeholders.
  • Strong problem-solving skills and experience working in an Agile or Scrum environment.
  • Ability to provide technical guidance and support to other team members on Databricks best practices.
  • Must have previous work experience in conducting Knowledge Transfer sessions, ensuring the resources will receive the required knowledge to support the system.
  • Must develop documentation and materials as part of a review and knowledge transfer to other members.

Must Have Skills:

  • 7+ years using ETL tools such as Microsoft SSIS, stored procedures, T-SQL
  • 2+ Delta Lake, Databricks and Azure Databricks pipelines
  • Strong knowledge of Delta Lake for data management and optimization.
  • Familiarity with Databricks Workflows for scheduling and orchestrating tasks.
  • 2+ years Python and PySpark
  • Solid understanding of the Medallion Architecture (Bronze, Silver, Gold) and experience implementing it in production environments.
  • Hands-on experience with CDC tools (e.g., GoldenGate) for managing real-time data.
  • SQL Server, Oracle

Expected salary

Location

Toronto, ON

Job date

Wed, 11 Dec 2024 04:25:11 GMT

To help us track our recruitment effort, please indicate in your email/cover letter where (jobsnearcanada.com) you saw this job posting.

To apply for this job please visit jobviewtrack.com.

Job Location