Data Engineer (SNS01636)

April 22, 2025

Job Description

Welcome to Sansaone, a dynamic force in the realm of ICT talent acquisition. Born out of a passion for excellence and a vision for connecting outstanding professionals with forward-thinking organizations, we stand as a beacon for strategic recruitment solutions in the Information and Communication Technology sector. With a commitment to excellence and a passion for connecting exceptional professionals with innovative organizations, we are your strategic partner in building transformative teams.
Responsibility
  • Design, develop, document, and maintain ETL/ELT processes, data integration, cleaning, transformation, dissemination and automation processes.
  • Create robust solutions to collect, integrate, and process structured, semi-structured, and unstructured data (e.g. JSON, Parquet, Delta).
  • Design, develop, document and maintain data architecture, data modelling and metadata.
  • Develop and support data warehouse/lakehouse architectures and data processing ensuring data quality, lineage, auditing, metadata, logging, linkage across datasets and impact assessments.
  • Work collaboratively with data providers to address data gaps and optimize source-system structures.
  • Develop and maintain business intelligence models, interactive dashboards, reports and analytics using tools such as Databricks, Jupyter Notebooks, and Power BI.
  • Design, develop, document, improve and maintain the Data Warehouse/Lakehouse ecosystem (e.g. the DataDevOps lifecycle, architecture).
  • Support the gathering and analysis of business requirements, translating them into scalable data collection and integration processes.
  • Contribute to the definition and documentation of data governance policies, procedures, standards, and metadata models.
  • Participate in meetings with project and data teams to align on strategic priorities, ensuring seamless integration and optimal data management practices.
Requirements
  • Have a minimum a High School diploma.
  • At least 6 years of professional experience.
  • Proficiency in data development and processing using Python, SQL, Power M, and DAX.
  • Experience working with structured, semi-structured, and unstructured data types and file formats (e.g., JSON, Parquet, Delta).
  • Ability to gather and translate business requirements into scalable data collection, integration, and analysis processes.
  • Hands-on experience with Microsoft On-Premise and Azure Data Platform tools, including: Azure Data Factory, Azure Functions, Azure Logic Apps, SQL Server, Azure Data Lake Storage (ADLS), Azure Databricks, Microsoft Fabric/Power BI, Azure DevOps, Azure AI Services, PowerShell).
  • Familiarity with CI/CD workflows, particularly using Azure DevOps.
  • Strong knowledge of the Databricks ecosystem, Apache Spark, and Python data libraries.
  • Solid understanding of data modeling principles and techniques.
  • Experience with Data Lakes and Lakehouse architectures, including governance and lifecycle management.
  • Knowledge of data integration and data warehouse/lakehouse modeling techniques, such as: Slowly Changing Dimensions (SCD), Functional engineering, Data Vault, Data streaming.
  • Understanding of data governance and data management standards, including metadata, data quality, policies, and compliance.
  • Familiarity with Web APIs and the OpenAPI specification.
  • Proficiency in English language at a minimum B2 level.

Hiring Team Member