Data Architect #6632
1950Labs
data-architecture
data-engineering
cloud-engineering
azure
databricks
sql
python
apache-spark
About this role
Client Description
The client is a global organization in the tourism industry, offering river, ocean, and expedition cruises for passengers worldwide and operating a large fleet of vessels.
The company is currently undergoing an extensive cloud data modernization and unification program. We support them across data architecture, BI, migration, and data platform development.
A key focus area is the migration to Databricks Unity Catalog, including:
- Migrating all data layers (landing, raw, prepared, reporting, services) from Hive Metastore to Unity Catalog
- Migrating DLT (Delta Live Tables) and Python/SQL jobs into Databricks
- Migrating pipelines in Azure Synapse/ADF
- Rebuilding and adapting metadata frameworks
- Standardizing access, lineage, governance, and overall Lakehouse structure
The client has very high technical expectations and is looking for top-level specialists capable of leading complex architectural initiatives.
Technical Requirements
- Advanced knowledge of Microsoft Azure (data infrastructure, networking, authorization, cloud design)
- Experience with Azure Synapse (especially Synapse Serverless and pipelines)
- Strong expertise in Databricks (DLT, workflows, workspace administration)
- Ability to design Data Lakehouse architectures (Medallion Architecture, Metadata-Driven ETL)
- Very good knowledge of Python and code optimization
- Strong SQL skills, including query optimization and SQL Server experience
- Experience with Apache Spark (data processing workflows)
- Experience building ETL/ELT processes and data warehouses
- Experience with CI/CD processes (Azure DevOps, Git, branching strategies)
- Experience implementing logging, monitoring, and optimization of data processes
- Familiarity with Power BI and analytics workflows
- Strong communication skills and documentation ability
- Experience as a Lead Engineer (technical leadership, decision-making, stakeholder collaboration)
Scope of Responsibilities
- Design and develop data architecture in Azure and Databricks environments
- Participate in the Unity Catalog transformation (infrastructure, pipelines, frameworks, standards)
- Migrate and modernize ETL/ELT processes (DLT, Python/SQL jobs, Synapse/ADF pipelines)
- Design and implement Data Lakehouse solutions using Medallion architecture
- Optimize data processing workflows (Python, SQL, Spark)
- Build CI/CD processes and automation in Azure DevOps
- Implement standards for logging, monitoring, and data quality
- Lead the project from a technical perspective (Lead Engineer role)
- Collaborate with business and technical stakeholders
- Document solutions and mentor team members
Originally posted on Himalayas
Similar jobs
RQ08913 - Software Developer - ETL - Senior
Rubicon Path·
Canada·
Remote
1d ago
Data
data-engineering
etl-development
software-development
via Himalayas
1d ago

Senior Data Engineer
Military, Veterans and Diverse Job Seekers·
United States·
Remote
1d ago
Data
data-engineering
senior-data-engineer
big-data
via Himalayas
1d ago

Trainee Data Engineering*
Haeger Consulting GmbH·
Bonn
4w ago
Data
software-development
python
sql
via Arbeitnow
🇩🇪DE
4w ago