Our client is a global technology and digital transformation organization specializing in software engineering and IT consulting, serving international clients across multiple industries. The company operates in a fast-paced, collaborative environment, with a strong focus on innovation, engineering excellence, and continuous professional development.
Primary / Required Skills
- 5+ years of experience in Data Engineering or related engineering roles
- Strong hands-on experience with Azure Fabric
- Proficiency in Python
- Solid experience with PySpark and SparkSQL
- Experience with Batching
- Experience with Spark Streaming
- Hands-on experience with OneLake / Delta Lake (OpenLake concepts)
- Knowledge of DF Gen2 and M-code
- Experience with CI/CD pipelines (Azure DevOps or equivalent)
- Good understanding of Azure services
- Experience integrating data solutions with Power BI
- Experience with code generation, including non-AI and AI-assisted approaches
- Expertise with Cosmos DB and variants (Mongo, Cassandra, Table APIs)
- Exposure to Azure AI Foundry
- Experience with Data Science workflows
- Strong background in Big Data and Spark ecosystems
- Knowledge of financial instruments and financial services data
- Hands-on experience with industry-standard LLMs (including GPT, Claude, or similar)
Primary / Required Skills
- 5+ years of experience in Data Engineering or related engineering roles
- Strong hands-on experience with Azure Fabric
- Proficiency in Python
- Solid experience with PySpark and SparkSQL
- Experience with Batching
- Experience with Spark Streaming
- Hands-on experience with OneLake / Delta Lake (OpenLake concepts)
- Knowledge of DF Gen2 and M-code
- Experience with CI/CD pipelines (Azure DevOps or equivalent)
- Good understanding of Azure services
- Experience integrating data solutions with Power BI
- Experience with code generation, including non-AI and AI-assisted approaches
- Expertise with Cosmos DB and variants (Mongo, Cassandra, Table APIs)
- Exposure to Azure AI Foundry
- Experience with Data Science workflows
- Strong background in Big Data and Spark ecosystems
- Knowledge of financial instruments and financial services data
- Hands-on experience with industry-standard LLMs (including GPT, Claude, or similar)