We are seeking a skilled Graph Database Data Engineer / Senior Data Engineer to join our dynamic team. The ideal candidate will have a strong background in data engineering, graph querying languages, and data modelling, with a keen interest in leveraging cutting-edge technologies like vector databases and LLMs to drive functional objectives.Role Overview:Job Title: Graph Database Data EngineerDuration: 12 months initiallyResponsibilities:Develop and implement scalable data pipelines using Azure Databricks and Azure Data Lake.Create and maintain ETL workflows to ensure data quality, integrity, and availability.Manage and analyze data at scale within the Azure ecosystem.Develop and optimize graph database solutions using querying languages such as Cypher, SPARQL, or GQL. Neo4J DB experience is preferred.Build and maintain ontologies and knowledge graphs, ensuring efficient and scalable data modeling.Work with Large Language Models (LLMs) to achieve functional and business objectives.Ensure data quality, integrity, and security while delivering robust and scalable solutions.Data Modeling: Strong skills in creating ontologies and knowledge graphs.Vector Databases: Understanding of similarity search techniques and RAG implementations.Coach junior data engineers on technologies like Azure Databricks and Azure Cosmos database.Implement CI/CD pipelines using Azure DevOps to streamline deployment processes.Conduct code reviews, provide mentorship, and facilitate knowledge sharing sessions within the team.Integrate data from various sources into a centralized data lake, enabling unified access and analysis.Perform data cleansing, transformation, and migration tasks as required.Monitor and troubleshoot data pipelines and resolve any issues promptly.Requirements:Bachelor’s or Master’s degree in Computer Science, Information Technology, or a related field.Highly experienced hands-on in data engineering, with experience working with Graph DB.Highly experienced in Azure Databricks, Azure Data Lake, and Azure Cosmos DB & proven track record of building data engineering solutions.Proficient in programming languages such as Python, SQL, and Scala.Experience with big data technologies including Apache Spark, Hadoop, and Hive.Knowledge of tools and technologies such as Azure Data Factory, Azure Synapse Analytics, and Apache Kafka.Solid understanding of CI/CD pipelines and hands-on experience with DevOps tools such as Git for continuous integration and deployment.Excellent problem-solving skills and the ability to troubleshoot complex data issues.Strong coaching abilities, excellent communication and collaboration skills, with the ability to work effectively with cross-functional teams.Preferred Qualifications:Databricks Certified Associate Developer for Apache Spark.Experience with data governance and data security practices.Knowledge of Docker, Podman, and Kubernetes for containerization and orchestration.Please feel free to contact Daisy Nguyen at dnguyen@gibbsconsulting.com for a confidential chat to know more details about the role.Please note: Due to the volume of applications received for positions, it will not be possible to respond to all applications and only applicants who are considered suitable for interview will be contacted.Seniority levelMid-Senior levelEmployment typeFull-timeJob functionInformation TechnologyIndustriesHospitals and Health Care
#J-18808-Ljbffr