Job Description Job Specification: - Maintenance of large-scale big data infrastructures, optimizing system performance, and ensuring high availability while looking for improved and newer ways of improving stability, reliability, and prioritising automation-first practices.
- Mentoring and emphasising on best practices and be the change-agents around processes, building documentations and projects that may involve some pain points or even go against legacy norms.
- Experience and knowledge of continuous delivery and/or continuous integration.
- Constantly seek to optimize data processing pipelines for efficiency, scalability, and cost-effectiveness.
- Collaborate with cross-functional teams to design, build, and maintain robust big data solutions using Hadoop ecosystem, Apache suite of tools (Spark), and related technologies.
- Hands-on experience with configuration management tools (Chef, Ansible, Puppet, etc) - Familiarity with monitoring and observability tools (Prometheus, Grafana, ELK Stack).
- Strong knowledge of cloud platforms (AWS, Azure, GCP).
- Experience in Implementing and managing automation tools and infrastructure as code (IaC) for deployment, monitoring, and scaling of big data clusters.
- Implement security best practices to safeguard sensitive data within the big data ecosystem.
- Collaborate with data engineers and developers to understand data requirements to ensure data quality.
- Can create automated tests to continuously monitor system health, performance, and functionality, managing meaningful Service Level Indicators, designing tests to measure them and ensuring SLOs are met.
- Understanding performance testing and capacity planning to simulate loads and assess system performance under various conditions to prevent resource exhaustion.
- Excellent knowledge of using QA techniques to perform root cause analysis, understand the impact of incidents, and develop strategies to prevent similar incidents in the future.
- Continuously evaluate emerging technologies and best practices in big data and DevOps to enhance system performance and reliability.
- Develops and implement Performance Testing standards, processes, tools and methodologies; partners with other functions to establish Non-functional requirements - Work with the application development and testing teams to assist with problem resolution and troubleshooting during performance issue analysis.
All About You - Bachelor/Master Degree in Computer Science or equivalent work experience with hand on technical and quality engineering skills.
- Experience in the Performance Testing.
- Excellent technical acumen, strong organizational and problem-solving skills with great attention to critical thinking and solid communication.
Strong experience in Automated testing/software development.
- Experience in creating and running performance scripts using tools & technologies: Jmeter / Blazemeter / Load runner.
(Must) - Experience performance monitoring tools such as Dynatrace and Splunk.
(Must) - Exposure of test and defect management tools such as RALLY/JIRA.
- Experience of working with other developers to ensure that systems are designed for reliability, performance and testability.
- Strong organizational and problem-solving skills with great attention to detail, critical thinking, solid communication skills.
(Must) Strong documentation skills and presentations skills Experience in Pivotal Cloud Foundry or other cloud platforms Experience with Spring Framework Skills: SDET Test automation blazemeter jmeter Java api testing nifi