Skip to main contentA logo with &quat;the muse&quat; in dark blue text.

Senior Data DevOps Engineer

AT EPAM Systems
EPAM Systems

Senior Data DevOps Engineer

Barra do Garças, Brazil

We're seeking a remote Senior Data DevOps Engineer to join our dynamic team for a new project focused on developing and managing data infrastructure in the cloud, primarily using AWS, Azure, or GCP.
In this role, you will be responsible for designing, deploying, and managing data systems, developing automation scripts and workflows for infrastructure provisioning, deployment, and monitoring, and optimizing performance, scalability, and reliability of data platforms and systems.
You will work closely with the data engineering team to ensure efficient data pipelines and processes, automating data workflows using Python. You will also be responsible for setting up and maintaining continuous integration and delivery (CI/CD) pipelines using tools such as Jenkins, GitHub Actions, or similar cloud-based CI/CD tools.

Want more jobs like this?

Get jobs in Barra do Garças, Brazil delivered to your inbox every week.

By signing up, you agree to our Terms of Service & Privacy Policy.


#LI-DNI#LI-AP13

Responsibilities
  • Design, deploy, and manage data infrastructure in the cloud, primarily using AWS, Azure, or GCP
  • Develop and implement automation scripts and workflows for infrastructure provisioning, deployment, and monitoring using tools like Terraform or similar Infrastructure as Code (IaC) tools
  • Ensure efficient data pipelines and processes, automating data workflows using Python
  • Work closely with the data engineering team to ensure efficient data pipelines and processes, automating data workflows using Python
  • Set up and maintain continuous integration and delivery (CI/CD) pipelines using tools such as Jenkins, GitHub Actions, or similar cloud-based CI/CD tools
  • Collaborate with cross-functional teams to optimize the performance, scalability, and reliability of data platforms and systems
  • Install, configure, and maintain data tools such as Apache Spark, Apache Kafka, ELK Stack, Apache NiFi, Apache Airflow, or similar tools in both on-premises and cloud environments
  • Monitor and troubleshoot data systems, proactively identifying and resolving performance, scalability, and reliability issues
Requirements
  • Minimum of 3 years of experience in data infrastructure management and DevOps
  • Strong proficiency in Python, and Batch experience
  • Professional mastery of the Linux operating system
  • Strong knowledge of Cloud technologies (AWS, GCP or Azure)
  • Solid understanding of network protocols and mechanisms such as TCP, UDP, ICMP, DHCP, DNS, and NAT
  • Hands-on experience using or setting up data tools such as Spark, Airflow, R
  • Proficiency with SQL
  • Experience with Infrastructure as Code (IaC) tools
  • Proficiency with setting up and managing CI/CD pipelines using tools like Jenkins, Bamboo, TeamCity, GitLab CI, GitHub Actions, or similar cloud-based CI/CD tools
  • Experience installing and configuring data tools such as Apache Spark, Apache Kafka, ELK Stack, Apache NiFi, Apache Airflow, or similar tools
  • Good verbal and written communication skills in English at a B2+ level
Nice to have
  • Expertise in AWS CloudFormation
  • Knowledge of Terraform and Ansible
  • Azure DevOps skills

Client-provided location(s): Brazil
Job ID: EPAM-epamgdo_blt67076f32d787f5b7_en-us_Other_Brazil
Employment Type: Other