Job Description

Summary

We are looking for a seasoned SRE Engineer to design and improve our central Big Data infrastructure/services to the next stage, to ensure the data, services, and infrastructures are reliable, fault-tolerant, efficiently scalable, and cost-effective.

Responsibilities

  1. Engage in and improve the whole lifecycle of service, from inception and design, through to deployment, operation, and refinement.
  2. Develop and maintain tools, re-designing capacity planning infrastructure for greater scalability.
  3. Troubleshooting, diagnosing, fixing software issues, and ensuring data security.
  4. Define architecture improvements, and push for changes that improve reliability.

Requirements

  1. Have source code understanding of open-source data groups, such as HDFS, HBase, YARN, Spark, Flink, Airflow, Kyuubi, ZK, Kafka, etc.
  2. Have used at least one automation component tool: Ansible, Terraform, etc.
  3. In-depth understanding of Linux and computer networks.
  4. Experience in at least one language (Python/Golang/Java, etc.).
  5. Experience in the public cloud (AWS, GCP, Azure, etc.) management and use is preferred.
  6. Minimum of 5 years of hands-on experience on backend or big data ecosystem.

Skills
  • AWS
  • Development
  • Java
  • Python
  • Software Engineering
© 2025 cryptojobs.com. All right reserved.