Job Description
We are seeking a skilled professional to join our team as a Data Infrastructure Engineer. The ideal candidate will have experience designing and improving central Big Data infrastructure/services, ensuring reliability, fault-tolerance, scalability, and cost-effectiveness.
This role requires the ability to engage in the entire lifecycle of services, from inception and design through deployment, operation, and refinement. You will develop and maintain tools, redesign capacity planning infrastructure for greater scalability, troubleshoot software issues, ensure data security, build production LLM systems, and optimize workloads in on-prem environments.
The successful candidate will have source code understanding of open-source data groups, such as HDFS, HBase, YARN, Spark, Flink, Airflow, Kyuubi, ZK, Kafka, etc., and an in-depth understanding of Linux and computer networks.
A minimum of 5 years of hands-on experience in backend or big data ecosystems is required, along with the ability to thrive in a high-velocity startup environment with evolving goals and systems.