The position requires a Bachelor's degree in Computer Science, Engineering, Information Technology, Mathematics, Physics, or a related field. Candidates must possess proficiency in programming and scripting languages including Python, SQL, Java, Scala, C, C++, or Go. The role necessitates experience in building software system tools, infrastructure, and ETL data pipelines. Furthermore, expertise in data warehousing, data modeling, and the Big Data technology stack, specifically HDFS, Hive, Oozie, Airflow, MapReduce, or Spark, is required.