Job Description :
• Basic to intermediate experience with Spark
• Good experience with SQL. Should be able to understand and implement performance optimization.
• Memory management experience
• Queue allocation, distribution experience in Hadoop/Cloud era environments
• Should be able to scale clusters in production and have experience with 18/5 or 24/5 production environments.
• Good with one programming language.
• Good exposure to Hive and Hadoop systems with exposure to monitoring tools. Sqoop, Oozie and other external tools.
• Familiar with patch upgrades.
Key skills required :
• Hadoop (preferably Hortonworks/Open Distribution), HDFS, Hive, Spark, Oozie/Airflow, HBase
• Intermediate proficiency in SQL & HQL, Kafka - Good to have
• Solid understanding of Linux, and scripting skills would be advantageous
• Experience with Kerberos, TLS, Ranger, and data encryption