Prospective candidates are required to have a proficient command of specific skills that are deemed essential for the position. These skills encompass a range of competencies and expertise that are crucial for successfully fulfilling the responsibilities associated with the role. The expectation is that applicants not only possess a basic understanding of these skills but also demonstrate a practical ability to apply them effectively in real-world scenarios.

  • 2-4 years of involvement with the Hadoop environment. 
  • Bachelor Degree in Mathematics, Statistics, Computer Science or investigative system with solid learning of and involvement with measurements.
  • 2-3 years of profound abilities in execution of the Big Data scene.
  • 2-3 years of profound working learning being used of open source devices, for example, Hadoop innovation Python and bash.
  • 2+ years of hands on creating Sqoop scripts and bunch preparing.
  • 3 + years of hands on encounters and intensive comprehension of UNIX/Linux working frameworks.
  • 2-5 years involvement in planning huge information distribution centers, working learning of outline methods, for example, star diagram, stream chip.
  • 2-5 years of involvement with different data demonstrating methods, (for example, information stream outlines, third ordinary structure, element relationship graphs or make/read/upgrade/erase frameworks).
  • Familiarity with business insight and information warehousing advancement procedures and strategies.
  • Hands-on involvement with the Hadoop stack (HDFS, MapReduce, Oozie, Pig, Hive, Sqoop, Spark).
  • Scripting and mechanization abilities in Java or another dialect.
  • Experience sending new Hadoop foundation, Hadoop group overhauls, and bunch upkeep.
  • Experience creating Hadoop combinations for information ingestion, information mapping and information. handling abilities