Overview
Degree in software engineering, computer science, informatics or a similar field.
2-7 years of relevant work experience in Big Data or distributed computing projects.
Experience in designing and implementing Big Data/ML applications (data ingestion, real-time data processing and batch analytics) in Spark Streaming, Kafka, Hadoop.
Solid server-side programming skills (Scala, Nodejs, or Java), and hands-on experience in OOAD, design patterns, and SQL.
Experience with microservice architecture, and design.
Experience on Hadoop-ecosystem technologies in particular MapReduce, Spark, Hive, YARN.
Experience working on any one distributed database system like Hadoop (Hive/HDFS), Qubole, Teradata, Redshift, or DB2.
Solid experience building APIs (REST), Java services, or Docker Microservices.
Good knowledge of database structures, theories, principles, and practices.
Familiarity with data loading tools like Flume, Sqoop.
Knowledge of workflow/schedulers like Oozie.
Analytical and problem solving skills, applied to Big Data domain.
Proven understanding with Hadoop, Hive, Pig, and HBase.
Good aptitude in multi-threading and concurrency concepts.