- Selecting and integrating any Hadoop Big Data tools and frameworks required to provide requested capabilities - Build automated and scalable data pipeline - Implementing ETL process with Shell Scripts, Sqoop, Hive, etc. - Build real-time systems and tools in NiFi to support all platforms - Monitoring performance and advising any necessary infrastructure changes - Experience with integration of data from multiple data sources(Hana,MySQL,MsSQL, Kafka, Rabbit,etc.) - Design and build scalable infrastructure and platform to collect and process very large amounts of data, including streaming real-time data - Good knowledge of Big Data querying tools, such as Hive. - Experience with performance/scalability tuning, algorithms and computational complexity - Proven ability to work cross functional teams to complete solution design, development and delivery - Version management with GitHub