Almost 4+Years of IT experience in complete life cycle of software development using Object Oriented analysis and design using big data Technologies / Hadoop ecosystem, SQL, Java, J2EE technologies. Almost 4+ years of strong industry experience in Designing and Development, implementing and testing of various client/ server, web-based, distributed application. Last 2.6+ years working on Big Data and Data Science building Advanced Customer Insight and Product Analytic Platforms using Big Data and Open Source Technologies. Wide experience on Data Mining, Real time Analytics, Business Intelligence, Machine Learning and Web Development. Experienced in working with Hadoop/Big-Data storage and analytical frameworks over Amazon AWS cloud using tools like SSH, Putty and Mind-Term Experienced in collecting metrics for Hadoop clusters using Ambari & Cloudera Manager. Experienced on YARN environment with Storm, Spark, Kafka and Avro. Experienced with the Scala, Spark improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Pair RDD's, Spark YARN. Leveraged strong Skills in developing applications involving Big Data technologies like Hadoop, Spark, Elastic Search, Map Reduce, Yarn, Flume, Hive, Pig, Kafka, Storm, Sqoop, HBase, Hortonworks, Cloudera, Mahout, Avro and Scala. Skilled programming in Map-Reduce framework and Hadoop ecosystems. Very good experience in designing and implementing MapReduce jobs to support distributed data processing and process large data sets utilizing the Hadoop cluster. Experience in implementing Inverted Indexing algorithm using MapReduce. Extensive experience in creating Hive tables, loading them with data and writing hive queries which will run internally in MapReduce way. Hands on experience in migrating complex MapReduce programs into Apache Spark RDD transformations. Experience in setting up standards and processes for Hadoop based application design and implementation. Implemented Large scale Hadoop (Hortonworks HDP 2.4 Stack) enterprise Data lake and HDF Nifi cluster for SIT, DEV, UAT, CERT and PROD Environment. Upgrade Hortonworks Ambari and HDP Stack from 2.3 to 2.4 Version in Dev, DR and Prod Environment. Responsible for provisioning and managing Hadoop clusters on public cloud environment Amazon Web Services (AWS) -EC2 for Product POCs. Diligently teaming with the infrastructure, network, database, application and Platform teams to guarantee high data quality and availability. Configured the Chef and Ansible for Hadoop package deployments and other configuration push Good Exposure on Apache Hadoop MapReduce programming, PIG Scripting and HDFS4. Worked on developing ETL processes to load data from multiple data sources to HDFS using FLUME and SQOOP, perform structural modifications using Map-Reduce, HIVE and analyse data using visualization/reporting tools.