Expertise in designing and implementing Big Data Solution for DoD agencies.
- Requirements and Use Case Analysis
- Enabling Statistical Analytics ‘R’ capability
- Sizing the Big Data Platform
- Creating the technical specifications
- Installing and configuring the Hadoop ecosystem components
- Installing and configuring Data Virtualization components
- Deploy Atlas for governance
- Deploy Kafka for streaming data
- Proof of Concept ; followed by complete life-cycle development of the big data solution
- Expertise in Hive, Spark, Pig, Sqoop, Flume, Oozie, MapReduce, HDFS, Kafka, Ranger, Atlas, Accumulo and more
- Integration expertise with NoSQL and relational databases such as Cassandra, HBase, MongoDB, Oracle Database and Microsoft SQL Server
- Hadoop performance tuning
- Kerborizing Hadoop cluster (MIT KDC)
- Implementing Role Based Access Control through Ranger
- Configuring Knox and single sign-on
- Installing and configuring Visualization tools like Qlik, Tableau and Power BI
- Enabling Python analytics modules for Hadoop
- Running Predictive Analytics algorithms for Logistics Datasets