From:                              route@monster.com

Sent:                               Friday, May 06, 2016 1:35 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Cloud

 

This resume has been forwarded to you at the request of Monster User xapeix03

Abhishek Kumar 

Last updated:  04/26/16

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received


New York City, NY  10005
US

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: Abhishek Kumar - Sr. Big Data/Hadoop Architect

Resume Value: hubi6qbz3na5wqdg   

  

 

              Page 1

 

SUMMARY:

·           More than seven years of comprehensive IT experience in Big Data domain with tools like Hadoop, Hive and other open source tools/technologies in Banking, Healthcare and Energy. 

·           Substantial experience writing MapReduce jobs in Java, PIG, Flume, Zookeeper and Hive and Storm.

·           Experience in development of Big Data projects using Hadoop, Hive, HDP, PIG, Flume, Storm and MapReduce open source tools/technologies.

·           Extensive Knowledge on automation tools such as Puppet and Chef.

·           Hands on experience in installing, configuring and using ecosystem components like Hadoop MapReduce, HDFS, Hbase, AVRO, Zoo Keeper, Oozie, Hive, HDP, Cassandra, Sqoop, PIG, Flume. 

·           Extensive experience in SQL and NoSQL development. 

·           In-depth understanding of Data Structure and Algorithms. 

·           Experience in web-based languages such as HTML, CSS, PHP, XML and other web methodologies including Web Services and SOAP. 

·           Extensive experience in all the phases of the software development lifecycle (SDLC).

·           Experience in deploying applications in heterogeneous Application Servers TOMCAT, Web Logic and Oracle Application. Server. 

·           Extensive knowledge of NoSQL databases such as HBase.

·           Worked on Multi Clustered environment and setting up Cloudera Hadoop echo System. 

·           Background with traditional databases such as Oracle, Teradata, Netezza, SQL Server, ETL tools / processes and data warehousing architectures. 

 

EDUCATION:

BS in Computer Science UPTU, Lucknow, India 2007

 

PROFESSIONAL EXPERIENCE:

 

Morgan Stanley, New York City, NY                                                                      Feb 2015-Present

Sr. Big Data/Hadoop Architect

 

Responsibilities:

·           Delivered Working Widget Software using EXTJS4, HTML5, RESTFUL Web services, JSON Store, Linux, Hadoop, ZOOKEEPER, NO SQL databases, JAVA, SPRING Security, JBOSS Application Server for Big Data analytics.

·           Developed a custom AVRO Framework capable of solving small files problem in Hadoop and also extended PIG and Hive tools to work with it. 

·           Working on use cases, data requirements and business value for implementing a Big Data Analytics platform.

·           Working on configuring and Maintaining Hadoop environment on AWS.

·           Developed application component interacting with MongoDB.

·           Working on Modifying Chef Recipes used to configure the Hadoop stack.

·           Evaluate Puppet framework and tools to automate the cloud deployment and operations.

·           Working on Installing and configuring Hive, HDP, PIG, Sqoop, Flume, Storm and Oozie on the Hadoop cluster.

·           Working in analyzing data using Hive, PIG, Storm and custom MapReduce programs in Java.

·           Developed Use cases and Technical prototyping for implementing PIG, HDP, HIVE and HBASE.

·           Working as a Big Data/Hadoop Architect on Integration and Analytics based on Hadoop, SOLR and web Methods technologies.

·           Working in implementing Hadoop with the AWS EC2 system using a few instances in gathering and analyzing data log files.

·           Working on data using Sqoop from HDFS to Relational Database Systems and vice-versa.

·           Working on GE on loading files to HIVE and HDFS from MongoDB.

·           Founded and developed environmental search engine engine using PHP5, JAVA, Lucene/SOLR, Apache and MYSQL.

·           Led the evaluation of Big Data software like Splunk, Hadoop for augmenting the warehouse, identified use cases and led Big Data Analytics solution development for Customer Insights and Customer Engagement teams.

·           Worked on Distributed/Cloud Computing (MapReduce/Hadoop, PIG, HBase, AVRO, Zookeeper, etc.), Amazon Web Services (S3, EC2, EMR, etc.), Oracle SQL Performance Tuning and ETL, Java 2 Enterprise and Web Development.

·           Designed techniques and wrote effective and successful programs in JAVA, Linux shell scripting to push the large data including the Text and Byte type of data to successfully migrate to NO SQL Stores using various Data Parser techniques in addition to Map Reduce jobs.  

·           Involved in scheduling Oozie workflow engine to run multiple Hive and PIG jobs.

·           Worked on TOAD for Data Analysis, ETL/Informatica for data mapping and the data transformation between the source and the target database. 

·           Working on Hive/Hbase vs RDBMS, imported data to Hive, HDP created tables, partitions, indexes, views, queries and reports for BI data analysis.

·           Developing data pipeline using Flume, Sqoop, PIG and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.

·           Currently working on XML parsing using PIG, Hive, HDP and Redshift.

·           Working on architected solutions that process massive amounts of data on corporate and AWS cloud based servers.

·           Uses Splunk to detect any malicious activity against webservers 

·           Tuned the Hadoop Clusters and Monitored for the memory management and for the Map Reduce jobs, to enable healthy operation of Map Reduce jobs to push the data from SQL to No SQL store.

·           Working in importing streaming logs and aggregating the data to HDFS through Flume.

 

JP Morgan Chase, New York City, NY                                                                           Aug 2014-Jan 2015

Big Data/Hadoop Developer/Architect

 

Responsibilities:

·           Worked on NoSQL databases including Hbase and MongoDB.

·           Used data stores included Accumulo/Hadoop and graph database.

·           Exploited Hadoop MySQL–Connector to store Map Reduce results in RDBMS.

·           Used Dell Crowbar as a wrapper to Chef to deploy Hadoop.

·           Collected data from different databases (i.e. Teradata, Oracle, MySQL) to Hadoop

·           Used Oozie and Zookeeper for workflow scheduling and monitoring.

·           Worked on Designing and Developing ETL Workflows using Java for processing data in HDFS/Hbase using Oozie.

·           Created Hbase tables to store various data formats of PII data coming from different portfolios

·           Conduct vulnerability analyses; reviewing, analyzing and correlating threat data from available sources such as Splunk.  

·           Designed, planned and delivered proof of concept and business function/division based implementation of Big Data roadmap and strategy project (Apache Hadoop stack with Tableau) in JP Morgan Chase using Hadoop.

·           Developed MapReduce jobs in Java for data cleaning and preprocessing. 

·           Importing and exporting data into HDFS and Hive using Sqoop.

·           Used Bash Shell Scripting, Sqoop, AVRO, Hive, HDP, Redshift, PIG and Java Map/Reduce daily to develop ETL, batch processing, and data storage functionality.   

·           Responsible for developing data pipeline using flume, Sqoop and PIG to extract the data from weblogs and store in HDFS.

·           Working on extracting files from MongoDB through Sqoop and placed in HDFS and processed.

·           Worked on Hadoop installation & configuration of multiple nodes on AWS EC2 system.

·           Experienced in running Hadoop streaming jobs to process terabytes of XML format data. 

·           Cluster coordination services through Zoo Keeper. 

·           Involved in loading data from UNIX file system to HDFS 

·           Installed and configured Hive and also written Hive UDFs. 

·           Worked on setting up PIG, Hive, Redshift and Hbase on multiple nodes and developed using PIG, Hive, Hbase, MapReduce and Storm. 

·           Installed and configured Hadoop through Amazon Web Services in cloud. 

·           Design and implement data processing using AWS Data Pipeline.

·           Drove holistic tech transformation to Big Data platform for JP Morgan Chase, create strategy, define blueprint, design roadmap, build end-to-end stack, evaluate leading technology options, benchmark selected products, migrate products, reconstruct information architecture, introduce metadata management, leverage machine learning, productionize consolidated data store: Hadoop, MR, Hive, HDP and MapReduce.

·           Developed Simple to complex MapReduce Jobs using Hive and PIG.

·           Worked on automate monitoring and optimizing large volume data transfer processes between Hadoop clusters and AWS.  

·           Strong knowledge on Data Warehousing experience using Informatica Power Center.

·           Configure and manage Splunk Forwarders, Splunk Indexers and Splunk Search Heads. 

·           Automated all the jobs for pulling data from FTP server to load data into Hive tables using Oozie workflows. 

 

Dell Inc., Round Rock, TX                                                                                                                                              Jan 2013-July 2014

Hadoop Developer

 

Responsibilities:

·           Worked on analyzing Hadoop cluster and different Big Data analytic tools including PIG, Hbase database and Sqoop. 

·           Responsible for building scalable distributed data solutions using Hadoop. 

·           Built Analytics KPI engine using Python and PIG. 

·           Worked on the data to import into AWS instance and Mapreduce jobs will be executed to analyze the data.

·           Developed fielded search dataset using Hadoop and Accumulo.

·           Working with HP to scale their existing data pipeline to handle 10x the data to match our growth trajectory.

·           Worked in Big Data Analytics Initiative from system design to live operation in production.

·           Created customizations report using pentaho and MongoDB/ Cassandra provisioning/ decommission.

·           Installed Oozie workflow engine to run multiple Hive, Redshift and PIG Jobs.

·           Production and pre-production cloud infrastructure based on AWS cloud.

·           Worked on supporting and managing Chef Server.

·           Worked on Hadoop AVRO Files to Network File System for recording Audit data.

·           Worked on supporting MapReduce Programs those are running on the cluster

·           Help assisting finding bugs and learning Tableau Desktop for Splunk.

·           Created HBase tables to store variable data formats of PII data coming from different portfolios. 

·           Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop. 

·           Implemented best income logic using PIG scripts and UDFs. 

·           Worked on tuning the performance PIG queries. 

·           Involved in loading data from UNIX file system to HDFS. 

·           Worked on Apache web log data into Hadoop, transform it into a standard AVRO format, and output it to a Snappy compressed AVRO file.

·           Installed and Configured AWS Data Pipeline to run multiple AWS.

·           Expertise in developing and deploying Splunk, installing Splunk forwarders and developing dashboards. 

·           Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files. 

·           Built firmware check-in service using MongoDB, Java/Spring as both hosted and cloud service.

·           Develop reusable tools to efficiently handle the AWS migration pipeline, mostly scripts.

·           Installed Oozie workflow engine to run multiple Hive, HDP, Redshift and PIG jobs. 

·            Supported in setting up QA environment and updating configurations for implementing scripts with PIG and Sqoop.

 

Hewlett-Packard, Palo Alto, CA                                                                                                                                  June 2008-Nov 2012                                     

JAVA Developer

 

Responsibilities:

·           Database analyzing, design and implementation.

·           Worked on entire data pipeline for automating using Flume and for the jobs scheduled periodically using Oozie.

·           Used JavaScript for client side validation. 

·           Database connections and code implementation.

·           Used Python because supports multiple programming paradigms, including object-oriented, imperative and functional programming styles.

·           Expertise in writing shell scripts and Oozie workflows

·           Worked on developing and extending serialization frameworks like AVRO.

·           Development and maintenance of splunk dashboards based on the requirements. 

·           Worked on configure AWS EMR (Elastic MapReduce).

·           Written the Apache PIG scripts to process the HDFS data.

·           Reviewing the existing Hadoop environment and make recommendations of new features that may be available and performance tuning with the other tools like Hive, PIG, MapReduce, Storm and Flume.

·           Worked on Monitoring, Replication and Sharding Techniques in MongoDB.

·           Performed Manual Testing, reported defects in JIRA and was responsible to keep track of them.

·           Used BI solution as a custom application build using OBIEE as front end and ODI and PL/SQL used for ETL. Developed HTML and JSP pages using Struts.

·           Designed GUI Components using Tiles frame work and Validation frame work.

·           Worked in Installation, Configuration and Management of Hadoop Cluster spanning multiple racks using automated tools like puppet and chef.

·           Worked on providing support for AWS Data Pipeline.

·           Worked on automating the jobs using Oozie in the project.

·           Used sequence and AVRO file formats and snappy compressions while storing data in HDFS.

·           Developed MapReduce application using Hadoop, Redshift, MapReduce programming and Hbase.

·           Monitoring Hadoop scripts which take the input from HDFS and load the data into Hive.

 

 



Experience

BACK TO TOP

 

Job Title

Company

Experience

Sr. Big Data/Hadoop Architect

Morgan Stanley & Co Inc

- Present

 

Additional Info

BACK TO TOP

 

Current Career Level:

Experienced (Non-Manager)

Work Status:

US - I am authorized to work in this country for any employer.

 

 

Target Job:

Target Job Title:

Sr. Big Data/Hadoop Architect

 

Target Company:

Company Size:

Occupation:

IT/Software Development

·         General/Other: IT/Software Development

 

Target Locations:

Selected Locations:

US-NY-New York City