From:                              route@monster.com

Sent:                               Monday, September 28, 2015 1:02 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Talend

 

This resume has been forwarded to you at the request of Monster User xapeix03

ARJUN KUMAR 

Last updated:  03/13/15

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: Big Data Consultant

Resume Value: yp9hnejf8a9mmgnf   

  

 

!
B !ig Data Lead / Architect.
Arjun Kumar
Phone Number
Email-Id
Skype id
!
: 678 900 8575
Linkedin Profile
: kumararjun.k@gmail.com
: kumararjun.k
Current Location : Plano, TX
:www.linkedin.com/in/kumararjunk1
Profile
• 11+ years of Total IT Experience.
• 5+ years of Big Data, Hadoop Experience.
• 10+ years of DataStage Experience.
• 10+ years of Teradata, Oracle, SQL Server.
• 8+ Year in Finance/Bank, Insurance, Retail and Aviation Domain.
• 4 years of Experience in Spark,R, Python, Scala, Java, Talend, Abinitio,DB2, Mainframe.
!
Summary
• 11 years of IT experience with extensive knowledge in Software Development Life Cycle
(SDLC) involving Requirements Gathering, Design Analysis, Development, Maintenance, Implementation
and Testing.
!
• 5+ Year Exp. On BIG DATA Hadoop HDFS, Hive, Beeline, Pig and Sqoop , Cloudera Hue,
java.
• 2+ year of Exp. on moving data from hdfs to Spark and building the application using
Scala / Spark Cluster.Used Scala Programs to analyze the data on Spark (Hive Tables/HDFS
files).
• Installation of zookeeper,spark,hive,hbase and other hadoop tools.
• Performance tuning of Existing Hadoop cluster and Guiding the team to use the best
practice to achieve the optimal performance.
• 4+ Year of Exp on Installation and Configuration of Hadoop Ecosystem using Ambari and
Cloudera Manager.
• Expertise on HDP Operation Tools, Puppet, Ambari , Zookeeper & Oozie.
• Experience in Design/ Development and Implementation of Big Data Application.
• Performance Tuning of Hive SQLS, Pig Scripts and R Analysis.
• Extensive experience in Cloudera 4/5 Tools and Products.
• Experience in Handling Huge volume of data in/out from Teradata/Big Data.
• Strong Etl experience using Datastage, Talend/Abinitio and other ETL tools.
• Good Experience on MongoDB , Hbase and Cassandra.
• Experience in Creating Audit control system for ETL process for Big Data and Data warehouse
Application.
• Strong skills in ETL (DataStage 9.1/Talend/Teradata) Architect/Design and development
and Performance Tuning the Data warehouse/Big Data.
• Extensive experience in using Teradata BTEQ, FLOAD, MLOAD, FASTXPORT utilities.
• Extensive experience in Performance tuning of Datastage Jobs and Teradata SQL
• Experience in Version Control/ Upgrading Datastage Version and Code Migration using
GitHub.
• Good Knowledge of Big Data and Data Warehouse Architecture and Designing Star Schema,
Snow flake Schema, Fact and Dimensional Tables, Physical and Logical Data Modeling using
Erwin.
• Designed Mapping documents/Process Flow chart for Big Data Application.
• Extensive experience in loading high volume data and performance tuning.
• Ability to work autonomously and in Agile Environment.
!!!!! !
!
TECHNICAL SKILLS
Big Data / Hadoop:
ETL Tools:
Languages:
BI Analytics
Database/File System:
Tools:
Operating Systems:
IDE Tools
!
Professional Experience
Cli ! ent: Capital One
Location: Plano, TX
Project: COAF
Role: Big Data Lead / Architect.
Project !
Respon !
Capital One is diversified bank that offer a broad array of financial products and services to
consumer, small business and commercial clients. Capital One has one of the most widely
recognized brand in America. As one of the Nation's top 10 largest banks based on deposit
and transaction. Objective of the Project was to migration of Teradata Tables to HDFS.As a
part of Fill the Lake initiative, I was involved in Data Migration of home loan, crdb, cards
data from Historical/Daily data from Teradata/Oracle to HDFS.
Description:

sibilities:
Creating Scala Programs to analyze the data on Spark (Hive Tables/HDFS
files).
Spark, Cloudera 4/5, Apache Hadoop,Horton Work HDP, Hive,
Pig, Sqoop , Cloudera Hue, Ambari, Zookeeper & Oozie
Pig, Talend 5.5, DataStage, Bteq, Fast load, Mload,TPT.
SQL, PL/SQL, Python, Java, Shell Script
R, Microstrategy, Impala, Beeline, Hive,Spark,Scala
Teradata 14/V2R6/V2R5, Oracle 10g/9i, SQL Server
2000/2005/2008,DB2,Hadoop HDFS
Teradata View Point, Teradata Sql Assistant, SQL* Plus 9.2
Windows, Linux, IBM AIX, CentOS
Eclipse, Scala IDE, IntelliJ
Apr 13 to Till Date










En! !
Creating ETL Process to Move Data from Landing Phase/Key/Split to Hive
Tables.
Historical loads (one-time) from Teradata using an unload script built on
Teradata Parallel Transporter (TPT) and FastLoad Scripts.
Creating Talend Jobs to apply business rule and to load into hdfs.
Recurring loads from the Oracle source system using an sFTP pull framework
and Landing of files into the unix edge node.
Registration of the datasets in a metadata registry that controls admittance
into Hadoop.
Designed criteria for progressive classification of the datasets within Hadoop
(from raw to validated, refined and archived)
Designed the data processing approach within Hadoop using Pig.
Keying to insert keys to support the data lineage based on certain ids. This is
done through a java program that looks up the data against a master file
ELT, using Talend Big Data for processing the data. Transformations done
include Tjoins, Treformats, Tsorts, if-else conditional transforms etc.
vironment: Cloudera Hue, HDFS, Pig, Hive, Teradata, Oracle 10g and UNIX.
Cli ! ent: McCAIN Foods
Location: Florenceville, NB
Project: Sales Performance
Role: Big Data Lead / Architect.
Project Description:
The McCain brand is one of the leading companies in Foodservice. McCain supplies product,
foodservice in more than 130 countries around the world. The development of data
warehouse is required to Extract data from 35+ source system and load it to a Teradata
Tables. Their sales data comes various heterogeneous source system i.e. Oracle, Sql Server,
Flat Files and AS/400.Hadoop was used as one of the ETL tool to load data into HDFS from
various source to Teradata Tables/SAP/Flat Files Sql server.
Hive was used as BI tool to analyses the Historical Sales Data and Daily Sales Data.
BI was used to predict the Future Sales and to set the Target Production volumes of Each
Manufacturing Plant.
!
• Leading Development team, In Developing / DBA and Business to Develop the ETL
solutions for various subject areas.
Responsibilities:
• Extensively used Cloudera (Hive) to generate the Reports.
• Provide Technical guidance and leadership on Etl methodologies for projects, such as
technical requirements, source system analysis, Etl design topics, ETL control system,
Etl Jobs, code reviews and testing procedures and Integration Testing.
• Worked extensively on Big Data, extensively used DataStage an ETL tool to design
mappings to move data from Source to Target database-using Stages.
• Interacting with Business/Domain exert to gather ETL requirements & analysis off
various data sources, Flat files, oracle, sql server, mainframe system.
Splitting to shard the dataset into NPI, Credit and Anonymous classifications
as per federal regulations. This is accomplished through a python script where
the schema of the split is passed as a parameter.
Mar`10 to Apr 13
• Interacting with DBA team to convert Logical Data Model into the Physical Data Model
and recommending primary index/secondary based on nature of query submitted by
the end user.
• Experience in Handling NLS Characters, i.e. loading CHN/JPN and Latin European
language special characters.
• Design and Development of Datastage ETL jobs using various stages like Aggregator,
Join, Transformer, Sort, Merge, Filter and Lookup, Sequence, Modify, Peek etc. stages
to load Daily and Historical data into star schema data model.
• Created Generic CDC script in UNIX/Bteq to capture the new records, update and
deletes.
• Used Teradata utilities like Sql assistant, BTEQ, FASTLOAD and MLOAD.
• Broadly involved in Data Extraction, Transformation and Loading (ETL process) from
Source to target systems using DataStage PX.
• Successfully Integrated data across multiple source with high volumes of data.
• Performance Tuning of existing Etl Jobs and Automation of ETL processes using
DataStage Job Sequencer, Job Control routines and Transform functions.
• Extensively used DataStage Designer, Administrator, Manager, and Director for creating
and implementing jobs.
En ! vironment: Cloudera HDFS, Pig,Hive, IBM Info Sphere Information Server V8.1 Suite
DataStage,Teradata DB2, Oracle, UNIX Shell Scripting, Windows XP, Erwin.
Cli ! ent: Royal Bank of Canada
Jan`09 to Mar`10
Location: Toronto, ON
Project: CP Migration
Role: Sr. ETL Developer.
Project !
Royal Bank of Canada is one of the leading Banks in Canada (North America).Royal Bank has
the largest Data warehouse in North America. Royal Bank of Canada had Clarton Passport
Program for most of Data Transformation.The purpose of the project was to replace the
existing Clarton Passport program, which was used for the data transformation and creating
load set. DataStage was used as the ETL tool to produce the same SPDS load set produced by
the Clarton Passport program. These SPDS load sets were used by the downstream process to
load data into Teradata Tables. This Project involves migrating data from the Binary
Mainframe files to SPDS load sets.
Description:
• Designing the ETL jobs using DataStage 7.5 tool to load data from Mainframe Binary
file to SPDS load sets and Used various Stages like Lookup, Joiner, Sorter, Sequential
file and other Stages.
Responsibilities:
• Preparing the technical documents for ETLs & preparing the Unit Test Cases
• Performing User Acceptance test & Analyzing the Data and Defect fixing, (PIV issue).
• Preparing the Technical Design documentation for every DataStage module completed.
• Compare the results from the existing process and the DataStage process by JCLs, File
Aid.
En ! vironment: IBM Info Sphere Information Server V7.5.2 Suite [DataStage], Teradata,
Clarton Passport 5.5,JCL-COBOL, Mainframe Binary EBCDIC files, ZEKE.
!
• Providing production support and resolving issues in small turnaround time.
• Decommission the Passport Program after successful results.
• Performing Impact analysis of SCR, Giving time estimate and Implementing SCR’s
• Migration of Entire ETL jobs/script/bteq/Shell Scripts/DSN connection to production
environment.
Cli! ent: American Airlines.
Mar`08 to Dec`08
Project: Mosaic
Role: Sr. DataStage Developer
Location: Teradata Corporation, Mumbai.
Project !
Respon!! !
American Airlines is one of the largest Airlines in United States. Client had a need for an
analytical data warehouse to improve their ticket booking process, Passenger information,
Tracking of Payments. To give them Real time information regarding their Ticket booking
Process thus helps to understand the trends. Data warehouse is developed to replace the
Current Operational System to perform the analysis of their Ticket booking Process. This
Project involves migrating data from the Mainframe to Teradata. It had four Data marts.
Description:
• Analyzed the existing ETL process and came up with an ETL design document that
listed the jobs to load, the logic to load and the frequency of load of all the tables.
sibilities:
• Designing the ETL jobs using DataStage 7.5.2 tool to load data from Mainframe Binary
file to Teradata Database and Parallel jobs to load the data into the Target Schema.
• Used various Stages like Teradata Enterprise, Multiload Stage, Lookup, Joiner, Sorter
and Remove Duplicate Stages, Transformer, Dataset, Row Generator, Column Generator,
Datasets, Sequential File, Aggregator and Modify Stages.
• Preparing the technical documents for ETLs.
• Testing the Jobs and preparing the Unit Test Cases & performing User Acceptance test
using Tivoli.
• Converted complex job designs to different job segments and executed through job
sequencer for better performance and easy maintenance.
• Extensively worked on Data Acquisition and Data Integration of the source data.
• Design Technical specifications and Mapping documents with Transformation rules.
• Defined projects and tuned parameters for fine-tuning of the projects.
• Defined & implemented DataStage jobs process monitoring.
• Involved in creating UNIX shell scripts for database connectivity and executing queries
in parallel job execution.
• Performed Unit testing on the module designed.
• Used the ETL DataStage Director to schedule, running the solution, testing and
debugging its components & monitor the resulting executable versions.
• Teradata utilizes like BTEQ, FASTLOAD, MLOAD, and FASTEXPORT.
En ! vironment: Ascential DataStage 7.5.2, IBM DB2, Teradata, Erwin, Linux, XP.
Cli!!! !
• Defined production support methodologies and strategies.
• Involved in 24/7 ETL Production Support, maintenance, troubleshooting, problem
fixing and ongoing enhancements to the Data mart.
ent: St. Paul
Nov` 06 to March` 08
Project: CLMI Warehouse Loss Subject Area:
Role: DataStage Developer
Location: Teradata Corporation, Mumbai.
Project!
St. Paul Travelers is one of the largest Auto Insurance Company in United States. St. Paul
Traveler had to develop the data warehouse to improve their decision making system to give
them Real time information regarding their Claim Process, to understand the trends, Fraud
detection, False Claims. Data warehouse is developed to replace the Current Operational
System to perform the analysis of their Claim Process. This Project involves migrating the data
from the Mainframe to Teradata and.DataStage was used as the ETL tool to load data from
Mainframe to Teradata tables. Warehouse had three subject areas, namely CLAIM, COVERAGE
and AMOUNTS.
Description:
Respon !
• Designed jobs involving various cross reference lookups and joins, shared containers which
can be used in multiple jobs.
sibilities:
• Sequencers are created at job level to include multiple jobs and a layer level sequence
which include all job level sequences.
• Extensively employed Datastage Director to validate, run, schedule, monitor the jobs and
followed job log carefully to debug the jobs.
• Carefully monitored the performance statistics and involved in fine tuning of jobs for the
improved processing time.
• Involved in developing UNIX scripts to call Datastage jobs.
• Involved in fine tuning, trouble shooting, bug fixing, defect analysis and enhancement of
the multiple admin systems Datastage jobs.
• Involved in the designing of marts and dimensional and fact tables.
En ! vironment: Data stage 7.5, Teradata, Mainframe system.
Cli ! ent: St. Paul
Project: MIETL – Historical Conversion:
Role: ETL Developer
Location: LNT INFOTEC, Pune.
Project !
Respon !
St. Paul Travelers is one of the largest Auto Insurance Company in United States. The
development of data warehouse is inextricably linked to the business needs of the St. Paul
Traveler, Client had a need for an analytical data warehouse to improve their decision
making system to give them Real time information regarding their Claim Process thus help
to understand the trends. Data warehouse is developed to replace the Current Operational
System to perform the analysis of their Claim Process. This Project involves migrating data
from the Mainframe files to Teradata tables.
Description:
As an ETL developer, I am involved in
• Designing the ETL jobs using DataStage 7.5.1 tool to load data from Mainframe to
Teradata Database.
sibilities:
• Developing Parallel jobs to load the data into the Target Schema.
• Created sequence for running ETLs using new job sequencer, job activity, nested
condition, notification activity and sequencer.
• Preparing the technical documents for ETLs.
• Testing the Jobs and preparing the Unit Test Cases.
• Used job sequence to create job sequencer, to call various parallel job and send message
in case of process failures through the execution command.
March` 06 to Nov` 06
En! vironment: Data stage 7.5, Teradata, Mainframe system.
C !lient: HAL Banglore.
July`04 to Feb`06
Project: Standardization of Committee of Management Minutes:
Role: Software Trainee
Location: HAL Banglore
Project !
HAL is one of the leading Aeronautics government organizations in India. They have 12+
divisions in India. Committee of Management Minutes project was required to create the
information system to meet the requirement of the entire divisons.Front end was FoxPro and
backend was in oracle. Forms and reports were created in fox pro to collect the data from
various divisions and data was stored in oracle tables and later on these fox pro is integrated
with the web service. It provided the detailed information on the stock on inventory, spare
parts, time taken for each repair and cost associated with each repair.
Description:
Respon !
• Analyzed the existing system to load data into oracle through fox pro.
• Creating Forms to accept data in the fox pro.
• Interacting with the end user to gather requirement and designing PDM table’s
structure.
sibilities:
• Loading data through fox pro and integrating fox pro and oracle.
• Data validation and bug fixing.
Environment: Oracle 8i and Fox pro.
TECHNICAL CERTIFICATION
!
!
!
!
!
• Teradata Master Professional V2R5.
• IBM Certified Solution Developer - InfoSphere Data StageDataStage v8.0
• IBM Certified Solution Developer - Web Sphere IIS Data Stage Enterprise(7.5)
EDUCATION
• Bachelor’s in Engineering, VTU, Belgaum 2004.
REFERENCE
Available on Request



Experience

BACK TO TOP

 

Job Title

Company

Experience

Big Data Consultant

Capital One Services Inc

- Present

 

Additional Info

BACK TO TOP

 

Current Career Level:

Experienced (Non-Manager)

Years of relevant work experience:

2+ to 5 Years

Date of Availability:

Immediately

Work Status:

US - I am authorized to work in this country for my present employer only.

Active Security Clearance:

None

US Military Service:

Citizenship:

Other

 

 

Target Job:

Target Job Title:

Big Data Consultant

Desired Job Type:

Employee
Temporary/Contract/Project

Desired Status:

Full-Time

 

Target Company:

Company Size:

Industry:

Retail
Banking

Occupation:

IT/Software Development

·         Enterprise Software Implementation & Consulting

·         Software/System Architecture

 

Target Locations:

Selected Locations:

US

Relocate:

Yes

Willingness to travel:

Up to 100%

 

Languages:

Languages

Proficiency Level

English

Fluent