From: route@monster.com
Sent: Monday, September 28, 2015 12:59 PM
To: hg@apeironinc.com
Subject: Please review this candidate for: Talend
This resume has been forwarded to
you at the request of Monster User xapeix03
|
|||||||
|
|||||||
|
|
|
||||||
|
||||||
|
Mari.
M
480-704-4626 ilaa28.m@gmail.com Professional Summary: ·
8 Years of
extensive experience including two plus years of Big Data
and BigData analytics. ·
In depth
and extensive knowledge of Hadoop architecture and various components ·
Familiar
with components of Hadoop Ecosystem: HDFS, Hive, HBase, Pig. ·
Over 2
years of experience in development of Big Data projects using Hadoop, Hive
and other open source tools/technologies ·
Expertise
in Hadoop Application Development, Designed and deployed Hadoop
cluster production environment that can scale to petabytes. ·
Used Talend
open studio 5.5 with Bigdata to load the data into HDFS format. ·
Developed
applications for Distributed Environment using Hadoop, Mapreduce
and Python in Linux Environment. ·
Worked with
big data developers, designers and scientists in troubleshooting
map reduce job failures and issues with Hive,
Pig, HBASE etc ·
Created Hive
queries that helped market analysts spot emerging
trends by comparing fresh data with Data warehouse reference tables
and historical metrics ·
Created HBase
tables to load large sets of structured, semi-structured and unstructured
data coming from UNIX, NoSQL and a variety of portfolios ·
Move data
from different sources in to Hadoop and define detailed technical
processes for data acquisition. ·
Advance analytics
and interpretation skill on large data. ·
Demonstrate
analytical and problem solving skills, particularly those that apply to a
"Big Data" environment ·
Strong
experience in Extraction, Transformation and Loading (ETL) data from various
sources into Data Warehouses and Data Marts using Informatica Power Center
(Repository Manager, Designer, Workflow Manager, Workflow Monitor, Metadata
Manger), Power Exchange, Power Connect as ETL tool on Oracle, DB2 and SQL
Server Databases ·
Expertise
in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations teamed with
project scope, Analysis, requirements gathering, data modeling, ETL Design,
development, System testing, Implementation and production support. ·
Strong
understanding of Data warehouse concepts, ETL, Star Schema, Snowflake, data
modeling experience using Normalization, Business Process Analysis,
Reengineering, Dimensional Data modeling, FACT& dimensions tables,
physical & logical data modeling. ·
Provided
high level solutions, partnering with end-users in the
resolution of reporting problems and in database design. ·
Extensive
experience in writing UNIX shell scripts and automation of the ETL processes
using UNIX shell scripting. ·
Experience
in using Automation job Scheduling tool Control-M ·
Excellent
interpersonal and communication skills, and is experienced in working with
senior level managers, business people and developers across multiple
disciplines ·
Experience
in integration of various data sources like Oracle, SQL server and MS access
and non-relational sources like flat files into staging area. · Expertise in developing Test Strategy, Test Plan, Test
Scenarios, Test Case Design, Test Reports using business/functional
requirements. ·
Thorough
knowledge in Functionality Testing, GUI Testing, Regression Testing,
Integration Testing, database testing, System Testing. · Responsibilities include identification of test scenarios,
test case creation, identifying test data requirements, test data creation, Quality
Center upload and test coverage activity vs. requirements, SIT
execution, defect tracking and resolution. · Have experience in Tools QTP. · Excellent Knowledge in SDLC, STLC, Defect Life Cycle. Technical Skills: ·
Hadoop : HDFS, Map Reduce, Hive, Pig, Sqoop, cassandra, Hbase ·
Languages
: SQL,T-SQL,UNIX, Python, NoSQL · ETL Tools
: Informatica Power Center 9.1/8.6,
Talend open studio 5.5 Bigdata ·
RDBMS
: MS SQL Server 2005/2008/2012,
Oracle 9i/10g, MS Access ·
BI
Tools
: Business Object XI 3.1, SSRS2008.
·
Scheduling
Tools : Control-M ·
Testing
Tools : Quality Control ·
Version
Control : VSS, Share point server. Education Qualifications: Qualification: Bachelor of computer Application, TN, India. Professional Experience: Client: The Home Depot, Atlanta, GA
May12 - Till date Role : Hadoop Developer Merchandising Management Project Description: The Home Depot
was founded in 1978 and Home Depot's proposition was to build
home-improvement superstores, larger than any of their competitors'
facilities. The Home Depot has commitment "to the environment and pledge
to continue to be an industry leader in looking for products and services
that are respectful of our world. The Home Depot introduced a label on nearly
3,000 products in 2007. The label promotes energy conservation, sustainable
forestry and clean water. Responsibilities: ·
Experienced
in running Hadoop streaming jobs to process terabytes
of xml format data ·
Load and transform large sets of structured, semi
structured and unstructured data ·
As the application
involves petabytes of data, efficiently processed
billions of rows using HIVE and mapreduce. ·
Developed TalenD
jobs based on the requirement using tmap, tjoin, tLogRow
etc., ·
Developed
standard and reusable jobs using TalenD Context Reference. ·
Load the
data from RDBMS to HDFS by using Talend5.5 open studio
with Bigdata. ·
Involved in
creating Hive tables, loading with data and writing hive
queries which will run internally in map reduce way. ·
Worked on partitioning
the HIVE table and running the scripts in parallel to reduce
the run time of the scripts. ·
Designed and deployed Hadoop cluster production
environment that can scale to petabytes. ·
Developed scripts
to run the queries and the Hadoop M/R program in a
workflow fashion ·
Experienced in managing and reviewing Hadoop log files ·
Experienced in running Hadoop
streaming jobs to process terabytes. ·
Developed
and executed Hive Queries for de-normalizing the
data ·
Developed
mappings, sessions and workflows in Informatica Power Center. Environment: Hadoop 2.0, Linux/UNIX, Talend open
studio 5.5 with Big Data, HIVE 0.7, PIG, Python, Control-M, Tortoise SVN. Client: CITI,
NJ
Feb10 to Apr12 Role : ETL Informatica Developer /
Report Developer Project
Description: Citibank is the consumer banking division of
financial services multinational Citigroup. Citibank was founded in 1812 as
the City Bank of New York, later First National City Bank of New York. As of
March 2010[update], Citigroup is the third largest bank holding company in
the United States by total assets, after Bank of America and JPMorgan Chase. Responsibilities: ·
Using
Informatica PowerCenter Designer analyzed the source data to Extract &
Transform from various source systems (oracle 10g,DB2, SQL server and flat
files) by incorporating business rules using different objects and functions
that the tool supports. ·
Developed
mappings, sessions and workflows in Informatica Power Center. ·
Identified
performance issues in existing sources, targets and mappings by analyzing the
data flow, evaluating transformations and tuned accordingly for better
performance. ·
Worked with
heterogeneous source to Extracted data from Oracle database, XML and flat
files and loaded to a relational Oracle warehouse. ·
Developed
standard and reusable mappings and mapplets using various transformations
like Expression, Aggregator, Joiner, Router, Lookup (Connected and
Unconnected) and Filter. ·
Interacted
with Users to know their Business views while gathering the Report
requirements and provided Several Report Mock-ups to finalize the
requirements. ·
Developed
flexible queries using filters, navigational attributes in BEx Analyzer to
facilitate data analysis in a drill down or summarized way to give detailed
levels of information. ·
Have done POC
in Big Data to load ETL to HDFS. ·
Involved in
meetings with production team for issues related to Deployment, maintenance,
future enhancements, backup and crisis management of DW Environment: Informatica PowerCenter 8.6,
TalenD, SSRS 2008, oracle 11g, DB2, SQL server, Control-M and UNIX. Client: Esurance, CA
Nov08 to Jan10 Role : ETL Informatica Developer /
Report Developer Project Description: Esurance Insurance Services, Inc. is an American
auto insurance provider. The company, which is a wholly owned subsidiary of
Allstate, sells car insurance direct to consumers online and over the phone
offering services to nearly 90 percent of the U.S. population in 40 states.
Esurance has more than 2,500 employees in 16 offices nationwide. Responsibilities: ·
Designed
and developed end-to-end ETL process from various source systems to Staging
area, from staging to Data Marts. ·
Created
reusable transformations and mapplets to prevent the redundancy
of
·
Extensively
used Source Qualifier Transformation to filter data at Source level rather
than at Transformation level. Created different transformations such as
Source Qualifier, Joiner, Expression, Aggregator, Rank, Lookups, Filters,
Stored Procedures, Update Strategy and Sequence Generator. ·
Extensively
worked on Informatica to extract data from Flat files, Excel files, and MS
Access database to load the data into the target database. ·
Implemented
the Incremental loading of Dimension and Fact tables. ·
Created Tasks,
Workflows, Sessions to move the data at specific intervals on demand using
Workflow Manager ·
Designed
and developed end-to-end ETL process from various source systems to Staging
area, from staging to Data Marts. ·
Worked with
production team to resolve data issues in Production database of OLAP and
OLTP systems. ·
Maintained
source and target mappings, transformation logic and processes to reflect the
changing business environment over time. Environment: Power
Exchange 8.1, SQL Server, UNIX. Client:Medstart,
VA
Oct 06 - Oct 08 Role : ETL Informatica Developer /
Report Developer Project
Description: MedStart Health
is a not-for-profit healthcare organization. It operates more than 120
entities, including ten hospitals in the Baltimore–Washington metropolitan
area of the United States. It is the employer with the largest number of
local employees in that area. Responsibilities: ·
Created
Tasks, Workflows, Sessions to move the data at specific intervals on demand
using Workflow Manager ·
Using
Informatica PowerCenter created mappings and mapplets to transform the data
according to the business rules ·
Worked on
different tasks in Workflow Manager like Sessions, Events raise, Event wait,
E-mail, Command, Worklets, Assignment, Timer and Scheduling of the workflow. ·
Extensively
used workflow variables, mapping parameters and mapping variables. ·
Extracted
data from Flat files, DB2, SQL and Oracle to build an Operation Data Source.
Applied business logic to load the data into Global Data Warehouse ·
Maintained
source and target mappings, transformation logic and processes to reflect the
changing business environment over time. ·
Worked on
different tasks in Workflow Manager like Sessions, Events raise, Event
wait, E-mail, Command,
Worklets, Assignment, Timer and Scheduling of the workflow. ·
Actively
involved in production support. Implemented fixes/solutions to issues/tickets
raised by user community. ·
Designed
and developed end-to-end ETL process from various source systems to Staging
area, from staging to Data Marts. ·
Implemented
CDC by tracking the changes in critical fields required by the user. ·
Developed Manual
test cases by decomposing use cases for specific functional requirements. ·
Using HP
Quality Center for storing, maintaining the test repository, defect
tracking and reporting. ·
Responsibilities
include identification of test scenarios, test case
creation, identifying test data requirements, test
data creation, Quality Center upload and test
coverage activity vs. requirements, SIT execution, defect
tracking and resolution. ·
Writing the functional
test cases, Prepare Test Reports. ·
Strong with
relational database design concepts. Environment: PowerCenter
7.1, Oracle 9i, DB2, SQL, UNIX, Quality Center 10.0, MS Office Suite, Windows XP. |
|
|
||||||||
|
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
|