From:                              route@monster.com

Sent:                               Monday, September 28, 2015 12:59 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Talend

 

This resume has been forwarded to you at the request of Monster User xapeix03

Mari M 

Last updated:  09/12/14

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received


Phoenix, AZ  85001
US

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: Hadoop developer

Resume Value: mx3x92d32m65aya4   

  

 

 

Mari M

 

E Mail ID: marim8033@gmail.com     

Professional Summary:

 

·   8 Years of extensive experience including two plus years of Big Data and BigData analytics.

·   In depth and extensive knowledge of Hadoop architecture and various components

·   Familiar with components of Hadoop Ecosystem: HDFS, Hive, HBase, Pig.

·   Over 2 years of experience in development of Big Data projects using Hadoop, Hive and other open source tools/technologies

·   Expertise in Hadoop Application Development, Designed and deployed Hadoop cluster production environment that can scale to petabytes.

·   Used Talend open studio 5.5 with Bigdata to load the data into HDFS format.

·   Developed applications for Distributed Environment using Hadoop, Mapreduce and      Python in Linux Environment.

·   Worked with big data developers, designers and scientists in troubleshooting map reduce job failures and issues with Hive, Pig, HBASE etc

·   Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with Data warehouse reference tables and historical metrics

·   Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios

·   Move data from different sources in to Hadoop and define detailed technical processes for data acquisition.

·   Advance analytics and interpretation skill on large data.

·   Demonstrate analytical and problem solving skills, particularly those that apply to a "Big Data" environment

·   Strong experience in Extraction, Transformation and Loading (ETL) data from various sources into Data Warehouses and Data Marts using Informatica Power Center (Repository Manager, Designer, Workflow Manager, Workflow Monitor, Metadata Manger), Power Exchange, Power Connect as ETL tool on Oracle, DB2 and SQL Server Databases

·   Expertise in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations teamed with project scope, Analysis, requirements gathering, data modeling, ETL Design, development, System testing, Implementation and production support.

·   Strong understanding of Data warehouse concepts, ETL, Star Schema, Snowflake, data modeling experience using Normalization, Business Process Analysis, Reengineering, Dimensional Data modeling, FACT& dimensions tables, physical & logical data modeling.

·   Provided high level solutions, partnering with end-users in the resolution of reporting problems and in database design.

·   Extensive experience in writing UNIX shell scripts and automation of the ETL processes using UNIX shell scripting.

·   Experience in using Automation job Scheduling tool Control-M

·   Excellent interpersonal and communication skills, and is experienced in working with senior level managers, business people and developers across multiple disciplines

·   Experience in integration of various data sources like Oracle, SQL server and MS access and non-relational sources like flat files into staging area.

·   Expertise in developing Test Strategy, Test Plan, Test Scenarios, Test Case Design, Test Reports using business/functional requirements.

·   Thorough knowledge in Functionality Testing, GUI Testing, Regression Testing, Integration Testing, database testing, System Testing.

·   Responsibilities include identification of test scenarios, test case creation, identifying test data requirements, test data creation, Quality Center upload and test coverage activity vs. requirements, SIT execution, defect tracking and resolution.

·   Have experience in Tools QTP.

·   Excellent Knowledge in SDLC, STLC, Defect Life Cycle.

 

 

Technical Skills:

·   Hadoop     : HDFS, Map Reduce, Hive, Pig, Sqoop, cassandra, Hbase

·   Languages              : SQL,T-SQL,UNIX, Python, NoSQL

·   ETL Tools           : Informatica Power Center 9.1/8.6, Talend open studio 5.5 Bigdata

·              RDBMS                 : MS SQL Server 2005/2008/2012, Oracle 9i/10g, MS Access

·              BI Tools                 : Business Object XI 3.1, SSRS2008.   

·              Scheduling Tools : Control-M

·              Testing Tools        : Quality Control

·              Version Control   : VSS, Share point server.

 

Education Qualifications:

     Qualification: Bachelor of computer Application, TN, India.

 

 

Professional Experience:

 

Client: The Home Depot, Atlanta, GA                                                                      May12 - Till date

Role    : Hadoop Developer

Merchandising Management

 

Project Description:

 

The Home Depot was founded in 1978 and Home Depot's proposition was to build home-improvement superstores, larger than any of their competitors' facilities. The Home Depot has commitment "to the environment and pledge to continue to be an industry leader in looking for products and services that are respectful of our world. The Home Depot introduced a label on nearly 3,000 products in 2007. The label promotes energy conservation, sustainable forestry and clean water.

Responsibilities:

·   Experienced in running Hadoop streaming jobs to process terabytes of xml format data

·   Load and transform large sets of structured, semi structured and unstructured data

·   As the application involves petabytes of data, efficiently processed billions of rows using HIVE and mapreduce.

·   Developed TalenD jobs based on the requirement using  tmap, tjoin, tLogRow etc.,

·   Developed standard and reusable jobs using TalenD Context Reference.

·   Load the data from RDBMS to HDFS by using Talend5.5 open studio with Bigdata.

·   Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.

·   Worked on partitioning the HIVE table and running the scripts in parallel to reduce the run time of the scripts.

·   Designed and deployed Hadoop cluster production environment that can scale to petabytes.

·   Developed scripts to run the queries and the Hadoop M/R program in a workflow fashion

  •    Experienced in managing and reviewing Hadoop log files
  •    Experienced in running Hadoop streaming jobs to process terabytes.

·   Developed and executed Hive Queries for de-normalizing the data

·   Developed mappings, sessions and workflows in Informatica Power Center.

 

Environment: Hadoop 2.0, Linux/UNIX, Talend open studio 5.5 with Big Data, HIVE 0.7, PIG, Python, Control-M, Tortoise SVN, Informatica 9.1.

 

 

Client: CITI, NJ                                                                                                 Feb10 to Apr12       

Role    : ETL Informatica Developer / Report Developer

 

Project Description:

 

Citibank is the consumer banking division of financial services multinational Citigroup. Citibank was founded in 1812 as the City Bank of New York, later First National City Bank of New York. As of March 2010[update], Citigroup is the third largest bank holding company in the United States by total assets, after Bank of America and JPMorgan Chase.

 

Responsibilities:

 

·   Using Informatica PowerCenter Designer analyzed the source data to Extract & Transform from various source systems (oracle 10g,DB2, SQL server and flat files) by incorporating business rules using different objects and functions that the tool supports.

·   Developed mappings, sessions and workflows in Informatica Power Center.

·   Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.

·   Worked with heterogeneous source to Extracted data from Oracle database, XML and flat files and loaded to a relational Oracle warehouse.

·   Developed standard and reusable mappings and mapplets using various transformations like Expression, Aggregator, Joiner, Router, Lookup (Connected and Unconnected) and Filter.

·   Interacted with Users to know their Business views while gathering the Report requirements and provided Several Report Mock-ups to finalize the requirements.

·   Developed flexible queries using filters, navigational attributes in BEx Analyzer to facilitate data analysis in a drill down or summarized way to give detailed levels of information.

·   Have done POC in Big Data to load ETL to HDFS.

·   Involved in meetings with production team for issues related to Deployment, maintenance, future enhancements, backup and crisis management of DW

 

Environment: Informatica  PowerCenter 8.6, TalenD, SSRS 2008, oracle 11g, DB2, SQL server, Control-M and UNIX.

 

 

Client: Esurance, CA                                                                                Nov08 to Jan10       

Role    : ETL Informatica Developer / Report Developer

 

Project Description:

Esurance Insurance Services, Inc. is an American auto insurance provider. The company, which is a wholly owned subsidiary of Allstate, sells car insurance direct to consumers online and over the phone offering services to nearly 90 percent of the U.S. population in 40 states. Esurance has more than 2,500 employees in 16 offices nationwide.

 

Responsibilities:

 

·   Designed and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.

·   Created reusable transformations and mapplets to prevent the redundancy of                          

·   Extensively used Source Qualifier Transformation to filter data at Source level rather than at Transformation level. Created different transformations such as Source Qualifier, Joiner, Expression, Aggregator, Rank, Lookups, Filters, Stored Procedures, Update Strategy and Sequence Generator.

·   Extensively worked on Informatica to extract data from Flat files, Excel files, and MS Access database to load the data into the target database.

·   Implemented the Incremental loading of Dimension and Fact tables.

·   Created Tasks, Workflows, Sessions to move the data at specific intervals on demand using Workflow Manager

·   Designed and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.

·   Worked with production team to resolve data issues in Production database of OLAP and OLTP systems.

·   Maintained source and target mappings, transformation logic and processes to reflect the changing business environment over time.

 

 

Environment: Power Exchange 8.1, SQL Server, UNIX.

 

Client:Medstart, VA                                                                                                           Oct 06 - Oct 08

Role    : ETL Informatica Developer / Report Developer

 

Project Description:

 

MedStart Health is a not-for-profit healthcare organization. It operates more than 120 entities, including ten hospitals in the Baltimore–Washington metropolitan area of the United States. It is the employer with the largest number of local employees in that area.

 

Responsibilities:

 

·   Created Tasks, Workflows, Sessions to move the data at specific intervals on demand using Workflow Manager

·   Using Informatica PowerCenter created mappings and mapplets to transform the data according to the business rules

·   Worked on different tasks in Workflow Manager like Sessions, Events raise, Event wait, E-mail, Command, Worklets, Assignment, Timer and Scheduling of the workflow.

·   Extensively used workflow variables, mapping parameters and mapping variables.

·   Extracted data from Flat files, DB2, SQL and Oracle to build an Operation Data Source. Applied business logic to load the data into Global Data Warehouse

·   Maintained source and target mappings, transformation logic and processes to reflect the changing business environment over time.

·   Worked on different tasks in Workflow Manager like Sessions, Events raise, Event wait,         E-mail, Command, Worklets, Assignment, Timer and Scheduling of the workflow.

·   Actively involved in production support. Implemented fixes/solutions to issues/tickets raised by user community.

·   Designed and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.

·   Implemented CDC by tracking the changes in critical fields required by the user.

·  Developed Manual test cases by decomposing use cases for specific functional requirements.

·  Using HP Quality Center for storing, maintaining the test repository, defect tracking and reporting.

·  Responsibilities include identification of test scenarios, test case creation, identifying test data requirements, test data creation, Quality Center upload and test coverage activity vs. requirements, SIT execution, defect tracking and resolution.

·  Writing the functional test cases, Prepare Test Reports.

·   Strong with relational database design concepts.

 

Environment: PowerCenter 7.1, Oracle 9i, DB2, SQL, UNIX, Quality Center 10.0, MS Office Suite, Windows XP.

 



Additional Info

BACK TO TOP

 

Current Career Level:

Experienced (Non-Manager)

Work Status:

US - I am authorized to work in this country for any employer.

 

 

Target Company:

Company Size:

 

Target Locations:

Selected Locations:

US-AZ-Phoenix