From: route@monster.com
Sent: Monday, September 28, 2015 1:01 PM
To: hg@apeironinc.com
Subject: Please review this candidate for: Talend
This resume has been forwarded to
you at the request of Monster User xapeix03
|
|||||||
|
|||||||
|
|
|
||||||
|
||||||
|
To
be considered for: Hadoop Big Data Architecture/development/Data data Science
positions Experienced IT
professional seeking a leadership position driving strategy, management,
planning, development, team building and end-to-end solution with focus on
leveraging information technology and data science to maximize business
values. Developer / Engineer /
Architect with over 8 years of development, administration and
programming experience with primary focus in "Big Data" and
emphasis on Hadoop Summary
Hadoop/BigData/Analytic Experience •
Spark
RDD and Spark DF development •
Worked
with Hadoop, HBase, Cascading, Zookeeper, Oozie, Hive, HiveQL, MapR, MogoDB,
Pentaho & Pig •
Worked
with AWS Cloud, EC2, EMR, RedShift, S3 etc •
Supported
Hadoop and Verita as administrator about 7 to 700 nodes •
Supported
and administrated users accounts and security rules with BigData domain •
Worked
with Application Servers, Tomcat, Oracle and MySQL •
Experienced
with distributed systems, large scale non-relational data stores, map-reduce
systems, data modeling, database performance, and multi-terabyte data
warehouses •
Experienced
in a SaaS environment that has an agile development process •
Experienced
in Java, Python or other object oriented programming language •
Experienced
in Mobile, Linux, Unix, Android, MAC platform •
Experienced
with server management, server operating systems(Windows, Linux, Unix), and
VMware •
Experienced
in testing APIs, Restful APIs •
Experienced
with full SW Lifecycle and Agile practice •
Experienced
with scripting language, in Perl, Python, Ruby, QTP for automated and
performance tests •
Experienced
with testing DNA sequence machines SW for regulatory approval •
Experienced
testing networking and storage technologies, protocols and hardware •
Experienced
with web services and SOAP UI testing •
Experienced
with Oracle, TOAD, and SQL •
Experienced
in testing UI technologies such as HTML5 •
Experienced
in Big Data/ Hadoop / Cassandra/Memcached/NoSQL/MapReduce •
Lead
and coached about 2 to 20 engineers onshore, offshore and nearshore •
Lead
the hiring process Skills: •
Big/data
and analytics platform administrator (Vertica, Greenplum, Hadoop, Hive) •
Experienced
with Display Advertising, Behavioral Ad Networks •
Recommended
Engines, Personalization •
Experienced
with Data Analytics, Data mining, Predictive Modeling •
QA
automation and testing Work Experience Sr. Big data
Architecture/development for FRB (SF) – April – Sept 2015 -Spark RDD and Spark DF
development -Docker Master and
install/deployment -NoSQL database design and
Spark SQL development -Unix/Hadoop administrator
and development -Cassandra, MongoDB,
Hbase, Hive development -Cloudera Hue and Solr
development -Data governance
policy/management and big data security -Big data security and
Kerberos development -Multi-tenant set up and
development Sr. DataBase Architecture
for GAP project (SF) - Jan - March 2015 -
Coverting large SAS SQL scripts to HiveQL for Data Science -
Coverting Oracle scripts to HiveQL programming for Data Science - Architecture
No-SQL databases and coding the MapReduce functions for over 10 years data
about emails campaigning results analytics •
Working
with data science to prototype data model and programming the business logic
with Hive, Pig •
Building
database applications on Hive, Pig, Hbase, MongoDB, Canssdra,Spark, Solr and
shark on a large hadoop nodes cluster in house •
Backend
No-SQL database design, development, architecture and testing •
-
Data gaverment Backend Hadoop Engineer
for SS Network (Redwood City) project Oct – Dec 20014 •
Spark
RDD and Spark SQL/Hive SQL •
No-SQL
database modeling •
AWS
cloud computing and architecture with Hadoop on Big Data •
Build
cloud application on AWS, S3, EMR, Hive, Pig, Hbase, MongoDB, Canssdra,Spark •
Backend
No-SQL database design, development, architecture and testing Big/Data and BI/BW
consulting for Cisco project (San Jose) March – Aug 2014 •
Create
and maintain scripts and programs to Extract, Transform and Load data •
Automate
ETL and aggregation processes using BASH/PHP and/or other scripting languages
•
Create
logging and monitoring elements to all Data Warehouse and ETL processes to
allow Operations team to monitor •
Respond
to Data Warehouse and ETL process alerts, with support from operations team. •
Create,
maintain and automate processes to distribute data warehouse extracts to
various users •
Continually
monitor measure and improve all data warehouse and ETL processes for speed,
reliability and accuracy. •
Perform
analyses and execute improvements regarding overall data quality, including
making recommendations to the business regarding inputs. •
Source
and/or create tools to deliver monitoring metrics and dashboard reports to
allow the rest of the company to understand the quality and timeliness of
data warehouse data. •
Install
and integrate 3rd party data maintenance tools such as address cleaning
software, etc •
Use
new technologies to service data warehousing needs such as hadoop, column
databases, etc. •
- Cisco big data analytics platform development,
testing and deployment Big/Data and BI/ consulting for Intel projects(Santa Clara)
July - Dec 2013 •
Spark RDD development and No-SQL database modeling •
Design/Arcteticture Big Data Hadoop Testing Frameworks •
Building Hadoop cluster in Cloud and local •
Working with AWS cloud environment: EMR, Redshift, S3 •
Working with Scala, Java, Python, Kafka, akka, SQL, MongoDB,
JSON, Avro, Tableau •
Working with Git, SBT, Ant, Maven, Ganglia, Jenkins, •
Working with Hadoop, Hbase, Zookeeper, Oozie, Scalding, Spark, Shark •
Testing automation and coding in Scala •
Worked with ETL and reporting tools (OBIEE, SAP etc.) Big/Data and BI consulting for Cisco projects (Milpitas) ca
Sep 2012- Jun 2013 •
Hadoop /MapR BW/BI data warehouse project. •
Big data move from Informatica and TD. •
Load data from different source to new build Hadoop analyst
platform •
Build Hadoop QA team •
Using Hadoop /Hive/Sqoop/bash to deployment the data load and
query data. •
Monitor QA environment for Hadoop problems •
Testing report form MapR report tool •
Provide the ideas/working process to other team •
Review other reports and code to understand coding logic •
Managed a technical team or functioning as a team lead. •
Worked with Hadoop stack (e.g. MapReduce, Sqoop, Pig, Hive,
Hbase, Flume). •
related/complementary open source software platforms and
languages (e.g. Java, Linux, Apache, •
Perl/Python/PHP, Chef). •
Worked with ETL (Extract-Transform-Load) tools
(e.gInformatica, Talend, Pentaho). •
Worked with BI tools and reporting software (e.g. Microstrategy,
Cognos,OBIEE, Pentaho). •
Worked with analytical tools, languages, or libraries (e.g.
SAS, SPSS, R, Mahout). •
Supported business development activities to shape and
communicate proposed solutions to client executives - Implemented of ETL
applications: •
Implemented of reporting applications •
Application/implementation of custom analytics support •
Administrated of relational databases •
Data migration from existing data stores •
Infrastructure and storage design •
Developed capacity plans for new and existing systems Big/Data and BI consulting
for multiple projects (SF) Mar 2009 – Jul 2012 through Accenture 1. ATT SAP BW/BI/BigData development
support a.
Conducted
detailed design applications developed on Hadoop platforms (Feature
testing, Regression Testing, Acceptance Testing and Sanity Testing) b.
Implemented Business analysis tools with Hadoop,MapReduce
scripts from ETL data to data warehouse for BI and enterprise analysis
platform c.
Administrated Log analysis scripts for Business
Analyst tool with HDFS Hadoop (file system level) d.
Advised the Hadoop and analytica work loads IO optimizing
solutions e.
Running benchmarked Hadoop/HBase clusters f.
Supported development of Hadoop and Vertica Analytics
Platform activities 2. Storage architecture/QA Netapp
Hadoop ecosystem administrator and support a. Tested NetApp OpenSolution for
Hadoop(NOSH) settings b. Supported and scaled Hadoop
systems c. Supported cluster failed over test
and document the results with various configurations d. Administrated enterprise-grade
storage arrays and eliminated Hadoop network bottlenecks e. Supported hot-pluggable disk
shelves, added storage and administrated services f.
Supported
NFS and HDFS file systems g. Loaded network-free hardware RAID h. Day-to-day support Hadoop hardware
and software issues i.
Used
automation scripts to run performance testing of the new storage OS design 5. Consulting for Bank of America
TCOE (Concord) Oct 2006 – Feb 2009 Supported
about over 100 projects and many functional teams •
Supported
analytic massive amounts of data and BI •
Communicated
and tracked defects to closure •
Administrated
with large data set testing with BigData NoSQL Hadoop and Oracle, DB2 scripts •
Wrote
and reviewed scripts and executed scripts, tracked defects •
Used
scripts languages in Perl and Python for Automations testing and test data
condition •
Used
scripts languages for ETL load with BW Middleware
Architecture employee with Kaiser (Walnut Creek) (2004 -2006) •
Designed
all middleware with 3000 interfaces for BigData •
IBM
MQ design/ architecture support •
Designed
QA environment and production support •
Designed
Message broker and production support |
|
|
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Languages: |
Languages |
Proficiency Level |
|
English |
Fluent |
|
|
|