From:                              route@monster.com

Sent:                               Monday, September 28, 2015 1:01 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Talend

 

This resume has been forwarded to you at the request of Monster User xapeix03

Sherry Ba 

Last updated:  09/21/15

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received


Pleasanton, CA  94588
US

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: SherryBa2015

Resume Value: tsukjwjee5egyzzg   

  

 

 

First Name

Sherry

Last Name

Ba

Current Location

San Francisco CA/Bay Area

E-mail Id

sherryxba@hotmail.com

Contact Number

510-224-4388

Skype Id

hadoopsherry

Skills

Sr. Hadoop Architect/ Developer/Data Science

Total US IT Experience

8+

Relevant US Big-data Experience

4

Name of the Degree / Course /Certification

Bachelors in Information Systems

University / Institute

University of San Francisco

Education and Start Date and End Date

2004

Visa Status and Validity

US Citizen

Availability to start

1 week

Available time slots for telephonic interview  

Any time 9 -5 pm PST

Gender

Female

Salary

Open

To be considered for: Hadoop Big Data Architecture/development/Data data Science positions

 

Experienced IT professional seeking a leadership position driving strategy, management, planning, development, team building and end-to-end solution with focus on leveraging information technology and data science to maximize business values.

Developer / Engineer / Architect with over 8  years of development, administration and programming experience with  primary focus in "Big Data" and emphasis on Hadoop

Summary Hadoop/BigData/Analytic Experience

           Spark RDD and Spark DF development

           Worked with Hadoop, HBase, Cascading, Zookeeper, Oozie, Hive, HiveQL, MapR, MogoDB, Pentaho & Pig

           Worked with AWS Cloud, EC2, EMR, RedShift, S3 etc

           Supported Hadoop and Verita as administrator about 7 to 700 nodes

           Supported and administrated users accounts and security rules with  BigData domain

           Worked with Application Servers, Tomcat, Oracle and MySQL

           Experienced with distributed systems, large scale non-relational data stores, map-reduce systems, data modeling, database performance, and multi-terabyte data warehouses

            Experienced in a SaaS environment that has an agile development process

            Experienced in Java, Python or other object oriented programming language

            Experienced in Mobile, Linux, Unix, Android, MAC platform

            Experienced with server management, server operating systems(Windows, Linux, Unix), and VMware

            Experienced in testing APIs, Restful APIs

            Experienced with full SW Lifecycle and Agile practice

            Experienced with scripting language, in Perl, Python, Ruby, QTP for automated and performance tests

            Experienced with testing DNA sequence machines SW for regulatory approval

            Experienced testing networking and storage technologies, protocols and hardware

            Experienced with web services and SOAP UI testing

            Experienced with Oracle, TOAD, and SQL

            Experienced in testing UI technologies such as HTML5

            Experienced in Big Data/ Hadoop / Cassandra/Memcached/NoSQL/MapReduce

            Lead and coached about 2 to 20 engineers onshore, offshore and nearshore

            Lead the hiring process

Skills:

           Big/data and analytics platform administrator (Vertica, Greenplum, Hadoop, Hive)

           Experienced with Display Advertising, Behavioral Ad Networks

           Recommended Engines, Personalization

           Experienced with Data Analytics, Data mining, Predictive Modeling

           QA automation and testing

Work Experience

Sr. Big data Architecture/development for FRB (SF) – April – Sept 2015

-Spark RDD and Spark DF development

-Docker Master and install/deployment

-NoSQL database design and Spark SQL development

-Unix/Hadoop administrator and development
- AWS EMR / Redshift administrator/development
- Cloudera Administrator/development
- Cloudera Navigator Administrator/development
- Kerbero Administrator
- Zoomdata administrator
- Big data platform design and architecture/development
- Hive and Hbase development
- Security office for big data
- POC all other big data tools
- Work with scripts for Oozie and Hue systems

-Cassandra, MongoDB, Hbase, Hive development

-Cloudera Hue and Solr development

-Data governance policy/management and big data security

-Big data security and Kerberos development

-Multi-tenant set up and development

Sr. DataBase Architecture for GAP project (SF) -  Jan - March 2015

- Coverting large SAS SQL scripts to HiveQL for Data Science

- Coverting Oracle scripts to HiveQL programming for Data Science

- Architecture No-SQL databases and coding the MapReduce functions for over 10 years data about emails campaigning results analytics

           Working with data science to prototype data model and programming the business logic with Hive, Pig

           Building database applications on Hive, Pig, Hbase, MongoDB, Canssdra,Spark, Solr and shark on a large hadoop nodes cluster in house

           Backend No-SQL database design, development, architecture and testing

           - Data gaverment

 

Backend Hadoop Engineer for SS Network (Redwood City) project Oct Dec 20014

           Spark RDD and Spark SQL/Hive SQL

           No-SQL database modeling

           AWS cloud computing and architecture with Hadoop on Big Data

           Build cloud application on AWS, S3, EMR, Hive, Pig, Hbase, MongoDB, Canssdra,Spark

           Backend No-SQL database design, development, architecture and testing

Big/Data and BI/BW consulting for Cisco project (San Jose) March Aug 2014

           Create and maintain scripts and programs to Extract, Transform and Load data

           Automate ETL and aggregation processes using BASH/PHP and/or other scripting languages

           Create logging and monitoring elements to all Data Warehouse and ETL processes to allow Operations team to monitor

           Respond to Data Warehouse and ETL process alerts, with support from operations team.

           Create, maintain and automate processes to distribute data warehouse extracts to various users

           Continually monitor measure and improve all data warehouse and ETL processes for speed, reliability and accuracy.

           Perform analyses and execute improvements regarding overall data quality, including making recommendations to the business regarding inputs.

           Source and/or create tools to deliver monitoring metrics and dashboard reports to allow the rest of the company to understand the quality and timeliness of data warehouse data.

           Install and integrate 3rd party data maintenance tools such as address cleaning software, etc

           Use new technologies to service data warehousing needs such as hadoop, column databases, etc.

           - Cisco big data analytics platform development, testing and deployment
- Working on data integration application design, development, and testing
- Using Pentaho data integration tool to create data integration jobs 
- Using MapR to run Hadoop / YARN jobs
- Moving enterprise data from Oracle to Hive for big data platform
- Moving WebEx, voice, phone, video and email enterprise data to Hive and build data mart 
- Writing use story, task, plan and testing cases
- Writing scripts for automating testing
- Writing design document for enterprise data on board
- Analysis deployment log for errors
- Monitoring big data staging environment
- Monitoring hive staging environment

 

 

Big/Data and BI/ consulting for Intel projects(Santa Clara) July - Dec 2013

 

     Spark RDD development and No-SQL database modeling

     Design/Arcteticture Big Data Hadoop Testing Frameworks

     Building Hadoop cluster in Cloud and local

     Working with AWS cloud environment: EMR, Redshift, S3

     Working with Scala, Java, Python, Kafka, akka, SQL, MongoDB, JSON, Avro, Tableau

     Working with Git, SBT, Ant, Maven, Ganglia, Jenkins,

     Working with Hadoop, Hbase, Zookeeper, Oozie, Scalding Spark Shark

     Testing automation and coding in Scala

     Worked with ETL and reporting tools (OBIEE, SAP etc.)

 

Big/Data and BI consulting for Cisco projects (Milpitas) ca Sep 2012- Jun 2013

 

           Hadoop /MapR BW/BI data warehouse project.

           Big data move from Informatica and TD.

           Load data from different source to new build Hadoop analyst platform

           Build Hadoop QA team

           Using Hadoop /Hive/Sqoop/bash to deployment the data load and query data.

           Monitor QA environment for Hadoop problems

           Testing report form MapR report tool

           Provide the ideas/working process to other team

           Review other reports and code to understand coding logic

           Managed a technical team or functioning as a team lead.

           Worked with Hadoop stack (e.g. MapReduce, Sqoop, Pig, Hive, Hbase, Flume).

           related/complementary open source software platforms and languages (e.g. Java, Linux, Apache,

           Perl/Python/PHP, Chef).

           Worked with ETL (Extract-Transform-Load) tools (e.gInformatica, Talend, Pentaho).

           Worked with BI tools and reporting software (e.g. Microstrategy, Cognos,OBIEE, Pentaho).

           Worked with analytical tools, languages, or libraries (e.g. SAS, SPSS, R, Mahout).

           Supported business development activities to shape and communicate proposed solutions to client executives - Implemented of ETL applications:

           Implemented of reporting applications

           Application/implementation of custom analytics support

           Administrated of relational databases

           Data migration from existing data stores

           Infrastructure and storage design

           Developed capacity plans for new and existing systems

 

Big/Data and BI consulting for multiple projects (SF) Mar 2009 Jul 2012 through Accenture

1.       ATT SAP BW/BI/BigData development support

a.         Conducted detailed design  applications developed on Hadoop platforms (Feature testing, Regression Testing, Acceptance Testing and Sanity Testing)

b.        Implemented Business analysis tools with HadoopMapReduce scripts from ETL data to data warehouse for BI and enterprise analysis platform

c.         Administrated  Log analysis scripts for Business Analyst tool with HDFS Hadoop  (file system level)

d.        Advised the Hadoop and analytica work loads IO optimizing solutions

e.        Running  benchmarked Hadoop/HBase clusters

f.           Supported development of Hadoop and Vertica Analytics Platform activities

 

2.       Storage architecture/QA Netapp Hadoop ecosystem administrator and support

 

a.       Tested NetApp OpenSolution for Hadoop(NOSH) settings

b.       Supported and scaled Hadoop systems

c.       Supported cluster failed over test and document the results with various configurations

d.       Administrated enterprise-grade storage arrays and eliminated Hadoop network bottlenecks

e.       Supported hot-pluggable disk shelves, added storage and administrated services

f.         Supported NFS and HDFS file systems

g.       Loaded network-free hardware RAID

h.       Day-to-day support Hadoop hardware and software issues

i.          Used automation scripts to run performance testing of the new storage OS design

 

5.       Consulting for Bank of America TCOE (Concord) Oct 2006 Feb 2009

Supported about over 100 projects and many functional teams

           Supported analytic massive amounts of data and BI

           Communicated and tracked defects to closure

           Administrated with large data set testing with BigData NoSQL Hadoop and Oracle, DB2 scripts

           Wrote and reviewed scripts and executed scripts, tracked defects

           Used scripts languages in Perl and Python for Automations testing and test data condition 

           Used scripts languages for ETL load with BW

 

Middleware Architecture employee with Kaiser (Walnut Creek) (2004 -2006)

 

           Designed all middleware with 3000 interfaces for BigData

           IBM MQ design/ architecture support

           Designed QA environment and production support

           Designed Message broker and production support

 

 

 



Additional Info

BACK TO TOP

 

Current Career Level:

Manager (Manager/Supervisor of Staff)

Date of Availability:

Immediately

Active Security Clearance:

None

US Military Service:

Citizenship:

None

 

 

Target Company:

Company Size:

Occupation:

IT/Software Development

·         IT Project Management

·         General/Other: IT/Software Development

·         Software/Web Development

 

Target Locations:

Selected Locations:

US-CA-Oakland/East Bay

Relocate:

No

Willingness to travel:

Up to 50% travel

 

Languages:

Languages

Proficiency Level

English

Fluent