From:                              route@monster.com

Sent:                               Monday, September 28, 2015 12:59 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Talend

 

This resume has been forwarded to you at the request of Monster User xapeix03

Preetpal Singh 

Last updated:  07/13/14

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received


Alpharetta, GA  30004
US

Work: 678-807-4456   
pphazrah@yahoo.co.in
http://binarylore.com
Contact Preference:  Email

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: Big Data / Hadoop Enterprise Architect

Resume Value: tggzmmk7jt4khhae   

  

 

Big Data Enterprise Architect

Cell: 678-807-4456

Email: pphazrah@yahoo.co.in

 

Summary:

Over 12 years of experience in delivering turnkey enterprise solutions in Big data, middleware and data warehousing arena. Hands on (coding) professional with unique ability to identify and articulate business problems, proposing technical solutions using a diverse set of framework/products (e.g. Data sciences, Real time, NOSQL, Analytics, ETL, ESBs, REST, Virtualization etc. frameworks).

 

Expertise in latest big data (Hadoop) architectures for Hadoop (Cloudera and Hortonworks distribution), Spark, Lambda architectures, Middleware, virtualization, real time and batch analytics, complex event processing, data warehousing, Infrastructure design and performance benchmarking to satisfy peculiar and complex business problem statements. Excellent analytical, communication and leadership skills to deliver work using Big data teams.

 

Currently leading a team of Big data/Hadoop professionals to deliver multiple use-cases at multiple clients. In-depth exposure to web, analytics, supply chain, healthcare, retail and telecom business verticals.

 

Technical Skills:

 

Big Data: Hbase, Spark, Spark SQL, Spark streaming, Impala, sentry, M/R, hive, oozie, pig, flume, kafka, storm, Mahout, R, Netezza, Teradata, Avro, Cassandra, NEO4J, MLbase, YARN, HDFS 2.0, Sqoop, parquet, Giraph, oryx, cascading, crunch

DW/BI: Talend, Informatica 9, syncsort, pentaho, OBIEE, Tableau, Microstrategy, D3

Java Technologies: J2EE 1.8, XML, JMS, Webservices, Hibernate 3.0, Mybatis/Ibatis, Spring 3.0, Ajax, Xpath, Dojo, Adobe Flex, JSCH, Quartz, JSON

 

SOA Technologies: Axis 2.0, Servicemix 2.0, Mule, WS-BPEL, Camel, Apache CXF, Oracle Fusion – BPEL, BPM, BAM, AIA, Oracle service bus

Languages: Scala, Java, JSP, JSF, JavaScript, C/C++, Html, CSS, Cobol, Unix Shell scripting

Database: DB2, Oracle 11g RAC, IMS, Teradata, Netezza

Designing Tool: Rational Rose XDE, UML, OOAD

Web Tools: WSAD, Jbuilder, ant, maven, Jenkins, eclipse, Myeclipse, JBuilder

App Servers: Web sphere, Weblogic 8.1/9.2/10.0, Tomcat, Apache

Performance profiling/monitoring tools: Ganglia, Nagios, Introscope, HF run monitor, perfstats (oracle9i).

Testing Tools: Win Runner, Test Director, Load runner, Wireshark, TCPMON.

Configuration Management tools: Visual Source Safe 6.0, PVCS, Dimensions,CVS, SVN

Operating systems: Unix, Linux (Ubuntu, RHEL, centos etc), Windows, Kerberos

 


Professional Experience:

 

Working as President – Big data and leading big data division. Delivering strategy, turn key solutions to customers. Also incubating a training company engaged in big data training (binarylore.com).

 

A Large Healthcare company

Role: Hadoop / Big data Enterprise architect - Spark

 

Evangelize, design and build an enterprise data warehouse on a big data platform for ETL offload from Teradata, new analytical capabilities including data sciences etc., multi-format ingest for different schemas, file types, databases using Hadoop ecosystem and spark streaming etc.

 

Responsibilities:

1.              Layout security policy and architect a security solution for big data platform to confirm to HIPAA requirements.

2.              Virtualization strategy for Hadoop using Mesos for Multi-tenancy and multi workload solution on same infrastructure.

3.              DR and archive strategy for Hadoop.

4.              Build and implement a framework Ingesting HL7, EDI messages in HBASE using spark streaming

5.              Implement HIVE and Pig UDF for data access from the HBASE repository

6.              Designing and mentoring developers to code Map reduce jobs including coding some jobs.

7.              Layout job dependencies using Oozie.

8.              Executive level presentation for the intended big data strategy.

9.              Work on 8 data sciences use case for cost containment and efficiency discovery strategy.

 

Environment: Spark, Spark-Streaming, Avro, Hbase, Accumulo, CDH, Mesos, HIVE, Impala, Phoenix

 

US Top 5 bank

Role: Hadoop / Big data Enterprise architect

 

Take the existing Oracle big data appliance and develop data strategies including data lake, BI on Hadoop, Mainframe integration with Hadoop etc. Assisting the bank in laying out the overall enterprise strategy for Hadoop at the bank.

 

Responsibilities:

1.              Implement a mainframe to Hadoop data migration strategy.

2.              Build out the infrastructure and design for a Big data platform.

3.              Implement ETL processes using Storm, Kafka and Syncsort to ingest large volumes data from various source System.

4.              Load NEO4J with customer data and relationship of the products to customer profile.

5.              Assist the data sciences team by identifying production (Mahout, Giraph, oryx, SAS, R, MLBASE etc) for purpose of building predictive data models.

6.              Layout job dependencies using Oozie on hadoop ecosystem.

7.              Implement a security policy using various tools and frameworks such as Kerberos, Sentry, Dataguise, Gazzang etc.

8.              Proof of concept of putting tableau on top of Hadoop by leveraging ODBC connectors.

9.              Performance tuning of cluster and putting recommendation documents for other developers to follow for delivery.

10.          Implement a CI/CD pipeline using Maven, Jenkins, automated testing etc. features.

 

 

Environment: Oracle big data appliance and many different BI/ETL toolsets.

 

Leading online vehicle advertiser

Role: Big data strategist and Enterprise architect (Hadoop)

 

Evangelized web log processing use case for real time analytics, event processing, machine learning and batch analytics on 24 months of log data (500gigs of raw data). POCed multiple BI tools.

 

Responsibilities:

1.              Implement a Netezza replacement strategy with a big data platform using HIVE, HBASE, Flume and other ETL products.

2.              Build out the infrastructure and design for a Big data platform.

3.              Executive Dashboarding using Platfora, Datameer, tableau etc. as toolsets.

4.              Build data sciences models using Mahout and other open sources libraries.

5.              Implement a CI/CD pipeline using Maven, Jenkins, automated testing etc. features.

6.              Implement real-time streaming engine to serve multiple purposes for data ingestion.

 

 

Environment: Hadoop, HBASE, Flume-ng, Hive, Map/Reduce, Datameer, platfora, tableau etc.

 

A large group purchasing Organization

Role: Big data Enterprise architect (Hadoop, Spark)

 

Purchasing analytics for multiple hospital chains to find price advantage, MDM, data standardization across all procurement activity at hospitals.

Chalking out a big data strategy for Healthtrust for data sciences, logical data warehouse, real-time use cases etc.

 

Responsibilities:

1.              Lay down architecture, roadmap for the company on introduction of a big data strategy.

2.              Implement a big data platform using HIVE, HBASE, Flume and other ETL products.

3.              Build out the infrastructure and design for a Big data platform.

4.              Executive Dash-boarding using Microstrategy etc. as toolsets.

5.              Build data sciences models using Graphlab and other open sources libraries.

6.              Implement NEO4J for loading data into the cluster.

7.              Implement a CI/CD pipeline using Maven, Jenkins, automated testing etc. features.

8.              Introduce spark, spark streaming and other engines for purpose of ingesting large amounts of data and applying predictive models while the data is being ingested.

9.              Implementing a data lake pattern for ingestion of data from multiple sources.

10.          Data virtualization strategy using APIgee, open source products etc.

11.          Introduction of NOSQL platforms into the customer’s application portfolio.

 

 

Environment: Hadoop, Hive, Spark, Shark, Spark streaming, HBASE, Flume-ng, hibernate 3.0, Spring-data, NEO4J, Cassandra

 


 

Insight global

Role: Big data Enterprise architect (Solr, Lucene, Nutch)

 

Evangelize, design and build an enterprise data warehouse on a big data platform for intelligent sourcing, recruiter performance management, company financial analytics etc.

 

Resume management system for storing large amounts of data on a big data platform for various candidate profiles including – resumes, emails etc.

 

Responsibilities:

1.              Implement a big data platform using HBASE, Solr and Solr cloud for full indexing support.

2.              Build out the infrastructure and design for a Big data platform.

3.              Executive Dash-boarding using Jaspersoft etc. as toolsets.

4.              Proposing the next generation product set for saving 40 million dollars in annual spend by not using competing products.

 

Environment: SOLR, Lucence, Hbase, Big data stack.

 

AT&T                            

Role: Big data Enterprise architect

 

Description: AT&T wireless is a communication company. Responsible for technical solution for enterprise applications such as Device lifecycle, Order track, DSC reflash & YODA etc. Provide architecture roadmap, proof of concept(s), coding guidelines, introduction to new technologies etc for these applications. Lend expertise to other groups for architecture, coding, performance tuning exercises.

 

While working at AT&T involved in multiple POC’s and applications around Web platforms, NOSQL technologies, CI/CD, Big data. Wrote some technical papers around Cassandra.

 

Responsibilities:

DSC Reflash (Device support Center)

·   Solution architect to design a Mobile device reflashing system to wipe and re-install OS and programs on user’s mobile devices. This system interfaces with equipment manufacturer software to Reflash the devices.

·   Designed and implemented a framework as proof of concept to invoke OEM (Samsung, Motorola, RIM, Apple) provided application, which reflash mobile devices.

·   Over the web user interaction capture while performing these updates.

·   Design solution for user activity tracking and analytics of AT&T customers.

·   Design Ajax based user interfaces for the online invocation using the java webstart.

 

Environment: Weblogic 10.0, Oracle 11g, J2ee – Spring 3.0, Java mail, DOJO, Spring 3.0, Struts, Hibernate, JSCH, Quartz, Java Web start etc.

 

DLC

·   Design a 2 billion dollar revenue generating system for fraud detection, GO Phone sales, supply chain enabler system. This system is 30TB system with huge volume processing on a daily basis.

·   Interface with business executives for providing solutions for future business problems, gathering current requirements.

·   Design and mentor a big data solution for making information available to other applications at AT&T.

·   Design applications for performance, horizontal and vertical scalability.

·   Design and implement a highly available and scalable system as an operational warehouse being fed from various systems on a big data platform.

·   Application performance profiling and tuning using advanced tools.

·   Design complex ETL components (mappings, workflows etc) to pull source data into this Hadoop platform.

·   Design scripts to push the synthesized facts into Teradata warehouse.

·   Design hardware and application server layout with implemention of clustering, failover, availabity and performance criterias (8 million dollar investment).

Environment: Cassandra, Oracle 9i/10g RAC, Informatica, Teradata, Hadoop

 

Ordertrack

·   Design a real time order tracking system for reporting on all orders submitted in the ERP systems.

·   Solution owner and architect for order track with a team of 9 people.

·   Provide overall coding guidelines, performance directives to the development team for a highly scalable application.

·  Design a stateless application based on rich interface architecture based on struts/Ajax and hibernate, spring/Ejb2.1 using latest design patterns.

·  Database modeling for better ETL architecture and reporting needs.

·  ETL architecture using informatica for fast batch systems with low performance loads.

·  Interface with internal teams for problem resolution and requirement capture for the order track system.

·  Assist PMO team to effectively manage the development for the product.

·  POC on HBASE for quicker data ingestion.

 

Environment: Weblogic 9.2/10.0, Oracle 10g, J2ee – Spring 2.5, EJB 2.1, hibernate 3.0, Ibatis, Struts, Ajax, Ajax taglibs, RIA, informatica power center etc.

 

YODA

·   Solution architect for a proof of concept for an order management system acting as an intelligent middleware for the AT&T mobility supply chain system.

·   Proof of concept owner, designer and implementer for YODA using latest open ESB architecture (servicemix), spring, ws-bpel, activemq, latest integration pattern such as wiretap, hibernate and oracle 10g RAC database.

·   Act as a hardware architect for laying out the server architecture, performance tuning for application, infrastructure and performance profiling using different tools.

·   Design and develop code snippets for developers in order to orchestrate complicated business process flows using the middleware.

·   Design Ajax based user interfaces for the online ERP business users.

·   Design reporting and archiving strategies for YODA.

 

Environment: Weblogic 9.2/10.0, Oracle 10g RAC, Teradata, J2ee – Spring 2.5, Servicemix 3.0, Apache camel, Ajax taglibs, Activemq, integration patterns, Hibernate 3.0, Webservices, JMS, email processing, Teradata, java 1.5 etc

 

 

Target Corporation       

Role: Java programmer and team lead.

 

Description: Target is Fortune 10 company in the retail domain. Responsible for working in two projects namely Target Integration Support, ITA Development/ Enhancements.

 

Responsibilities:

ITA

·   Design, Develop and test the ITA Application. The project is using j2ee Framework with MVC, Business Delegate, Session Façade and DAO design patterns.

·   Code presentation layer using the Struts, business Layer using EJB.

·   Write stored procedures and SQL’s, Triggers, Views for the DB2 and oracle.

·   Write Testing Automation scripts using Win Runner.

Target Integration Support

·   Mainframe and Web environment to be maintained for all projects coming into the integration. Integration environment is setup to simulate production region like activities in order to ensure lower Defect rates in Production.

·   Developing Methodologies for Quantifying Defect occurrences across application releases. Come up with vital statistics about performance and implementation.

·   Evaluation and Trouble shooting applications on Web sphere and ATG Dynamo servers for design, coding or environmental defects. Specifically Application using Design patterns, Struts, Jsp, application servers came under my purview.

·   DB2, MQSeries Administration for the integration environment.

·   Data loading tasks for Oracle and DB2 using platinum, SQL Loader.

·   Automation of various processes using Win Runner Scripting for testers.

·   Designing and coding small applications using struts, jsp for automation of some tasks.

·   Unix shell scripting and general tasks on the unix boxes.

Environment:

DB2, ATG DYNAMO, Web sphere, Win Runner, Java/J2EE, EJB, JSP, Html, Java Script, Struts, MQSeries, PVCS Dimensions, Unix shell scripting, CICS, JCL, Cobol, VSAM, CA-7,Oracle

 

 

FedEx

 

Future state compensation (FSC)                                                          

 

·   Design and code applications for Management of Sales professional Network of Fed Ex using RUP (rational unified process) SDLC process. 

·   Participate in Use Case modeling for the requirements. Participate in Design process for the applications.

·   Code JSP’S, EJB’S implementing various design patterns such as Factory pattern, Value Object, Business delegate, DAO, MVC, Singleton etc. Deploy code on Web sphere 5.0.

·   Requirement Trace ability and implementing CMMI level processes for the projects. Come up with metrics based on above.

·   Test plan and test case preparation for the Developers.

 

Environment:

Web Sphere 5.0, Java/J2EE, EJB, JSP, Servlets, JavaScript, Html, CSS, Struts, Jbuilder, toad, Visual source safe

 

 

Daimler Chrysler

Rite Source Project (US Imports)                                                         

 

·   Writing applications for the US Imports system, a system for managing the process of importing parts and vehicles into the US.

·   Use case modeling from the requirement document for developing new systems like Ocean shipment. Converting the use cases into a high level design using Design patterns like MVC, Business delegate, DAO, Session Façade and then classes.

·   Code EJB’s, JSP’s using Struts in WSAD, Dream Weaver. JavaScript, CSS and html scripting was also done in order to develop the application.

·   Deployment tasks for Websphere application server. Customize the resource pooling parameters etc.

·   Writing Complex SQL for DB2 to check data integrity.

·   Coding programs in CICS, JCL, DB2.

·   Developing Test Plans, Test Cases and do unit, system and load Testing.  

·   Data loading tasks from IMS to DB2, DB2 mainframe to db2 UDB (creation of test bed).

Environment:

Ca-7, DB2 UDB, CICS, JCl, IMS, Web sphere 5.0, Java/J2EE, EJB, Struts, MQSeries, DB2 Stored procedures.

 

 

Vertex software                                                                                                     

Remote server configuration framework over the network/ Internet

 

·   Do feasibility study for a project aimed at providing a robust and secure tool for the network Administrator. This tool allows the network administrator to manage his servers independent of location. Proof of concept was presented to the Company management.

·   Team lead for a team of six people.

·   Propose design based on scalability, modularity, security and performance.

·   Work load planning and delegation for the team.

·   Code using JSP, JavaScript, Servlets, RMI

 

Environment:

Window 2000, Linux, java, SSL, JNDI, LDAP, struts, APACHE, TOMCAT, JSP, Servlets, Struts

 

 

Education:

 

Bachelor of Engineering in Computer Science            



Experience

BACK TO TOP

 

Job Title

Company

Experience

Big Data / Hadoop Enterprise Architect

Hashmap inc

- Present

 

Additional Info

BACK TO TOP

 

Current Career Level:

Executive (SVP, VP, Department Head, etc)

Years of relevant work experience:

10+ to 15 Years

Date of Availability:

Within 2 weeks

Work Status:

US - I am authorized to work in this country for any employer.

Active Security Clearance:

None

US Military Service:

Citizenship:

None

 

 

Target Job:

Target Job Title:

Big Data / Hadoop Enterprise Architect

Desired Job Type:

Employee
Temporary/Contract/Project

 

Target Company:

Company Size:

Occupation:

IT/Software Development

·         Enterprise Software Implementation & Consulting

·         General/Other: IT/Software Development

·         Software/System Architecture

 

Target Locations:

Selected Locations:

US-GA-Atlanta North

Relocate:

Yes

Willingness to travel:

Up to 100%

 

Languages:

Languages

Proficiency Level

English

Fluent