From: route@monster.com
Sent: Monday, September 28, 2015 12:59 PM
To: hg@apeironinc.com
Subject: Please review this candidate for: Talend
This resume has been forwarded to
you at the request of Monster User xapeix03
|
|||||||
|
|||||||
|
|
|
||||||
|
||||||
Summary:
Over 12 years of experience in delivering turnkey enterprise solutions
in Big data, middleware and data warehousing arena. Hands on (coding)
professional with unique ability to identify and articulate business
problems, proposing technical solutions using a diverse set of
framework/products (e.g. Data sciences, Real time, NOSQL, Analytics, ETL,
ESBs, REST, Virtualization etc. frameworks). Expertise in latest big data (Hadoop) architectures for Hadoop
(Cloudera and Hortonworks distribution), Spark, Lambda architectures,
Middleware, virtualization, real time and batch analytics, complex event
processing, data warehousing, Infrastructure design and performance
benchmarking to satisfy peculiar and complex business problem statements.
Excellent analytical, communication and leadership skills to deliver work
using Big data teams. Currently
leading a team of Big data/Hadoop professionals to deliver multiple use-cases
at multiple clients. In-depth exposure to web, analytics, supply chain,
healthcare, retail and telecom business verticals. Technical Skills:
Big Data: Hbase, Spark, Spark SQL, Spark streaming, Impala,
sentry, M/R, hive, oozie, pig, flume, kafka, storm, Mahout, R, Netezza,
Teradata, Avro, Cassandra, NEO4J, MLbase, YARN, HDFS 2.0, Sqoop, parquet,
Giraph, oryx, cascading, crunch DW/BI:
Talend, Informatica 9, syncsort, pentaho, OBIEE, Tableau,
Microstrategy, D3 Java
Technologies: J2EE 1.8, XML, JMS, Webservices, Hibernate 3.0, Mybatis/Ibatis,
Spring 3.0, Ajax, Xpath, Dojo, Adobe Flex, JSCH, Quartz, JSON SOA Technologies: Axis 2.0, Servicemix
2.0, Mule, WS-BPEL, Camel, Apache CXF, Oracle Fusion – BPEL, BPM, BAM, AIA,
Oracle service bus Languages: Scala, Java, JSP, JSF, JavaScript, C/C++, Html, CSS, Cobol, Unix
Shell scripting Database: DB2, Oracle 11g RAC, IMS, Teradata, Netezza Designing Tool: Rational Rose XDE, UML, OOAD Web Tools: WSAD, Jbuilder, ant, maven, Jenkins, eclipse, Myeclipse, JBuilder App Servers: Web sphere, Weblogic 8.1/9.2/10.0, Tomcat, Apache Performance
profiling/monitoring tools: Ganglia,
Nagios, Introscope, HF run monitor, perfstats (oracle9i). Testing Tools: Win Runner, Test Director, Load runner, Wireshark, TCPMON. Configuration
Management tools: Visual Source Safe 6.0, PVCS, Dimensions,CVS,
SVN Operating
systems: Unix, Linux (Ubuntu, RHEL, centos etc),
Windows, Kerberos Professional Experience:
Working as
President – Big data and leading big data division. Delivering strategy, turn
key solutions to customers. Also incubating a training company engaged in big
data training (binarylore.com). A Large Healthcare company
Role: Hadoop /
Big data Enterprise architect - Spark Evangelize,
design and build an enterprise data warehouse on a big data platform for ETL
offload from Teradata, new analytical capabilities including data sciences
etc., multi-format ingest for different schemas, file types, databases using
Hadoop ecosystem and spark streaming etc. Responsibilities: 1.
Layout
security policy and architect a security solution for big data platform to
confirm to HIPAA requirements. 2.
Virtualization
strategy for Hadoop using Mesos for Multi-tenancy and multi workload solution
on same infrastructure. 3.
DR and
archive strategy for Hadoop. 4.
Build
and implement a framework Ingesting HL7, EDI messages in HBASE using spark
streaming 5.
Implement
HIVE and Pig UDF for data access from the HBASE repository 6.
Designing
and mentoring developers to code Map reduce jobs including coding some jobs. 7.
Layout
job dependencies using Oozie. 8.
Executive
level presentation for the intended big data strategy. 9.
Work on
8 data sciences use case for cost containment and efficiency discovery
strategy. Environment:
Spark, Spark-Streaming, Avro, Hbase, Accumulo, CDH, Mesos, HIVE,
Impala, Phoenix US Top 5 bank
Role: Hadoop /
Big data Enterprise architect Take the
existing Oracle big data appliance and develop data strategies including data
lake, BI on Hadoop, Mainframe integration with Hadoop etc. Assisting the bank
in laying out the overall enterprise strategy for Hadoop at the bank. Responsibilities: 1.
Implement
a mainframe to Hadoop data migration strategy. 2.
Build
out the infrastructure and design for a Big data platform. 3.
Implement
ETL processes using Storm, Kafka and Syncsort to ingest large volumes data
from various source System. 4.
Load
NEO4J with customer data and relationship of the products to customer
profile. 5.
Assist
the data sciences team by identifying production (Mahout, Giraph, oryx, SAS,
R, MLBASE etc) for purpose of building predictive data models. 6.
Layout
job dependencies using Oozie on hadoop ecosystem. 7.
Implement
a security policy using various tools and frameworks such as Kerberos,
Sentry, Dataguise, Gazzang etc. 8.
Proof
of concept of putting tableau on top of Hadoop by leveraging ODBC connectors. 9.
Performance
tuning of cluster and putting recommendation documents for other developers
to follow for delivery. 10.
Implement
a CI/CD pipeline using Maven, Jenkins, automated testing etc. features. Environment:
Oracle big data appliance and many different BI/ETL toolsets. Leading online vehicle
advertiser
Role: Big data
strategist and Enterprise architect (Hadoop) Evangelized
web log processing use case for real time analytics, event processing,
machine learning and batch analytics on 24 months of log data (500gigs of raw
data). POCed multiple BI tools. Responsibilities: 1.
Implement
a Netezza replacement strategy with a big data platform using HIVE, HBASE,
Flume and other ETL products. 2.
Build
out the infrastructure and design for a Big data platform. 3.
Executive
Dashboarding using Platfora, Datameer, tableau etc. as toolsets. 4.
Build
data sciences models using Mahout and other open sources libraries. 5.
Implement
a CI/CD pipeline using Maven, Jenkins, automated testing etc. features. 6.
Implement
real-time streaming engine to serve multiple purposes for data ingestion. Environment: Hadoop, HBASE, Flume-ng, Hive, Map/Reduce, Datameer, platfora,
tableau etc. A large group purchasing
Organization
Role: Big data
Enterprise architect (Hadoop, Spark) Purchasing
analytics for multiple hospital chains to find price advantage, MDM, data standardization
across all procurement activity at hospitals. Chalking out a
big data strategy for Healthtrust for data sciences, logical data warehouse,
real-time use cases etc. Responsibilities: 1.
Lay
down architecture, roadmap for the company on introduction of a big data
strategy. 2.
Implement
a big data platform using HIVE, HBASE, Flume and other ETL products. 3.
Build
out the infrastructure and design for a Big data platform. 4.
Executive
Dash-boarding using Microstrategy etc. as toolsets. 5.
Build
data sciences models using Graphlab and other open sources libraries. 6.
Implement
NEO4J for loading data into the cluster. 7.
Implement
a CI/CD pipeline using Maven, Jenkins, automated testing etc. features. 8.
Introduce
spark, spark streaming and other engines for purpose of ingesting large
amounts of data and applying predictive models while the data is being
ingested. 9.
Implementing
a data lake pattern for ingestion of data from multiple sources. 10.
Data
virtualization strategy using APIgee, open source products etc. 11.
Introduction
of NOSQL platforms into the customer’s application portfolio. Environment: Hadoop, Hive, Spark, Shark, Spark streaming, HBASE, Flume-ng,
hibernate 3.0, Spring-data, NEO4J, Cassandra Insight global
Role: Big data
Enterprise architect (Solr, Lucene, Nutch) Evangelize,
design and build an enterprise data warehouse on a big data platform for
intelligent sourcing, recruiter performance management, company financial
analytics etc. Resume
management system for storing large amounts of data on a big data platform
for various candidate profiles including – resumes, emails etc. Responsibilities: 1.
Implement
a big data platform using HBASE, Solr and Solr cloud for full indexing
support. 2.
Build
out the infrastructure and design for a Big data platform. 3.
Executive
Dash-boarding using Jaspersoft etc. as toolsets. 4.
Proposing
the next generation product set for saving 40 million dollars in annual spend
by not using competing products. Environment:
SOLR, Lucence, Hbase, Big data stack. AT&T
Role: Big data
Enterprise architect Description:
AT&T wireless is a communication company. Responsible for technical
solution for enterprise applications such as Device lifecycle, Order track,
DSC reflash & YODA etc. Provide architecture roadmap, proof of
concept(s), coding guidelines, introduction to new technologies etc for these
applications. Lend expertise to other groups for architecture, coding,
performance tuning exercises. While working
at AT&T involved in multiple POC’s and applications around Web platforms,
NOSQL technologies, CI/CD, Big data. Wrote some technical papers around
Cassandra. Responsibilities: DSC Reflash
(Device support Center) ·
Solution architect to design a
Mobile device reflashing system to wipe and re-install OS and programs on
user’s mobile devices. This system interfaces with equipment manufacturer
software to Reflash the devices. ·
Designed and implemented a
framework as proof of concept to invoke OEM (Samsung, Motorola, RIM, Apple)
provided application, which reflash mobile devices. ·
Over the web user interaction
capture while performing these updates. ·
Design solution for user activity
tracking and analytics of AT&T customers. ·
Design Ajax based user interfaces
for the online invocation using the java webstart. Environment: Weblogic 10.0, Oracle 11g, J2ee – Spring 3.0,
Java mail, DOJO, Spring 3.0, Struts, Hibernate, JSCH, Quartz, Java Web start
etc. DLC · Design a 2 billion dollar revenue generating system for
fraud detection, GO Phone sales, supply chain enabler system. This system is
30TB system with huge volume processing on a daily basis. · Interface with business executives for providing solutions
for future business problems, gathering current requirements. · Design and mentor a big data solution for making
information available to other applications at AT&T. · Design applications for performance, horizontal and
vertical scalability. · Design and implement a highly available and scalable
system as an operational warehouse being fed from various systems on a big
data platform. · Application performance profiling and tuning using
advanced tools. · Design complex ETL components (mappings, workflows etc) to
pull source data into this Hadoop platform. · Design scripts to push the synthesized facts into Teradata
warehouse. · Design hardware and application server layout with
implemention of clustering, failover, availabity and performance criterias (8
million dollar investment). Environment: Cassandra, Oracle 9i/10g RAC, Informatica,
Teradata, Hadoop Ordertrack ·
Design a real time order tracking
system for reporting on all orders submitted in the ERP systems. ·
Solution owner and architect for
order track with a team of 9 people. ·
Provide overall coding guidelines,
performance directives to the development team for a highly scalable
application. · Design a stateless application based on rich interface
architecture based on struts/Ajax and hibernate, spring/Ejb2.1 using latest
design patterns. · Database modeling for better ETL architecture and
reporting needs. · ETL architecture using informatica for fast batch systems
with low performance loads. · Interface with internal teams for problem resolution and
requirement capture for the order track system. · Assist PMO team to effectively manage the development for
the product. · POC on HBASE for quicker data ingestion. Environment: Weblogic 9.2/10.0, Oracle 10g, J2ee – Spring 2.5,
EJB 2.1, hibernate 3.0, Ibatis, Struts, Ajax, Ajax taglibs, RIA, informatica
power center etc. YODA ·
Solution architect for a proof of
concept for an order management system acting as an intelligent middleware
for the AT&T mobility supply chain system. ·
Proof of concept owner, designer
and implementer for YODA using latest open ESB architecture (servicemix),
spring, ws-bpel, activemq, latest integration pattern such as wiretap,
hibernate and oracle 10g RAC database. ·
Act as a hardware architect for
laying out the server architecture, performance tuning for application,
infrastructure and performance profiling using different tools. ·
Design and develop code snippets
for developers in order to orchestrate complicated business process flows
using the middleware. ·
Design Ajax based user interfaces
for the online ERP business users. ·
Design reporting and archiving
strategies for YODA. Environment: Weblogic 9.2/10.0, Oracle 10g RAC, Teradata, J2ee
– Spring 2.5, Servicemix 3.0, Apache camel, Ajax taglibs, Activemq,
integration patterns, Hibernate 3.0, Webservices, JMS, email processing,
Teradata, java 1.5 etc Target Corporation
Role: Java
programmer and team lead. Description:
Target is Fortune 10 company in the retail domain. Responsible for working in
two projects namely Target Integration Support, ITA Development/
Enhancements. Responsibilities: ITA · Design, Develop and test the ITA Application. The project
is using j2ee Framework with MVC, Business Delegate, Session Façade and DAO
design patterns. · Code presentation layer using the Struts, business Layer
using EJB. · Write stored procedures and SQL’s, Triggers, Views for the
DB2 and oracle. · Write Testing Automation scripts using Win Runner. Target
Integration Support · Mainframe and Web environment to be maintained for all
projects coming into the integration. Integration environment is setup to
simulate production region like activities in order to ensure lower Defect
rates in Production. · Developing Methodologies for Quantifying Defect
occurrences across application releases. Come up with vital statistics about
performance and implementation. ·
Evaluation and Trouble shooting
applications on Web sphere and ATG Dynamo servers for design, coding
or environmental defects. Specifically Application using Design
patterns, Struts, Jsp, application servers came under my purview. ·
DB2, MQSeries Administration for
the integration environment. ·
Data loading tasks for Oracle and
DB2 using platinum, SQL Loader. ·
Automation of various processes
using Win Runner Scripting for testers. ·
Designing and coding small
applications using struts, jsp for automation of some tasks. ·
Unix shell scripting and general
tasks on the unix boxes. Environment:
DB2, ATG DYNAMO, Web sphere, Win Runner,
Java/J2EE, EJB, JSP, Html, Java Script, Struts, MQSeries, PVCS Dimensions,
Unix shell scripting, CICS, JCL, Cobol, VSAM, CA-7,Oracle FedEx
Future state compensation
(FSC)
· Design and code applications for Management of Sales
professional Network of Fed Ex using RUP (rational unified process) SDLC
process. · Participate in Use Case modeling for the requirements.
Participate in Design process for the applications. · Code JSP’S, EJB’S implementing various design patterns
such as Factory pattern, Value Object, Business delegate, DAO, MVC, Singleton
etc. Deploy code on Web sphere 5.0. · Requirement Trace ability and implementing CMMI level
processes for the projects. Come up with metrics based on above. · Test plan and test case preparation for the Developers. Environment:
Web Sphere 5.0, Java/J2EE, EJB, JSP, Servlets,
JavaScript, Html, CSS, Struts, Jbuilder, toad, Visual source safe Daimler Chrysler
Rite Source Project (US
Imports)
· Writing applications for the US Imports system, a system
for managing the process of importing parts and vehicles into the US. · Use case modeling from the requirement document for
developing new systems like Ocean shipment. Converting the use cases into a
high level design using Design patterns like MVC, Business delegate, DAO,
Session Façade and then classes. · Code EJB’s, JSP’s using Struts in WSAD, Dream Weaver.
JavaScript, CSS and html scripting was also done in order to develop the
application. · Deployment tasks for Websphere application server.
Customize the resource pooling parameters etc. · Writing Complex SQL for DB2 to check data integrity. · Coding programs in CICS, JCL, DB2. · Developing Test Plans, Test Cases and do unit, system and
load Testing. · Data loading tasks from IMS to DB2, DB2 mainframe to db2
UDB (creation of test bed). Environment:
Ca-7, DB2 UDB, CICS, JCl, IMS, Web sphere 5.0,
Java/J2EE, EJB, Struts, MQSeries, DB2 Stored procedures. Vertex software
Remote server configuration
framework over the network/ Internet
· Do feasibility study for a project aimed at providing a
robust and secure tool for the network Administrator. This tool allows the
network administrator to manage his servers independent of location. Proof of
concept was presented to the Company management. · Team lead for a team of six people. · Propose design based on scalability, modularity, security
and performance. · Work load planning and delegation for the team. · Code using JSP, JavaScript, Servlets, RMI Environment:
Window 2000,
Linux, java, SSL, JNDI, LDAP, struts, APACHE, TOMCAT, JSP, Servlets, Struts Education:
Bachelor of
Engineering in Computer Science
|
|
|
||||||||
|
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Languages: |
Languages |
Proficiency Level |
|
English |
Fluent |
|
|
|