Narender
Reddy Kanuganti.
Big Data Architect /IoT/ Cyber
Security Specialist (CompTIA-Certified)
naren.bigdatapro@gmail.com
470-564-4000
•BIG DATA Practice •HADOOP • BIG DATA
ARCHITECTURE• IoT-INTERNET OF THINGS
DATA
INTEGRATION • DATA SCIENCE• E-COMMERCE •
PREDICTIVE ANALYTICS• POC
PRE-SALES /
SALES DEVELOPMENT / CLIENT MEETINGS & PRESENTATIONS.
SUMMARY
·
Tech Savvy Techno-Functional Data
Wrangler / Analyst turned BIG DATA Architect with specializations in IoT
Technology Platform & Hadoop Eco-system with 14 + years of Cradle to
Grave Experience in IT Landscape ranging from Broad-spectrum of Industries-
like IT/Retail/Telecom/Healthcare/Insurance/Logistics performing Predictive
Data Analysis, Data Science Analytics, Requirements Elicitation, Business
Modeling, and Use Case development, Internet of Things (IoT) Inception (POC)
to Production.
·
IOT platform strategy,
architecture, and planning, M2M Platform Services: Connectivity Management,
Application Enablement, and Device Management offering world leading security
solutions for device protection, encryption, authentication, key management,
and code signing. HADOOP: Create a Blueprint and Roadmap for
Clients for their journeys into Big Data.
·
“Get things done” attitude with a
strong Boots-on-the-ground mentality to steer projects into
Right portfolios and create Win-Win Situation for
the Client and Employer.
·
Knowledge of Data Center
Architectures and Services incorporating Applications, Servers, Storage, and
Data Center Switching technologies and ability to manage a long and complex
sales cycle. Ability to leverage and converge Big Data with IoT Platform Technology.
·
TOGAF Certified Architect with 3
years of Hands-on Experience in Big Data Technologies like Hadoop and its
eco-system and still making my hands dirty and shaking hands with Top
Management folks /Stakeholders on front end and helping the fellow Developers
at the back end Cracking, Debugging and fixing the Codes.
·
Extensive experience in Software
Development Cycle from Inception to Launch , and strong passion for new
emerging Technologies like Hadoop and Internet of Things(IoT)
·
Ability to guide the lifecycle
of Hadoop and other Big Data solutions, including requirements analysis,
platform selection, Logical and Physical architectural designs, application
design and development, testing, construct the Big data pipelines and
seamlessly integrate the Legacy platforms/Apps into new emerging Technologies
and SME for IoT Platform strategy.
·
Strong understanding of what’s
“Under the Hood” of Big data World to offer Best practices / recommendations
to the clients and drive the Big Data /IoT Initiatives from Pilot
–to- Production”. Self-motivated
and sell audiences on IoT solutions by performing technical product
presentations and demos. Possess a balance of technical and
sales aptitude, including handling customer technical questions and
issues, creating product positioning and translating business needs into new
product features
·
Blessed with Techniques to seek
Sign-off from Key & tough Stakeholders through Diplomacy acting as
a Liaison and Bridging the Gap between Business World and Technical World and
keep all Stakeholders in Sync and involve myself in all phases of the Product
Life Cycle right from Inception to launching and ensure the project
deliverables within Time limits and Scope.
·
Effective Communicator With Front
End Stakeholders And Good Listener And A Team Player With Strong Technical
Documentation Skills And Analytical Bent Of Mind Possessing Strong
Willingness To Learn New Systems, Good coordinating skills ability to Ramp up
to the Speed and hit the Ground Running.
·
Team-Oriented & Self-starter
Professional working with Cross functional teams in a Techno functional
environment playing multiple Roles as Big Data Analyst/Admin
/Developer Role with strong understanding of Timelines, Budgets and
deliverables and add Value to the company.
·
Proactive to enter Developer’s
world and willing to wrangle with lines of Code and make my hands dirty
enough with hardware & Software and indulge in troubleshooting to resolve
technical issues and take ownership of the Product deliverable and drive the
project from end to end.
·
A good Deal puncher with customer
focused Negotiating abilities/solution finding skills wearing different Hats
at different Scenarios. Ability to Read the Code and do the critical Analysis
of Bugs and Defects and accomplish the End user’s/Business requirements.
·
Self-Starter and Self-motivated
Team leader, DETAIL ORIENTED and ability to think OUTSIDE THE BOX and thirst
for learning new systems and adapt with changing corporate environments and
increase the Bottom line of the Company. Strong in AGILE and WATERFALL
Methodologies and expert in Managing “Contracts Backlog” and prioritize
according to project deadlines
·
Experienced in Full Life cycle
implementations of Big Data solutions (HADOOP) & gathering Requirements
from Clients and perform customizations/ setting up configurations and
testing for Performance. Ability to de-mystify predictive analytics and
create actionable insights from the output of that algorithm to add business
value to the enterprise.
·
Strong understanding of Software
Development Life Cycle (SDLC), including good knowledge of RUP methodology.
Reproduce, Expertise in SOA, investigate and debug software issues. Knowledge
of both Agile and Waterfall Software development methodologies. ENTERPRISE
WEB DEVELOPMENT Experience in Accu Process, SaaS (Ajax), and Informatica
Cloud and Extensive knowledge of Data warehouse Concepts. HP (PPM) and
Business Objects (SAP).
·
Extensive experience in
interacting with Offshore Team/Virtual team & Stakeholders, Problem
Management Analysis, Eliciting Requirements and creating Business Requirement
Documents (BRD), User Requirement Specifications, Functional Requirement
Documents (FRD), System Requirement Specification (SRS), Test Plan, analyzing
and Creating Use Cases, Use Case Diagrams, PROCESS FLOW Diagrams, BPMN
Diagrams, Activity diagrams, System Workflow. Strong Understanding of all
Versions of SharePoint.
·
Experience in applying Rational
Unified Process (RUP) methodology using Modeling and requirement
documentation tools such as MS Office, MS VISIO, and MS Project.
·
Conducting and facilitating JAD
Sessions and communicating the Concepts with Key Stakeholders, Development
team, SMEs, System Analysts, Business Analysts and Project Managers and
External Vendors.
·
Experience in customizing the
portal sites on Share point 2007 and 2010, Rational Requisite Pro, Team
Foundation Server and Variety of e-Commerce applications, Cloud Computing and
Deployment of Apps. Strong understanding of SOA architecture designs and
concepts
·
“Cradle to grave” experience in
big data initiatives and strategic planning from concept to launching.
·
Big Data Practice / Big Data
Architecture / SME for IoT-Internet Of Things strategies.
·
Strong understanding of IoT
Platform like Axeda, Thingworx, Microsoft Azure. Working knowledge on IoT
Strategy, standards, protocols like MQTT, COAP. Exposed to Embedded device
programming Device protocols and wearable technology Exposed to Sensor
analytics
·
Working Knowledge and demonstrated
experience in IoT, Integration or Cloud integration Architectures
·
Data Integration & Warehousing
Data Science E-Commerce Web Analytics
COMMERCIAL Distributions /Platforms of HADOOP
worked.
§ CLOUDERA
AWS –Cloud Computing /
Deployments
§ HDP-
Hortonworks
Microsoft AZURE
§ MAPR
IoT – Internet of Things Platform Strategy Implementation
§ HD
Insight
(IBM)
LINUX & WINDOWS Environment
TECHNICAL SKILLS
· Design
Blue print and create a Road map for Big data journey incorporating Big Data
Concepts Cradle –to-Grave understanding of the Hadoop eco-system and
Reference Architecture, Hadoop-(HDFS), Write MapReduce Jobs & Algorithms
Using Various Tools Pig, Hive, Hbase, Impala, NoSQL, Cassandra, IBM Data
Explorer. Zookeeper, Sqoop, Flume /R Program/Pentaho / Vertica/Informatica
/Talend/ Teradata Aster.
· BIG
DATA TECHNOLOGY STACK: HDFS,YARN, HIVE,PIG, SPARK,STORM,
CASSANDRA, HBase
MONGO db, ‘R ”Machine Language, PENTAHO, TALEND, TABLEAU,
HYPERION.
· Hadoop
Cluster Administration, Configuration, Monitoring, Debugging, and Performance
Tuning. Ability To Implement Hadoop Based Solutions And Offer Best Practices
In Big Data Space.
· Hadoop
Eco-System: Setting Up Clusters/Nodes/Maintenance And Tuning Of The Cluster
Nodes End-To-End, Troubleshoot The Technical Issues And Offer Solutions.
· Develop
Solutions For (IoT) Internet Of Things And Provide End-To-End Support From
Concept To Production Phase Of The Product Life Cycle.
· Strong
Understanding Of Hadoop Eco-System And Other Business Intelligence And ETL
Tools On Top Of Hadoop Like Vertica, Pentaho, Sqoop, Oozie, Flume,
Hbase, Tableau, Teradata , Datastax Datameer And Mahout(Machine Learning) Web
Analytics (Omniture).Mpp (Massive Parallel Processing) In Teradata and Data
Modelling software ERwin 8.2.
· Ability
to write MapReduce Programs and create Business Intelligence reports from the
output.
· Daily
support of several Hadoop, data warehouse appliances, including monitoring
capacity, throughput, health, and usage and Clickstream Analysis out of the
web logs.
· Collaboration
with vendors and users to coordinate and accomplish repairs, upgrades,
patches, and other enhancements, additions, or replacements.
· Query
analysis and tuning advice for end users, to maintain throughput and reliable
operation.
· Scripting
to deploy monitors, checks, and other sys admin function automation.
· Production
Support for any problem leading to acceptable resolution, including daytime,
nighttime, and weekend support if required.
· Performs
Incident resolution, Problem Determination and Root Cause Analysis.
(i.e., hardware and software diagnostic tools to monitor performance and
perform problem determination). Familiarity with hardware and software
diagnostic tools to monitor performance and perform problem determination.
· Oversee
installations, monitoring and managing change to servers (Overall Change
Management for Servers).Oversees implementation of security guidelines in
order to prevent unauthorized access to servers and report any violations.
· Ability
to Monitor and tune operating systems to achieve optimum performance levels
in standalone and multi-tiered environments.
· Collaborate
with System Engineering, Network Engineering, solves complex and recurring
operational issues and develops corrective actions, as needed.
· Interact
regularly with Metrics team, developers, engineers, and the IT outsourcer to
ensure the Company’s Reliability, Availability and Serviceability (RAS)
metrics are sustained and improved from current level.
· Participate
in the evaluation, recommendation, and selection of hardware and software
solutions. - Reviews, evaluate, designs, implements and maintain internal and
external data.
· Identify
data sources, constructs data decomposition diagrams, provides data flow
diagrams and documents the process.
· Writes
codes for database access, modifications, and constructions including Map
Reduce Programs, Pig/Hive Scripts, SQL-H, Stored Procedures, etc.
· Developed
and reviewed project plans, identifies issues, resolves issues, and
communicates status of assigned projects to end-users and project
Stakeholders.
· Experience
in operational support and hands-on implementation Hadoop based Big Data
Platforms. Gathers requirements, builds logical models and provides
quality documentation of detailed user requirements for the design and
development of systems.
· Databases:
Hbase/NoSQL/SQL Server, Oracle/IMPALA /MongoDB, CASSANDRA.
· Cyber
Security Skills: (IAM,
Encryption, DLP, Risk and Compliance Management, IDS/IPS, UTM, Firewall,
Antivirus/Antimalware, SIEM, Disaster Recovery, DDOS Mitigation, Web
Filtering, and Security Services
BUSINESS SKILLS
· Technical
Pre-Sales, White Boarding and PowerPoint presentations to the Board members
and Lead the Team to provide thought process leadership and direction and
hold accountable for the project deliverables within the budgets and
timelines throughout the life cycle of the project and avoid the Scope creep
and add Clarity to the tasks and communicate the concepts and update the
status reports to the front-end stakeholders through white boarding and
PowerPoint point presentations. Speak both Technical and Non-Technical
languages and flip both sides of coin.
· People
Skills, Networking With People, Connect The Dots And Get Things Done.
Facilitate JAD Sessions, Power Point Presentations of New Products and
Services to wide spectrum of audience/ Clients and Business Users /
Stakeholders.
· Test
Cases, Vision, Scope, and SRS documents/UAT Sessions with Stakeholders.
· Gap
Analysis, Impact Analysis and SWOT Analysis/Feasibility Analysis, Product
Marketing &Sales.
· Focus
on End Game and Prioritize the Requirements and achieve goals within
Timelines and Budget.
· Business
Process Analysis & Research using i-Rise Software/Enterprise Architect.
· USE
CASE Modeling & Analysis, Troubleshooting software bugs and defects
management.
· Functional
Requirement Gathering & Technical Requirements Development and
Documentation.
· Prototyping
/ Wireframes & Mockup Screen Creations
BUSINESS DOCUMENTATION
· Business
Requirements / Functional and Non-Functional Documentation of Technical Specs
/ and sign off from the Stakeholders.
· Documentation
and Visualization using Reporting tools like /Tableau /Qlickview.
· Story
Boards, Scenarios, Personas. Test Plans, Test Scripts and Test Cases,
Prototypes.
· Setup
Workshops for Business sponsors / end users and Developers and educate and
make sure all the stakeholders are same page and understand the strategy and
concepts of the IT initiatives.
PROFESSIONAL
EXPERIENCE
BANK of AMERICA.
H.Q Charlotte,
NC.
05/2015 - Current
(Tata Consultancy Services -Employee ID # 887870)
Big Data Architect
PCI /Fraud Detection / Real-Time 380° Customer
View/ Personalized offerings /Risk Management / Contact Center efficiency
Optimization/ Customer Segregation/ Customer Churn Analysis / Sentiment
Analysis hashtag using social media.
Technologies: Hadoop HDFS
Data Lake, HIVE, SQOOP, SPARK, KAFKA, CASSANDRA, TABLEAU, PENTAHO Integrating
with SOCIAL MEDIA to create actionable insights and real-time alerts& and
create Personalized offers / portfolios to enhance customer loyalty.
· Client
facing role setting up Architecture Designs scoping and strategies /
Showcasing ROI.
· Team
Lead- Real-Time Data Ingestion using Big data stack of technologies(SPARK
SQL)
· Own
and establish a Reference Architecture for Big Data Blue print and create a
road map for a centralized operations in coordination with all verticals and
cross functional teams.
· End-to-End
ownership for Security /Firewall layers and Kerberisation of all 49 clusters.
· Prepare
documentations and arrange Power point presentations /white boarding and
Train and mentor developers towards solution development and POC execution.
· Participate
in Strategic discussions about Data Integration, Data Ingestion, Data ETL /
ELT
· Design
Real-time processing data Pipelines and trigger Proactive System
notifications in Fraud detection event processing.
· Created
personalized service offerings through data mining technology using R
Algorithms.
· Coordinate
and drive the project initiatives successfully with Off-Shore resources.
· Providing
technical leadership for enabling the successful design and delivery of
enterprise solutions through strategic business development and high level
system design.
Applied Materials.
H.Q – Santa Clara, CA
09/2014 – 05 / 2015
(Tata Consultancy Services- Employee ID # 887870)
Big data Solution Architect :
Cloudera / Amazon Web Services/ IoT-Internet of Things.
· Big
Data POC Project Role & Responsibilities. End to End Ownership and
accountability.
· Data
Migration. Took ownership of the Project from Pilot /Solution – To-
Production.
· Big
Data POC Development on AWS-Amazon Web services/Cloudera.
· Data
Ingestion / On-Premise Data Integration, Tech Support and Documentation.
· Created
a “Data Lake”-Data Migration of existing data from disparate systems and
sources into Hadoop Data Lake using various tools like Sqoop (Hadoop) into
HDFS/Hive. Tools: Sync sort.
· Design
and implement solutions to address business problems and showcase the
ROI/Business value proposition and consensus with the client requirements.
· Drive
Proof of Concept (POC) and Proof of Technology (POT) evaluation on
interoperable technology platforms and seamlessly migrate the Legacy Apps
into Big data platform.
· Train
and mentor developers towards solution development and POC/POT execution
· Enhancement
of SQL queries and making recommendations and best practices.
· POC
conducted for different Use cases and documented on AWS Platform.
· Documented
Functional and Non-Functional Requirements needed for Big data initiatives.
· Communicating
the AWS concepts /Business value /ROI to the top management
· Tech
support and Offshore team interaction and taking ownership of the project and
driving project deliverables from end to end.
· Vendor
selection process evaluation and presentations.
· Use
Cases: Discovery of Internet of Things (IoT), Regression
Analysis and created Predictive Models based on the Sensor Data from the
Production floors lead to Operational efficiency.
· Machine
language/Unstructured data to create insights and establish benchmarks for
production in real –time and laid down strategy for Low cost data storage
· Real-time
Query response for end-users
Walgreens Corporate
H.Q. – Deerfield, IL
04/2014 – 08/2014
(Tata Consultancy Services- Employee ID # 887870)
Big Data Architect
· Architect
Role / Hortonworks
· Setting
up Hadoop Cluster (100 Nodes) from ground-up and setup the visibility of the
data flow from end to end. Provide technical direction in a team that designs
and develops path breaking large-scale cluster data processing systems.
· Take
the ownership the Big data strategy and draw the roadmap and Design new data
pipelines from the legacy systems and recommend the Best practices and lay
down the blue prints for the project deliverables.
· Help
establish thought leadership in the big data space by contributing internal
papers, technical details/ recommendations and best practices to
stakeholders.
· Testing,
Fine-tuning and Diagnosis of Clusters, Applying fixes Configuring
Benchmarking Capacity planning Disaster/failure recovery automation
Detection/repair of data corruption and Optimize the Cluster for better
performance. Interact with the Vendor (Cloudera) for any Technical issues.
· Maintain
the cluster with detailed information to support the sales teams and then
identify trends, forecast from reports, understand and highlight anomalies
and improve performance within each sales division and be comfortable working
with both technical and non-technical groups.
· Hadoop
Production Support, Change Management, Maintenance, Capacity Planning,
Compression techniques, Performance Component verification Plan
production cut-over/deployment and recommend the best practices in the
industry and End-to-End execution of the project from conceptual beginning to
final output and seek the solutions for the Technical issues encountered
during the production phase.
· Gathered
requirements, built logical models, and providing documentation, Benchmark
systems, Analyze system bottlenecks and
UPS (United Parcel
Services)
Supply Chain
Solutions – Atlanta,
GA
07/2013 – 03/2014
Big Data Engineer : Hadoop
Admin- Developer (Cluster setup & Hadoop Map Reduce Team)
· Used
Cases: Efficient Truck Routing, On-Time Deliveries of Shipments,
Creating Models based on sensor data from the Trucks and equipment –Regression
Analysis(IoT)
· Big
Data Analysis & Optimization / Architecture /Machine Learning Algorithms.
· (HADOOP-Proof
of Concepts /ORION Big Data Project Implementation)
· Responsibilities:
Map Reduce Jobs & Hadoop Cluster Maintenance
· BIG
Data – JAVA/ Hadoop – ORION Software Development Project
· Project
name: ORION-(On-Road Integrated Optimization and Navigation). To Create
actionable insights using the unstructured data related to Logistic
telematics and crunching of big data package information, user preferences and
creating an efficient routing to drivers lead to a huge savings to the tune
of $50million a year @ one mile a day for every UPS driver
· Playing
a Hands-On Hybrid role of a Hadoop Admin-cum-Developer which required me to
pull large data sets into HDFS, as per the Use Case, and then write JAVA
based Algorithm/ MapReduce jobs to trigger into MapReduce framework and with
the emitted output, and analyze them to create Statistical/Graphical
visualization reports as per the Business Users needs using various BI tools
like Tableau/Clikview .
· Hands-On
Experience in creating MapReduce Jobs and making my hands dirty by entering
Development environment and troubleshooting and analyzing the end results to
create actionable insights and graphical Dashboards and feed to Business
Intelligence Reporting system/DSS (Decision Support Systems) using various
third party tools like Tableau, Pentaho and Big insights(IBM)/ IBM Data
Explorer/ Teradata Aster/SPLUNK.
· Strong
Understanding of the Hadoop and its Eco-system /Architecture and associated
sub-projects sitting on top of Hadoop like Hive/ Pig /H-Base /Sqoop etc.
· Testing,
Fine-tuning and Diagnosis of Clusters, Applying fixes Configuring
Benchmarking Capacity planning Disaster/failure recovery automation
Detection/repair of data corruption and Optimize the Cluster for better
performance. Interact with the Vendor (Cloudera)for any Technical issues.
· Gathered
requirements, built logical models, and providing documentation, Benchmark
systems, analyses system bottlenecks and
· Propose
solutions to eliminate them and interact with the Vendor to raise Tech
support Tickets to resolve the issues. Subdivide a complex application,
during design phase. Communicated the concepts to Back end Developers and
explain the dependencies
· Worked
directly with UPS clients to map out their existing Business Processes and
providing system-based Predictive Analytic solutions that increase efficiency
and reduce operating costs in setting up automation in their newly planned
system and Integrating with UPS IT environment using Big Data solutions for
increased productivity, customer satisfaction & avoid Customer churn.
· ETL
Jobs : Performed ETL Jobs with structured (transaction), semi-structured
(user behavior) and unstructured (text) data and develop algorithms and
systems before ingesting the data into HDFS using state-of-the-art
open-source platforms like Talend, Pentaho, Splunk, Hive and Pig.
· Deployed
multi-node Cloudera Distribution Hadoop clusters ( 60 nodes, version 1.x and
2.x) in order to prototype solutions using Mahout (0.7, 0.8) to build
predictive models with data from millions of the Retail EDI -856 Transactions
received in UPS database. This helps to reduce product Recalls & Shipment
specifications for Pharma clients of UPS and enhance Customer Satisfaction.
· Designed
BI dashboards, scorecards, charts/graphs, drill-downs, and dynamic reports to
meet the needs of the top management and decision makers.
· CASSANDRA:
Used DATASTAX brand of Cassandra (Peer-to-Peer) tools to handle a real-time
operational data store for online transactional applications and a
read-intensive database for large-scale business intelligence (BI) systems
and created Graphical BI Dashboards out of the ad-hoc query output for
the Top hierarchy management.
· STORM:
Expert in using Storm to perform Real-time Processing to perform Predictive
analytics by ingesting the telemetric/satellite data and triggering the
Operational Alerting systems and scheduled Announcement systems to the Driver
on Road in Real-time .
· Clickstream
Analysis out the web logs to create the actionable and meaningful insights.
· WEB
ANALYTICS: Measuring and collecting off-site and on-site web logs to do
analysis and reporting of internet data for purposes of understanding and
optimizing web usage and enhance the KPI’s and improve the customer web
browsing experience.
· Strong
understanding of Web Analytics like OMNITURE./ WebTrends.
· Cradle
to Grave understanding of HADOOP Eco-System HDFS/MapReduce, JAVA Related
Projects and other Hadoop related projects like Pig Hive NOSql, Zookeeper,
Sqoop, Mahout, and Cassandra. Expert in (MPP) Massive Parallel Processing
architecture in Teradata.
· HADOOP
Eco-System: Setting up Clusters/Multi-nodes/Maintenance/Troubleshooting and
Tuning of the clusters. Involved in integrating Hadoop into existing technology
stacks and software portfolios to achieve maximum Business value.
· Ability
to design solutions independently based on high level architecture
· Implemented
Hadoop based solutions and developed governance strategy and provided
architectural recommendations on integration standards.
· Architected
and Designed Solutions for the business to accomplish Business Value.
· Estimated
Workload Profiles (for analytical processing, Data Processing, Ad-hoc
processing etc.) ETL using various Tools like Pentaho, HP Vertica,
Informatica, Hive, and Pig.
· Determined
Workload Types, Data Landings, Estimate amount of data/ intervals, Determine
data retention periods, any transformations, types/number of integrations,
Plan compression levels.
· Strong
Requirements gathering tasks using JAD Sessions & Conducting User
Interviews to seek Clarity and avoid ambiguity, prepare functional documents
like BRD’s, Use Cases, Software Requirements Specifications (SRS) and setting
up design sessions with Backend Developers and make sure all team members are
in Sync with the Business expectations and engage all stakeholders throughout
the Project Life cycle.
Environment: HADOOP Eco-System, JAVA,.NET, Agile,
MS Office, Cloud Computing.
The Home Depot –
Atlanta, GA
(Tata Consultancy Services- Employee ID # 887870)
06/2012 – 06/2013
Big Data Consultant / Hadoop
Engineer
· Role
& Responsibilities: Admin- cum- Developer.
· HADOOP
Cluster Implementation Strategy/Big Data /WMOS Solutions /SOA Architecture
· Software
Enhancements/Business Intelligence/ Data Migration/SHAREPOINT.
· Big
Data Consultant for Big Data downstream projects.
· Experience
deploying best practices and methodologies to define Hadoop (Cloudera)
infrastructure to roll out releases into production.
· Clickstream
Analysis out the web logs to do Basket Analysis and create meaningful and
actionable insights like consumer buying patterns/predictable analytics to
prevent customer churn/pre-empt competitors by bringing the most desired
items to store shelves.
· Targeted
Marketing - Hadoop framework helped to deploy to increase sales volume and
conversion rates, reduce stock-outs and lead times, and more effectively
compete with alternative web-based E-Commerce options like HD.Com.
· Capacity
forecasting – Hadoop data output helped to get an updated view of order
inventory to enable real-time pricing tools which incorporate projections and
actual behavior to maximize high-fixed, low variable cost, inventory
forecasting and demand supply cycle helped save 28% of revenue to the company
and maintain a healthy growth trajectory cycle.
· Hands-on
experience with MapReduce Jobs on Hadoop based distributed systems (e.g.
MapReduce, Hive, Hbase, Pig, and Flume) Using JAVA program extensively.
· Responsible
for writing MapReduce programs. Import and export data into HDFS from other
RDMS using Sqoop/Hive. Involved in loading data from UNIX file system to HDFS
· Expert
level experience architecting, building, maintaining, and performance tuning
and Enterprise grade Hadoop commercial distribution-Cloudera CDH
· Worked
with large data sets, automate data extraction, build monitoring/reporting
and high-value, and automated clickstream Analysis and offering Business
Intelligence solutions.
· Build
monitoring solution(s) for the Big Data infrastructure to understand the
health of the infrastructure.
· Developed
data architectural strategies at the modeling, design and implementation
stages to address product requirements
· Setting
up the Hadoop Clusters & HDFS/MapReduce Jobs. Ability in Administering,
Installation, configuration, troubleshooting, Security, Backup, Performance
Monitoring and Fine-tuning of HADOOP Clusters. Experience in using Scoop,
ZooKeeper and Cloudera Manager. Good Knowledge on Hadoop Cluster
architecture and monitoring the clusters and Huge Data sets integration.
· Hadoop
MapReduce programs helped for better understanding of customer basket size
and structure, real-time access to inventory levels, and insight into trade
and promotion effectiveness to help refine future advertising campaigns and
align inventory levels by location. Further helps to get an updated
view of order inventory to enable real-time pricing tools which incorporate
projections and actual behavior to maximize high-fixed, low variable cost,
inventory.
· The
output results from Hadoop jobs helped to adjust the content to each
user, to attract and retain customers, and thereby improve sales/usage volume
and stop Customer Churn.
· Conducting
JAD Sessions/Standup meetings, setting up Workshops for Clients/ Technical
and Business Stakeholders. Act as a Liaison between Business
Stakeholders and Technical team.
Environment: Pentaho, Teradata Aster, Vertica,
Splunk, Talend, Tableau.
CVS Caremark –
Cumberland,
RI
05/2011 – 05/2012
(Tata Consultancy Services- Employee ID # 887870)
Big Data Analyst / Hadoop
Consultant
· Project:
Big Data- HADOOP - POC (Proof of Concept)DATA Analytics & Third Party
DATA Integration.
· USE
Cases: Fraud Detection, prevent Customer Churn, Patient Clinical notes
Integration.
· Role
& Responsibilities: Started my Hadoop journey here as I got involved in
Setting up of a New POC– in Amazon EMR’s and after successful POC, we
eventually collaborated with Cloudera distribution technicians to initiate a
Multi-Node Cluster setup, Configure and Test for Development and production.
· The
data is fed into ETL and then processed using Hive to de-normalize and
aggregate the disparate data sources. The customer profiles are categorized
and product profiles are built using Pig. The processed data is then moved
into Hive for real-time access using a REST-based API.
· Exported
data from SQL Server to HDFS using Sqoop and NFS mounts and created Graphical
Analytical reports using “BigInsights IBM”
· Importing
and exporting data into HDFS from RDBMS/ Hive using Sqoop.
· Loaded
and transform large sets of structured, semi structured and unstructured
data.
· Clustered
coordination services through Zoo Keeper. Exported the analyzed data to the
relational databases using Sqoop for visualization and to generate reports
for the BI team.
· Installed
and configured Flume, Hive, Pig, Sqoop, HBase on the Hadoop cluster.
Implemented 30 nodes Hadoop cluster on Amazon EMR and then Cloudera
commercial Distribution.
· Worked
on installing cluster, commissioning & decommissioning of data node, name
node recovery, capacity planning, and slots configuration. Setup Hadoop
cluster on Amazon EC2 using whirr for POC.
· Responsible
to manage data coming from different sources. Installed and configured Hive
and also written Hive UDFs.
IBM -
Express-Scripts Inc. Minneapolis,
MN
04/2010 – 04/2011
Data
Analyst
· Involved
as Data Analyst to perform DATA INTEGRATION using TIBCO Tool for this newly
merged company- Express Scripts + Medco.
· Conducted
GAP Analysis /Impact Analysis /Pricing Analysis and flushing out High level
Requirements from the Business users and other Stakeholders involved in this
Initiative.
Hewlett-Packard –
Windsor, CT
01/2010 – 03/2010
Business Data
Analyst
· Production
Support& Testing and Analysis of Defects Related to Transaction Sets 837/
835/277 CA Utilizing HP “Quality Center” Requirements/Defects Module,
/Edifecs Working Collaboratively with Technical Team to Fix the defects and
find appropriate Solutions to the satisfaction of the Business/Trading
partners/End users.
Motorola Mobility –
Horsham, PA
01/2006 – 12/2009
Data Integration Analyst / CRM
Admin
· Roles
& responsibilities: Data Integration using TIBCO. Third party software
integrations. Created Dashboards and scheduled automatic refresh and email.
· Responsible
for all dashboard, metrics & analytics for global operations.
· Analyzed
the needs of 1000+ users and updated requirements. Identify Risks and
involved management in decision making.
Frontier
Communications –
07/2000 – 12/2005
Business Systems Analyst
· Project:
Data Integration and Migration of Legacy Apps.
· Responsible
for creating and reviewing business requirements, functional specifications,
project schedules, technical documentation and test plans.
EDUCATION
· John
Hopkins University, Certificate of Completion in Data Science Toolbox,
· Diploma
in Computer Networking Applications,
· Master
of Sciences. MS Computer Science
· Bachelor
of -Sciences, BS Computer Science
· Pursuing
Graduate Certificate in CYBER SECURITY – MIT-Massachusetts
Institute of Tech.
· Pursuing
COMPTIA Security+ (Cyber Security- Internet of Things)
CERTIFICATION
· Certified
“Cloud Computing Consultant” RACKSPACE/SALESFORCE.COM/IBM
· HADOOP
Admin. (Big Data) & Cassandra Specialist
· CISSP
– Working on Certification(Certified Information Systems Professional)
· Member
of PMI- Project Management Institute. Membership .ID #: 2504042
· CISCO
Certified for IoT( IMINS ) “Managing Industrial Networking Technologies
“in Progress
PROFESSIONAL AFFILIATIONS & MEMBERSHIPS
· “TOGAF”
Level 1 Certified Architect.
· Rackspace
Certified “Cloud Solutions Consultant”
· Hadoop
Admin Certification.
· CISCO
Certification for IoT( IMINS ) “Managing Industrial Networking Technologies”
· AHIMA,
Member, (American Health Information Management Association)
· IIBA
Member, (International Institute of Business Analysts)
· PMI
Member #2504042 – Project Management Institute.
LinkedIn Account:
http://www.linkedin.com/in/narenkanuganti/