From:                              route@monster.com

Sent:                               Monday, September 28, 2015 1:02 PM

To:                                   hg@apeironinc.com

Subject:                          Please review this candidate for: Talend

 

This resume has been forwarded to you at the request of Monster User xapeix03

Narender Kanuganti 

Last updated:  08/14/15

Job Title:  no specified

Company:  no specified

Rating:  Not Rated

Screening score:  no specified

Status:  Resume Received


Alpharetta, GA  30004
US

Mobile: 470-564-4000   
naren.bigdatapro@gmail.com
Contact Preference:  Email

Quick View Links:

Resume Section

Summary Section

 

 

RESUME

  

Resume Headline: Kanuganti - Big Data / IoT Solutions Architect

Resume Value: mpufqyebnj8unnmj   

  

 


 

                        Narender Reddy Kanuganti.

     Big Data Architect /IoT/ Cyber Security Specialist (CompTIA-Certified)

    naren.bigdatapro@gmail.com                                        470-564-4000                                               

 

BIG DATA Practice •HADOOP BIG DATA ARCHITECTURE IoT-INTERNET OF THINGS

DATA INTEGRATION • DATA SCIENCEE-COMMERCE PREDICTIVE ANALYTICSPOC

PRE-SALES / SALES DEVELOPMENT / CLIENT MEETINGS & PRESENTATIONS.

 

SUMMARY   

 

·              Tech Savvy Techno-Functional Data Wrangler / Analyst turned BIG DATA Architect with specializations in IoT Technology Platform & Hadoop Eco-system with 14 + years of Cradle to Grave Experience in IT Landscape ranging from Broad-spectrum of Industries- like IT/Retail/Telecom/Healthcare/Insurance/Logistics performing Predictive Data Analysis, Data Science Analytics, Requirements Elicitation, Business Modeling, and Use Case development, Internet of Things (IoT) Inception (POC) to Production. 

·              IOT platform strategy, architecture, and planning, M2M Platform Services: Connectivity Management, Application Enablement, and Device Management offering world leading security solutions for device protection, encryption, authentication, key management, and code signing. HADOOP: Create a Blueprint and Roadmap for Clients for their journeys into Big Data.

·              “Get things done” attitude with a strong Boots-on-the-ground mentality to steer projects into

Right portfolios and create Win-Win Situation for the Client and Employer.

·              Knowledge of Data Center Architectures and Services incorporating Applications, Servers, Storage, and Data Center Switching technologies and ability to manage a long and complex sales cycle. Ability to leverage and converge Big Data with IoT Platform Technology.

·              TOGAF Certified Architect with 3 years of Hands-on Experience in Big Data Technologies like Hadoop and its eco-system and still making my hands dirty and shaking hands with Top Management folks /Stakeholders on front end and helping the fellow Developers at the back end Cracking, Debugging and  fixing the Codes.

·              Extensive experience in Software Development Cycle from Inception to Launch ,  and strong passion for new emerging Technologies like Hadoop and Internet of  Things(IoT)

·              Ability  to guide the lifecycle of Hadoop and other Big Data solutions, including requirements analysis, platform selection, Logical and Physical architectural designs, application design and development, testing, construct the Big data pipelines and seamlessly integrate the Legacy platforms/Apps into new emerging Technologies and SME for IoT Platform strategy.

·              Strong understanding of what’s “Under the Hood” of Big data World to offer Best practices / recommendations to the clients and drive the Big Data /IoT Initiatives from Pilot –to- Production”. Self-motivated and sell audiences on IoT solutions by performing technical product presentations and demos. Possess  a balance of technical and sales aptitude, including handling customer technical questions and issues, creating product positioning and translating business needs into new product features

·              Blessed with Techniques to seek Sign-off from Key & tough  Stakeholders through Diplomacy acting as a Liaison and Bridging the Gap between Business World and Technical World and keep all Stakeholders in Sync and involve myself in all phases of the Product Life Cycle right from Inception to launching and ensure the project deliverables within Time limits and Scope.

·              Effective Communicator With Front End Stakeholders And Good Listener And A Team Player With Strong Technical Documentation Skills And Analytical Bent Of Mind Possessing Strong Willingness To Learn New Systems, Good coordinating skills ability to Ramp up to the Speed and hit the Ground Running.

·              Team-Oriented & Self-starter Professional working with Cross functional teams in a  Techno functional environment playing multiple Roles as  Big Data  Analyst/Admin /Developer Role with strong understanding of Timelines, Budgets and deliverables and add Value to the company.

·              Proactive to enter Developer’s world and willing to wrangle with lines of Code and make my hands dirty enough with hardware & Software and indulge in troubleshooting to resolve technical issues and take ownership of the Product deliverable and drive the project from end to end.

·              A good Deal puncher with customer focused Negotiating abilities/solution finding skills wearing different Hats at different Scenarios. Ability to Read the Code and do the critical Analysis of Bugs and Defects and accomplish the End user’s/Business requirements.

·              Self-Starter and Self-motivated Team leader, DETAIL ORIENTED and ability to think OUTSIDE THE BOX and thirst for learning new systems and adapt with changing corporate environments and increase the Bottom line of the Company. Strong in AGILE and WATERFALL Methodologies  and expert in Managing “Contracts Backlog” and prioritize according to project deadlines

·              Experienced in Full Life cycle implementations of Big Data solutions (HADOOP) & gathering Requirements from Clients and perform customizations/ setting up configurations and testing for Performance. Ability to de-mystify predictive analytics and create actionable insights from the output of that algorithm to add business value to the enterprise.

·              Strong understanding of Software Development Life Cycle (SDLC), including good knowledge of RUP methodology. Reproduce, Expertise in SOA, investigate and debug software issues. Knowledge of both Agile and Waterfall Software development methodologies. ENTERPRISE WEB DEVELOPMENT Experience in Accu Process, SaaS (Ajax), and Informatica Cloud and Extensive knowledge of Data warehouse Concepts. HP (PPM) and Business Objects (SAP).

·              Extensive experience in interacting with Offshore Team/Virtual team & Stakeholders, Problem Management Analysis, Eliciting Requirements and creating Business Requirement Documents (BRD), User Requirement Specifications, Functional Requirement Documents (FRD), System Requirement Specification (SRS), Test Plan, analyzing and Creating Use Cases, Use Case Diagrams, PROCESS FLOW Diagrams, BPMN Diagrams, Activity diagrams, System Workflow. Strong Understanding of all Versions of SharePoint.

·              Experience in applying Rational Unified Process (RUP) methodology using Modeling and requirement documentation tools such as MS Office, MS VISIO, and MS Project.

·              Conducting and facilitating JAD Sessions and communicating the Concepts with Key Stakeholders, Development team, SMEs, System Analysts, Business Analysts and Project Managers and External Vendors.

·              Experience in customizing the portal sites on Share point 2007 and 2010, Rational Requisite Pro, Team Foundation Server and Variety of e-Commerce applications, Cloud Computing and Deployment of Apps. Strong understanding of SOA architecture designs and concepts

·              “Cradle to grave” experience in big data initiatives and strategic planning from concept to launching.

·              Big Data Practice / Big Data Architecture / SME for IoT-Internet Of Things strategies.

·              Strong understanding of IoT Platform like Axeda, Thingworx, Microsoft Azure. Working knowledge on IoT Strategy, standards, protocols like MQTT, COAP. Exposed to Embedded device programming Device protocols and wearable technology Exposed to Sensor analytics

·              Working Knowledge and demonstrated experience in IoT, Integration or Cloud integration Architectures

·              Data Integration & Warehousing Data Science E-Commerce Web Analytics

     COMMERCIAL Distributions /Platforms of HADOOP worked.

§    CLOUDERA                                       AWS –Cloud Computing / Deployments              

§    HDP- Hortonworks                      Microsoft AZURE

§    MAPR                                                 IoT – Internet of Things Platform Strategy Implementation

§    HD Insight (IBM)                           LINUX & WINDOWS Environment

 

 

TECHNICAL SKILLS

 

·  Design Blue print and create a Road map for Big data journey incorporating Big Data Concepts Cradle –to-Grave understanding of the Hadoop eco-system and Reference Architecture, Hadoop-(HDFS), Write MapReduce Jobs & Algorithms Using Various Tools Pig, Hive, Hbase, Impala, NoSQL, Cassandra, IBM Data Explorer. Zookeeper, Sqoop, Flume /R Program/Pentaho / Vertica/Informatica /Talend/ Teradata Aster.

·  BIG DATA TECHNOLOGY STACK: HDFS,YARN, HIVE,PIG, SPARK,STORM, CASSANDRA, HBase

MONGO db,  ‘R ”Machine Language, PENTAHO, TALEND, TABLEAU, HYPERION.

·  Hadoop Cluster Administration, Configuration, Monitoring, Debugging, and Performance Tuning. Ability To Implement Hadoop Based Solutions And Offer Best Practices In Big Data Space.

·  Hadoop Eco-System: Setting Up Clusters/Nodes/Maintenance And Tuning Of The Cluster Nodes End-To-End, Troubleshoot The Technical Issues And Offer Solutions.

·  Develop Solutions For (IoT) Internet Of Things And Provide End-To-End Support From Concept To Production Phase Of The Product Life Cycle.

·  Strong Understanding Of Hadoop Eco-System And Other Business Intelligence And ETL Tools  On Top Of Hadoop Like Vertica, Pentaho, Sqoop, Oozie, Flume, Hbase, Tableau, Teradata , Datastax Datameer And Mahout(Machine Learning) Web Analytics (Omniture).Mpp (Massive Parallel Processing) In Teradata and Data Modelling software ERwin 8.2.

·  Ability to write MapReduce Programs and create Business Intelligence reports from the output.

·  Daily support of several Hadoop, data warehouse appliances, including monitoring capacity, throughput, health, and usage and Clickstream Analysis out of the web logs.

·  Collaboration with vendors and users to coordinate and accomplish repairs, upgrades, patches, and other enhancements, additions, or replacements.

·  Query analysis and tuning advice for end users, to maintain throughput and reliable operation.

·  Scripting to deploy monitors, checks, and other sys admin function automation.

·  Production Support for any problem leading to acceptable resolution, including daytime, nighttime, and weekend support if required.

·  Performs Incident resolution, Problem Determination and Root Cause Analysis.  (i.e., hardware and software diagnostic tools to monitor performance and perform problem determination).  Familiarity with hardware and software diagnostic tools to monitor performance and perform problem determination.

·  Oversee installations, monitoring and managing change to servers (Overall Change Management for Servers).Oversees implementation of security guidelines in order to prevent unauthorized access to servers and report any violations.

·  Ability to Monitor and tune operating systems to achieve optimum performance levels in standalone and multi-tiered environments. 

·  Collaborate with System Engineering, Network Engineering, solves complex and recurring operational issues and develops corrective actions, as needed.

·  Interact regularly with Metrics team, developers, engineers, and the IT outsourcer to ensure the Company’s Reliability, Availability and Serviceability (RAS) metrics are sustained and improved from current level.

·  Participate in the evaluation, recommendation, and selection of hardware and software solutions. - Reviews, evaluate, designs, implements and maintain internal and external data.

·  Identify data sources, constructs data decomposition diagrams, provides data flow diagrams and documents the process. 

·  Writes codes for database access, modifications, and constructions including Map Reduce Programs, Pig/Hive Scripts, SQL-H, Stored Procedures, etc.

·  Developed and reviewed project plans, identifies issues, resolves issues, and communicates status of assigned projects to end-users and project Stakeholders.

·  Experience in operational support and hands-on implementation Hadoop based Big Data Platforms.  Gathers requirements, builds logical models and provides quality documentation of detailed user requirements for the design and development of systems.

·  Databases: Hbase/NoSQL/SQL Server, Oracle/IMPALA /MongoDB, CASSANDRA.

·  Cyber Security Skills: (IAM, Encryption, DLP, Risk and Compliance Management, IDS/IPS, UTM, Firewall, Antivirus/Antimalware, SIEM, Disaster Recovery, DDOS Mitigation, Web Filtering, and Security Services

 

 

BUSINESS SKILLS

 

·  Technical Pre-Sales, White Boarding and PowerPoint presentations to the Board members and Lead the Team to provide thought process leadership and direction and hold accountable for the project deliverables within the budgets and timelines throughout the life cycle of the project and avoid the Scope creep and add Clarity to the tasks and communicate the concepts and update the status reports to the front-end stakeholders through white boarding and PowerPoint point presentations. Speak both Technical and Non-Technical languages and flip both sides of coin.

·  People Skills, Networking With People, Connect The Dots And Get Things Done. Facilitate JAD Sessions, Power Point Presentations of New Products and Services to wide spectrum of audience/ Clients and Business Users / Stakeholders.

·  Test Cases, Vision, Scope, and SRS documents/UAT Sessions with Stakeholders.

·  Gap Analysis, Impact Analysis and SWOT Analysis/Feasibility Analysis, Product Marketing &Sales.

·  Focus on End Game and Prioritize the Requirements and achieve goals within Timelines and Budget.

·  Business Process Analysis & Research using i-Rise Software/Enterprise Architect.

·  USE CASE Modeling & Analysis, Troubleshooting software bugs and defects management.

·  Functional Requirement Gathering & Technical Requirements Development and Documentation.

·  Prototyping / Wireframes & Mockup Screen Creations

 

 

BUSINESS DOCUMENTATION

 

·  Business Requirements / Functional and Non-Functional Documentation of Technical Specs / and sign off from the Stakeholders.

·  Documentation and Visualization using Reporting tools like /Tableau /Qlickview.

·  Story Boards, Scenarios, Personas. Test Plans, Test Scripts and Test Cases, Prototypes.

·  Setup Workshops for Business sponsors / end users and Developers and educate and make sure all the stakeholders are same page and understand the strategy and concepts of the IT initiatives.

 

 

 

 

 

 

PROFESSIONAL EXPERIENCE                                                                

BANK of AMERICA. H.Q Charlotte, NC.                                 05/2015 - Current

(Tata Consultancy Services -Employee ID # 887870)

 

Big Data Architect

PCI /Fraud Detection / Real-Time 380° Customer View/ Personalized offerings /Risk Management / Contact Center efficiency Optimization/ Customer Segregation/ Customer Churn Analysis / Sentiment Analysis hashtag using social media.

Technologies: Hadoop HDFS Data Lake, HIVE, SQOOP, SPARK, KAFKA, CASSANDRA, TABLEAU, PENTAHO Integrating with SOCIAL MEDIA to create actionable insights and real-time alerts& and create Personalized offers / portfolios to enhance customer loyalty.

 

·  Client facing role setting up Architecture Designs scoping and strategies / Showcasing ROI.

·  Team Lead- Real-Time Data Ingestion using Big data stack of technologies(SPARK SQL)

·  Own and establish a Reference Architecture for Big Data Blue print and create a road map for a centralized operations in coordination with all verticals and cross functional teams.

·  End-to-End ownership for Security /Firewall layers and Kerberisation of all 49 clusters.

·  Prepare documentations and arrange Power point presentations /white boarding and Train and mentor developers towards solution development and POC execution.

·  Participate in Strategic discussions about Data Integration, Data Ingestion, Data ETL / ELT

·  Design Real-time processing data Pipelines and trigger Proactive System notifications in Fraud detection event processing.

·  Created personalized service offerings through data mining technology using R Algorithms.

·  Coordinate and drive the project initiatives successfully with Off-Shore resources.

·  Providing technical leadership for enabling the successful design and delivery of enterprise solutions through strategic business development and high level system design. 

 

 

Applied Materials. H.Q – Santa Clara, CA                            09/2014 – 05 / 2015

(Tata Consultancy Services- Employee ID # 887870)

Big data Solution Architect : Cloudera / Amazon Web Services/ IoT-Internet of Things.

·  Big Data POC Project Role & Responsibilities. End to End Ownership and accountability.

·  Data Migration. Took ownership of the Project from Pilot /Solution – To- Production.

·  Big Data POC Development on AWS-Amazon Web services/Cloudera.

·  Data Ingestion / On-Premise Data Integration, Tech Support and Documentation.

·  Created a “Data Lake”-Data Migration of existing data from disparate systems and sources into Hadoop Data Lake using various tools like Sqoop (Hadoop) into HDFS/Hive. Tools: Sync sort.

·  Design and implement solutions to address business problems and showcase the ROI/Business value proposition and consensus with the client requirements.

·  Drive Proof of Concept (POC) and Proof of Technology (POT) evaluation on interoperable technology platforms and seamlessly migrate the Legacy Apps into Big data platform.

·  Train and mentor developers towards solution development and POC/POT execution

·  Enhancement of SQL queries and making recommendations and best practices.

·  POC conducted for different Use cases and documented on AWS Platform.

·  Documented Functional and Non-Functional Requirements needed for Big data initiatives.

·  Communicating the AWS concepts /Business value /ROI to the top management

·  Tech support and Offshore team interaction and taking ownership of the project and driving project deliverables from end to end.

·  Vendor selection process evaluation and presentations.

·  Use Cases: Discovery of Internet of Things (IoT), Regression Analysis and created Predictive Models based on the Sensor Data from the Production floors lead to Operational efficiency.

·  Machine language/Unstructured data to create insights and establish benchmarks for production in real –time and laid down strategy for Low cost data storage

·  Real-time Query response for end-users

 

Walgreens Corporate H.Q. – Deerfield, IL                                          04/2014 – 08/2014

(Tata Consultancy Services- Employee ID # 887870)

Big Data Architect 

·  Architect  Role / Hortonworks

·  Setting up Hadoop Cluster (100 Nodes) from ground-up and setup the visibility of the data flow from end to end. Provide technical direction in a team that designs and develops path breaking large-scale cluster data processing systems.

·  Take the ownership the Big data strategy and draw the roadmap and Design new data pipelines from the legacy systems and recommend the Best practices and lay down the blue prints for the project deliverables.

·  Help establish thought leadership in the big data space by contributing internal papers, technical details/ recommendations and best practices to stakeholders.

·  Testing, Fine-tuning and Diagnosis of Clusters, Applying fixes Configuring Benchmarking Capacity planning Disaster/failure recovery automation Detection/repair of data corruption and Optimize the Cluster for better performance. Interact with the Vendor (Cloudera) for any Technical issues.

·  Maintain the cluster with detailed information to support the sales teams and then identify trends, forecast from reports, understand and highlight anomalies and improve performance within each sales division and be comfortable working with both technical and non-technical groups.

·  Hadoop Production Support, Change Management, Maintenance, Capacity Planning, Compression techniques, Performance Component verification  Plan production cut-over/deployment and recommend the best practices in the industry and End-to-End execution of the project from conceptual beginning to final output and seek the solutions for the Technical issues encountered during the production phase.

·  Gathered requirements, built logical models, and providing documentation, Benchmark systems, Analyze system bottlenecks and

 

UPS (United Parcel Services)

Supply Chain Solutions – Atlanta, GA                    07/2013 – 03/2014               

Big Data Engineer : Hadoop Admin- Developer (Cluster setup & Hadoop Map Reduce Team)

·  Used Cases: Efficient Truck Routing, On-Time Deliveries of Shipments, Creating Models based on sensor data from the Trucks and equipment –Regression Analysis(IoT)

·  Big Data Analysis & Optimization / Architecture /Machine Learning Algorithms.

·  (HADOOP-Proof of Concepts  /ORION Big Data Project Implementation)

·  Responsibilities: Map Reduce Jobs & Hadoop Cluster Maintenance

·  BIG Data – JAVA/ Hadoop – ORION Software Development Project

·  Project name: ORION-(On-Road Integrated Optimization and Navigation). To Create actionable insights using the unstructured data related to Logistic telematics and crunching of big data package information, user preferences and creating an efficient routing to drivers lead to a huge savings to the tune of $50million a year @ one mile a day for every UPS driver

·  Playing a Hands-On Hybrid role of a Hadoop Admin-cum-Developer which required me to pull large data sets into HDFS, as per the Use Case, and then write JAVA based Algorithm/ MapReduce jobs to trigger into MapReduce framework and with the emitted output, and analyze them to create Statistical/Graphical visualization reports as per the Business Users needs using various BI tools like Tableau/Clikview .

·  Hands-On Experience in creating MapReduce Jobs and making my hands dirty by entering Development environment and troubleshooting and analyzing the end results to create actionable insights and graphical Dashboards and feed to Business Intelligence Reporting system/DSS (Decision Support Systems) using various third party tools like Tableau, Pentaho and Big insights(IBM)/ IBM Data Explorer/ Teradata Aster/SPLUNK.

·  Strong Understanding of the Hadoop and its Eco-system /Architecture and associated sub-projects sitting on top of Hadoop like Hive/ Pig /H-Base /Sqoop etc.

·  Testing, Fine-tuning and Diagnosis of Clusters, Applying fixes Configuring Benchmarking Capacity planning Disaster/failure recovery automation Detection/repair of data corruption and Optimize the Cluster for better performance. Interact with the Vendor (Cloudera)for any Technical issues.

·  Gathered requirements, built logical models, and providing documentation, Benchmark systems, analyses system bottlenecks and

·  Propose solutions to eliminate them and interact with the Vendor to raise Tech support Tickets to resolve the issues. Subdivide a complex application, during design phase. Communicated the concepts to Back end Developers and explain the dependencies

·  Worked directly with UPS clients to map out their existing Business Processes and providing system-based Predictive Analytic solutions that increase efficiency and reduce operating costs in setting up automation in their newly planned system and Integrating with UPS IT environment using Big Data solutions for increased productivity, customer satisfaction & avoid Customer churn.

·  ETL Jobs : Performed ETL Jobs with structured (transaction), semi-structured (user behavior) and unstructured (text) data and develop algorithms and systems before ingesting the data into HDFS  using state-of-the-art open-source platforms like Talend, Pentaho, Splunk, Hive and Pig.

·  Deployed multi-node Cloudera Distribution Hadoop clusters ( 60 nodes, version 1.x and 2.x) in order to prototype solutions using Mahout (0.7, 0.8) to build predictive models with data from millions of the Retail EDI -856 Transactions received in UPS database. This helps to reduce product Recalls & Shipment specifications for Pharma clients of UPS and enhance Customer Satisfaction.

·  Designed BI dashboards, scorecards, charts/graphs, drill-downs, and dynamic reports to meet the needs of the top management and decision makers.

·  CASSANDRA: Used DATASTAX brand of Cassandra (Peer-to-Peer) tools to handle a real-time operational data store for online transactional applications and a read-intensive database for large-scale business intelligence (BI) systems and created Graphical  BI Dashboards out of the ad-hoc query output for the Top hierarchy management.

·  STORM: Expert in using Storm to perform Real-time Processing to perform Predictive analytics by ingesting the telemetric/satellite data and triggering the Operational Alerting systems and scheduled Announcement systems to the Driver on Road in Real-time .

·  Clickstream Analysis out the web logs to create the actionable and meaningful insights.

·  WEB ANALYTICS: Measuring and collecting off-site and on-site web logs to do analysis and reporting of internet data for purposes of understanding and optimizing web usage and enhance the KPI’s and improve the customer web browsing experience.

·  Strong understanding of Web Analytics like OMNITURE./ WebTrends.

·  Cradle to Grave understanding of HADOOP Eco-System HDFS/MapReduce, JAVA Related Projects and other Hadoop related projects like Pig Hive NOSql, Zookeeper, Sqoop, Mahout, and Cassandra. Expert in (MPP) Massive Parallel Processing architecture in Teradata.

·  HADOOP Eco-System: Setting up Clusters/Multi-nodes/Maintenance/Troubleshooting and Tuning of the clusters. Involved in integrating Hadoop into existing technology stacks and software portfolios to achieve maximum Business value.

·  Ability to design solutions independently based on high level architecture

·  Implemented Hadoop based solutions and developed governance strategy and provided architectural recommendations on integration standards.

·  Architected and Designed Solutions for the business to accomplish Business Value.

·  Estimated Workload Profiles (for analytical processing, Data Processing, Ad-hoc processing etc.) ETL using various Tools like Pentaho, HP Vertica, Informatica, Hive, and Pig.

·  Determined Workload Types, Data Landings, Estimate amount of data/ intervals, Determine data retention periods, any transformations, types/number of integrations, Plan compression levels.

·  Strong Requirements gathering tasks using JAD Sessions & Conducting User Interviews to seek Clarity and avoid ambiguity, prepare functional documents like BRD’s, Use Cases, Software Requirements Specifications (SRS) and setting up design sessions with Backend Developers and make sure all team members are in Sync with the Business expectations and engage all stakeholders throughout the Project Life cycle. 

Environment: HADOOP Eco-System, JAVA,.NET, Agile, MS Office, Cloud Computing.

 

The Home Depot – Atlanta, GA

(Tata Consultancy Services- Employee ID # 887870)                     06/2012 – 06/2013

Big Data Consultant / Hadoop Engineer

·  Role & Responsibilities:  Admin- cum- Developer.

·  HADOOP Cluster Implementation Strategy/Big Data /WMOS Solutions /SOA Architecture

·  Software Enhancements/Business Intelligence/ Data Migration/SHAREPOINT.

·  Big Data Consultant for Big Data downstream projects.

·  Experience deploying best practices and methodologies to define Hadoop (Cloudera) infrastructure to roll out releases into production.

·  Clickstream Analysis out the web logs to do Basket Analysis and create meaningful and actionable insights like consumer buying patterns/predictable analytics to prevent customer churn/pre-empt competitors by bringing the most desired items to store shelves.

·  Targeted Marketing - Hadoop framework helped to deploy to increase sales volume and conversion rates, reduce stock-outs and lead times, and more effectively compete with alternative web-based E-Commerce options like HD.Com.

·  Capacity forecasting – Hadoop data output helped to get an updated view of order inventory to enable real-time pricing tools which incorporate projections and actual behavior to maximize high-fixed, low variable cost, inventory forecasting and demand supply cycle helped save 28% of revenue to the company and maintain a healthy growth trajectory cycle.

·  Hands-on experience with MapReduce Jobs on Hadoop based distributed systems (e.g. MapReduce, Hive, Hbase, Pig, and Flume) Using JAVA program extensively.

·  Responsible for writing MapReduce programs. Import and export data into HDFS from other RDMS using Sqoop/Hive. Involved in loading data from UNIX file system to HDFS

·  Expert level experience architecting, building, maintaining, and performance tuning and Enterprise grade Hadoop commercial distribution-Cloudera CDH

·  Worked with large data sets, automate data extraction, build monitoring/reporting and high-value, and automated clickstream Analysis and offering Business Intelligence solutions.

·  Build monitoring solution(s) for the Big Data infrastructure to understand the health of the infrastructure.

·  Developed data architectural strategies at the modeling, design and implementation stages to address product requirements

·  Setting up the Hadoop Clusters & HDFS/MapReduce Jobs. Ability in Administering, Installation, configuration, troubleshooting, Security, Backup, Performance Monitoring and Fine-tuning of HADOOP Clusters. Experience in using Scoop, ZooKeeper and Cloudera Manager.  Good Knowledge on Hadoop Cluster architecture and monitoring the clusters and Huge Data sets integration.

·  Hadoop MapReduce programs helped for better understanding of customer basket size and structure, real-time access to inventory levels, and insight into trade and promotion effectiveness to help refine future advertising campaigns and align inventory levels by location.  Further helps to get an updated view of order inventory to enable real-time pricing tools which incorporate projections and actual behavior to maximize high-fixed, low variable cost, inventory.

·  The output results from Hadoop jobs helped  to adjust the content to each user, to attract and retain customers, and thereby improve sales/usage volume and stop Customer Churn.

·  Conducting JAD Sessions/Standup meetings, setting up Workshops for Clients/ Technical and Business   Stakeholders. Act as a Liaison between Business Stakeholders and Technical team.

Environment: Pentaho, Teradata Aster, Vertica, Splunk, Talend, Tableau.

 

 

CVS Caremark – Cumberland, RI                         05/2011 – 05/2012

(Tata Consultancy Services- Employee ID # 887870)     

Big Data Analyst / Hadoop Consultant

·  Project: Big Data- HADOOP - POC (Proof of Concept)DATA Analytics & Third Party DATA Integration. 

·  USE  Cases: Fraud Detection, prevent Customer Churn, Patient Clinical notes Integration.

·  Role & Responsibilities: Started my Hadoop journey here as I got involved in Setting up of a New POC– in Amazon EMR’s and after successful POC, we eventually collaborated with Cloudera distribution technicians to initiate a Multi-Node Cluster setup, Configure and Test for Development and production.

·  The data is fed into ETL and then processed using Hive to de-normalize and aggregate the disparate data sources. The customer profiles are categorized and product profiles are built using Pig. The processed data is then moved into Hive for real-time access using a REST-based API.

·  Exported data from SQL Server to HDFS using Sqoop and NFS mounts and created Graphical Analytical reports using “BigInsights IBM”

·  Importing and exporting data into HDFS from RDBMS/ Hive using Sqoop.

·  Loaded and transform large sets of structured, semi structured and unstructured data.

·  Clustered coordination services through Zoo Keeper. Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.

·  Installed and configured Flume, Hive, Pig, Sqoop, HBase on the Hadoop cluster.  Implemented 30 nodes Hadoop cluster on Amazon EMR and then Cloudera commercial Distribution.

·  Worked on installing cluster, commissioning & decommissioning of data node, name node recovery, capacity planning, and slots configuration.  Setup Hadoop cluster on Amazon EC2 using whirr for POC.

·  Responsible to manage data coming from different sources. Installed and configured Hive and also written Hive UDFs.

 

IBM - Express-Scripts Inc. Minneapolis, MN                 04/2010 – 04/2011 

Data Analyst                                                                                      

·  Involved as Data Analyst to perform DATA INTEGRATION using TIBCO Tool for this newly merged company- Express Scripts + Medco.

·  Conducted GAP Analysis /Impact Analysis /Pricing Analysis and flushing out High level Requirements from the Business users and other Stakeholders involved in this Initiative.

 

Hewlett-Packard – Windsor, CT                                            01/2010 – 03/2010 

Business Data Analyst        

·  Production Support& Testing and Analysis of Defects Related to Transaction Sets 837/ 835/277 CA Utilizing HP “Quality Center” Requirements/Defects Module, /Edifecs Working Collaboratively with Technical Team to Fix the defects and find appropriate Solutions to the satisfaction of the Business/Trading partners/End users.

 

Motorola Mobility – Horsham, PA                                        01/2006 – 12/2009                  

Data Integration Analyst / CRM Admin                                            

·  Roles & responsibilities: Data Integration using TIBCO. Third party software integrations. Created Dashboards and scheduled automatic refresh and email.

·  Responsible for all dashboard, metrics & analytics for global operations.

·  Analyzed the needs of 1000+ users and updated requirements. Identify Risks and involved management in decision making.

 

Frontier Communications –                                         07/2000 – 12/2005

Business Systems Analyst

·  Project: Data Integration and Migration of Legacy Apps.

·  Responsible for creating and reviewing business requirements, functional specifications, project schedules, technical documentation and test plans.

 

EDUCATION

 

·  John Hopkins University, Certificate of Completion in Data Science Toolbox,

·  Diploma in Computer Networking Applications,

·  Master of Sciences. MS Computer Science

·  Bachelor of  -Sciences, BS Computer Science

·  Pursuing Graduate Certificate in CYBER SECURITYMIT-Massachusetts Institute of Tech.

·  Pursuing COMPTIA Security+ (Cyber Security- Internet of Things)

 

CERTIFICATION

·  Certified “Cloud Computing Consultant” RACKSPACE/SALESFORCE.COM/IBM

·  HADOOP Admin. (Big Data) & Cassandra Specialist

·  CISSP – Working on Certification(Certified Information Systems Professional)

·  Member of PMI- Project Management Institute. Membership .ID #: 2504042

·  CISCO Certified for IoT( IMINS ) “Managing Industrial Networking  Technologies “in Progress

 

PROFESSIONAL AFFILIATIONS & MEMBERSHIPS

·  “TOGAF” Level 1 Certified Architect.

·  Rackspace Certified “Cloud Solutions Consultant”

·  Hadoop Admin Certification.

·  CISCO Certification for IoT( IMINS ) “Managing Industrial Networking Technologies”

·  AHIMA, Member, (American Health Information Management Association)

·  IIBA Member, (International Institute of Business Analysts)

·  PMI Member #2504042 – Project Management Institute.

 

LinkedIn Account:      http://www.linkedin.com/in/narenkanuganti/

 

 

 

 



Experience

BACK TO TOP

 

Job Title

Company

Experience

IT Solutions Architect

Tata Consultancy Services Limited

- Present

 

Additional Info

BACK TO TOP

 

Current Career Level:

Manager (Manager/Supervisor of Staff)

Date of Availability:

Immediately

Work Status:

US - I am authorized to work in this country for any employer.

Active Security Clearance:

None

US Military Service:

Citizenship:

US citizen

 

 

Target Job:

Target Job Title:

IT Solutions Architect

 

Target Company:

Company Size:

Occupation:

Security/Protective Services

·         Security Intelligence & Analysis

IT/Software Development

·         Computer/Network Security

·         Software/System Architecture

 

Target Locations:

Selected Locations:

US-GA-Atlanta North

Relocate:

Yes

Willingness to travel:

Up to 100%

 

Languages:

Languages

Proficiency Level

German

Beginner

Hindi

Beginner

Telugu

Beginner

Urdu

Beginner