Sei sulla pagina 1di 6

Ponnam R

Mobile: 248 987 6154


Gmail ID: krishna@krishradiants.com

Professional Summary

• Over 8 years of extensive IT experience in all phases of software development life cycle
(SDLC) including 4+ years of hands on experience working with Hadoop, HDFS, Map Reduce
frame work and Hadoop ecosystem like Hive, Hue, Pig, Sqoop, HBase, Zookeeper, OOzie,
Kafka and Apache spark.
• Excellent Programming skills at a higher level of abstraction using Scala and Spark.
• Thorough knowledge with the data extraction, transformation and load in Hive, Pig and
HBase.
• Managing and scheduling batch Jobs on a Hadoop Cluster using Oozie.
• Experienced using Sqoop to import data into HDFS from RDBMS and vice-versa.
• Managing and scheduling batch Jobs on a Hadoop Cluster using Oozie.
• Created Hive tables in raw, staging and presentation layer to store the data from HDFS.
• Used Informatica BDM to moved data from raw to stag and stag to presentation layer.
Excellent understanding of Hadoop architecture and various components such as HDFS, Job Tracker,
Task Tracker, NameNode, Data Node and MapReduce programming paradigm.

• Extensive experience in writing Map Reduce, Hive, PIG Scripting and HDFS.
• Good understanding and knowledge of NOSQL databases like MongoDB, HBase, and
Cassandra.
• Worked on Importing and exporting data into HDFS and Hive using Sqoop.
• Wrote Hive queries for data analysis to meet the requirements
• Strong working knowledge and ability to debug complex problems.
• Worked with Apache Spark which provides fast and general engine for large data
processing.
• Integrated with functional programming language Scala.
• Extending HIVE and PIG core functionality by using custom User Defined Function (UDF),
User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions
(UDAF) for Hive and Pig.
• Acquaintance in requirement extraction, analysis and design document preparation.
• Excellent oral and written communication skills.
• Collaborated well across technology groups.
• Major strengths include the ability to learn new technologies quickly and adapt to new
environments.
• Have a good experience working in agile development environment including Scrum
methodology.

Technical Skills

• Programming Languages • SQL, Java, J2EE, Scala and Unix shell scripting
• Big Data Ecosystem • Hadoop, HDFS, MapReduce, Hive, Pig, HBase, Impala, Hue, Sqoop,
Kafka, Oozie, Flume, Zookeeper, Spark, Cloudera and Hortonworks,
Informatica BDM 10.2.0, Ranger policy, putty, Linux servers, Hue
graphical user interface
• Databases & NoSQL • Oracle, Teradata, MySQL, SQL Server, DB2, Familiar with NoSQL-
HBase
• Scripting&Query Languages • UNIX Shell scripting, SQL and PL/SQL.
• Hadoop Paradigms • Map Reduce, YARN, In-memory computing, High Availability, Real-
time Streaming.
• Other Tools • Eclipse, IntelliJ, SVN, GitHub, Jira, BitBucket.
• Methodology • Agile, waterfall
Project Details:
Client : Eli Lilly, Indianapolis, IN
Role : Hadoop Developer
Duration : January 2018 – June 2019

Responsibilities:

• Experience in working with Informatica BDM 10.2.0.


• Experience in using putty, Ranger policy, Hive.
• Experience in importing and exporting data using Informatica mappings HDFS to Relational
Database Systems and vice-versa.
• Design and develop ETL code using Informatica Mappings to load data from heterogeneous
Source systems like flat files, XML’s, MS Access files, Oracle to target system Oracle under Stage,
then to data warehouse and then to Data Mart tables for reporting.
• Developed ETL with SCD’s, caches, complex joins with optimized SQL queries.
• Created workflow instances, and improved the performance with large data using round robin,
Key range partitions, pushdown optimization.

• Hands on experience in application development using Linux shell scripting and HQL.
• Knowledge in job workflow scheduling and monitoring tools like CA Workload Automation
(CAWA).
• Experience in working with service now as chain process management.

• Environment: Hadoop, HDFS, Hive, MySQL, Service Now, CAWA, Informatica


Admin console, Informatica BDM 10.2.0, shell scripting, Python, Spark.

Client : Aetna, Hartford, CT


Role : Hadoop Developer
Duration : Feb 2017 - December 2018

Aetna is one of the leading Insurance Provider in the United States. A key purpose of this project is to
generate various reports for Personal Insurance domain and my work duties encompasses performing
analysis on the effectiveness of huge data & storing terabytes of information received & generated by
policy issuing systems.
It’s an enhancement project where data is collected from xml’s & RDBMS and migrated to HDFS for
further processing, various major and minor enhancements were made in the system for storing data in
Hive tables.
Recently we have started using HBase for storing data.

Responsibilities:
• Coordinated with the BA team for finalization of requirements.
• Responsible for generating actionable insights from complex data to drive real business
results for various application teams.
• Provided solutions by writing scripts in PIG Latin to process HDFS data.
• Effectively used Sqoop to migrate data from RDBMS to HDFS.
• Effectively used SerDe to load data in Hive table.
• Developed Spark code by using Scala and Spark-SQL for faster processing and testing.
• Experience in retrieving data from databases like MYSQL and DB2 into HDFS using
Sqoop and ingesting them into HBase.
• Worked with Apache Spark which provides fast and general engine for large data
processing integrated with functional programming language Scala.
• Experience data processing like collecting, aggregating, moving from various sources
using Apache Flume and Kafka.
• Developed Hive Queries to analyze the data in HDFS to identify issues and behavioral
patterns.
• Experienced on loading and transforming of large sets of structured, semi and
unstructured data.
• Monitored Hadoop scripts which take the input from HDFS and load the data into Hive.
• Effectively used Oozie to develop automatic workflows of Sqoop, Mapreduce and Hive
jobs.
• Post implementation support and various production fixes, discrepancies.
• Experienced with handling administration activations using Cloudera manager.
• Extensively used TOAD to test, debug SQL and PL/SQL Scripts, packages, stored
procedures and functions. And TOAD for DB2 to manage the DB2 database.
• Created Autosys JILs, and scheduled the jobs to run the shell scripts to execute the SQL
and Workflows.
• Responsible for Code Migration in to various environments like Testing, staging and
Production.
• Maintained warehouse metadata, naming standards and data warehouse standards for
future application development
• Developed Shell scripts to automate routine DBA tasks.
• Extensively involved in Installation and configuration of Cloudera distribution Hadoop,
Name Node, Secondary Name Node, Job Tracker, Task Trackers, and Data Nodes.
• Delivered successfully various projects in this application within stipulated timeline.

Environment: Hadoop, MapReduce, HDFS, Pig, Hive, HBase, MySQL, Spark, Cassandra, Pig, Netezza,
Sqoop, Oozie, Version one, Shell, Map Reduce, SVN.

Client : Nike, Inc. Beaverton, OR


Role : Hadoop Developer
Duration : April 2016 – Nov 2016
Description: The purpose of the project is to analyze the data coming from the different sources into the
Hadoop data lake.

Responsibilities:
• Responsible for building scalable distributed data solutions using Hadoop.
• Interacted with different system groups for analysis of systems.
• Created tables, views in Teradata, according to the requirements.
• Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
• Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop.
• Implemented best income logic using Pig scripts and UDFs.
• Implemented test scripts to support test-driven development and continuous integration.
• Used Sqoop to efficiently transfer data between databases and HDFS and used Flume to stream
the log data from servers/sensors.
• Involved in Hadoop cluster task like Adding and Removing Nodes without any effect to running
jobs and data
• Managed and reviewed Hadoop Log files
• Hands on experience in AWS Cloud in various AWS services such as Redshift cluster, Route 53
domain configuration.
• Responsible for smooth error-free solution and Integration with Hadoop
• Designed a data warehouse using Hive
• Used Control-m and oozie scheduling tool to manage interdependent Hadoop jobs and to
automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as
system specific jobs
• Involved in Scrum calls, Grooming and Demo meeting, Very good experience with agile
methodology.

Environment: Hadoop, MapReduce, HDFS, Pig, Hive, HBase, MySQL, Control-M, Ubuntu, Oracle,
Spark, Java, Cassandra, Pig, Sqoop, Oozie, AWS,Version one, Shell, Map Reduce, SVN.

Client : Belk, Inc. Charlotte, NC


Role : Hadoop Developer
Duration : Mar 2014 – Dec 2015

Description: Designed and developed big data solutions involving Terabytes of data. The big data
solution consists of collecting large amounts of log data from distributed sources, transformations and
standardizations analysis, statistics, aggregations and reporting etc. Built an on-demand
elastic Hadoop cluster infrastructure to cater the needs of various Big Data projects, automated various
Big Data workflows to process and extracts analytics out of the data using MapReduce, Pig and Hive.

Responsibilities:

Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem.
• Used Sqoop to transfer data between databases (Oracle & Teradata) and HDFS and used Flume to
stream the log data from servers.
• Developed MapReduce programs for pre-processing and cleansing the data in HDFS obtained
from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
• Experienced in managing and reviewing Hadoop log files.
• Involved in loading data from UNIX file system to HDFS.
• Load and transform large sets of structured, semi structured and unstructured data.
• Extensively worked on creating combiners, Partitioning, distributed cache to improve the
performance of Map Reduce jobs.
• Implemented Different analytical algorithms using map reduce programs to apply on top of
HDFS data.
• Used Pig to perform data transformations, event joins, filter and some pre-aggregations before
storing the data onto HDFS.
• Implemented Partitions, Buckets in Hive for optimization.
• Involved in creating Hive tables, loading structured data and writing hive queries which will run
internally in map reduce way.

Environment: Apache Hadoop, Cloudera, Hive, Pig, Sqoop, Zookeeper, HBase, Java, Oozie,
Oracle, Teradata, and UNIX Shell Scripting.

Client : Zoom MD, HYD


Role : Java/J2EE developer
Duration : June 2008 – Aug 2011

Responsibilities:

Development of Java code to meet specifications and designs an d using best practices.
• Development of low level component base design documentation (UML).
• Developed the DAO layer for the application using Spring Hibernate Template support.
• Implemented Transactions using spring framework.
• Used Spring MVC and Web Flow to bind web parameters to business logic.
• Implemented Ant and Maven build tools to build jar and war files and deployed war files to target
servers.
• Maintained relationship between objects using Spring IOC.
• Extensively written COREJAVA & Multi-Threading code in application
• Implemented the JMS Topic to receive the input in the form of XML and parsed them through a
common XSD.
• Written JDBC statements, prepared statements, and callable statements in Java, JSPs and Servlets.
• Followed Scrum approach for the development process.
• Modified and added database functions, procedures and triggers pertaining to business logic of the
application.
• Used TOAD to check and verify all the database turnaround times and also tested the connections for
response times and query round trip behavior.
• Used ANT Builder to build the code for production line.
• Used Eclipse IDE for all recoding in Java, Servlets and JSPs.
• Used JSP Tag Libraries (JSTL) to implement the logic inside the JSPs.
• Extremely used plain JavaScript and JQuery, JavaScript Library to do the client side validations.
• Used AJAX to get the data from the server asynchronously by using JSON object.
• Used JIRA as a bug-reporting tool for updating the bug report.
• Focus on converting the existing features in the application towards globalization of the application,
which is internationalization of the web representation.
• Have worked on Oracle 10g data base for storing and retrieving the application data.
• Involved configuring JMS in application developer.
• Developed MQ JMS Queues for asynchronous messaging and Web Services using SOAP/WSDL.
• Involved in Web Logic administration like setting up Data Sources, deploying applications.
• Configured and Deployed the Web Application Achieve (WAR) in Web Logic Application Server.

Environment:Core Java, J2EE, Servlets, JavaScript, XML, Tomcat, Web logic, SOAP, Eclipse, AJAX,
SVN, JDBC, Web Services, XSLT, CSS, DOM, HTML, Maven, JSP, ANT, DB2, JUnit, Oracle.

Education Details:

Bachelors in computer science from JNTU HYDERABAD, 2008


Masters in Computer Science from University Of Newhaven, CT in 2013

Potrebbero piacerti anche