Sei sulla pagina 1di 8

Ranga Korivi

469-310-3365
annits1010@gmail.com

SUMMARY
?
Over7 Years of professional experience in designing, developing, integrating and testing
software applications, which includes 4+ year of experience in various Big Data technologies
of Hadoop like Map-Reduce, Hive, Pig, HBase, Spark, Sqoop and Flume and 3+ years of
experience in Java EE.
?
Hands on experience in programming and implementation of Java and Python codes
with strong knowledge in Object Oriented Concepts.
?
Good experience in Data warehousing using different relational database management
systems like Oracle, DB2, MySQL and Microsoft SQL server in-addition to sound knowledge of
using NoSQL systems like HBase.
?

Proficient in using various IDEs like Eclipse, MyEclipse and NetBeans.

?
Expertise inimporting and exporting different formats of Data into HDFS, HBase and Hive
from different RDBMS databases and vice-versa.
?
Hands on experience in working with User Defined Functions in Hive and Pig using Java
and Python scripts.
?
Efficient in writing Map-Reduce programs using Apache Hadoop API for analyzing
structured and unstructured data.
?
Skilled at extraction, transformation and analysis of Big Data using Sqoop, pig and hive
respectively.
?
Good at optimizing and debugging Hive-QL queries, Pig-Latin scripts and Map-Reduce
programs.
?

Expert understanding of design patterns with strong analytical skills.

Experience in Big Data solutions for traditional enterprise businesses.

?
Proficient in gathering requirements, analysis, validation, business requirements
specifications and functional specifications for schema creations and table creations.
?

Extensive experience in all phases of Software development life cycle (SDLC).

?
Hands on experience in tuning mappings with expertise in identifying and resolving
performance bottlenecks in various levels.

?
Excellent skills in analyzing system architecture usage, defining and implementing
procedures
?

A quick learner, punctual and trustworthy.

?
Motivated problem solver and resourceful team member with decent written and verbal
communication skills.

TECHNICAL SKILLS

Development Technologies

JDK, Python, Unix Shell Scripting and Oracle PL/SQL

IDEs

Eclipse, MyEclipse and NetBeans.

Big Data Ecosystems

Hadoop-0.2, Hadoop-2.0.0, Hadoop 2.3.0.

Hadoop Distributions

CDH 3, CDH 4.2.1, CDH 5.0.3.

Operating Systems

Linux, Unix and Windows.

Databases

MySQL, Oracle, MS Access, MS SQL server, Oracle and HBase.

File Formats

Compressed file, Text, XML and JSON.

EDUCATION
?

Bachelor of Technology, JNTU, India.

PROFESSIONAL EXPERIENCE

Senior Hadoop Developer


Progressive Insurance, Mayfield Village, OH

Jan 2015 Present

Designing how data in Hadoop was going to be processed to make doing BI analysis on the data easier;
wrote a set of SQL-like (Hive) jobs implementing parts of the design and developed code that ingests
gigabytes of data into Progressive's Hadoop cluster.
Responsibilities:
?
Analyze user requirements, procedures, and problems to automate or improve existing
systems and review computer system capabilities, workflow, and scheduling limitations.
?
Prepare technical design documents based on business requirements and prepare data flow
diagrams.
?

Implement new design as per technical specifications.

?
Integrated Hadoop with Oracle in order to load and then cleanse raw unstructured data in
Hadoop ecosystem to make it suitable for processing in Oracle using stored procedures and
functions.
?
Experience in using Map-Reduce programming model for Batch processing of data
stored in HDFS.
?
Developed Java Map-Reduce programs on log data to transform into structured way to find
user location, login /logout time and spending time, errors.
?

Load and transform large sets of structured, semi structured and unstructured data.

?
Used SQOOP for importing data into HDFS and exporting data from HDFS to oracle
database
?
Built re-usable Hive UDF libraries for business requirements which enabled users to use
these UDF's in Hive querying
?
Used Hive to analyze the partitioned and bucketed data and compute various metrics for
reporting.
?
Optimized the Hive tables using optimization techniques like partitions and bucketing to
provide better performance with Hive-QL queries.
?
Developed Pig Scripts for ETL kind of operation on captured data and delta record
processing between newly arrived data and already existing data in HDFS.
?

Extensively used Pig for data cleansing.

?
Used Pig to do transformations, event joins, filter boot traffic and some pre-aggregations
before storing the data onto HDFS.

?
Experienced in extending Hive and Pig core functionality by writing custom Pig UDFs using
Java and Python.
?

Performance Optimization for Map-Reduce, Pig, Hive scripts.

?
Worked extensively in performance optimization by adopting/deriving at appropriate design
patterns of the Map-Reduce jobs by analyzing the I/O latency, map time, combiner time, reduce
time etc.
?

Troubleshooting: Used Hadoop logs to debug the scripts.

Environment: Big Data Platform CDH 5.0.3, Hadoop HDFS, Map Reduce, Hive, Sqoop, Pig, Flume,
HBase, MRUnit, PigUnit, Junit, Java, Shell Scripts, Oracle 10g, Eclipse Advanced Sql and PL/SQL

Senior Hadoop Developer


Percento Technologies,Houston, TX

Jul 2013 Dec 2014

Percento Technologies is a networking company that transforms how people connect, communicate and
collaborate. At Percento Technologies, very large datasets about customers, products, and network
activity represent hidden business intelligence. This project is aimed for finding management insights to
unable decision making by analyzing the past sales in order to increase Percento Technologies revenues
by identifying hidden opportunities for partners to sell services. Which requires the processing of
unstructured data and very large data sets far more quickly and at far less cost in-addition to moving the
customer install base and service contracts to the Hadoop platform to provide more value in the future
because of the reusability to other Percento Technologies business teams for their own initiatives.
Responsibilities:
?
Integrated, managed and optimized utility systems, including assets, devices, networks,
servers, applications and data.
?
Ensured quality integration into the overall functions of smart meters into the system data
acquisition and processing.
?
Enabled the use of metering data for a variety of applications such as billing, outage
detection and recovery, fraud detection, finance, energy efficiency, customer care and a variety of
analytics.
?
Analyzed large amounts of raw data in an effort to create information. Compiled technical
specifications that allowed IT to create data systems, which supported the smart metering system.
?
Responsible for technical reviews and gave the quick-fix solution for the customer on
production defects.
?
Developed Map-Reduce programs to parse the raw data, populate staging tables and
store the refined data in partitioned tables in the enterprise data warehouse (EDW).

?
Developed custom UDFs in Pig Latin using Python scripts to extract the data from sensor
devices output files to load into HDFS
?
Worked with Importing and exporting the data using Sqoop from HDFS to Relational
Database systems and vice-versa.
?

Written Map-Reduce java programs to analyze the log data for large-scale weather data sets.

Involved in testing Map-Reduce programs using MRUnit and JUnit testing frameworks.

Customize parser loader application of Data migration to HBase.

Provide support for data analysts in running ad-hoc Pig and Hive queries

?
Developed PL/SQL Procedures, Functions, and Packages using Oracle Utilities like PL/SQL,
SQL Loader and Handled Exceptions to handle key business logic.
?
Utilized PL/SQL bulk collect feature to optimize the ETL performance. Fine-Tuned and
optimized number of SQL queries and performed code debugging.
?
Developed UNIX & SQL script to load large volume of data for Data Mining &Data
Warehousing.
Environment: Big Data Platform - CDH 4.2.1, Hadoop HDFS, Map Reduce, Hive, Sqoop, Pig, Flume,
IBM DB2, PL/SQL, UNIX, Python,Eclipse.

Hadoop Developer
BB&T, NC

Jan 2013 Jun 2013

Involved in design and development phases of Software Development Life Cycle (SDLC) using Scrum
methodology. Designed and coded application components in an agile environment utilizing a test driven
development approach.
Responsibilities:
?
Involved in design and development of server side layer using XML, JDBC and JDK patterns
using Eclipse IDE.
?

Involved in unit testing, system integration testing and enterprise user testing.

Extensively used Core Java, Servlets, and JDBC.

?
Developed data pipeline using Flume, Sqoop, Pig and Map Reduce to ingest customer
behavioral data and purchase histories into HDFS for analysis.
?
Worked with NoSQL databases like Hbase in creating tables to load large sets of semi
structured data coming from various sources.
?

Wrote MRUnit test cases to test and debug Map Reduce programs in local machine.

Involved in creating Hive tables, loading data and running hive queries in those data.

Imported data using Sqoop to load data from Oracle to HDFS on regular basis.

Developed scripts and Batch Jobs to schedule various Hadoop Program.

Written Hive queries for data analysis to meet the business requirements.

Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.

?
Written Hive queries to parse the logs and structure them in tabular format to facilitate
effective querying on the log data.
?

Developed Pig UDFs to pre-process data for analysis.

Developed Complex and Multi-Step data pipeline using Spark.

Written Spark SQL queries for data analysis.

Environment: Big Data Platform - CDH 4.0.1,XML, Hadoop HDFS, Pig, Hive, Sqoop, Flume, Spark,
Oracle 10g, Java, Eclipse.

Hadoop Developer
Self Reliance Federal Credit Union, NY
Oct 2011 Dec 2012
The genesis history container provides retail risk data integration from 160 countries banking data.
Primary retail bank products like credit card, Mortgages and loans account level data were collected
through standard template. All data from different countries were staged at Hadoop data clusters as part
of retail risk data ingestion process. Then data were profiled against tech checks, sanity checks, data type
checks, data range checks, domain checks, business checks and outliers checks as requirements from
FED banking regulatory standards. Data were processed, validated and profiles in Hadoop and profiling
results were captured in the Hive tables (External table) and persisted back to Oracle DB. The results
were submitted for business evaluation. Once approved from business, respective data were aggregated
at different level with data enrichments. Aggregation and enrichment of data were performed using Hive
Job process.
Responsibilities:
?
Involved in analysis, design and development of data collection, data ingestion, and
dataprofiling and data aggregation.
?

Working in development of controller, Batch and logging module using JDK 1.6.

?
Worked on development of data ingestion process using FS Shell and data loading into
HDFS.

?
Working in the definition of Hive query for different profiling rules like business checks,
outliers checks and domain and data range validation.
?

Working on the automating the generation of Hive query and Map-Reduce programs.

?
Developed User Defined Function in java and python to facilitate data analysis in Hive and
pig.
?

Managed the end-to-end delivery during the different phase of the software implementation.

?
Involved in initial POC implementation using Hadoop Map Reduce, PIG Latin Scripting,
and Hive Scripting.
?
Designed the framework for Data Ingestion, Data Profiling and generating the Risk
Aggregation report based various business entities.
?

Mapped the business requirements and rules with the Risk Aggregation System.

Used JDBC to invoke Stored Procedures and database connectivity to ORACLE.

Code debugging and creating Documentation for future use.

Environment: Big Data Platform - CDH 3, Map-Reduce, Hive, Pig LATIN Scripting, JDK 1.6, and Oracle.

Java Developer
Next Step Solutions, India
Aug 2009 Sep 2011
This is an e-commerce project for online shopping. It allows customers to make selective payments via
credit card/EFT (Electronic Fund Transfer). It also provides customers to Lookup for Phone parts,
shopping and ordering. Customers can view their open Orders, and Invoices. Customers can change their
default Credit Card and EFT information.
Responsibilities:
?
Involved in various stages of Projects from Architecture Designing, Business Analysis,
Development, Testing and finally Production Stage.
?

Designed and developed different modules as part of project using Java/J2EE.

?
Contributed to an effective order processing system and simplified the existing order
process, which proved to be more efficient.
?

Developed session beans as an enterprise business service object.

Used JDBC, Application server provided transaction API for accessing data from Oracle.

Involved in developing the unit test classes using J-Unit.

Used CVS for version control integrated with WSAD.

Used JavaScript for client side validations

Used Cascading Style Sheets in the application.

Involved in Development of User Interface using and JSPs.

Used Tomcat as the application server in the application.

Environment:Core Java, JDK 1.3, J-Unit, WSAD,Oracle, JavaScript, JDBC, EJB, CSS, HTML,Tomcat
application server.

Java Developer
Visual Soft, India
Jun 2008 Jul 2009
This project attempts to automate the income tax and the sales tax procedures. This project is not specific
to any country, but to display the working of the model, the guidelines and business rules specific to the
Indian governments tax procedures have been considered.
Responsibilities:
?

Gather user requirements and followed by analysis and design.

Worked on the technical design to conform the framework.

Developed JSPs, action classes, form beans, response beans, EJBs.

Coded Servlets for the Transactional Model to handle many requests.

?
Developed business objects and business object helpers, which interact with middleware
stubs.
?

Implemented business delegate pattern to separate view from business process.

Extensively used XML to code configuration files.

Developed PL/SQL stored procedures, triggers.

Developed complete Web tier of the application with Struts MVC framework.

Performed functional, integration, system, and validation testing.

Environment:JDK1.3, JSP, Apache Struts 1.0, Servlets, EJB 2.1, XML, JDBC, Eclipse, JBOSS, PL/SQL,
Oracle 9i, Rational Rose, UNIX, MVC framework, JUnit, Rational Clear Case.

Potrebbero piacerti anche