Hadoop Developer & Administrator Resume
Columbus, OH
SUMMARY:
- 2 1/2 years of experience in software development life cycle design, development, and support of systems.
- Extensive working experience on Hadoop Ecosystems like HDFS, Hive, Drill, Sqoop, Pig, Spark, MapReduce, HBase, Oozie, MongoDB and NoSQL (HBase, MongoDB)
- Implemented various optimization techniques such as Vectorization, Partitioning and Bucketing concepts using Hive(HQL) and have excellent skills on developing SQL scripts using Pig, Hive.
- In Depth understanding of Hadoop Architectures like Yarn and MapReduce.
- Good Knowledge on Streaming applications like Flume, Kafka and Spark Streaming.
- Equipped with other technical skills - SparkSQL, Scala, Cloudera 5.10.1, Hortonworks 2.3, Shell Scripting, Objects Oriented Programming concepts and Scrum Methodology.
- Well versed with data visualization tools like Tableau and PowerBi with various gauge meters, pipe graphs etc.
- Enthusiastic in learning new technologies.
- Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.
- Strong Communication skills of written, oral, interpersonal and presentation.
TECHNOLOGIES:
Big Data Ecosystems: Hadoop, MapReduce, HDFS, Zookeeper, Hive, Pig, Sqoop, Oozie, Flume, Kafka, Yarn, Spark, Cloudera, Hortonworks and MapR
Programming Languages: C, Core-Java 1.6, Scala, Shell Scripting
Databases: MySQL, PL SQL and Oracle.
NoSQL Databases: HBase, Cassandra, MongoDB
Data Visualization: Tableau, PowerBi
Operating Systems: Windows XP, 7, 10, IOS, Linux, Ubuntu 10.10, 14.04, 16.04
MS Office: MS Excel, MS Access, MS PowerPoint
Cloud Technologies: AWS, Azure
Application Servers: Tomcat, WebSphere, Jenkins
Version Controls: GitHub, CVS, Jira
Methodologies: Agile, Waterfall
PROFESSIONAL EXPERIENCE:
Confidential, Columbus, OH
HADOOP DEVELOPER & ADMINISTRATOR
Responsibilities:
- Involved in identifying various KPI’s for project and gathering requirement from customers, designed architecture and documenting the data flow of the project.
- Imported data into the cluster from dynamically generated files from various RDBMS located in different geographical locations into HDFS using Sqoop.
- Implemented various optimization techniques such as Vectorization, SerDe’s, Partitioning and Bucketing concepts using Hive(HQL) and developed SQL scripts using all type of joins and applied various operation on datasets.
- Configured big data workflows to run on the top of Hadoop using Oozie and these workflows comprises of heterogeneous jobs like loading new incoming files form staging area to Hive, impala tables and Sqoop.
- Involved in converting Hive/SQL queries into Spark transformations using Spark Confidential .
- Ingested hive/ impala views to Power Bi to generate dashboard with various gauge meters, pipe graphs etc.
- Worked in Agile environment, which uses Jira to maintain the story points.
- Created users and groups using Cloudera Hue Manager, provided access to local system and database.
- Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring, Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
Confidential
Jr. Java Developer
Responsibilities:
- Followed MVC architecture to develop the web applications.
- Designed service design document which contains UML use case diagrams, class diagrams, sequence diagrams and activity diagrams.
- Create database object such as Stored procedures, views, tables, triggers, functions, query tuning, optimization, Indexing.
- Developed various Java classes, SQL queries and procedures to retrieve and manipulate the data from backend Oracle database using JDBC.
- Analysis of business requirements and develop system architecture document for the enhancement project.