Hadoop Developer Resume
Phoenix, AZ
SUMMARY:
- 8+ years' experience in software development includes Analysis, Design, Development, Testing and Maintenance of web based, multi - tier, real time business applications using Java/J2EE technologies.
- 3+ years' experience in Big D technologies and Hadoop ecosystem projects like Map Reduce,YARN, HDFS, Apache Cassandra , NoSQL, HBase, Oozie, Hive, Tableau, Sqoop, Pig, Storm, Kafka, HCatalog, Zoo Keeper and Flume .
- Extensive experience in working with Struts and Spring MVC (Model View Controller) architecture for developing applications using various Java/J2EE technologies like Servlets, JSP, JDBC, JSTL.
- Proficiency in frameworks like Struts, Spring, Hibernate .
- Configured Splunk to perform the web analytics for the sites
- Hands-on knowledge of REST and SOAP web services, XSD, WSDL, XML .
- Experience in analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java
- Experience in client side Technologies such as HTML, DHTML, CSS, JavaScript, AJAX, jQuery, JSON.
- Knowledge in UI development, UX design , Frontend development, rich user interface design, development, documentation, web 2.0 specifications , standards, media (screen, mobile, touch interface) compatibilities, application architecture, visual design, wireframes, project and team management.
- Experienced in developing applications using HTML, HTML5, CSS, CSS3, JavaScript, Angularjs, Tiles, Adobe Flex4, DHTML, DOM, Flash, XHTML, jQuery, and Ajax.
- Expertise in building strong websites confirming Web 2.0 standards using Yahoo User Interface (YUI) Framework, JQuery, HTML5, XHTML, DHTML & CSS3 , PHP to develop valid code and table-free sites.
- Experience in implementing Auto Complete/Auto Suggest functionality using Ajax, JQuery, DHTML, Web Service call and JSON.
- Experience in Drupal Themes developing Drupal custom modules, resolving Drupal configuration issues.
- Excellent experience in creating Dynamic Wireframes, HTML Wireframes , Dynamic Screenshots and Sitemaps.
- Good Understanding of Document Object Model (DOM) and DOM Functions.
- Experience in designing UI patterns and UI applications with the help of Adobe products like Adobe Dreamweaver CS3 , Adobe Photoshop CS3/CS4 , Adobe Fireworks CS3 and Adobe Illustrator CS3 .
- Expertise in Client Side designing and validations using HTML, HTML5 and Java Script .
TECHNICAL SKILLS:
Operating Systems: Windows, LINUX, UNIX
Java Technologies: JDK, J2SE, J2EE, Spring, Hibernate, Struts, Servlet, JSP, EJB, JDBC and JNDI
Web Technologies: HTML, CSS, AJAX, JavaScript, AngularJS, jQuery
Web Services: SOAP, REST, WSDL, XML, XSD.
Databases: Oracle, MS SQL Server, HBase, Hive
Application / Web Servers: BEA Web Logic, Apache Tomcat, Websphere.
Big Data Technologies: Hadoop, HDFS, Apache Hive, Map Reduce, Pig, Sqoop, Flume, Zookeeper, Impala, NoSQL, HBase, Hue, HCatalog, Hortonworks, Amazon Web Services.
Messaging Services: ActiveMQ, Kafka.
SDLC methodologies: Agile, Waterfall.
IDEs and Tools: Eclipse, Net Beans, Jboss Dev Studio, MyEclipse
Build Tools: Apache-ANT, Maven
Logging Tools: Log4j
Frame Works: Jakarta Struts 1.x, Spring 2.x
Test Framework: JUnit
PROFESSIONAL EXPERIENCE:
Confidential, Phoenix, AZ
Hadoop Developer
Responsibilities:
- Being a ground up project, we have developed the entire application from scratch and I have worked mainly on writing code for Kafka Producer and Kafka Consumer as per our requirement.
- After persisting the data into Kafka brokers successfully, It is written to a flat file from where we load it into HIVE table.
- Defined and created the structure of Hive table on one side and Hbase table on the other side.
- Explored into Hive query for JsonSerde and making our data (message) compatible JsonSerde so it can be loaded successfully into hive.
- Faced many challenges unique to this project and got a chance to explore new and logical ways to handle them.
- Writing queries for real-time attribution and batch attribution.
- Coordinated with team mates and carried out many changes required in their codes in order to receive the data in format required in my section as a good team player.
- Attended meetings (Technical & Business requirement) and gave input at critical phases. Also gathered requirement directly from business team.
- In spite of fast paced Agile environment and aggressive timelines successfully managed to make last minute changes and delivered Release 1 deliverables.
- Wrote Test cases for Component testing and System Integration Testing (SIT).
- Worked with the infra-structure team to deploy the application successfully and overcame different environment related issues.
- Did E2E testing on local machine as well as on servers including the connectivity testing with another team.
- Researched on data available from Google’s “Double-Click” and put that to use for our requirement.
- Communicated back and forth with multiple teams like Business Team, End to End Team, UI Team and Testing team to align not just the project flow but also the timelines.
Environment : Maven project using, Jboss Dev Studio, Jboss server, ActiveMQ, AMQ Resource adaptor, SoapUI, Restful Web service, WSO2, Load Balancer, Kafka, Zookeeper, Mapr, Hbase, and Hive.
Confidential, Bowie, MD
Hadoop Developer
Responsibilities:
- Working with engineering Leads to strategize and develop data flow solutions using Hadoop, Hive, Java, Perl in order to address long-term technical and business needs.
- Developed data pipeline using Flume, Sqoop, Pig and Java map reduce and Spark to ingest customer behavioral data and purchase histories into HDFS for analysis.
- Develop and run Map-Reduce jobs on a multi Peta byte YARN and Hadoop clusters which processes billions of events every day, to generate daily and monthly reports as per user's need.
- Working with R&D, QA, and Operations teams to understand, design, develop and support ETL platform(s) and end-to-end data flow requirements.
- Used Storm to consume events coming through Kafka and generate sessions and publish them back to Kafka.
- Developing tools and leverage existing applications (open source) to ensure data management objectives are met in terms of data quality, data integrity, and data monitoring.
- Building data insightful metrics feeding reporting and other applications.
- Various activates supporting team, like mentoring and training new engineers joining our team and conducting code reviews for data flow/data application implementations.
- Implementing a technical solution on POC's, writing programming codes using technologies such as Hadoop, YARN, Python and Microsoft SQL Server.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Optimizing Map reduce code, pig scripts, user interface analysis, performance tuning and analysis.
- Planned, implemented, and managed Splunk for log management and analytics
- Hands-on experience on Hadoop tools like Mapreduce, Hive.
- Ability to understand and capture technical as well as business requirements.f
- Data analysis and design in order to handle huge amount of data.
- Delivering tuned, efficient and error free codes for new Big Data requirements using my technical knowledge in Hadoop and its Eco-system.
- Storing, processing and analyzing huge data-set for getting valuable insights from them.
Environment: Apache Cassandra, Hadoop, YARN, Cloudera Manager, Redhat Linux, Cent OS, Java, NoSQL-Kafka, Perl, Cloudera Navigator.
Confidential, Mountain View, CA
Hadoop Developer
Responsibilities:
- Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem
- Installed and configured Hive on the Hadoop cluster
- Work closely (face-to-face) with the accountants, financial analysts, data analysts, data scientists, statisticians, compliance, sales, marketing, pricing strategists, product development, and business analysts to create solutions for their issues.
- Developed Simple to complex Map/reduce streaming jobs using Java language that are implemented
- Using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Effectuate server relocation projects and participate in 3rd party audits when working with TuroTax.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
- Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior.
- Tested Apache(TM) Tez, an extensible framework for building high performance batch and interactive data processing applications, on Pig and Hive jobs.
- Used Impala to query the Hadoop data stored in HDFS.
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Exported the analyzed data to the relational databases using HIVE for visualization and to generate reports for the BI team
- Perform data analysis on large datasets and present results to risk, finance, accounting, pricing, sales, marketing, and compliance teams.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Written multiple MapReduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats
Environment: Hadoop - PIG, Hive, Cloudera manager, 30 Node cluster with Linux-Ubuntu
Confidential, Las Vegas, NV
Programmer Analyst
Responsibilities:
- Front-end UI is designed using JSP, HTML, CSS, JavaScript and AJAX.
- Validation for the user input in the advanced search was written using JavaScript.
- Used AJAX and JavaScript for developing some part of web pages.
- Generated dynamic Web pages with Java Server Pages (JSP)
- Implemented Struts Action Forms, Action classes and templates and created action mapping using Struts.
- Performed data validation in Form beans and Action Classes.
- Applied Model-View-Controller (MVC) design paradigm for designing the application Used JAXB to convert object model to XML and vice-versa for Web Service calls.
- Configuration and Deployment of Application on WebSphere Application Server.
- Used CSS for look and feel.
- Used JMS Tibco properties like Boolean property, String Property.
- Used Eclipse IDE.
Environment: Java 1.6, Eclipse, Hibernate, Struts, JSP, Maven, WebSphere Application Server, HTML, CSS, XML, JQuery, Java Script, SQuirrel, Tortoise, JAXB, JMS Tibco.
Confidential, Westbury, NY
JavaScript Developer
Responsibilities:
- Created and updated user interfaces for a public facing website, which generated $2 million, per year in sales.
- Made tweaks to existing webpages and built new webpages using HTML, CSS, JavaScript, ASP, and PHP.
- Developed a complex script in Classic ASP to retrieve Sales data, from the firm's FoxPro database. It generated a report showing which Orders were placed by Salespeople and which Orders were placed directly by customers online. Prior to this, the management was not able to tell which orders were placed by customers directly online
- Wrote and updated SQL Queries in ADO and PHP for processing Sales and Purchase Orders and also generate Sales reports from the firm's database in Excel.
- Developed a script which included SQL Queries against 7 tables to generate a Purchase Order Report that showed the shipment status of products