Hadoop Developer/administrator Lead Resume
Eagan, MN
SUMMARY
- Results - oriented individual with 9+ years of experience in the Big Data, Cloud & Application Design, Development / Web Development (Java, J2EE), Testing, Delivery and producing successful, high quality products for startup and high growth companies and have been consistently successful in producing products under budget and within time constraints.
- Mentored project teams on Big DataAnalytics using Hadoop for analyzing customer and product matrix enabling marketing personnel for targeting potential customers.
- Hands-on experience with HDFS, MapReduce, Pig, Hive, Flume, HBase and Sqoop.
- Experience in building, maintaining multiple Hadoop clusters (prod, Dev etc.,) of different sizes and configuration.
- Successfully loaded files to Hive and HDFS from Oracle, Netezza,Teradataand SQL Server using SQOOP
- Extending Hive and Pig core functionality by writing Custom UDFs.
- Expertise in creating Custom Serdes in Hive.
- Excellent knowledge in NoSQL databases like HBase, Cassandra and MongoDB.
- Evaluation of ETL and OLAP tools and recommend the most suitable solutions based on business needs.
- Worked on Kafka, Kafka-Mirroring to ensure that the data is replicated without any loss.
- Used tools like SQOOP, Kafka to ingest data into Hadoop.
- Good knowledge with NoSQL databases like HBase, Cassandra.
- Having good experience in the BigDATA - Hadoop, Oracle BigDATA. Good knowledge in Core Components MapReduce, YARN, HDFS. Also Hadoop EcoSystem - Pig, Hive, Spark, Avro, Parquet, Flume, Sqoop, Crunch, HBase, ZooKeeper, NoSQL databases like Cassandra, HBase,
- Handled Administration, installing, upgrading and managing distributions of Cassandra.
- Advanced knowledge in performance troubleshooting and tuning Cassandra clusters.
- Done Scaling Cassandra cluster based on lead patterns.
- Good understanding of Cassandra Data Modelling based on applications.
- Experience with Cassandra Performance tuning.
- Highly involved in development/implementation of Cassandra environment.
- Extensively worked on Database Applications using DB2, Oracle, MySQL, and PL/SQL.
- Hands on experience in application development using Java, RDBMS.
- Strong experience as a senior Java Developer in Web/intranet, Client/Server technologies using Java, J2EE, Servlets, JSP, EJB, JDBC.
- RDBMS experience includes Oracle, PostgreSQL, MySQL and programming using PL/SQL, SQL.
- Work closely with the project teams as outlined in the SDLC engagement model to provide guidance in implementing solutions at various stages of projects
- Architect solutions for key business initiatives ensuring alignment with future state Analytics architecture vision.
- Loaded data from flat data files into SQL Server 2008 database tables using bulk insert, and table export/import; loaded data from flat files and excel sheet into Oracle database using SQL*Loader.
- Provide Consulting to Customers in identifying Big Data use cases and guiding them to implementation
- Define and establish Enterprise Wide Strategic Planning and Architecture for Big Data solutions.
- Excellent understanding of software development methodologies and software development life cycle like Waterfall model, AGILE and SCRUM.
- Exceptional expertise in technologies such as Core Java, java script, HTML5, CSS3, AJAX, XHTML, CSS, jQuery, JQuery Mobile, Adobe Flex and Action Script.
- Troubleshooting with Linux/bash shell scripting.
- Knowledge on Hortonworks, spark.
- Knowledge on Spark Streaming, forFast failure and straggler recovery,Load balancing,unification of streaming, batch and interactive workloads and advanced analytics.
PROFESSIONAL EXPERIENCE
Hadoop developer/Administrator lead
Confidential, Eagan, MN
Responsibilities:
- Determining the viability of a business problem for a Big Datasolution.
- Handled Administration, installing, upgrading and managing distributions of Cassandra.
- Defining a logical architecture of the layers and components of abigdatasolution.
- Selecting the right products to implement abig datasolution.
- Setup and Install Hadoop (With YARN / MapReduce) cluster and EnterpriseDataWare House.
- Build High-Availability (HA) architectures and deployed withBigDataTechnologies.
- Plan and manage HDFS storage capacity. Advise a team on best tool selection, best practices, and optimal processes using Sqoop, Oozie, Hive, Hbase, Pig, Flume and linux/Bash Shell Scripting.
- Facilitate access / ETL to largedatasets utilizing Pig/Hive/Hbase/Impala on Hadoop Ecosystem.
- Some Java applications requirements make integration with a scripting language necessary (java scripting). If your users may need to write scripts that drive the application, extend it, or contain loops and other flow-control constructs.
- Successfully loaded files to Hive and HDFS from Oracle, Netezza and SQL Server using SQOOP
- Worked on Kafka, Kafka-Mirroring to ensure that the data is replicated without any loss.
- Set up Camus for reading the data from Kafka and storing it in HDFS.
- Using some simple (java script)language in pig scripting can drive more analytical reports.
- Install OS and administrated Hadoop stack with CDH5 (with YARN) Cloudera Distribution.Including configuration management, monitoring, debugging, and performance tuning.
- Manage Hadoop operations with multi-node HDFS cluster using Cloudera Manager.
- Manage Massive Parallel Processing with Impala with HBase and Hive.
- Worked on Qlik View to providedataintegration, reporting,java script,datamining and ETL.
- Manageddatasecurity and privacy with Kerberos and role based access.
- Along with the immediate architecture, support provision, monitor, evolve, supportand evangelize the chosen technology stack(s).
- Spark Streaming, forFast failure and straggler recovery,Load balancing,unification of streaming, batch and interactive workloads and advanced analytics.
- My responsibility involves in setting up the Hadoop cluster for the project and also working on the project usingHQL and SQL.
- Advanced knowledge in performance troubleshooting and tuning Cassandra clusters.
- Done Scaling Cassandra cluster based on lead patterns.
- Good understanding of Cassandra Data Modelling based on applications.
- Experience with Cassandra Performance tuning.
- Highly involved in development/implementation of Cassandra environment.
- Later on as per as business requirements the coding is done in Map Reduce by giving suggestion to the junior developers.
- Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and processing.
- Worked on Cluster maintenance, Adding and removing cluster nodes, Cluster Monitoring and
- Troubleshooting, Racks, Disk Topology, Manage and review data backups, Manage and review Hadoop logfiles.
- Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.
- Making the communication between the clusters to be secured I used to implement Kerberos Protocol for safer and secure communication. The data in the namenode is protected in such a way that no other employees can format the data fromthe namenode.
- Making the data replicated where one copy of the data is stored in the data node with password protected.
- Spark Streaming, forFast failure and straggler recovery,Load balancing,unification of streaming, batch and interactive workloads and advanced analytics.w
- Importing and exporting data into HDFS and Hive using Sqoop.
Environment: Hadoop 1.2.1, MapReduce, Sqoop 1.4.4, Hive 0.10.0, Flume 1.4.0, Oozie 3.3.0, Pig 0.11.1, Hbase 0.94.11, Scala,Hcatalog,Zookeeper 3.4.3, Talend Open Studio v1.10, Talend 5.5, Oracle 11g/10g, Apache Cassandra, Datalake, SQL Server 2008, Kafka, MySQL 5.6.2, Java, SQL, PL/SQL, UNIX shell script, Eclipse Kepler IDE
Senior Hadoop developer/Administrator
Confidential, Lake Forest, IL
Responsibilities:
- Interact with customers, business partners and all stakeholders to understand the business objective and drive solutions that effectively meet the needs of a client.
- Sketches thebigdatasolution architecture, then monitors and governs the implementation.
- Design strategies and programs to collect, store, analyze and visualizedatafrom various sources.
- Participated in development and execution of system and disaster recovery processes and actively collaborated in all Security Hardening processes on the Cluster.
- Upgrade the Hadoop Cluster from CDH 4.1 to CDH4.7Cloudera Distribution.
- Support thedataanalysts and developers of BI, java script and for Hive/Pig development.
- As Hadoop Developer, managed review, capacity planning, technical consultation, performance optimization for production clusters.
- Job duties involved the design, development of various modules in Hadoop Big Data Platform and processing data using MapReduce, Hive, Pig,Scoop, Oozie, Kafka and Storm.
- Integrated Apache Storm with Kafka to perform web analytics. Uploaded click stream data from Kafka to HDFS, Hbase and Hive by integrating with Storm.
- Responsible for Hadoop training for team and authored guidelines, best practices, patterns/anti-patterns, checklist, and FAQ documents for creating Hadoop.
- Lead technical solution blueprinting, training and adaption programs for Kerberos-Hadoop (version 0.20.100) and YARN (version 0.23).
- Managed Data Analytics team responsible for consulting and supporting Hadoop applications.
- Co-ordinated with offshore team members in completing the assigned tasks.
- Installing and configuring Hadoop ecosystem like Pig,Hive.
- Hands on experience in Installing, Configuring and managing the Hue andHCatalog.
- Experience in importing and exporting the data using Sqoop from HDFS to Relational Databasesystems/ mainframe andvice-versa.
- ExperiencedofServiceMonitoring,ServiceandLogManagement,AuditingandAlerts,HadoopPlatform Security, and ConfiguringKerberos
- Experience in importing and exporting the logs usingFlume.
- Optimizing performance of HBase/Hive/Pigjobs
Environment: Cloudera DistributionCHD 4.1/4.7/5, Hadoop 1.1 X/2.X,MapR 3.1, Sqoop, Oozie 3.2.0, Pig 0.9, Hbase 0.93.,Apache hive 0.9,Apache Zookeeper, Talend Open Studio 5.5.0, Oracle 11g/10g, Apache Cassandra, Datalake, SQL Server 2008, Kafka, MySQL 5.6.2, Java, SQL, PL/SQL, UNIX shell script, Eclipse Kepler IDE
Hadoop Developer
Confidential, Walnut Creek, CA
Responsibilities:
- Setup and Install Hadoop cluster and EnterpriseDataWare House.
- Build High-Availability (HA) architectures and deployed withBigDataTechnologies.
- Understand technical opportunities and limitations of the various technologies at disposal.
- Determining the viability of a business problem for NoSQL database solution and migration.
- Analyzed large data sets by runningjava script, Hive queries and Pig scripts.
- Worked with the Data Science team to gather requirements for various data mining projects.
- Developed Simple to complex MapReduce Jobs using Hive and Pig.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Created and maintained Technical documentation for launchingHADOOPClusters and for executing Hive queries andPig Scripts.
- Migrating data from Oracle, MySQL in to HDFS using Sqoop and importing various formats of flat files in to HDFS.
- Written java scriptcustomized Pig UDFs in Java where the functionality is too complex.
- Designing and creating Hive external tables using shared meta-store instead of derby with partitioning, dynamic partitioning and buckets.
- Generate final reporting data using QlikView for testing by connecting to the corresponding Hive tables using ImpalaODBC connector.
- Co-ordinated with offshore team members in completing the assigned tasks.
Environment: Apache Hadoop 1.0, MapR 3.0, Sqoop, Hive1.0, Flume, Oozie 3.1.0, Pig 0.9, Zookeeper 3.4.2, Talend Open Studio 5.4.0, Oracle 11g/10g, Apache Cassandra, Impala,SQL Server 2008, MySQL, Java, SQL, PL/SQL, UNIX shell script, Eclipse Kepler IDE
Confidential, Houston, TX
Web Developer/Java Developer
Responsibilities:
- Designed and implemented HTML 5 basedmobileweb applications and user interfaces formobileplatforms (iOS, Android, Windows Phone and Kindle).
- Designed, developed and supported HTML5 based applications, predominantly for use onmobile devices.
- Provided expertise for touch screen UI utilizingHTML5andJavaScriptincluding 2D Canvas, File system API, Asset Loading, Manifest Caching, Progress Bar Indicators, Local Storage and Web Workers for next generationHTML5Application to facilitate sales process.
- Architecture featuring Model-View-Controller modular object orientedJavaScript, AJAX and server-side JSP.
- Documented entire system for stake holder review, wireframe review and iterative enhancement and led the corresponding meetings.
- User Centered Design featuring Personas, Task Models, User Journeys, Content Requirements, Sitemaps and Usability Test Reports.
- Multiple REST Based Web Services with XML Payloads and large datasets normalized into JSON for application Integration. it's sensible to support a scripting language interpreter that can read user scripts, then run them against Java application's classes. To accomplish that task, running a Java-based scripting language interpreter in the same JVM as a application.
- Produced cross-browser compliant web applications based on client provided comps using HTML5, CSS, jQuery, JSON, and MySQL.
- Developed responsive layouts for different screen sizes and resolutions.
- Worked with graphic artists, manipulating images, and precisely matching UI mockups.
- Demonstrated expert-level understanding and proficiency of HTML 5, CSS, AJAX and JavaScript and a strong sense of aesthetics and UI/UX.
- Written semantic relevant HTML.
- Implemented AJAX interactions with back-end services.
- Used Front-end JS libraries like JQuery, angular.js.
- Extensively used JavaScript object-oriented programming.
- Produced top-notch code that maintains the integrity of the system design and ensures compliance of code to craftsmanship standards.
- Provide feedback on functionality requests regarding feasibility and complexity.
Environment: Java,J2EE,JavaScript,MyEclipse,WeblogicHTML5, CSS3, AJAX, XHTML, CSS, jQuery, JQuery Mobile, Adobe Flex and Action Script and XML.
Java Developer
Confidential
Responsibilities:
- Working effectively with Stakeholders (Business and Technical including Executive Management); liaising with third party vendors, global team and system integrators.
- Involved in understanding client requirements and translating them to technical requirements and prepared a detailed Function Specification document (FSD).
- Involved in performance testing&defects fixing.
- Adding Some Java applications requirements make integration with a scripting language necessary. If your users may need to write scripts that drive the application, extend it, or contain loops and other flow-control constructs. In such cases, it's sensible to support a scripting language interpreter that can read user scripts, then run them against Java application's classes. To accomplish that task, running a Java-based scripting language interpreter in the same JVM as a application.
- Implemented Cairngormframework to develop the dynamic user interfaces to provide messaging and Dependency Injection.
- Used services like the Remote object, HTTP Services and Web-Services for data communication using Blaze DS/LCDS.
- Developed Custom Events and Custom Components in various instances of the application.
- Used Hibernate to provide database connectivity to database tables in Oracle.
- Used various Core Java concepts such as Multithreading, Exception Handling, Collection APIs to implement various features and enhancements.
- Applied Java/J2EE Design Patterns like Intercepting Filter, Front Controller, Composite View, Dispatch View, Business Delegate, Service Locator, Value Objects, DAO, and Singleton etc.
- Deployment, Unit and Regression testing usingFlexUnit.
- Worked on critical defects like memory leakage, performance related issues and analyzing the log files to track the issues.
- Designed and debugged the system with the help of IDEs like Flash Builder and Eclipse.
- Used JIRA& Quality Center tools for bug tracking.
- Used SVN for version control.
- Worked on preparing test cases and executing unit testing and Integration testing.
Environment: Java 1.6, JSP, Servlets, Spring 3.0, Spring JDBC, Hibernate3, WebSphere Application Server 7.0, Hudson, Nexus, Eclipse, JQuery, JSON, JavaScript, CSS, Velocity Engine, JSP, JSTL, HTML, DB2, SQL/PL-SQL, XML, JUnit, Log4j, SVN, Maven 2.2, Windows.