We provide IT Staff Augmentation Services!

Hadoop/big Data/developer Resume

3.00/5 (Submit Your Rating)

Plano, TexaS

SUMMARY

  • 10 years of experience in IT industry covering Development, Analysis, Design, Testing and System Maintenance using JAVA/J2EE Technologies and BigData/Hadoop development wif 5+yrs.
  • Operative expertise sound noledge in using Hadoopecosystem components like HadoopMap Reduce, HDFS, HBase, Hive. Hands on experience in installing, configuring Sqoop, Pig, Zookeeper and Flume.
  • Cross - functional Exposure on ApacheHadoopMap Reduce programming, PIG Scripting and Distribute Application and HDFS. Solid understanding of Hadoop MRV1 and Hadoop MRV2 (or) YARN Architecture
  • Extensive involvement in Hadoop cluster architecture noledgebase and monitoring clusters.
  • Experienced in building highly scalable Big-data solutions using Hadoop and multiple distributions i.e.Cloudera and NoSQL platforms(Hbase & Cassandra).
  • Updated noledge on Amazon AWS concepts like EMR & EC2 web services which provides fast and efficient processing of Big Data.
  • Experience in managing and reviewing Hadoop log files. Experience in writing Map Reduce programs and using Apache Hadoop API for analyzing teh data.
  • Strong experience in developing, debugging and tuning Map Reduce jobs in Hadoop environment. Used Compression Techniques (snappy) wif file formats to leverage teh storage in HDFS.
  • Implemented in setting up standards and processes forHadoopbased application design and implementation. Expertise in developing PIG and HIVE scripts for data analysis.
  • Hands on experience in data mining process, implementing complex business logic and optimizing teh query using HiveQL and controlling teh data distribution by partitioning and bucketing techniques to enhance performance.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Experience in managing Hadoopclusters using Cloudera Manager Tool.
  • Experience working wif Flume to handle large volume of streaming data.
  • Good working noledge on Hadoop hue ecosystems.
  • Extensive experience in migrating ETL operations into HDFS systems using Pig Scripts.
  • Detailed noledge of big data analytics libraries (MLlib) and utilization of data exploration tools like Spark-SQL.
  • Expert in implementing advanced procedures like text analytics and processing using teh in-memory computing capabilities like Apache Spark written in scala.
  • Knowledge on Apache Spark and its stack.
  • Experience in creating and designing data ingest pipelines using technologies such as spring Integration, Apache Storm wif Kafka.
  • Experience wif Oozie Workflow Engine in running workflow jobs wif actions that run Hadoop Map Reduce and Pig jobs.
  • Worked on power exchange in Informatica. Scheduled and Monitored jobs on Informatica Monitoring Tool.
  • Developed applications using Core Java, Multithreading, Collections, JDBC, Swing, Networking, Reflections.
  • Java/J2EE Software Developer wif experience of Core Java and Web based applications wif expertise in reviewing client requirement; prioritize requirements, creating project proposal (scope, estimation) and baseline project plan.
  • Implemented core modules in large cross-platform applications using JAVA, J2EE, Hibernate, Python, Spring, JSP, Servlets, EJB, JDBC, JavaScript, XML, and HTML.
  • Devised continuous integration of Java projects wif build tools Maven and ANT.
  • Working Knowledge in configuring and monitoring tools like Ganglia and Nagios.
  • Hands-on expertise in using relational databases like Oracle, MySQL, PostgreSQL and MS-SQL Server.
  • Extensive experience in developing and deploying applications using Web Logic, Apache Tomcat and JBOSS.
  • Developed Unit test cases using Junit, Easy Mock and MRUnit testing frameworks.
  • Experienced wif version controller systems like SVN, Clear case, Git, bitbucket.
  • Experience using IDEs tools Eclipse 3.0, My Eclipse, RAD and NetBean.
  • Experience in designing Use Cases, Class diagrams, Sequence and Collaboration diagrams for multi-tiered object-oriented system architectures.
  • Extensive experience wif design and development of J2EE based applications involving technologies such as Java Server Pages (JSP), Java Messaging Service (JMS), Java Data Base Connectivity (JDBC).

TECHNICAL SKILLS

Big Data Ecosystem: Visual Source Safe, Win CVS, MKS,Putty,WINSCP, Rational -Rose, Eclipse, Rational Application Developer and HDFS, HBase,HadoopMap Reduce, Zookeeper, Hive, Pig, Sqoop, Oozie, Cassandra, Spark, Apache Storm, Hadoop Hue

Programming Languages: Java 8, Java 7 (J2EE), Cobol, Focus, C, C++, python, perl, JavaScript, Cuda.

Technologies: J2EE (JSP, Servlets, EJB, JMS, JDBC), Multi-Threading, Collections, Spring, Java Scripting, Hibernate, Struts, design patterns, JUnit,JDK 1.5,log4j.

Databases: Oracle 9i, 10G,DB2,Sybase,Netezza

Operating systems: Windows 95/NT/2000, Red hat LINUX.

Servers: Web sphere, Tomcat, JBoss.

User interfaces: HTML, JSP, XML, CSS, PHP

PROFESSIONAL EXPERIENCE

Confidential, Plano, Texas

Hadoop/Big Data/Developer

Responsibilities:

  • Created an end to end fault tolerant mailing solution leveraging AWS EC2/AMR instances and Amazon tools like STS and SQS queues using Java 8.
  • Used Lambdas to connect AWS tools and fault tolerant connectivity.
  • Developed UI Node-js compatible angular-js and angular components such as pagination, Sorting and range slider wif realtime HTTPS GET update for Squadron process monitoring framework.
  • Worked on angular integration and developing some angular modules for teh front end compatibility.
  • Integrated Team code and maintained it on teh proprietary Github
  • Leveraged Scala and Scallop a Scala based argument parsing Library for Controls Management framework.
  • Utilized Scala based frameworks like Play and Json4s frameworks for Json validations and manipulations.
  • Utilized Snowflake Sql and Spark and Scala snowflake Connectors to query and transform tables
  • Leveraged Spark dataframes for data manupulation and fast transformations for on teh fly table updates and views generation
  • Worked on multi level data validation between Oracle and corresponding Snowflake Data tables using partition and slicing queries to match data from both platforms.

Confidential, Chicago, IL

Sr. Hadoop/Big Data/Developer

Responsibilities:

  • Extracted and updated teh data into HDFS using sqoop import and export command line utility interface.
  • Responsible for developing data pipeline using flume, Sqoop and pig to extract teh data from weblogs and store in HDFS.
  • Involved in using HCATALOG to access Hive table metadata from Map Reduce or Pig code.
  • Responsible for managing data from multiple sources. Implemented best income logic using pig scripts.
  • Involved in developing Hive UDFs for teh needed functionality.
  • Involved in creating Hive tables, loading wif data and writing hive queries. Used Hive to analyze teh partitioned and bucketed data and compute various metrics for reporting.
  • Involved in creating hive tables loading and analyzing data using hive queries.
  • Created and managed data indexing, developing custom tokenizers and relevance tuning, creating filters, adding functionality.
  • Used Hive to analyze teh partitioned and bucketed data and compute various metrics for reporting. Used Hive connections to analyze data from Oracle.
  • Used Pig to do transformations, event joins, filter boot traffic and some pre-aggregations before storing teh data onto HDFS.
  • Developed data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
  • Involved in emitting processed data from Hadoop to relational databases or external file systems using SQOOP.
  • Developed multiple Map Reduce jobs in java 8 libraries for data cleaning and pre-processing.
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Assisted in exporting analyzed data to relational databases using sqoop.
  • Involved in loading data from UNIX file system to HDFS.
  • Created HBase tables to store different data formats.
  • Experience in managing and reviewing Hadoop log files.
  • Export teh analyzed data to teh relational databases using sqoop for visualization and to generate reports for teh BI team.
  • Monitored jobs on Informatica monitoring tool.
  • Extensive noledge on debugging MapReduce programs, Hive UDF’s using Eclipse

Environment: Hadoop, HDFS, Map Reduce, Hive, Flume, HBase, Sqoop, PIG, Java (JDK 1.6), Eclipse, MySQL and Ubuntu, Zookeeper, Oozie, Apache Kafka, Apache Storm.

Confidential, Kansas, MO

Hadoop/Java/Big Data Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on teh Hadoop cluster.
  • Developed pig scripts for analyzing large data sets in teh HDFS.
  • Collected teh logs from teh physical machines and teh OpenStack controller and integrated into HDFS using Flume.
  • Good noledge analyzing data using Python development and scripting for HadoopStreaming. Setup and benchmarked Hadoop/HBase clusters for internal use.
  • Developed Simple to complex Map/reduce Jobs using Java 7 and Java 8 programming language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted teh data from MySQL into HDFS using Sqoop.
  • Analyzed teh data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior.
  • Used UDF s to implement business logic in Hadoop.
  • Organized query ready and write and query teh Hadoop data in HDFS or HBase or Cassandra using impala.
  • Developed programs in Spark based on teh application for faster data processing than standard MapReduce programs.
  • Develop Spark code using Scala and Spark-SQL for faster testing and data processing.
  • Experience wif batch processing of data sources using Apache Spark.
  • Implemented business logic by writing UDFs in Java and used various UDFs from other sources.
  • Continuous monitoring and managing teh Hadoop cluster using Cloudera Manager.
  • Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
  • Supported in setting up QA environment and updating configurations for implementing scripts wif pig and sqoop.
  • Implemented Daily Cron jobs that automate parallel tasks of loading teh data into HDFS using autosys and Oozie coordinator jobs. Installed Oozie workflow engine to run multiple Hive and Pig jobs.
  • Established connections to ingest data in and from HDFS
  • Worked wif application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Exported teh analyzed data to teh relational databases using Sqoop for visualization and to generate reports for teh BI team
  • Diverse Experience in loading and transforming of large sets of structured, semi structured and unstructured data.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Written multiple MapReduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.

Environment: Hadoop, MapReduce, HDFS, Hive, Spark, Pig, Java, SQL, Cloudera Manager, Sqoop, Strom, Solr, Flume, Cassandra, Oozie, Java (jdk 1.6), Eclipse

Confidential, Phoenix, AZ

Java Developer

Responsibilities:

  • Defining teh requirements for new capabilities for teh platform (product) after due research and study of teh prospective client requirements.
  • Defined architecture and provided optimal design approach for new capabilities during product cycle.
  • Conduct reviews (code/artifacts) during teh implementation phase.
  • Analyze and identify teh rectification of teh issues identified during product security reviews.
  • Provide technical support to teh product engineering team during teh implementation phase.
  • Preparing client presentations for prospective clients to showcase teh existing capabilities and features of teh product.
  • Involved in analysis, specification, design, and implementation and testing phases of Software Development Life CycleSDLC.
  • Implemented service layer classes usingSpring IOCandAOPclasses.
  • Actively interacted wif Business Analysts for requirement gatherings and analysis. Developed designedspecificationsusing UML including Use case, Class and Sequence diagrams.
  • Implemented and maintainedAJAXbased rich client for improved customer experience.
  • Developed teh Presentation and Controller layers usingJSP, HTML, Java Script,Business logic usingSpringIOC, AOP, DTO, JTA, and Persistent layerDAO, Hibernatefor all modules.
  • Developed teh application using industry standard design patterns likeService Locator, Singleton, Business Delegate, MVC, and Factory Patternetc for re-usability.
  • Developed Java Messaging ServiceJMSwifMessage Driven Beansby configuringJMSQueues, Topics, and Connection Factories.
  • UsedJavaScriptfor Client Side validation inJSPpages.
  • Developed teh code environment using IDE asEclipse3.2and deployed intoTomcat ServerDeveloped ANT build.xml to achieve more functionality for build process.
  • Implemented Web services componentsSOAP, WSDL, and UDDIto interact wif external systems.
  • Developed JUnit test framework and executed unit test cases by usingJUNITfor fixes.

Environment: MySQL, Jdk1.5, AJAX, JavaScript, JSP, Spring 3.0, DAO, Hibernate 3.2, UML, Design patterns, JMS, Eclipse3.2, Oracle 10g, ANT, JUNIT, HTML, DHTML, XML, slf4, XSL, CSS, JMeter, Windows XP and UNIX

Confidential, New York

Java Developer

Responsibilities:

  • Organized requirement gathering discussions wif all teh stake holders.
  • Undertook distributing and tracking and communicated issues to developers and reporting status to manager on daily basis.
  • Involved in High Level Design and Low Level Design document preparation.
  • Development according to teh specified design.
  • Published SOAP based web services using JAX-WS, JAXB, XSD, XML Bean and XML.
  • Front end is developed based on struts MVC architecture
  • SOAPUI TEMPhas been used to test teh web services.
  • Struts and spring frameworks TEMPhas been used for teh newly designed UI Infrastructure services to interact wif teh legacy application systems.
  • Developed Action classes, Action forms, Validate methods, struts-config.xml file using struts and also used various struts tag libraries.
  • Used Enterprise Java Beans (EJB session beans) in developing business layer APIs.
  • Hibernate is used as ORM.
  • Applied J2EE design patterns like Business Delegate, DAO and Singleton.
  • Deployed and tested teh application using Tomcat web server.
  • Usingjavascripts did client side validation.
  • Involved in developing DAO's using JDBC.
  • HQL and Criteria API has been used extensively.
  • Developed complex SQL queries, stored procedures, functions, triggers and created indexes wherever applicable in Oracle database.
  • Co-ordination wif Onshore development team
  • Involved in debugging and testing teh application for teh change requests
  • Preparing weekly status reports /Monthly status reports
  • Coordinating wif complete offshore team on filling weekly time sheets on Clarity and Field glass.
  • Given teh code walk through to teh newly joined team members on teh deliverables
  • Planning teh forecast for teh individuals on their task sheets.
  • Prepared teh test case documents for enhancements
  • JUNIT is used for unit testing and prepared JUNIT Test cases document.

Environment: JDK 1.5/1.4, J2EE, Servlets, Strut, Spring, Hibernate 3/3.5/4.0, HQL, Maven 3.0, JAX-WX, JAXB, XML, XSD, SOAPUI, JQuery, CSS, JUNIT, Oracle 9i/10g, SQL, PL/SQL, Quality Center, SSH shell, SSH Client, Putty, VSS, WAS, Web Sphere, Visual Studio, Microsoft Visio, Microsoft Project, UML, Share point, Windows XP and UNIX.

We'd love your feedback!