We provide IT Staff Augmentation Services!

Big Data/ Hadoop Developer Resume

3.00/5 (Submit Your Rating)

SeattlE

SUMMARY:

  • Software professional with 8+ years of experience in JAVA and Hadoop
  • Exposure to Object - Oriented Design, methodologies and design patterns
  • Hands on experience in installing, configuring and using Hadoop ecosystem components like
  • Knowledge and experience with NoSQL platforms like MongoDB, HBase
  • Responsible for building a scalable, fault-tolerant and high available big data platform
  • Proficient in Designing and developing the data ingestion pipelines into the Hadoop cluster for a variety of big data use cases
  • Experience in processing large sets of structured, semi structured and unstructured data
  • Experience in handling Terabytes of data every day
  • Experience with Cloudera components like Hue
  • Experience in managing and reviewing Hadoop log files
  • Proficient in developing PigLatin and HiveQL scripts for data analysis and ETL purpose
  • Proficient with CDH3, CDH4, CDH5
  • Good Knowledge of Software Development Life Cycle ( SDLC )
  • Solid SQL skills, strong database design and development capabilities
  • Conversant with cloud computing services like Amazon web services and Google app engine
  • Strong understanding of RDBMS concepts and experience with relational databases like MySQL
  • Strong knowledge of object oriented programming concepts
  • Experience developing and consuming RESTful Web Services
  • Strong experience applying design patterns
  • Well versed with design patterns like Singleton, Dependency Injection, Factory, Builder, Observer
  • Experience in deploying applications on Apache Tomcat sever and Amazon EC2 instances
  • Good aptitude in multi-threading and concurrency concepts
  • Conversant with test frameworks and tools like JUnit, TestNG
  • Proficient with web technologies like JavaScript, JQuery, AJAX
  • Experience in developing and testing data intensive Android applications
  • Agile/Scrum and Test-Driven Development Experience
  • Strong problem solving and debugging skills
  • Experience designing and building applications from ground up
  • Experience working both in a start-up and corporate

TECHNICAL SKILLS:

Languages: Java, SQL, JavaScript, Objective C

Big data: Hadoop, Hive, HBase, Map Reduce, Spark, Kafka, Flume, Sqoop, Pig, Oozie, ZooKeeper

Mobile Development: Android, iOS

Web Development: CSS, HTML, JavaScript, AJAX, JQuery, JSON, REST, Bootstrap

Test framework and tools: JIRA, JUnit, Bugzilla, TestNG, Selenium, Appium, UIAutomator

Databases: MySQL, MongoDB, SQLite

Mobile Analytics: Google Mobile Analytics, Mixpanel, Amplitude, Fabric

UX Design tools: Sketch, Marvel, Adobe Photoshop, InVision

Cloud Computing Services: AWS, Google

IDE: Eclipse, Android Studio, Visual Studio, Xcode

Version Control: Mercurial, Git, GitHub, BitBucket

Build Automation: Ant, Maven, Gradle

PROFESSIONAL EXPERIENCE:

Confidential, CA

Big Data/ Hadoop Developer

Responsibilities:

  • Involved in design and development of technical specifications using Hadoop components
  • Using Spark for streaming and analyzing data
  • Performing data cleansing by writing Map Reduce jobs in Java
  • Implementing optimization techniques like partitioning and bucketing in Hive
  • Creating external tables in Hive in Sequence and AVRO file formats
  • Moving all data generated from various sources into HDFS for further processing
  • Using Sqoop to connect to database and move the pivoted data to Hive tables or Avro files
  • Using Flume for ingesting streaming data into HDFS
  • Loading data into Spark schema and querying using Spark-SQL
  • Incorporated Spark-Streaming APIs to process data obtained from Kafka
  • Developing Spark scripts by writing custom RDDs
  • Designing and implementing column family schemas of Hive and HBase within HDFS
  • Developing custom UDFs in Java to extend functionality of Pig scripts
  • Developing unit test cases and automating the scripts
  • Implementing map reduce programs to perform joins on the Map side using Distributed Cache in Java
  • Managing and reviewing Hadoop log files
  • Working with NoSQL databases like HBase
  • Orchestrating the workflow using Oozie
  • Using Hue to manage and monitor the Hadoop clusters

Confidential, Seattle

Big Data/Hadoop Engineer

Responsibilities:

  • Monitored and managed Hadoop cluster using Cloudera Manager
  • Integrated Apache Kafka for data ingestion
  • Loaded and transformed large sets of structured, semi structured and unstructured data
  • Used Sqoop for importing and exporting data to and from HDFS
  • Analyzing data with Hive and Pig
  • Assess the quality of datasets for Hadoop data lake
  • Created Sqoop jobs with incremental load to populate Hive tables
  • Migrated MapReduce programs into Spark transformations
  • Used Hue web interface to monitor Hadoop clusters
  • Engaged in performance tuning for Hive queries
  • Proficient in writing PIG scripts to process data
  • Converting Hive/SQL queries into Spark transformations using Spark RDDs
  • Optimized the Hive queries using Partitioning and bucketing techniques
  • Conversant with both managed and external tables and writing Hive UDFs
  • Engaged Sequence, Avro, Parquet file formats
  • Configured Flume for efficiently collecting, aggregating and moving large amounts of log data
  • Job scheduling and monitoring through Oozie and ZooKeeper
  • Developed Map reduce programs in Java for data processing
  • Moved data efficiently between clusters using Distributed Copy
  • Responsible for streaming log files from various sources into HDFS using Flume

Confidential, Denver

Hadoop/Java Consultant

Responsibilities:

  • Configured and tested Hadoop ecosystem components
  • Implemented CDH3 Hadoop cluster on CentOS
  • Gathered business requirements and use cases
  • Analyzed stored data by writing MapReduce jobs based on business requirements
  • Created custom UDFs in Java
  • Developed multiple MapReduce jobs in Java for data cleansing
  • Used Sqoop to import and export data to and from HDFS
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data
  • Wrote efficient Hive queries by using dynamic partition and bucketing techniques
  • Created external tables and managed tables
  • Used Hue to monitor Hadoop Cluster
  • Managed and reviewed Hadoop log files
  • Used Oozie to automate data loading into HDFS

Confidential

Java Developer

Responsibilities:

  • Developed various modules for service desk ticketing portal
  • Employed technologies like Java, Spring MVC, Hibernate, Jackson
  • Built RESTful web services
  • Exposed and consumed web services
  • Implemented DAO to integrate with Hibernate and process the database operations
  • Developed various mapping files using Hibernate
  • Used Jackson to serialize Java objects to JSON and vice versa
  • Performed server side validations using Annotations
  • Written efficient unit test cases
  • Deployed application on Apache Tomcat server
  • Implemented design patterns like Singleton, Factory, Observer
  • Successfully carried out transitions and roll out of the ticketing portal

Confidential

Java Developer

Responsibilities:

  • Developed the web interfaces using JSP.
  • Developed business logic codes using Servlets at the back end of the system.
  • Involved in developing the database tables to hold lender information
  • Responsible for designing the front-end using HTML / Java Servlets.
  • Developed JSP's and Servlets to provide dynamic content to the HTML pages.
  • Developed data access components and multilingual screen generator classes.
  • Developed JSPs, for client side validations.
  • Participate in software development building custom components and doing performance tuning.
  • Used JSP and Servlets for server side transactions. Login authentication is done by JSP by verifying with database security tables.
  • Designed and developed Entity beans and Session beans.
  • Configured and Deployed EJB's Entity and Session beans on Weblogic Server
  • Developed the interface to automatically forward quote requests to qualified lenders using SMTP.
  • Developed test cases to test the business logic.

We'd love your feedback!