Big Data/ Hadoop Developer Resume
SeattlE
SUMMARY:
- Software professional with 8+ years of experience in JAVA and Hadoop
- Exposure to Object - Oriented Design, methodologies and design patterns
- Hands on experience in installing, configuring and using Hadoop ecosystem components like
- Knowledge and experience with NoSQL platforms like MongoDB, HBase
- Responsible for building a scalable, fault-tolerant and high available big data platform
- Proficient in Designing and developing the data ingestion pipelines into the Hadoop cluster for a variety of big data use cases
- Experience in processing large sets of structured, semi structured and unstructured data
- Experience in handling Terabytes of data every day
- Experience with Cloudera components like Hue
- Experience in managing and reviewing Hadoop log files
- Proficient in developing PigLatin and HiveQL scripts for data analysis and ETL purpose
- Proficient with CDH3, CDH4, CDH5
- Good Knowledge of Software Development Life Cycle ( SDLC )
- Solid SQL skills, strong database design and development capabilities
- Conversant with cloud computing services like Amazon web services and Google app engine
- Strong understanding of RDBMS concepts and experience with relational databases like MySQL
- Strong knowledge of object oriented programming concepts
- Experience developing and consuming RESTful Web Services
- Strong experience applying design patterns
- Well versed with design patterns like Singleton, Dependency Injection, Factory, Builder, Observer
- Experience in deploying applications on Apache Tomcat sever and Amazon EC2 instances
- Good aptitude in multi-threading and concurrency concepts
- Conversant with test frameworks and tools like JUnit, TestNG
- Proficient with web technologies like JavaScript, JQuery, AJAX
- Experience in developing and testing data intensive Android applications
- Agile/Scrum and Test-Driven Development Experience
- Strong problem solving and debugging skills
- Experience designing and building applications from ground up
- Experience working both in a start-up and corporate
TECHNICAL SKILLS:
Languages: Java, SQL, JavaScript, Objective C
Big data: Hadoop, Hive, HBase, Map Reduce, Spark, Kafka, Flume, Sqoop, Pig, Oozie, ZooKeeper
Mobile Development: Android, iOS
Web Development: CSS, HTML, JavaScript, AJAX, JQuery, JSON, REST, Bootstrap
Test framework and tools: JIRA, JUnit, Bugzilla, TestNG, Selenium, Appium, UIAutomator
Databases: MySQL, MongoDB, SQLite
Mobile Analytics: Google Mobile Analytics, Mixpanel, Amplitude, Fabric
UX Design tools: Sketch, Marvel, Adobe Photoshop, InVision
Cloud Computing Services: AWS, Google
IDE: Eclipse, Android Studio, Visual Studio, Xcode
Version Control: Mercurial, Git, GitHub, BitBucket
Build Automation: Ant, Maven, Gradle
PROFESSIONAL EXPERIENCE:
Confidential, CA
Big Data/ Hadoop Developer
Responsibilities:
- Involved in design and development of technical specifications using Hadoop components
- Using Spark for streaming and analyzing data
- Performing data cleansing by writing Map Reduce jobs in Java
- Implementing optimization techniques like partitioning and bucketing in Hive
- Creating external tables in Hive in Sequence and AVRO file formats
- Moving all data generated from various sources into HDFS for further processing
- Using Sqoop to connect to database and move the pivoted data to Hive tables or Avro files
- Using Flume for ingesting streaming data into HDFS
- Loading data into Spark schema and querying using Spark-SQL
- Incorporated Spark-Streaming APIs to process data obtained from Kafka
- Developing Spark scripts by writing custom RDDs
- Designing and implementing column family schemas of Hive and HBase within HDFS
- Developing custom UDFs in Java to extend functionality of Pig scripts
- Developing unit test cases and automating the scripts
- Implementing map reduce programs to perform joins on the Map side using Distributed Cache in Java
- Managing and reviewing Hadoop log files
- Working with NoSQL databases like HBase
- Orchestrating the workflow using Oozie
- Using Hue to manage and monitor the Hadoop clusters
Confidential, Seattle
Big Data/Hadoop Engineer
Responsibilities:
- Monitored and managed Hadoop cluster using Cloudera Manager
- Integrated Apache Kafka for data ingestion
- Loaded and transformed large sets of structured, semi structured and unstructured data
- Used Sqoop for importing and exporting data to and from HDFS
- Analyzing data with Hive and Pig
- Assess the quality of datasets for Hadoop data lake
- Created Sqoop jobs with incremental load to populate Hive tables
- Migrated MapReduce programs into Spark transformations
- Used Hue web interface to monitor Hadoop clusters
- Engaged in performance tuning for Hive queries
- Proficient in writing PIG scripts to process data
- Converting Hive/SQL queries into Spark transformations using Spark RDDs
- Optimized the Hive queries using Partitioning and bucketing techniques
- Conversant with both managed and external tables and writing Hive UDFs
- Engaged Sequence, Avro, Parquet file formats
- Configured Flume for efficiently collecting, aggregating and moving large amounts of log data
- Job scheduling and monitoring through Oozie and ZooKeeper
- Developed Map reduce programs in Java for data processing
- Moved data efficiently between clusters using Distributed Copy
- Responsible for streaming log files from various sources into HDFS using Flume
Confidential, Denver
Hadoop/Java Consultant
Responsibilities:
- Configured and tested Hadoop ecosystem components
- Implemented CDH3 Hadoop cluster on CentOS
- Gathered business requirements and use cases
- Analyzed stored data by writing MapReduce jobs based on business requirements
- Created custom UDFs in Java
- Developed multiple MapReduce jobs in Java for data cleansing
- Used Sqoop to import and export data to and from HDFS
- Created HBase tables to load large sets of structured, semi-structured and unstructured data
- Wrote efficient Hive queries by using dynamic partition and bucketing techniques
- Created external tables and managed tables
- Used Hue to monitor Hadoop Cluster
- Managed and reviewed Hadoop log files
- Used Oozie to automate data loading into HDFS
Confidential
Java Developer
Responsibilities:
- Developed various modules for service desk ticketing portal
- Employed technologies like Java, Spring MVC, Hibernate, Jackson
- Built RESTful web services
- Exposed and consumed web services
- Implemented DAO to integrate with Hibernate and process the database operations
- Developed various mapping files using Hibernate
- Used Jackson to serialize Java objects to JSON and vice versa
- Performed server side validations using Annotations
- Written efficient unit test cases
- Deployed application on Apache Tomcat server
- Implemented design patterns like Singleton, Factory, Observer
- Successfully carried out transitions and roll out of the ticketing portal
Confidential
Java Developer
Responsibilities:
- Developed the web interfaces using JSP.
- Developed business logic codes using Servlets at the back end of the system.
- Involved in developing the database tables to hold lender information
- Responsible for designing the front-end using HTML / Java Servlets.
- Developed JSP's and Servlets to provide dynamic content to the HTML pages.
- Developed data access components and multilingual screen generator classes.
- Developed JSPs, for client side validations.
- Participate in software development building custom components and doing performance tuning.
- Used JSP and Servlets for server side transactions. Login authentication is done by JSP by verifying with database security tables.
- Designed and developed Entity beans and Session beans.
- Configured and Deployed EJB's Entity and Session beans on Weblogic Server
- Developed the interface to automatically forward quote requests to qualified lenders using SMTP.
- Developed test cases to test the business logic.