Hadoop/spark/java Developer Resume
SUMMARY:
- Over 7 years of professional IT experience which includes experience in Big data , Spark, Hadoop ecosystem and Java and related technologies .
- Oracle Certified Professional (Java)
- Strong experienceinHadoop Architecture and Daemons such as HDFS, JOB Tracker, Task Tracker Name Node, Data Node and Map Reduceprograming concepts.
- Strong knowledge in writing Map Reduce programs using Java to handle different data sets using Map and Reduce tasks
- WrittenHIVE queries & knowledgein Pig scripts to perform data analysis, data transfer and table design.
- Implemented Ad - hoc query using Hive to perform analytics on structured data.
- Strong knowledge in writing Hive UDF, Generic UDF's to in corporate complex business logic into Hive Queries.
- Implemented SQOOP for large dataset transfer between Hadoop and RDBMS.
- Experience in meeting expectations with Hadoop clusters using Horton Works.
- Knowledge of real time data analytics using Spark Streaming, Kafka and Flume.
- Good experience in writing Spark applications using Scala.
- Implemented pre-defined operators in spark such as map, flat Map, filter, reduceByKey, groupByKey, aggregateByKey and combineByKey etc.
- Used Scala IDE to develop Scala coded spark projects and executed using spark-submit
- Worked on different file formats (ORCFILE, TEXTFILE) and different Compression Codecs (GZIP, SNAPPY, LZO).
- Experience in importing and exporting data using SQOOP(HIVE table) from HDFS to Relational Database Systems and vice - versa
- In-depth understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Spark MLib
- Experience in using Accumulator variables, Broadcast variables, RDD caching for Spark Streaming.
- Expertise in using Spark-SQL with various data sources like JSON, Parquet and Hive.
- Experience in creating tables, partitioning, bucketing, loading and aggregating data using Hive.
- Migrating the code from Hive to Apache Spark and Scala using Spark SQL, RDD.
- Hands-on knowledge on core Java concepts like Exceptions, Collections, Data-structures, Multi-threading, Serialization and deserialization.
- Good knowledge on developing micro service APIs using Java 8, Spring Boot 2.x.
TECHNICAL SKILLS:
Application Programming: Scala, Java 8, SQL, PL/SQL
RDBMS/NoSQL DB: Oracle 10g and Mysql, Big Data,HBase, Redis
Frameworks: Spark, spring (Boot, core,web), Restful Web-Services
Software: Eclipse, Scala IDE, Spring echo system
SCM Tools: GIT (Bitbucket), SVN
Operating Systems: Windows, Linux
PROFESSIONAL EXPERIENCE:
Confidential
Hadoop/Spark/Java Developer
Responsibilities:
- Good knowledge and worked on Spark SQL, Spark Core topics such as Resilient Distributed Dataset (RDD) and Data Frames.
- Developing Spark programs using Scala API's to compare the performance of Spark with Hive and SQL.
- Used Spark API over Hortonworks Hadoop YARN to perform analytics on data in Hive.
- Implemented Spark using Scala and SparkSQL for faster testing and processing of data.
- Worked on converting Hive queries into Spark transformations using Spark RDDs.
- Imported data from AWS S3 and into Spark RDD and performed transformations and actions on RDD's.
- Used Spark-SQL to Load JSON data and create Schema RDD and loaded it into Hive Tables and handled structured data using SparkSQL.
- Load the data into Spark RDD and do in memory data Computation to generate the Output response.
- Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs using Scala
- Worked on improving Hive queries performance by rewriting in Spark.
- Implemented Spark using Scala and utilizing Data frames and Spark SQL API for faster processing of data.
- Developed Spark code using Scala/java and Spark-SQL for faster testing and Processing of data.
- Import the data from different sources like HDFS/Hbase into Spark RDD.
- Involved in the development of API for Tax Engine, CARS Module and Admin module as java/API developer
- Developed several REST webservices supporting JSON to perform tasks such calculate/return tax.
- Involved in developments of service-oriented architecture to integrate with 3rd party systems while maintaining loose coupling.
- Created fully functional REST web services supporting JSON message transformationusing spring technology.
- Maintained high level of unit test coverage through test-driven development.
Environment: Java 1.8, Spring Boot 2.x, RESTful Web Services, Eclipse, MySQL, Maven, Bit Bucket (Git), Hadoop, HDFS, Spark, MapReduce, Hive, Sqoop, HBase, Scala, AWS, Java, JSON, SQL Scripting and Linux Shell Scripting, Avro, Parquet, Hortonworks.JIRA, Agile Scrum methodology .
Confidential, Boston
Hadoop Developer
Responsibilities:
- Worked on big data tools including Hadoop,HDFS,Hive and SQOOP.
- Importing and exporting data into HDFS and HIVE using SQOOP.
- Involved in creating Hive tables,loading with data and writing Hive queries that will run internally in map reduce way.
- Created Hive tables and worked on them using HiveQL.
- Implemented Partitioning,Dynamic Partitions and Bucketing in Hive for efficient data access.
- Written multiple MapReduce programs in java for data extraction,transformation and aggregation from multiple file formats including XML,JSON,CSV and other compressed file formats.
Environment: Hadoop, HDFS, MapReduce, Hive, Sqoop, HBase, Oozie, Flume, AWS, Java, JSON, SQL Scripting and Linux Shell Scripting, Avro, Parquet, Hortonworks.
Confidential
Java Developer
Responsibilities:
- The application is developed using Apache Struts framework to handle the requests and error handling.
- Developed Oracle stored procedures / triggers to automate the transaction updated while any type of transactions occurred in the bank database.
- Used XML to get the data from some of the legacy system.
- Involved in database modeling and design using ERWin tool.
- Supported for System test and UAT and Involved in pre & post implementing support.
- Involved in production implementation planning/strategy along with client.
Environment: Java 1.4, J2EE, Tomcat 5.0, Apache Struts1.1 Oracle 9i, Visio, Visual Source Safe 6.0