Senior Software/big Data Engineer Resume
5.00/5 (Submit Your Rating)
SUMMARY
- Over 12+ years of Software Development Experience in Java/J2EE Technologies and Big data Hadoop / Spark platform.
- Over 4+ years of experience in implementing complete big data ETL solutions, including data acquisition, data validation, data profiling, data transformation, data analytics using Big Data technology stack that includes Spark, Spark - SQL, Spark Streaming, KAFKA, Presto, Google Big Query, Google Big table, Google Data Flow, HDFS, HBase, Hive, Pig, Sqoop, YARN, Apache Crunch.
- Experience in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
- Experience in implementing Recommendation, Entity resolution/record Linkage in Big data platform.
- Experience in Setting up distributed Search Engine like Elastic Search, indexing data and integrating it with Spark for populating and querying data.
- Expertise and Knowledge in using job scheduling and monitoring tools like Azkaban and AirFlow.
- Strong hands on experience in development of Java/J2EE Enterprise Applications and RESTFUL Web API using Frameworks/Web technologies like Spring Boot, Spring MVC, Spring Core, Spring Cloud, Spring data JPA/JDBC, Angular JS, Struts, EJB 3.0, Hibernate, Jersey, Ajax, JavaScript, Servlets, JDBC, Junit, KODO JPA.
- Experience in project Build management tools like Maven, Gradle, ANT and Container based deployment using Docker .
- Experience with various SDLC methodologies like Waterfall and Agile and Object Oriented Analysis and Design (OOAD).
- Excellent Knowledge in DataStructure, Algorithms and mathematics.
- Have theatrical knowledge of Machine learning/Data Mining, Supervised/Unsupervised learning , ML algorithms by completing courses in coursera/kaggle.
TECHNICAL SKILLS
- Java, Python, Perl, C
- Amazon aws/EMR, Cloudera CDH, Google cloud DataFlow.
- Spark, Spark SQL, Spark Streaming, KAFKA, PRESTO, HIVE, HDFS, MapReduce, YARN, Hive, HBase, Pig, Sqoop, Azkaban, Airflow, Zookeeper, Apache Crunch,
- Kite SDK, Google Big query, google dataflow, Google Big table
- MongoDB, Hbase, google big table
- Oracle, DB2, PostGres. Sybase
- Coherence, Gemfire,MemCache, Elastic Search
- Windows, LINUX
- Amazon AWS, Google Cloud.
- Jboss 4.0, WebSphere 6.1, Weblogic 10.3, Apache-Tomcat 7.0
- SOAP,REST, WSDL, JAXB
- Hibernate, Spring Boot, Spring MVC, Spring Cloud, Spring data JPA/JDBC, Struts, JMS, EJB, XStream, Jackson, Junit.
- HTML5, CSS3, AngularJS JavaScript, JQuery, AJAX, Servlets, JSP,JSON, XML, XHTML, JSF
- Maven, Gradle, Docker, Terraform
- Git, CVS, SVN, Clearcase
- Eclipse, Intellij, RSA
PROFESSIONAL EXPERIENCE
Confidential
Senior Software/Big data engineer
Responsibilities:
- Design and develop data pipeline using Google Dataflow to perform data processing, publishing, transformation and analytics.
- Populate data to Google Bigquery for adhoc analysis and business intelligence.
- Design and develop google Bigquery to extract insights and intelligence out of huge transactional dataset.
- Design and develop framework to generate multiple dimensions of aggregations using Big query on huge transactional dataset and populate it to Google Bigtable using google dataflow .
- Write Airflow job to orchestrate and automate the data pipeline process.
- Build and manage java application using Maven / Gradle.
- Experience in container based deployment using Docker technology.
Confidential
Senior Software/Big data engineer
Responsibilities:
- Design and develop data pipeline using Java Spark API in AWS EMR to perform data transformation and data analytics .
- Experience in processing large datasets using Spark RDD/Dataframes, and apply Spark transformations like map, filter, reduceByKey, join, cogroup during ingestion process and publish it to Datalake.
- Experience in migrating the data using Sqoop from Relational Database System to S3 and vice-versa according to client's requirement.
- Write Azkaban job to orchestrate and automate the data pipeline process.
- Develop tables and Schema on top of datalake using spark sql to perform adhoc data analytics.
- Record Linkage/Entity resolution of different Big data sources using Spark platform and Elastic search engine.
- Experience in setting up distributed Search Engine like Elastic Search using terraform scripts, indexing data, adding custom search algorithm using duke's plugin, querying for matching different big data feeds from Spark using java API.
- Build UI using Angular JS and Restful web services for CRUD operations using Spring Boot Spring MVC, Spring data JPA/JDBC, Spring cloud for record linkage exception process.
- Publish and consume data from High throughput messaging system KAFKA using Spark/Spark Streaming .
Confidential
Senior Software/data Engineer
Responsibilities:
- Design and develop data loading unit using Kite SDK to populate data to HDFS directory.
- Develop data processing/transformation using Apache Crunch API in hadoop platform and load processed data to Hbase for API accessing.
- Experienced in handling ETL transformations using Pig Latin scripts, expressions, join operations and Custom UDF's for evaluation, filtering and storing data.
- Use Avro serialization to reduce the footprint/size of data.
Confidential
Senior Java/J2EE Developer
Responsibilities:
- Design and develop batch job using Spring Batch framework to process the XML feed from dealmaker.
- Configure the XML -> Java mapping module using Digester 3 Rules/pattern.
- Integrate Spring framework with Hibernate ORM framework and setup hibernate configuration files.
- Persist the Java object to the database using Hibernate 3 ORM Framework.
- Develop Oracle stored procedures to handle report tasks.
Confidential
Senior Java/J2EE Developer
Responsibilities:
- Analyze and identify the possible impacts on technology rationalization.
- Migrate the application from Java 1.4 to 1.6, Kodo 3 to Kodo 4.1 and upgrade the coherence cache..
- Migrate the application from windows to Linux and also modify ANT build scripts
- Write Junit test case to test the various components in the application.
- Deploy/configure the application EAR in the Linux server.