Software Engineer Resume
SUMMARY
- 9.5 years of progressive experience in design and development of distributed systems and data platforms along with web applications using BigDATA Technologies like Hadoop, MapReduce, SPARK,YARN, Streaming, Presto pig, hive, Kafka, zookeeper, Cascading, Driven, Core Java, Spring, Spring - Batch, BIRT. And cloud technologies of AWS along with visualization tools like tableau.
- Demonstrated capability in understanding business requirements and translated them into workable code.
- Experience working on Public and private cloud like AWS, RackSpace,Telstra,Terramark,Azure.
- Working as a Senior BigData Developer in my current organization.
- Cloudera Certified Hadoop Developer, Cloudera Certified HBASE Developer, Hortonwroks Certified Hadoop Developer, Map-R Certified Spark Developer
- Hands-on with Apache, Cloudera, Map-R and Hortonworks distribution.
TECHNICAL SKILLS
BigData Technologies: NoSQL Databases, Storm, Kafka, Hadoop, Streaming, MapReduce, YARN, Zookeeper, Cascading, Driven, Amazon EMR, EC2 etc, SPARK, Impala, Spark SQL, Presto, Athena etc
Development Tools: Eclipse, IntelliJ,Maven.
Database: MySQL, NoSQL Cassandra, Hbase, Solr, Lucene.
Programming Languages: C++, Java
Reporting Tool: BIRT,Pentaho
Scripting: Basic Java Script, Ajax
Application or Web Server: Apache Tomcat, JBoss.
Web Technology: HTML, XML.
Framework: Spring, Spring-Batch.
Tools & Concepts: Basic Data Structures, Design Patterns, Soap UI.
Operating Systems: Windows, Linux, Mac.
Documentation Related Tools: Star UML, Microsoft Visio
Build Deployment Tools: Jenkins, Hudson, Sonar, checkstyle
PROFESSIONAL EXPERIENCE
Confidential
Software Engineer
Skills: Java, Spark Streaming, snowflake, spark SQL,Tableau
Responsibilities:
- Writing Java code for BDI to integrate Hadoop file systems.
- Writing MR, Hive UDAF and spark applications For data and address quality
- Exploring and doing tasks for workflow integration with Hue and oozie
- Setting up the apache, CDH and HDP clusters for dev and product certification tasks
- Collaborations with Cloudera and Horton for certifying the spectrum product, on their latest version, included new security features like Sentry, Ranger and encryption zones as per their requirement along with Kerberoes.
Advertising Intelligence
Confidential
Skills: Kafka, zookeeper,Avro, Cascading, Driven, AWS, Junit, Amazon s3, amazon EC2 and EMR
Responsibilities:
- Setup the complete ingestion cluster setup over kafka and zookeeper
- Cluster Design for ingestion
- Data processing from avro being read and analyzed using Cascading and Driven
- Cluster deployment over EMR and writing Junit Testing for processing flow.
- Using S3 as filesystem and migration of processing over AWS to hadoop2.
Confidential
Skills: Hadoop, Hadoop 2.0, YARN, Storm, Kafka (0.72,0.8), HDFS,Oracle NoSQL, ParAccel
Responsibilities:
- Wrote java code using multithreaded application for application and kafka and DB Configuration
- Deployment on Tomcat
- YARN Performance Benchmarking
- Used STORM spouts and bolts for Real time Data analytics
Confidential
Skills: Spring, Spring-Batch, Web Services, Cassandra, JDBC, BIRT, MySQL, Localization, XML Parser, File SFTP, Solandra, Core Java
Responsibilities:
- Using Java API to write subscription Services and other Categorization services for events, which get triggered and persisted through putting a message on Queue and listening by Consumers.
- Exposing the API as Rest services.
- Using Spring Batch to Process Persisted Events through various steps and producing output files and getting it delivered to the Subscriber’s SFTP location.
- Writing Purge Batch Jobs to delete processed events or expired events from Cassandra with Multiple Threads and using Transactions to speed up the deletion Process.
- Development and testing of Services to subscribe and the code to read and authenticate Input files and Deliver the output file as per matched by supplied input.
- Explored and configured solandra in the system and thorough testing on the cluster for Replication factor and consistency level, when used in cluster.
- Java API for handling all kinds of Data and its Access to The handlers of reports.
- Customizing BIRT reports.
- Deployment and Testing of reports on Server.
- Applying Localization in to the code.