Hadoop/bigdata Developer/architect/admin Resume
Md, NY
SUMMARY:
- I am a software engineer with 4+ Hadoop Big Data, 10 yrs in Java experience, Admin, relational database. I am a Java developer who has both knowledge and practical experience with Hadoop, Big Data Amazon webservice and NOSQL, J2EE components, such as Servlets, Java Sever Pages (JSP) Java
- Messenger Service (JMS), Spring, Ejb, Struts, Webservcices, cloud computing, JavaBeans, Java Cryptology Extensions (JCE), Java Data Base Connectivity (JDBC), Web services, and web - client programming such as XML, XSLT, HTML, DHTML and JavaScript. I have extensive experience in the development of software systems. Recent experience has included design, development and coding in JAVA, Oracle PL/SQL, and C++. Worked with everything from Web/Windows GUI front ends to Postgre/Oracle back ends, Worked in all phase of Software development life cycle in "BFSI" domain Project.
- Involved in loading data from MS - SQL to Hive
- Hands on experience in working with Echo-systems like Hive, Pig, HCatalog, Sqoop, Map-Reduce, Oozie.
- Developed Kafka programs to process MSBI Server(ETL) logs.
- Involved in creating dynamic Oozie workflow using Spark.
- Developed pig scripts to process HDFS Data.
- Involved in creating external tables in Hive and also loading the using Sqoop.
- Loading the data into hive tables for blocking and matching.
- Converted many SQL Procedures and SSIS packages into Hadoop.
- Pig Scripts are utilized the Sequence File and HCatalog for better performance.
- Development of MapReduce programs and data migration from existing data source using sqoop.
- Suggested several big data use cases to upper-management to reduce cost, consolidate data sets and provide new data services.
- Migration of some ETL processes from Microsoft SQL Server to Hadoop utilizing pig as data pipe line for easy data manipulation
- Importing and exporting the data using Sqoop from HDFS to Microsoft SQL Server.
- Worked on large data sets of structured data processed using distributed processing environment with Hadoop eco system mostly thru Pig Scripting
- Experience in Writing PIG User Define Function and Hive UDFS.
- Hands on experience on CDH 5.3 platform.
- Scheduled the jobs on Oozie coordinator when the client loaded the data in to SAN
TECHNICAL SKILLS:
File System: HDFS, GFS
Big Data Technology: Hadoop, Google Big Query
Data processing Language: Map-Reduce, SQL-MR, Impala
Databases and Big Data ETL / LET: DB2, Tera Data Hive, HBase, Pig
Relational Database: Oracle 10g and 11g (SQL and PL/SQL) • Operating System Windows, Ubuntu, CentOs, RedHat.
Tools: Used: Eclipse, IBM InfoSphereBigInsight, putty, cygwin
Hadoop on Cloud: Amazon Web Service (AWS), IBM SMART cloud, Google cloud
BI and Reporting Tool: Pentaho, Tableau
Testing: Hadoop Testing, Hive Testing, Quality Center (QC)
Java Skills: JFC/Swing, AWT, Applets, JavaBeans, JMS, Java Mail, RMI, Servlets, JSP, EJB, JNDI, JMS, JDBC, SOAP, Multi-threading, Java Networking, Socket Programming, JUnit
Object Persistence Tools: Hibernate, Castor JDO, ObjectRelationalBridge
RDBMS: Oracle, MS SQL Server, MySQL, PostgeSQL, DB2, Cloudscape
WEB frameworks: WebWork, Spring, Jakarta Struts
Web/Application Servers: Apache, Jakarta Tomcat, Bea WebLogic, IBM Web Sphere, JBoss, Resin
CASE Tools: Rational Rose, WithClass
Version Control: CVS, WinCVS
Operating Systems: MS Windows 7/XP / 2000 / NT4 / 9x, Linux, Solaris
PROFESSIONAL EXPERIENCE:
Confidential, MD/NY
Hadoop/BigData Developer/Architect/Admin
Environment: CDH5.3, Pig 0.12, Hive 0.13, HBase, Oozie, Sqoop, MapReduce, Impala, Kafka, Spark, TFS, Linux
Responsibilities:
- Developed a scalable, cost effective, and fault tolerant data ware house system on Amazon EC2 Cloud. Developed MapReduce/EMR jobs to analyze the data and provide heuristics and reports. The heuristics were used for improving campaign targeting and efficiency.
- Successfully virtualized Hadoop
- Developed a custom File System plugin for Hadoop so it can access files on Hitachi Data Platform. This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly. The plugin also provided data locality for Hadoop across host nodes and virtual machines.
- Advised file system team on optimizing IO for Hadoop / analytics work loads.
- Setup and benchmarked Hadoop/HBase clusters for internal use
- Wrote data ingesters and map reduce programs, scripting to provision and spin up virtualized hadoop clusters
- Implemented Hadoop and big data analytics.
Confidential, NY
Java Developer
Responsibilities:
- Analyzed users needs and planed information streams (UML, Rational Roses);
- Developed and implemented a database structure and software modules (SQL, Ms SQL Server, MS Access);
- Designed the server applications for a financial analysis and representation of the stock quotation in real time
- Developed client environment for financial analysis (Java 1.1, Applets, HTML, JavaScript, IIS 3.0 -4.0);