Hadoop Lead Resume
Newyork, NY
SUMMARY
- Certified Java programmer with 9+ Years of extensive experience in IT including few years of Big Data related technologies.
- Currently Researcher and developer Technical lead of data engineering team, team works with data scientists in developing insights
- Good exposure in following all the process in a production environment like change management, incident management and managing escalations
- Hands - on experience on major components in Hadoop Ecosystem including Hive, HBase, HBase-Hive Integration, PIG, Sqoop, Flume & knowledge of Mapper/Reducer/HDFS Framework.
- Hands on experience Installation, configuration, maintenance, monitoring, performance and tuning, and troubleshooting Hadoopclusters in different environments such as Development Cluster, Test Cluster and Production
- Defined file system layout and data set permissions
- Monitor local file system disk space usage, log files, cleaning log files with auto script
- Extensive knowledge of Front End technologies like HTML, CSS, Java Script.
- Good working Knowledge in OOA & OOD using UML and designing use cases.
- Good communication skills, work ethics and the ability to work in a team efficiently with good leadership skills.
TECHNICAL SKILLS
Big Data: Hadoop, HDFS, MapReduce, Hive, Sqoop, Pig, HBase, MongoDB, Flume, Zookeeper, Oozie.
Operating Systems: Windows, Ubuntu, Red Hat Linux, Linux, UNIX
Java Technologies: JDBC, JAVA, SQL, JavaScript, J2EE, C, JDBC, SQL, PL/SQL
Programming or Scripting Languages: Java, SQL, Unix Shell Scripting, C.,Python
Database: MS-SQL, MySQL, Oracle, MS-Access
Middleware: Web Sphere, TIBCO
IDE’s & Utilities: Eclipse and JCreator, NetBeans
Protocols: TCP/IP, HTTP and HTTPS.
Testing: Quality Center, Win Runner, Load Runner, QTP
Frameworks: Hadoop,py-spark,Cassendra
PROFESSIONAL EXPERIENCE
Confidential, NewYork, NY
Hadoop Lead
Responsibilities:
- HDP 2.3 distribution for development Cluster
- Hadoop eco systems Hive, Map reduce to process data Contribution
- Writing Map reduce for processing xmls and flat files
- Experience in Scala and Cassendra database for processing jobs
- Provided production support for cluster maintenance
- Provide strategic direction to the team
- Assigning work to subordinates
- Track risk and report to project manager
- Provide project status to senior management
- There were 10 node cluster with Hortonworks data platform with 550 GB RAM, 10 TB SSDs and 8 cores
- Worked on analyzing Hadoop stack and different big data analytic tools including Pig and Hive, Hbase database and Sqoop
- Taking trainings for new joiners into project
- Triggered workflows based on timeoravailability of data using the Oozie Coordinator
Confidential, Louisville, KY
Hadoop Lead
Responsibilities:
- HDP 2.0 distribution for development Cluster
- All the datasets was loaded from two different source such as Oracle, MySQL to HDFS and Hive respectively on daily basis
- We were getting on an average of 80 GB on daily basis on the whole the data warehouse. We used 12 node cluster to process the data
- Involved in loading data from UNIX file system to HDFS
- Hadoop eco systems hive, Map reduce, Pyspark to process data Implemented capacity scheduler to share the resources of the cluster and perform Hadoop admin responsibilities as needed
- Writing Map reduce and Pyspark jobs for cleansing and applying algorithms
- Cassendra database was use to transform queries to Hadoop HDFS
- Designed scalable big data cluster solutions
- Monitored job status through email received from cluster health monitoring tools
- Responsible to manage data coming from different sources.
Confidential, Buffalo Grove, IL
Hadoop Lead
Responsibilities:
- Worked on Importing and exporting data into HDFS in financial sector
- Involved as a team in reviewing of functional and non-functional requirements for writing debit processing in Atlanta location.
- Implemented Oozie workflows to perform Ingestion & Merging of data in the MapReduce jobs for credit card fraud detection.
- Extracted files from Cassendra Database through Sqoop and placed in HDFS and processed.
- Hands on experience in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way to administer transactions.
- Developed a custom File system plug in for Hadoop so it can access files on Data Platform.
- This plug-in allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
- Expertise in server-side and J2EE technologies including Java, J2SE, JSP, Servlets, XML, Hibernate, Struts, Struts2, JDBC, and JavaScript development.
- Design of GUI using Model View Architecture (STRUTS Frame Work).
- Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.
Environment: Hadoop 1x, Hive, Pig, HBASE, Sqoop and Flume, Spring, Jquery, Java, J2EE, HTML, Javascript, Hibernate
Confidential, Mooresville, NC
Sr. Java Developer
Responsibilities:
- Developed new DAOs methods using Hibernate as ORM for application.
- Used DOM Parser to parse XML 1.1 data from file.
- Used JAXB 2.0 annotations to convert Java object to/from XML 1.1 file.
- Created a SOAP 1.2 web service and then got its WSDL 2.0 generated.
- Created a Web Service Client and invoked the web service using the client
- Developed a REST based service which reads the JSON 2.0 file and passed it as an argument to the Controller which handles the multiple HTML UI files.
- Used Struts MVC framework for user authentication by using Ping Federate Server for single sign on (SSO)
- Performed multi threading and java scripting
- Designing algorithms, patterns and structure for java framework
- Used SAML to use many services by entering into the system for one service
- Involved in coding front end using Swing, HTML, JSP, JSF, Struts Framework
- Involved in all software development life cycle phases like development, unit testing, regression testing, performance testing, deployment
- Responsible for developing, configuring, or modifying REST and SOAP web services using technologies like JAX-RS, JAX-WS, Jersey, Spring MVC.
- Used Spring JDBC as data layer to query databases DB2 and Cassandra.
- Worked UNIX batch applications that generates product feeds and XML files.
- Worked with Rest API automation using RestAssured and Testing framework.
- Participated in scrum meetings, daily stand-ups, grooming sessions.
- Used technologies like Spring, REST, JAX-RS, Jersy, JSON, Junit, Testing, Mockito, EasyMock, RestAssured, Ehcache, Maven, DB2, JDBC, Batch Scripting, DB2, WebSphere commerce, websphere.
Environment: Java, J2EE, JSP, ExtJS, Servlets, Struts, JDBC, Java Script, LifeRay, Google Web Toolkit, EJB (SSB, MDB), Ajax, Websphere