We provide IT Staff Augmentation Services!

Hadoop Lead Resume

0/5 (Submit Your Rating)

Newyork, NY

SUMMARY

  • Certified Java programmer with 9+ Years of extensive experience in IT including few years of Big Data related technologies.
  • Currently Researcher and developer Technical lead of data engineering team, team works with data scientists in developing insights
  • Good exposure in following all the process in a production environment like change management, incident management and managing escalations
  • Hands - on experience on major components in Hadoop Ecosystem including Hive, HBase, HBase-Hive Integration, PIG, Sqoop, Flume & knowledge of Mapper/Reducer/HDFS Framework.
  • Hands on experience Installation, configuration, maintenance, monitoring, performance and tuning, and troubleshooting Hadoopclusters in different environments such as Development Cluster, Test Cluster and Production
  • Defined file system layout and data set permissions
  • Monitor local file system disk space usage, log files, cleaning log files with auto script
  • Extensive knowledge of Front End technologies like HTML, CSS, Java Script.
  • Good working Knowledge in OOA & OOD using UML and designing use cases.
  • Good communication skills, work ethics and the ability to work in a team efficiently with good leadership skills.

TECHNICAL SKILLS

Big Data: Hadoop, HDFS, MapReduce, Hive, Sqoop, Pig, HBase, MongoDB, Flume, Zookeeper, Oozie.

Operating Systems: Windows, Ubuntu, Red Hat Linux, Linux, UNIX

Java Technologies: JDBC, JAVA, SQL, JavaScript, J2EE, C, JDBC, SQL, PL/SQL

Programming or Scripting Languages: Java, SQL, Unix Shell Scripting, C.,Python

Database: MS-SQL, MySQL, Oracle, MS-Access

Middleware: Web Sphere, TIBCO

IDE’s & Utilities: Eclipse and JCreator, NetBeans

Protocols: TCP/IP, HTTP and HTTPS.

Testing: Quality Center, Win Runner, Load Runner, QTP

Frameworks: Hadoop,py-spark,Cassendra

PROFESSIONAL EXPERIENCE

Confidential, NewYork, NY

Hadoop Lead

Responsibilities:

  • HDP 2.3 distribution for development Cluster
  • Hadoop eco systems Hive, Map reduce to process data Contribution
  • Writing Map reduce for processing xmls and flat files
  • Experience in Scala and Cassendra database for processing jobs
  • Provided production support for cluster maintenance
  • Provide strategic direction to the team
  • Assigning work to subordinates
  • Track risk and report to project manager
  • Provide project status to senior management
  • There were 10 node cluster with Hortonworks data platform with 550 GB RAM, 10 TB SSDs and 8 cores
  • Worked on analyzing Hadoop stack and different big data analytic tools including Pig and Hive, Hbase database and Sqoop
  • Taking trainings for new joiners into project
  • Triggered workflows based on timeoravailability of data using the Oozie Coordinator

Confidential, Louisville, KY

Hadoop Lead

Responsibilities:

  • HDP 2.0 distribution for development Cluster
  • All the datasets was loaded from two different source such as Oracle, MySQL to HDFS and Hive respectively on daily basis
  • We were getting on an average of 80 GB on daily basis on the whole the data warehouse. We used 12 node cluster to process the data
  • Involved in loading data from UNIX file system to HDFS
  • Hadoop eco systems hive, Map reduce, Pyspark to process data Implemented capacity scheduler to share the resources of the cluster and perform Hadoop admin responsibilities as needed
  • Writing Map reduce and Pyspark jobs for cleansing and applying algorithms
  • Cassendra database was use to transform queries to Hadoop HDFS
  • Designed scalable big data cluster solutions
  • Monitored job status through email received from cluster health monitoring tools
  • Responsible to manage data coming from different sources.

Confidential, Buffalo Grove, IL

Hadoop Lead

Responsibilities:

  • Worked on Importing and exporting data into HDFS in financial sector
  • Involved as a team in reviewing of functional and non-functional requirements for writing debit processing in Atlanta location.
  • Implemented Oozie workflows to perform Ingestion & Merging of data in the MapReduce jobs for credit card fraud detection.
  • Extracted files from Cassendra Database through Sqoop and placed in HDFS and processed.
  • Hands on experience in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way to administer transactions.
  • Developed a custom File system plug in for Hadoop so it can access files on Data Platform.
  • This plug-in allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
  • Expertise in server-side and J2EE technologies including Java, J2SE, JSP, Servlets, XML, Hibernate, Struts, Struts2, JDBC, and JavaScript development.
  • Design of GUI using Model View Architecture (STRUTS Frame Work).
  • Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.

Environment: Hadoop 1x, Hive, Pig, HBASE, Sqoop and Flume, Spring, Jquery, Java, J2EE, HTML, Javascript, Hibernate

Confidential, Mooresville, NC

Sr. Java Developer

Responsibilities:

  • Developed new DAOs methods using Hibernate as ORM for application.
  • Used DOM Parser to parse XML 1.1 data from file.
  • Used JAXB 2.0 annotations to convert Java object to/from XML 1.1 file.
  • Created a SOAP 1.2 web service and then got its WSDL 2.0 generated.
  • Created a Web Service Client and invoked the web service using the client
  • Developed a REST based service which reads the JSON 2.0 file and passed it as an argument to the Controller which handles the multiple HTML UI files.
  • Used Struts MVC framework for user authentication by using Ping Federate Server for single sign on (SSO)
  • Performed multi threading and java scripting
  • Designing algorithms, patterns and structure for java framework
  • Used SAML to use many services by entering into the system for one service
  • Involved in coding front end using Swing, HTML, JSP, JSF, Struts Framework
  • Involved in all software development life cycle phases like development, unit testing, regression testing, performance testing, deployment
  • Responsible for developing, configuring, or modifying REST and SOAP web services using technologies like JAX-RS, JAX-WS, Jersey, Spring MVC.
  • Used Spring JDBC as data layer to query databases DB2 and Cassandra.
  • Worked UNIX batch applications that generates product feeds and XML files.
  • Worked with Rest API automation using RestAssured and Testing framework.
  • Participated in scrum meetings, daily stand-ups, grooming sessions.
  • Used technologies like Spring, REST, JAX-RS, Jersy, JSON, Junit, Testing, Mockito, EasyMock, RestAssured, Ehcache, Maven, DB2, JDBC, Batch Scripting, DB2, WebSphere commerce, websphere.

Environment: Java, J2EE, JSP, ExtJS, Servlets, Struts, JDBC, Java Script, LifeRay, Google Web Toolkit, EJB (SSB, MDB), Ajax, Websphere

We'd love your feedback!