Talend/hadoop Developer Resume
Minnetonka, MN
SUMMARY
- 8 Years of extensive experience working as a developer and 3+ years of experience with Big Data and BigData analytics
- Good experience in Apache Hadoop Framework, HDFS, Map/Reduce, Pig, Hive, Sqoop and MapR Hadoop distribution.
- Extensive experience with Talend and Query big data tools like Pig Latin and Hive QL.
- Proficient in big data ingestion and streaming tools like Sqoop and Kafka.
- Has hands on experience in Sequence files, AVRO and JSON file formats and compression.
- Good Knowledge on all Talend components.
- Worked with application teams to install operating system, Hadoop updates, patches and version upgrades as required.
- Excellent understanding of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Hands on experience in installing, configuring and using ecosystem components like Hadoop, Map Reduce, HDFS, HBase, Oozie, Sqoop, Pig & Hive.
- Excellent hands on experience in analyzing data using Pig Latin, HQL, Hbase, Scala and MapReduce programs in Java
- Extending Hive and Pig core functionality by writing custom UDFs.
- Good experience in Oozie Framework and Automating daily import jobs.
- Has good experience in Sqoop and Apache Flume for collecting, aggregating and moving large amounts of Relational and Streaming data from application servers.
- In - depth understanding of designing and coding using SQL, Linux /UNIX technologies.
TECHNICAL SKILLS
Big Data & Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Sqoop, Cassandra, Kafka,talend, Oozie, Flume.
Java Technologies: Java 6, Java 7, JAXP, AJAX, I18N, JFC Swing, Log4j, Java Help API
J2EE Technologies: JSP 2.1 Servlets 2.3, JDBC 2.0,JNDI, XML, JAXP, Java Beans
Methodologies: Agile, UML, Design Patterns (Core & J2EE)
Framework: MVC, Struts, Hibernate
Database: Oracle 10g, DB2, MySQL
Application Server: Apache Tomcat 5.x 6.0, Jboss 4.0
Web Tools: HTML, Java Script, XML, DTD, Schemas, XSL, XSLT, XPath, DOMTools Sql developer, DB visualize
IDE / Testing Tools: NetBeans, MyEclipse, Eclipse, WSAD, RAD
Operating System: Windows, Linux
Scripts: UNIX, Java Script, ANT
Testing API: JUINT
PROFESSIONAL EXPERIENCE
Confidential, Minnetonka, MN
Talend/Hadoop Developer
Responsibilities:
- Developed Pig scripts to load data from files to Hbase
- Developed Hive scripts to pull data from Data Lake to our tenant.
- Proficient with UNIX shell scripting.
- Developed custom UDFs in Hive.
- Involved in design discussions for the ingestion process
- Work with customers to resolve the incidents created in ServiceNow as per the SLA.
- Monitor the ingestion for different sources and resolve any issues causing a delay in the process.
- Developed Talend jobs to identify the gaps in the data and reload the data to clear the gaps.
- Performed data manipulations using various Talend components like tMap, tJavarow, tjava, tOracleRow, tOracleInput, tOracleOutput, tMSSQLInput and many more.
- Worked on improving the performance of Talend jobs.
- Deployed Talend jobs on TAC and created triggers to run the jobs.
- Leading a team of 4 and tracking the defects to closure.
- Developed Talend jobs to configure RabbitMQ for batch load on data.
- Configured Splunk dashboard to view ingestion details.
Environment: MapR,Hadoop Distributed File System (HDFS), Hive, Pig, TAC, Talend, HBase,Splunk,RabbitMQ, SVN.
Confidential, Milwaukee, WI
Talend Developer
Responsibilities:
- Developed Data Integration jobs to load data from Hive to MDM.
- Created custom Talend components in Java.
- Involved in discussions with Product Owners and designing the data model for MDM.
- Used SQL queries and other data analysis methods, as well as TalendEnterprise Data Quality Platform for profiling and comparison of data, which will be used to make decisions regarding how to measure business rules andquality of the data.
- Developed Routinesin Talend.
- Performed data cleansing using TalendData Preparation.
- Participate in designing the overall logical & physical Data warehouse/Data-mart data model and data architectures to support business requirements
- UsedGitas Version Control Software.
- Trained team members on Talend.
Environment: JDK 1.8,Core java,Hadoop Distributed File System (HDFS), Talend, Hive,Oracle 11g, Eclipse IDE.
Confidential, Eden Prairie, MN
Talend/Hadoop Developer
Responsibilities:
- Worked directly with clients in managing their requests, understanding their requirements and translating functional requirements to technical specifications through hands-on analysis and coordinating with the onsite and offshore team for development.
- Developed scripts to convert raw data to AVRO and load the data into Hive tables.
- Developed Map Reduce scripts to make web service calls and convert data into XML.
- Developed and maintained large scale distributed data platforms with experienced in data warehouses, data marts and data lakes.
- Import/export data from Oracle data base to/from HDFS using Sqoop.
- Worked extensively on HBase to track the ingestion process.
- Created external partitioned tables in Hiveto provision the ingested data.
- Developed jobs to read messagesfromRabbitMQqueue and load the messages intoHBase.
- Created custom serde in hive to load data from different types of files.
- Developed Talend jobs to ingest data into HDFS and hive tables.
- Developed Talend jobs to make REST web service calls and update the response in HBase tables.
- Developed Talend jobs to extract data from traditional databases and create data files.
- Performed data quality checks like schema validation, record count validation, timestamp check on the raw data received.
- Created workflows using Bedrock.
- UsedSVNas Version Control Software.
- Developed solutions utilizing the Hadoop ecosystem such as Hive, HBASE, Pig, Sqoop, Oozie, Zookeeper etc.
- Provide support to other internal teams which are using our framework.
Environment: JDK 1.7,Core java, Hadoop, Hadoop Distributed File System (HDFS),Talend, Map Reduce,Hive 0.8.1,HBase, SQOOP, Pig,Log4J, Junit, Linux, Eclipse IDE, XML.
Confidential, Eagan, MN
Hadoop Developer
Responsibilities:
- Developed Map Reduce jobs for data cleaning and transformation.
- Created Producersfor different sources and Consumers for different topics inKafka.
- Configured log4j properties to write logs to files.
- Developed consumer in Kafka to create indices in Elasticsearch for the logs.
- Good Experience in managing and reviewing Hadoop log files
- Used Pig as ETL tool to do transformations, joins and pre-aggregations before loading data onto HDFS.
- Worked on large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Installed and configured Hive and also developed Hive UDFs to extend core functionality of hive
- Responsible for loading data from UNIX file systems to HDFS.
- Used HCatalog to share schema’sbetween hive and Pig tables.
- Collected logs and clickstream data from the physical machines and integrated into HDFS using Flume.
- JSONserde was used to de serialize data obtained from the java API while creating Hive tables.
- Defined workflows using Oozie.
- Imported and Exported Data from Different Relational Data Sources like DB2,SQL Server,Teradata to HDFS using Sqoop.
- Developed PIG Latin scripts to extract the data from the web server output files.
- Documentation and execution of Test Cases for JUnit Testing.
- Used Gradle for project build.
- Involved in story-driven Agile development methodology and actively participated in daily scrum meetings.
Environment: JDK 1.7,Core java, Hadoop, Hadoop Distributed File System (HDFS),Map/Reduce,Sqoop 2,Hive 0.8.1,Log4J, Junit, Linux, Eclipse IDE, XML,Kafka, Pig, Flume, Oozie.
Confidential IA
Hadoop Developer
Responsibilities:
- Worked extensively in creating map reduce jobs for data cleansing and preprocessing
- Designed a datawarehouse using Hive
- Worked extensively with sqoop for importing metadata from Oracle
- Extensively used pig for data cleansing.
- Worked on tuning the performance Pig queries.
- Worked with different file formats such as AVRO and Json and picked AVRO as the preferred file format
- Created partitioned tables in Hive
- Worked with business teams and created Hivequeries for adhoc access.
- Evaluated usage of OOZIE for Workflow Orchestration.
- Experience building and scaling highly available systems at Amazon Web Services (AWS)
- Experience deploying and managing infrastructures within AWS
- Created data collection pipeline leveraging Java and using Spring framework.
- Created data models in NoSQL databases (Cassandra) and leveraged Cassandra as preferred data store
- Installedhadoop and configured entire eco system (hive, pig) on test environments
- Created Talend jobs to connect to Quality Stage using FTP connection and process data received from Quality Stage.
- Used IBatis for Data Access POJOs
- Mentored analyst and test team for writing Hive Queries
- Created OLAP cubes in SSAS on hive.
Environment: Hadoop, MapReduce, Talend, HDFS, Hive, Java (jdk1.6), Hadoop distribution of HortonWorks.
Confidential, Richmond, VA
Java Developer
Responsibilities:
- Involved in writing programs for XA transaction management on multiple databases of the application.
- Developed java programs, JSP pages and servlets using ContataStruts framework.
- Involved in creating database tables, writing complex TSQL queries and stored procedures in the SQL server.
- Used JAXB to read and manipulate the xml properties
- Analyzing the business requirements and preparing use-case diagrams usingStar UML.
- Responsible for developing conceptual designs from requirements.
- Responsible for writing detail design documents andclass diagrams and sequence diagrams.
- Design patterns used to achieve clean separation of different layers.Used different patterns likeMVC pattern,Session Façade, and singleton.
- Provide EAI concepts in integrating business with partners like BizTalk and Xcbl.
- UsedSVNas Version Control Software.
- Involved in developingcustomcomponents in Presentation tier usingJSF,Ice facestags.
- Prepared the test plans and executed test cases for unit testing.
- Performed Defect Tracking on a regular basis to make sure every defect TEMPhas traceability back to the source by using tools such asQuality Center.
Environment: JDK 1.6,Core java,JSF,ANT, Log4J, Junit, SQL, Linux, Eclipse 3.1 IDE, XML,Apache Axis2, Oracle WebLogic Server 11gR1 (10.3.2), Windows XP,SVN,SQL Developer, Oracle11g.
Confidential, Des Moines, IA
Java/J2EE Developer
Responsibilities:
- Responsible for understanding the scope of the project and requirement gathering.
- Designed a data-warehouse using dimensional modelling.
- Developed the web tier using JSP, Struts MVC, and Tiles to show account details and summary.
- Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
- Created and maintained the configuration of the Spring Application Framework (IoC).
- Implemented various design patterns - Singleton, Business Delegate, Value Object and Spring DAO.
- Used Spring JDBC to write some DAO classes to interact with the database to access account information.
- Mapped business objects to database using Hibernate.
- Involved in writing Spring Configuration XML files that contains declarations and other dependent objects declaration.
- Used Tomcat web server for development purpose.
- Involved in creation of Test Cases for JUnit Testing.
- Used CVS, Perforce as configuration management tool for code versioning and release.
- Developed application using Eclipse and used build and deploy tool as Maven.
- Used Log4J to print the logging, debugging, warning, info on the server console.
Environment: Java1.5, J2EE Servlet, JSP, XML, Spring 3.0, Struts 1.1, Hibernate 3.0, Design Patterns, Log4j, CVS, Maven, Eclipse, Apache Tomcat 6, Windows XP, Oracle 11g/10g/9i, SQL, PL/SQL, Erwin 7.0