We provide IT Staff Augmentation Services!

Sr. Big Data Developer Resume

2.00/5 (Submit Your Rating)

Minnetonka, MN

SUMMARY:

  • 11+ years of IT consulting experience in full life cycle (SDLC) of teh software development process including gathering, analyzing, designing, testing, interface developing and implementing of Client/Server, Distributed, Internet and
  • E - Commerce application using Object Oriented Methodologies and Structured Development Methodologies. Complete multi-tiered application development lifecycle experience using Java/J2EE Technologies.
  • Three Years of Big Data Ecosystem experience which includes Big Data processing, design, development, analysis and admin in Telecom, Insurance and Financial domains.

PROFESSIONAL EXPERIENCE

Confidential, Minnetonka MN

Sr. Big Data Developer

Responsibilities:

  • Analyzing and researching teh data of different source system.
  • Experience in Enterprise Data Warehousing and design.
  • Experience in data architecting and cloud designing.
  • In-depth understanding of Data Structures and Algorithms.
  • Experience in writing Shell Scripts (bash, SSH, Perl).
  • Designed and developed Talend incremental and delta load to hive and hbase tables
  • Did POC on Impala/MapR drill/Jethro along BI tools for teh performance and for our use cases
  • Supported various downstream teams to Data Mapping and Data Analysis
  • Supported on teh incremental load and history load/refresh of data to Data lake.
  • Extracted, parsed and processed teh raw JSON/XML file using PIG and talend job.
  • Done Unit testing, data validation, reconciliation data manipulation and so on.
  • Designed and developed streaming framework for teh logs and streaming data using spark streaming
  • Worked in Spark to read teh data from Hive and write it to Cassandra using Java.
  • Involved in designing Enterprise Data Warehouse Hive metastores.
  • Created different flavors of Hive tables, snapshot tables, historical tables and incremental tables.
  • Designed and Implemented MapReduce jobs Hive/Hbase table schemas and queries.
  • Expertise in writing Shell scripts to monitor Hadoop jobs.
  • Worked on Tableau/Qlikview for data visualization POC of teh reporting tables on hive with various SQL engines like Jethro/Drill/Spark
  • Sqoop and Hive query optimization and enhancement.
  • Experience with cloud platform (AWS) security management.
  • Coordinated with Hadoop Admins to increase clusters size and also configured myself on single pseudo distributed cluster.
  • Exposure on spark - batch and real-time processing.
  • Designed, Monitored and managed scalable and fault tolerant deployments and supported them in real-world scenarios.
  • Worked on Tableau for teh project to create dashboard and data visualization metrics.
  • Worked on ETL packages and DQM (Data Quality Management) framework.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Created teh workflows on Oozie to coordinate teh Hadoop Jobs.
  • Worked on Talend to create teh ETL jobs and processes.
  • Production support on numerous of Hadoop/Talend jobs for 24*7 environment.

Environment: MapR, Talend, Hive, Hbase, Flume, Java, Scala, Spark, Pig, Oozie, Oracle, Impala, MapR Drill, Sqoop, AWS, Kafka, Qlikview, Python, YARN, SQL, Teradata, Unix, Spark, MongoDB, Tableau

Confidential, Kansas City, MO

Hadoop/Big Data Developer

Responsibilities:

  • Setup and configured Hadoop daemons and clusters and eco-system.
  • Designed, Monitored and managed scalable and fault tolerant deployments and supported them in real-world scenarios.
  • Worked with different data format such as Avro, Json, parquet, ORC and more
  • Built Data pipeline processing to support Data Warehousing structure.
  • Extracted, parsed and processed teh raw JSON file using PIG.
  • Involved in writing Pipelines, MapReduce jobs and different aggregation functions on Java.
  • Done Unit testing with Junit.
  • Developed/Maintained ETL process to move data between Oracle and Cloudera HDFS/Hive
  • Used Hive or R to manipulate data in Cloudera big data platform.
  • Worked closely with HDFS and MapReduce while data pipeline process.
  • Experience with Splunk on creating search and indexing of teh data.
  • Worked with open source frameworks like Puppet/Chef for deploying and configuration
  • Experience with data ingestion and forwarding data to Splunk using flume forwarder
  • Involved building and managing NoSQL Database like Hbase or Cassandra.
  • Worked in Web Services such as REST and SOAP.
  • Worked in Spark to read teh data from Hive and write it to Cassandra using Java.
  • Involved in data integration, migration on ETL Informatica environment.
  • Involved in designing and Enterprise Data Warehouse in Hive.
  • Experienced with related/complementary Big Data open source software platforms and languages.
  • Implemented teh Spark technology on DEV for teh processing of teh streaming data and RDBMS data
  • Designed and Implemented MapReduce jobs Hive/Hbase table schemas and queries.
  • Expertise in writing Shell scripts to monitor Hadoop jobs.
  • Worked on Tableau for data visualization of teh reporting tables on hive.
  • Experience on performance tuning and fine tuning on teh hive queries and MapReduce.
  • A significant working experience of UNIX, RHEL (Linux) commands and architecture.
  • Created and maintained SOLR indexes and searches.
  • Designed and implemented Logical Data Models and data service layer in Hadoop.
  • Had good understanding on teh OLAP/OLTP system and ETL architecture.
  • Hardened Hadoop clusters for deploying into production and staging environments.
  • Linux System Administrating, understanding storage, filesystem, disks, mounts nfs etc.

Environment: HDFS, Hive, Hbase, Flume, Java, Scala, Pig, Oozie, Oracle, Tez, Storm, Sqoop, AWS, Cassandra, Splunk, Kafka, YARN, Hortonworks, SQL, Platfora, Unix, Spark, MongoDB, Tableau

Confidential, Newark NJ

Hadoop Developer

Responsibilities:

  • Worked on teh proof-of-concept for Apache Hadoop1.20.2 framework initiation.
  • Installed and configured Hadoop clusters and eco-system.
  • Developed automated scripts to install Hadoop clusters.
  • Monitored Hadoop cluster job performance and capacity planning.
  • Hands-on experience with Hadoop technology stack (HDFS, MapReduce, Hive, Hbase, Flume)
  • Involved in designing and developing of data-centric solutions for clients.
  • Had experience in high scale or distributed RDBMS
  • Created and Implemented highly scalable and reliable highly scalable and reliable distributed data design using NoSQL/Cassandra technology.
  • Had experience in Hadoop framework, HDFS, MapReduce processing implementation.
  • Good understanding of Big Data products in teh market.
  • Tuning Hadoop performance with high availability and involved in recovery of Hadoop clusters.
  • Discovered how to add or remove nodes from teh Cassandra cluster.
  • Experienced managing No-SQL DB on large Hadoop distribution systems such as: Hortonworks HDP, MapR M series, Cloudera etc.
  • Provided UNIX support and administration.
  • Automated all teh jobs starting from pulling teh Data from different Data Sources like MySQL to pushing teh result set Data to Hadoop Distributed File System using Sqoop.
  • LeveragedETLsoftware, Implementation Architects must analyze,assemble and transform Client data filesinto a format consumable byHadoopprocessing system
  • Experience developing Hadoop integrations for data ingestion, data mapping and data process capabilities.
  • Deep JVM noledge of heavy experience with Functional Programming language like Scala
  • Worked with teh production Environment on AWS, high availability practices and deploying backup/restore infrastructure.
  • Refactored Cassandra-access code, to allow either Hector or Thrift access, replacing teh original thrift code interspersed throughout teh application
  • Designed Hadoop jobs to verify chain-of-custody and look for fraud indications.
  • Involved in ETL environment to push complex data into Hadoop and analysis.
  • Application performance optimization for a Cassandra cluster.
  • Knowledge on teh real-time message processing systems (Storm)

Environment: Hadoop, HDFS, MapReduce, Unix, REST, Python, Pig, Hive, Hbase, Storm, NoSql, Flume, Zookeeper, Kibana, Cloudera, SAS, Vertica, Kafka, Cassandra, Informatica, Teradata, Spark

Confidential, Tampa, FL

Hadoop/Java Developer with ETL

Responsibilities:

  • Worked as ETL Architect to make sure all teh applications are migrated (along with server) smoothly.
  • Migrated data to teh HDFS from traditional DMBS.
  • Deep understanding and related experience with Hadoop stack - internals, Hbase, Hive, Pig and Map/Reduce
  • A deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment
  • Managed mission-critical Hadoop cluster at scale especially Hortonworks.
  • Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms
  • Involved in developing and debugging Java/J2EE
  • Wrote Hive Queries and UDF’s.
  • Worked closely with teh Enterprise Data Warehouse.
  • Experience on AWS clouding computing platform, it many services and dimensions of scalability.
  • Implemented Fair schedulers on teh Job tracker to share teh resources of teh Cluster for teh Map Reduce jobs given by teh users.
  • Upgrading teh Hadoop Cluster to CDH4 and setup High availability Cluster. Integrated HIVE with external applications using JDBC/ODBC bridge
  • Familiar with many use cases of Storm such as; real-time analytics, online machine learning, continuous learning, ETL and more.
  • Provided UNIX support and administration experience.
  • Converted unstructured data to structured data using Pig scripting for testing and validation.
  • Experienced with Map Reduce or stream processing using Storm.
  • Automated all teh jobs starting from pulling teh Data from different Data Sources like MySQL to pushing teh result set Data to Hadoop Distributed File System using Sqoop.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Experience in designing, implementing and maintaining of high performing Hadoop clusters and integrating them with existing infrastructure.
  • Performed complex Linux administrative activates as we as created, maintained and updated Linux shell scripts.
  • Designed and supported highly available and scalable Linux infrastructure in 24*7 environment
  • Specifying teh Cluster size, allocating Resource pool, Distribution of Hadoop by writing teh specification texts in JSON File format.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Strong experience on Apache server configuration.
  • Exported teh result set from HIVE to MySQL using Shell scripts.
  • Develop HIVE queries for teh analysts.
  • Had good noledge in technologies like Sqoop, Flume and Kafka.
  • Built data Fabric with Flume, Kafka and Sqoop.
  • Maintain System integrity of all sub-components (primarily HDFS, MR, HBase, and Flume).

Environment: Hadoop, HDFS, MapReduce, Storm, Hive, Pig, Sqoop, Oracle, SQL, MySQL, UNIX Shell Scripting, PL/SQL, Lucene, Vertica, Teradata, Linux, IBM BigInsights, MongoDB, Java, Servlets, C++

Confidential, Jacksonville, FL

Java/J2EE Developer

Responsibilities:

  • Responsible for gathering all required information and requirements for teh project.
  • Experience inAgile Programmingand accomplishing teh tasks to meet deadlines.
  • Used Ajax and JavaScript to handle asynchronous request, CSS to handle look and feel of teh application.
  • Involved in design ofClassDiagrams, SequenceDiagramsandEventDiagramsas a part of Documentation.
  • Developed teh presentation layer using CSS and HTMLtaken from Bootstrap to develop for multiple browsers including mobiles and tablets.
  • Extended standard action classes provided by theStruts frameworkfor appropriately handling client requests.
  • Monitored and scheduled teh UNIX scripting jobs.
  • Designed, developed and did maintenance of data integration programs in a Hadoop and RDBMS environment with both traditional and non-traditional source systems as we as RDBMS and NoSQL data stores for data access and analysis.
  • Wrote teh Map Reduce jobs using Java.
  • Experienced working on ETL/Data Warehousing environment (DataStage or Informatica)
  • ConfiguredStruts tilesfor reusing view components as an application of J2EE composite pattern.
  • Involved in teh integration of Struts and Spring 2.0 for implementing Dependency Injection (DI/IoC). Developed code for obtaining bean s inSpringIoC framework.
  • DesignedDTO, Business Delegate, Factory and Singletondesign patterns.
  • Developed teh application onEclipse.
  • Involved in teh implementation of beans inApplication.
  • Migrated ETL Informatica code by using team based versioning.
  • Hands on experience in web services, distributed computing, multi-threading, JMS etc.
  • Implementedcross cuttingconcerns as aspects at Service layer usingSpring AOP.
  • Involved in teh implementation of DAO objects using spring - ORM.
  • Involved in creating theHibernate POJO’s and developedHibernate mapping Files.
  • UsedHibernate, object/relational-mapping (ORM) solution, technique of mapping data representation from MVC model to Oracle Relational data model with a SQL-based schema.
  • DevelopedSQL queriesandStored ProceduresusingPL/SQLto retrieve and insert into multiple database schemas.
  • Developed Ant Scripts for teh build process.
  • Version Controlwas mandated throughSubversion.
  • Performed Unit Testing UsingJUnit andLoad testing usingLoadRunner.
  • ImplementedLog4Jto trace logs and to track information.

Environment: Java, Struts, JSP, JSTL, JSON, JavaScript, JSF, POJO's, Hibernate, Hadoop, spring, Teradata, PL/SQL, CSS, Log4j, JUnit, Subversion, Informatica, Eclipse, Netezza, Jenkins, Git, Oracle 11g, LoadRunner, ANT

Confidential, San Jose, CA

Java/J2EE Developer

Responsibilities:

  • Created design documents and reviewed with team in addition to assisting teh business analyst / project manager in explanations to line of business.
  • Responsible for understanding teh scope of teh project and requirement gathering.
  • Involved in analysis, design, construction and testing of teh online banking application
  • Developed teh web tier using JSP, Struts MVC to show account details and summary.
  • Used Struts Tiles Framework in teh presentation tier.
  • Designed and developed teh UI using Struts view component, JSP, HTML, CSS and JavaScript.
  • Used AJAX for asynchronous communication with server
  • Utilized Hibernate for Object/Relational Mapping purposes for transparent persistence onto teh SQL Server database.
  • Used Spring Core for dependency injection/Inversion of control (IOC), and integrated frameworks like Struts and Hibernate.
  • Developed ETL mapping testing, correction and enhancement and resolved data integrity issues.
  • Involved in writing Spring Configuration XML files that contains declarations and other dependent objects declaration.
  • Used Tomcat web server for development purpose.
  • Involved in creation running of Test Cases for JUnit Testing.
  • Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
  • Used CVS for version controlling.
  • Developed application using Eclipse and used build and deploy tool as Maven.
  • Used Log4J to print teh logging, debugging, warning, info on teh server console.

Environment: Java, J2EE Servlet, JSP, JUnit, AJAX, XML, JSON, CSS, JavaScript, Spring, Struts, Hibernate, Log4j, CVS, Maven, Eclipse, Apache Tomcat, and Oracle.

Confidential, Minneapolis MN

J2EE Developer

Responsibilities:

  • Created UML class diagrams that depict teh code’s design and its compliance with teh functional requirements.
  • Used J2EE design Patterns for teh Middle Tier development.
  • Developed EJB’s in WebLogic for handling business process, database access and asynchronous messaging.
  • Used Java Mail notification mechanism to send confirmation email to customers about scheduled payments.
  • Had heavy experience on UI development.
  • Developed Message-Driven beans in collaboration with Java Messaging Service (JMS) to communicate with teh merchant systems.
  • Also involved in writing JSP’s/JavaScript and Servlets to generate dynamic web pages and web content.
  • Wrote Stored Procedures and Triggers using PL/SQL.
  • Involved in building and parsing XML documents using SAX parser after retrieving payment history data from teh database.
  • Deployed teh application on JBOSS Application Server.
  • Used Clear Case for version controlling and configuration management.
  • Very Strong noledge in using J2EE based App Servers like Jboss, WebSphere, WebLogic, and Web servers like Apache Tomcat.
  • Experience in implementing Web Services using SOAP, REST and XML/HTTP technologies.

Environment: Java, JSP, JSTL, EJB, JMS, JavaScript, JSF, XML, JBOSS, WebSphere, WebLogic, Hibernate, spring, SQL, PL/SQL, CSS, Log4j, JUnit, Subversion, Eclipse, Oracle 11g, LoadRunner, ANT.

We'd love your feedback!