We provide IT Staff Augmentation Services!

Big Data Architect Resume

4.00/5 (Submit Your Rating)

- Atlanta, GA

SUMMARY:

  • Above 8+ years of experience as Big Data Engineer Analysis, Design, Development, Deployment and Maintenance of software in Java/J2EE technologies and BigData applications.
  • Expertise in Data Development in Hortonworks HDP platform & Hadoop ecosystem tools like Hadoop, HDFS, Spark, Zeppelin, Hive, HBase, Sqoop, flume, Atlas, SOLR, Pig, Falcon, Oozie, Hue, Tez, Apache NiFi, Kafka.
  • Built streaming applications using SPARK Streaming.
  • Knowledge on big - data database HBase and NoSQL databases Mongo DB and Cassandra
  • Expertise in Java Script, JavaScript MVC patterns, Object Oriented JavaScript Design Patterns and AJAX calls.
  • Experience includes Requirements Gathering, Design, Development, Integration, Documentation, Testing and Build.
  • Experience in working with MapReduce programs, Pig Scripts and Hive commands to deliver the best results.
  • Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services which provides fast and efficient processing of Teradata Big Data Analytics.
  • Experienced in collection of Log Data and JSON data into HDFS using Flume and processed the data using Hive/Pig.
  • Strong knowledge and experience in Object Oriented Programming using Java.
  • Extensively worked on development and optimization of MapReduce programs, PIG Scripts and HIVE queries to create structured data for data mining.
  • Expertise in developing the presentation layer components like HTML, CSS, JavaScript, JQuery, XML, JSON, AJAX and D3.
  • Good knowledge of coding using SQL, SQL Plus, T-SQL, PL/SQL, Stored Procedures/Functions
  • Worked on Bootstrap, Angular JS and Node JS, knockout, ember, Java Persistence Architecture (JPA).
  • Hands on experience in advanced Big-Data technologies like Spark Ecosystem (Spark SQL, MLlib, SparkR and Spark Streaming), Kafka and Predictive analytics
  • Knowledge of the software Development Life Cycle (SDLC), Agile and Waterfall Methodologies.
  • Strong experience in developing Enterprise and Web applications on n-tier architecture using Java/ J2EE based technologies such as Servlets, JSP, Spring, Hibernate, Struts, EJBs, Web Services, XML, JPA, JMS, JNDI and JDBC.
  • Developed applications based on Model-View-Controller (MVC)
  • Working knowledge on Oozie, a workflow scheduler system to manage the jobs that run on PIG, HIVE and SQOOP.
  • Expertise in developing test cases for Unit testing, Integration testing and System testing.
  • Extensively development experience in different IDE like Eclipse, Net Beans, IntelliJ and STS.
  • Experienced with programming language such as C, C++, Xpath, Core Java and JavaScript.
  • Good experience in Installing, Upgrading and Configuring Redhat Linux using Kickstart Servers and Interactive Installation.
  • Good experience in Tableau for Data Visualization and analysis on large data sets, drawing various conclusions.
  • Extensive experience in building and deploying applications on Web/Application Servers like Web logic, Web sphere, and Tomcat.
  • Expertise in core Java, J2EE, Multithreading, JDBC, Hibernate, Spring, Shell Scripting and proficient in using Java API's for application development.
  • Good at problem-solving skills to identify areas of improvement and incorporating best practices for delivering quality deliverables
  • Have good experience, excellent communication and interpersonal skills which contribute to timely completion of project deliverable well ahead of schedule.

TECHNICAL SKILLS:

JAVA (7 years), APACHE (6 years), JAVASCRIPT (5 years), MODEL VIEW CONTROLLER (5 years), MODEL-VIEW-CONTROLLER (5 years)

PROFESSIONAL EXPERIENCE:

Big data architect

Confidential - Atlanta, GA

Responsibilities:

  • Provide technical leadership and contribute to the definition, development, integration, test, documentation, and support across multiple platforms.
  • Design/ architected and implemented complex projects dealing with the considerable data size (GB/ PB) and with high complexity.
  • Provide deployment solutions based on customer needs with Sound knowledge about the clustered deployment architecture.
  • Able to guide / partner with VP / Directors for architecting solutions for the Big data Organization
  • Created detailed AWS Security groups which behaved as virtual firewalls that controlled the traffic allowed reaching one or more AWS EC2 instances.
  • Data modeling, Design, implement, and deploy high-performance, custom applications at scale on Hadoop /Spark.
  • Data processing with MapReduce and Spark.
  • Stream processing on Spark/storm thru Kafka message broker.
  • Review and audit of existing solution, design and system architecture.
  • Perform profiling, troubleshooting of existing solutions.
  • Create technical and designing documentation.
  • Creation of a User Interface to search and/or view content within the cluster by using solar cloud.
  • Worked on AWS provisioning EC2 Infrastructure and deploying applications in Elastic load balancing.
  • Cluster management and analytic in Cloudera and Horton work.
  • Distributed database Design, Data modeling, Development and Support in Datastax Cassandra distribution.
  • Cassandra products strengths and weakness to produce efficient schema designs that serves effective and high performance queries.
  • Maintain and work with our data pipeline that transfers and processes several terabytes of data using Spark, Scala, Python, Apache Kafka, Pig/ Hive & Impala
  • Apply data analysis, data mining and data engineering to present data clearly.
  • Ensure high-quality data and understand how data is generated out experimental design and how these experiments can produce actionable, trustworthy conclusions.
  • Full life cycle of Data Lake, Data Warehouse with Big data technologies like Spark, Hadoop, Cassandra
  • Working with Spark, RDD, Data Frames, Data Pipelines.
  • Building complex ETLs, Data Warehousing or custom pipelines from multiple data sources.
  • Setting up connector for security logs and Splunk data use cases.
  • Building the Hadoop cluster (MTS) to host the three use cases
  • Analyzing the data using Tableau.
  • Extract and analysis the data before load into cluster.
  • Review and understand data architecture, data models, Source to target mapping rules and Match and merge rules.
  • Evaluate Hadoop infrastructure requirements and design/deploy solutions (high availability, big data clusters, elastic load tolerance, etc.).
  • Hadoop ecosystem components in our open source infrastructure stack specifically: HBase, HDFS, Map/Reduce, Yarn, Oozie, Pig, Hive, Kafka, Storm, Spark, Spark-SQL and Flume.
  • Estimate and obtain management support for the time, resources and budget required to perform in different projects.
  • Keep track of the new requirements / change in requirements of the Project.
  • Understand Inbound and outbound data flow requirements, data models for Landing, Staging and base objects, Mapping documents, Match and Merge rules.
  • Proof of Concept (POC) and Proof of Technology(POT) execution and evaluation on MTS platforms.
  • Installing and Configuring required ecosystem tools for each use case

Environment: Big Data, Spark, YARN, HIVE, Pig, Scala, Python, Hadoop, AWS, Dynamo DB, Kibana, Cloudera, EMR, JDBC, Redshift, NOSQL, Sqoop, MYSQL.

Sr. Big Data/ Hadoop Developer

Confidential - Florhan - Park, NJ, US

Responsibilities:

  • Coordinated with business customers to gather business requirements. And also interact with other technical peers to derive Technical requirements and delivered the BRD and TDD documents.
  • Implement Big Data systems in distributed cloud environment (AWS) using Amazon EMR
  • Extensively involved in Design phase and delivered Design documents. Experience in Hadoop eco system with HDFS, HIVE, PIG, SQOOP and SPARK with SCALA.
  • Worked on analyzing Hadoop cluster and different Big Data Components including Pig, Hive, Spark, HBase, Kafka, Elastic Search, database and SQOOP. Installed Hadoop, Map Reduce, HDFS, and developed multiple Map-Reduce jobs in PIG and Hive for data cleaning and pre-processing.
  • Importing and exporting data into HDFS and Hive using SQOOP.
  • Migration of 100+ TBs of data from different databases (i.e. Netezza, Oracle, SQL Server) to Hadoop. Wring code in different applications of Hadoop Ecosystem to achieve the required output in a sprint time period
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Involved in creating Hive tables, loading with data and writing Hive queries that will run internally in map reduce way. Generate OBIEE reports to verify the Hive tables data.
  • Experienced in defining job flows. Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting. Experienced in managing and reviewing the Hadoop log files.
  • Used Pig as ETL tool to do Transformations with joins and pre-aggregations before storing the dataonto HDFS. Responsible to develop data pipelines from different sources.
  • Wrote Hive and Impala queries to load and processing data in Hadoop File system.
  • Utilized Apache Hadoop environment by Cloud era Distribution. Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
  • Worked on Oozie workflow engine for job scheduling. Involved in Unit testing and delivered Unit test plans and results documents.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Hue, Ganglia, Nagios, Java, Kafka, Elastic Search, SQL, Scala, Oracle, Netezza, Ambari, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse.

Confidential - Orlando, FL

Full Stack Java Devoloper

Responsibilities:

  • Done the design, development and testing phases of Software Development using AGILE methodology and Test-Driven Development (TDD)
  • Designed the application using Front Controller, Service Controller, MVC, Spring DAO, Factory, Data Access Object, Service Locator and Session Facade Design Patterns.
  • Involved in development of the applications using Spring Web MVC and other components of the Spring Framework, the controller being Spring 3.0 Core (Dispatcher Servlet).
  • Used Hibernate for Object relational model for handling server side / database object data.
  • Work closely with our partners and clients to develop and support ongoing API integrations.
  • Used SOAP and REST based web service protocol.
  • Bootstrap used along with Angular 2 and EcmaScript7 with TypeScript in creating the Application.
  • Provide leadership in developing the company's IT software and ongoing strategy for internal and external purposes.
  • Ensure Client & Subsidiaries are in compliance with all domestic and international data privacy lAWS ensuring that the flow of confidential data is secured and in compliance with local jurisdictions.
  • Developed Linux bash & MS DOS Scripts for internal use.
  • Used PostgreSQL as back end and developed Stored procedures, Batch jobs, triggers.
  • Used Jenkins to perform software build with Gradle, run shell Scripts and worked on integration tests.
  • Develop customized reports for clients or internal customers using Pentaho - ETL tools.
  • Utilized React for its efficient data flow architecture to create a lightweight and render efficient web app that searched projects via the GitHub API through keywords
  • Designed Frontend with in object oriented JavaScript Framework like Bootstrap, Node.js, Express.js and Angular.js, Redux.
  • Wrote a Python module to connect and view the status of an Apache Cassandra instance.
  • Implemented log4j by enabling logging at runtime without modifying the application binary.
  • Developed Mean Stack from scratch including Mongo DB server setup and Express JS server development.
  • Worked in using React JS components, Forms, Events, Keys, Router, Animations, and Flux concept.
  • Used popular Node.js frameworks like Express and RESTify to create a RESTful Mock API.
  • Provide estimates, designs, and specifications for AEM templates, components, and workflows.

Environment: MVC, Factory, Session Facade Design Patterns, Spring, SOAP, RESTful web services, Angular.js, Linux bash, MS DOS, Hibernate, PostgreSQL, Dynatrace, Git, Github, Bootstrap, Node.js, log4j, Rally, AWS.

Sr. Java Developer

Confidential - Germantown, MD

Responsibilities:

  • Worked on the existing application, wireframes, FDN and BRD documents to get the requirements and analyzed.
  • Hands-on Experience with Cassandra to provide Scalability along with NoSQL.
  • Developed Agile processes using Groovy, JUnit to use continuous integration.
  • Integrated Automated functional tests (Groovy) with Continuous-Integration in Jenkins.
  • Parse requests and built response data using Groovy's JSON tools, and Grails web services.
  • Imported data from various resources to the Cassandra cluster using Java APIs.
  • Used Eclipse SWT for developing the applications.
  • Involved in preparation of TSD documents using UML diagrams - Class, Sequence and Use case diagrams using Microsoft VISIO tool.
  • Wrote RESTful services on the server in NodeJS to listen to requests from devices.
  • Built a Grails web application that allows admin users to manage detailed data for all types of Target locations
  • Have worked with Standard Widget Toolkit (SWT).
  • Conversion of major Openwork's components in to Eclipse RCP/SWT platform along with support of Swing-SWT components.
  • Involved in to develop view pages of desktop portal using HTML, Java Script, JSP, Struts Tag libraries, AJAX, JQUERY, GWT, DOJO, XML, and XSLT.
  • Developed and deployed Web services to interact with partner interfaces, and client interfaces to consume the web services using CXF, WSDL, SOAP, AXIS and JAX-WS technologies.
  • Integrating third party libraries to augment those lacking or inefficient in ExtJS.
  • Used RESTful web services using JERSEY tool to develop web services easily and to be invoked by different channels.
  • Developed service objects as beans by using Spring IOC/DI.
  • Developed Web API using NodeJS and hosted on multiple load balanced API instances.
  • Implementation of enterprise application with jQuery, angularJS, node.js and SpringMVC.
  • Used Spring Beans to encapsulate business logic and Implemented Application MVC Architecture using Spring MVC framework.
  • Implemented Hibernate (ORM Mapping tool) framework to interact with the database to update, retrieve, insert and delete values effectively.
  • Used Java Swing for few components in accordance with SWT application with multithreading environment with Concurrency and Java Collections.
  • Used EH Cache for second level cache in Hibernate for the application.
  • Involved in to pass messages like payload to track different statuses and milestones using EJB, JMS.
  • Involved in unit testing, integration testing, SOAP UI testing, smoketesting, system testing and user acceptance testing of the application.
  • Used Spring programmatic transaction management for Java Persistence.
  • Involved in integration of Spring and Hibernate frameworks.
  • Involved in setting server properties, DSs, JNDI, queues & deploying app in WebSphere Application Server.
  • Followed the test driven development using the JUNIT and Mockito framework.
  • Created continuous integration builds using Maven.
  • Involved in fixing QA/UAT/Production issues and tracked them using QC.

Environment: Java, JSP, Servlets, JavaScript, Spring DI, Spring IOC, Spring AOP, Hibernate 3.0, AJAX, XML, XSLT, JAXP, JAXB, AXIS, CSS, CXF, WSDL

Java Developer

Confidential - Plano, TX

Responsabilites:

  • Developed Controller Servlets and Action Servlets to handle the request and responses.
  • Developed and coordinated complex high quality solutions to clients using IBM Products/Tools, Apache Tools, J2SE, J2EE, EJB, Servlets, JSP, HTML, JavaScript, JQuery, JSON and XML.
  • Developing the web applications using Spring Framework, Hibernate
  • Applying Spring Framework for transaction Management and Spring JDBC for building ORM, and for AOP and Dependency Injection
  • Responsible for using AJAX framework with JQuery, Dojo, ExtJs implementation for Widgets and Events handling.
  • Customizing Log4J for maintaining information and debugging.
  • Customizing third party vendor information using Web services (SOAP and WSDL).
  • Developed Request Cash Message and get Cash plugins using Java Beans.
  • Worked with development of data access beans using Hibernate, middle ware web service components.
  • Develop the GUI using JSP, Spring web flow following Spring web MVC pattern.
  • Implemented persistence layer using Hibernate that use the POJOs to represent the persistence database tables.
  • Used SVN for version control across common source code used by developers.
  • Written the JUNIT test cases for the functionalities.

Environment: J2EE/J2SE, Java1.5, JSP, Ajax4JSF, JSF 1.2, Spring Frame Work 3, Hibernate, JMS, CSS3, Apache CXF, XML, HTML, Oracle.

We'd love your feedback!