We provide IT Staff Augmentation Services!

Big Data Architect/developer Resume

3.00/5 (Submit Your Rating)

Segundo, CA

SUMMARY

  • Around 7+ years' of IT experience in software Development and support with experience in developing strategic methods for deploying big data technologies to efficiently solve Big Data processing requirement.
  • Expert consulting and training services to Fortune 500 clients and startups.
  • Verticals: Financial, Legal / eDiscovery, Online Advertising, Appliances
  • Implemented Hadoop based data warehouses, integrated Hadoop with Enterprise Data Warehouse systems
  • Built real - time Big Data solutions using HBase handling billions of records
  • Built scalable, cost-effective solutions using Cloud technologies
  • Implemented Big Data analytical solutions that 'close the loop' and provide actionable intelligence
  • Developed free text search solution with Hadoop and Solr. Analyzing emails for compliance and eDiscovery.
  • Taught courses and seminars on Big Data and Cloud technologies
  • Expert knowledge in Hadoop/HDFS, MapReduce, HBase, Pig, Sqoop, Amazon Elastic Map Reduce (EMR), Accumulo.
  • Cloud: Amazon EC2, EMR, Rackspace, Google Cloud.
  • Experience with distributed systems, large-scale non-relational data stores, RDBMS, NoSQL map-reduce systems, data modeling, database performance, and multi-terabyte data warehouses.
  • Working experience in Hadoop framework, Hadoop Distributed file system and Parallel processing implementation.
  • 3 years of hands on experience in Hadoop Framework and its ecosystem including but not limited to HDFS Architecture, MapReduce Programming, Hive, Pig, Sqoop, Hbase, Oozie etc.
  • Working experience with large scale Hadoop environments build and support including design, configuration, installation, performance tuning and monitoring.
  • Experience in application development using the technologies Java, RDBMS, Linux/Unix shell scripting and Linux internals.Excellent knowledge on Hadoop Architecture and ecosystems such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce.
  • Experience in strong and analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Experience in installation, configuration and management of development, testing and production Hadoop Cluster.
  • Performed Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in working with Flume to load the log data from multiple sources directly into HDFS.
  • Experience in designing both time driven and data driven automated workflows using Oozie.
  • Experience in writing UNIX shell scripts.
  • Quick learning skills and effective team spirit with good communication skills
  • Strong analytical and Problem solving skills.
  • Good Inter personnel skills and ability to work as part of a team. Exceptional ability to learn and master new technologies and to deliver outputs in short deadlines.

TECHNICAL SKILLS

Hadoop: HDFS, Map Reducing, HDFS, Oozie, Hive, Pig, Sqoop, Flume, Zookeeper and Hbase, Cassandra.

Server SideScripting: UNIX Shell Scripting

Database: Oracle 10g, Microsoft SQL Server, MySQL, DB2

Programming Languages: Core Java, JSP, Impala, JDBC, SQL, PL/SQL, HQL, PIG Latin and Python J2EE technologies Web Tier (JSF, Struts, JQuery, DOJO, JSON, AJAX, HTML, CSS, Icefaces) J2EE Technologies Business Tier (EJB, JMS, Web services JAX-WS) J2EE Technologies Persistence (Hibernate, JPA, Spring JDBC)

Web Servers: WebLogic, Apache, Tomcat, WebSphere, JBoss

OS/Platforms: Windows 2008/Vista/2003/XP/2000/NT, Linux, Unix.

Client side: JavaScript, CSS, HTML, JQuery

XML: XML, HTML, DTD, XML Schema

Methodologies: Agile, UML, Design Patterns

PROFESSIONAL EXPERIENCE

Confidential, Segundo, CA

Big Data Architect/Developer

Responsibilities:

  • Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Provide proof-of-concepts to reduce engineering churn.
  • Give extensive presentations about the Hadoop ecosystem, best practices, data architecture in Hadoop
  • Provide mentorship and guidance to other architects to help them become independent.
  • Provide review and feedback for existing physical architecture, data architecture and individual code.
  • Debug and solve issues with Hadoop as on-the-ground subject matter expert. This could include everything from patching components to post-mortem analysis of errors.
  • Experience in Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced in defining Ooziejob flows and make them automata.
  • Provided proof of concepts converting Avro data into parquet format to faster the query processing by using Impala.
  • Experienced in managing and reviewingHadooplog files.
  • Participated in development/implementation of Cloudera Hadoop environment.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Experience in working with various kinds of data sources such as MongoDb and Oracle.
  • Successfully loaded files to Hive and HDFS from Mongo DB.
  • Installed Oozie workflow engine to run multiple map-reduce programs which run independently with time and data.
  • Performed Data scrubbing and processing with Oozie.
  • Responsible for managing data coming from different sources.
  • Gained good experience with NOSQL database.
  • Experience in working with Flume to load the log data from multiple sources directly into HDFS.
  • Supported Map Reduce Programs those are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.
  • Worked in installing cluster, commissioning & decommissioning of data node, name node recovery, capacity planning, and slots configuration.
  • Implemented best income logic using Pig scripts.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.

Environment: Apache Hadoop, Map Reduce, HDFS, Hive, Java, SQL, PIG, Zookeeper, Cassandra, Java (jdk1.6), Flat files, Oracle 11g/10g, MySQL, Windows NT, UNIX, Sqoop, Hive, Oozie.

Confidential, Houston, TX

Hadoop Developer

Responsibilities:

  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbasedatabase andSqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from LINUX file system to HDFS.
  • Devised and lead the implementation of the next generation architecture for more efficient data ingestion and processing.
  • Proficiency with mentoring and on-boarding new engineers who are not proficient in Hadoop and getting them up to speed quickly.
  • Experience with being a technical lead of a team of engineers.
  • Proficiency with modern natural language processing and general machine learning techniques and approaches.
  • Extensive experience with Hadoop and HBase, including multiple public presentations about these technologies.
  • Experience with hands on data analysis and performing under pressure.
  • Designed and wrote a layer on top of MapReduce to make the task of writing MapReduce jobs easier and safer for Junior Engineers.
  • Analyzed large data sets by running Hive queries and Pig scripts
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Developed Simple to complex MapReduce Jobs using Hive and Pig
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Mentored analyst and test team for writing Hive Queries.
  • Develop and maintains complex outbound notification applications that run on custom architectures, using diverse technologies including Core Java, J2EE, SOAP, XML, JMS, JBoss and Web Services.
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Good knowledge onAgile Methodologyand the scrum process.
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing
  • Implemented a script to transmit sysprin information from Oracle toHbase using Sqoop.
  • Implemented best income logic using Pig scripts and UDFs.
  • Implemented test scripts to support test driven development and continuous integration.
  • Worked on tuning the performance Pig queries.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Responsible to manage data coming from different sources.
  • Involved in loading data from UNIX file system to HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data
  • Cluster coordination services through Zookeeper.
  • Experience in managing and reviewing Hadoop log files.
  • Job management using Fair scheduler.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reportsfor the BI team.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and Troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Installed Oozie workflow engine to run multiple Hive and pig jobs.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig andSqoop.
  • Crawled some websites usingPython and collected information about users, questions asked and the answers posted.
  • Hands- on experience in developing web applications using Python on Linux and UNIX platform.
  • Experience inAutomation Testing, Software Development Life Cycle (SDLC)using theWaterfall Modeland good understanding ofAgileMethodology.

Environment: Java 6, Eclipse, Oracle 10g, Sub Version, Hadoop, Hive, Hbase, Linux,, MapReduce, HDFSHive, Java (JDK 1.6), Hadoop Distribution of HortonWorks, Cloudera, MapReduce, Data Stax, IBM Data Stage 8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting

Confidential, NYC, NY

Hadoop Developer

Responsibilities:

  • Exported data from DB2 to HDFS using Sqoop.
  • Developed Map Reduce jobs using Java API.
  • Installed and configured Pig and also wrote Pig Latin scripts.
  • Wrote Map Reduce jobs using Pig Latin.
  • Developed workflow using Oozie for running Map Reduce jobs and Hive Queries.
  • Worked on Cluster coordination services through Zookeeper.
  • Worked on loading log data directly into HDFS using Flume.
  • Involved in loading data from LINUX file system to HDFS.
  • Responsible for managing data from multiple sources.
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Responsible to manage data coming from different sources.
  • Assisted in exporting analyzed data to relational databases using Sqoop.
  • ImplementedJMSfor asynchronous auditing purposes.
  • Involved in developingMessage Driven and Session beansfor claimant information integration with MQ based JMS queues.
  • Created and maintained Technical documentation for launching Cloudera Hadoop Clusters and for executing Hive queries and Pig Scripts
  • Experience indefining, designing and developing Java applications, specially using Hadoop Map/Reduce by leveraging frameworks such as Cascading and Hive.
  • Worked on installing cluster, commissioning & decommissioning of data node, name node recovery, capacity planning, and slots configuration.
  • Created Hbase tables to store variable data formats of PII data coming from different portfolios.
  • Experience in Architect and build Turn's multi-petabyte scale big data Hadoop infrastructure
  • Experience in Develop monitoring and performance metrics for Hadoop clusters.
  • Experience in Document designs and procedures for building and managing Hadoop clusters.
  • Experience in Partner with Hadoop developers in building best practices for Warehouse and Analytics environment.
  • Strong Experience in troubleshooting the operating system, maintaining the cluster issues and also java related bugs.
  • Experience in working with various kinds of data sources such as MongoDb Solar and Oracle.
  • Successfully loaded files to Hive and HDFS from Mongo DB Solar.
  • Experience with installing and Configuring on Open LDAP with SSL/TLS on Debian.
  • Experience in configuring Kerberos with Open LDAP at back end.
  • Led initiatives to automated application builds and deployments using Hudson/Jenkins.
  • Experience in Automate deployment, management and self-serve troubleshooting applications.
  • Define and evolve existing architecture to scale with growth data volume, users and usage.
  • Investigate emerging technologies relevant to our needs and also Mentoring Sr. Hadoop Administrators on Hadoop Cloudera cluster management best practices.
  • Design and develop JAVA API (Commerce API) which provides functionality to connect to the Cassandra through Java services.
  • Installed and configured Hive and also written Hive UDFs.
  • Familiarity with a NoSQL database such as MongoDb Solar.
  • Experience in managing the CVS and migrating into Subversion.
  • Experience in managing development time, bug tracking, project releases, development speed, release forecast, scheduling and many more. Using a custom framework of Nodes and MongoDb to take care of the back-end calls with a lightning fast speed. Intensive Object-Oriented JavaScript, jQuery and plug-ins are used to work on dynamic user interface.

Environment: Hadoop, HDFS, Pig, Sqoop, HBase, Shell Scripting, Maven, Hudson/Jenkins, Ubuntu, Linux Red Hat, Mongo DB.

Confidential, Atlanta, GA

JAVA Developer

Responsibilities:

  • Worked with sprint planning, sprint demo, status and daily standup meeting.
  • Developed the application using Spring Web MVC framework.
  • Worked with Spring Configuration files to add new content to the website.
  • Worked on the Spring DAO module and ORM using Hibernate. Used Hibernate Template and HibernateDaoSupport for Spring-Hibernate Communication.
  • Extensively used Spring's features such as Dependency Injection/Inversion of Control to allow loose coupling between business classes (POJOs)
  • Used Hibernate framework to retrieve and update information and dependency Injection is achieved by Spring MVC Framework.
  • Configured Association Mappings such as one-one and one-many in Hibernate
  • Worked with JavaScript calls as the Search is triggered through JS calls when a Search key is entered inthe Search window
  • Worked on analyzing other Search engines to make use of best practices.
  • Collaborated with the Business team to fix defects.
  • Worked on XML, XSL and XHTML files.
  • Interacted with project management to understand, learn and to perform analysis of the Search Techniques.
  • Used Ivy for dependency management.
  • As part of the team to develop and maintain an advanced search engine, would be able to attain expertiseon a variety of new software technologies.

Environment: Java 1.6, J2EE, Eclipse SDK 3.3.2, Java Spring 3.x, JQuery, Oracle 10i, Hibernate, JPA, JsonApache Ivy, SQL, stored procedures, Shell Scripting, JQuery, XML, HTML and JUnit, TFS, Ant, VisualStudio Premium 2010, Rational Clear quest

Confidential

Java Developer

Responsibilities:

  • Played an active role in the team by interacting with welfare business analyst/program specialists and converted business requirements into system requirements.
  • Developed analysis level documentation such as Use Case, Business Domain Model, Activity & Sequence and Class Diagrams.
  • Conducted Design reviews and Technical reviews with other project stakeholders.
  • Implemented Services using Core Java.
  • Developed and deployed UI layer logics of sites using JSP.
  • Struts (MVC) is used for implementation of business model logic.
  • Worked with Struts MVC objects like Action Servlets, Controllers, and Validators, Web Application Context, Handler Mapping, Message Resource Bundles and JNDI for look-up for J2EE components.
  • Developed dynamic JSP pages with Struts.
  • Used built-in/custom Interceptors and Validators of Struts.
  • Developed the XML data object to generate the PDF documents and other reports.
  • Used Hibernate, DAO, and JDBC for data retrieval and medications from database.
  • Messaging and interaction of Web Services is done using SOAP.
  • Developed JUnit Test cases for Unit Test cases and as well as System and User test scenarios
  • Involved in Unit Testing, User Acceptance Testing and Bug Fixing.
  • Implemented mid-tier business services to integrate UI requests to DAO layer commands.

Environment: J2EE, JDBC, Java 1.4, Servlets, JSP, Struts, Hibernate, Web services, SOAP, WSDL, Design Patterns, MVC, HTML, JavaScript 1.2, WebLogic 8.0, XML, JUnit, Oracle 10g, My Eclipse.

We'd love your feedback!