Big Data Engineer Resume
Atlanta, GA
PROFILE:
To pursue a challenging career and be a part of a progressive organization that gives scope to enhance my knowledge, skills and reach the pinnacle in Software Engineering with sheer determination, dedication and hard work.
PROFESSIONAL SUMMARY:
- Over 10 years of IT professional experience with 3+ Year of Big Data Hadoop Ecosystems experience in ingestion, storage, querying, processing and analysis of big data.
- Experienced in installing, configuring, and administrating Hadoop cluster of major Hadoop distribution.
- Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4), Yarn distributions.
- Experience in building, maintaining multiple Hadoop clusters (prod, dev etc) of different sizes.
- Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop HDFS, Yarn, MapReduce, HBase, Oozie, Hive, Pig, Flume, Storm, Kafka, and Knox.
- In depth knowledge of Hadoop Architecture and Hadoop daemons such as Name Node, Secondary Name Node, Data Node, Job Tracker and Task Tracker.
- Experience in writing Map Reduce programs using Apache Hadoop for analyzing Big Data.
- Good working experience in PySpark and SparkSql.
- Experience in using Zookeeper and Oozie Operational Services for coordinating the cluster and scheduling workflows.
- Good experience on Design, configure and manage the backup and disaster recovery for Hadoop data.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience in understanding the security requirements for Hadoop and integrating with Kerberos authentication infrastructure - KDC server setup, creating realm /domain, managing.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
- Experience in implementations of Service Oriented Architecture (SOA) using Web Services (SOAP, WSDL and UDDI).
- Configured Servers like DNS, DHCP, Samba, NFS, NIS, send mail, http, daemon, Ftp, Telnet, SSH, Firewalls and package management using RPM patch management.
- Experience with VMware, vSphereClient, Oracle VM and data center migration
- Experienced in Logical Volume Management on Linux with LVM, Weblogic server, Apache, tomcat.
- Highly motivated, self-starter with a positive attitude, willingness to learn new concepts and acceptance of challenges.
TECHNICAL SKILLS:
Big Data Ecosystem: HDFS, MapReduce, Spark, Spark SQL, Scala, Impala, YARN, Hive, Pig, Zookeeper, Sqoop, Apache Nifi, Oozie, Flume, Kafka.
Web Technologies: HTML, XML, JDBC, JSP, JavaScript, AJAX, REST, SOAP, jQuery.
NOSQL Technologies: HBase, MongoDB, Cassandra
Web/Application Servers: Apache Tomcat 5.5/6.0, Glass Fish 3.0, WebLogic 10.3, Web Sphere 6.1
Databases: Oracle 11g/10g/9i, DB2, MS-SQL Server, Teradata, IBM, MySQL
Languages: C, C++, Java, SQL, PL/SQL, Python, Scala, Unix shell scripting, VB
Web Servers: Web Logic 10.3, Web Sphere 6.1, Apache Tomcat 5.5/6.0.
Cloud: AWS, AZURE and ISILON
Operating Systems: Linux RHEL/Ubuntu/CentOS/Solaris, VMware ESX, Windows (XP/7/8)
Networking Tools: NIS, NFS, DNS, TFTP, Ethernet, Fast Ethernet switches, Load balancers, Network monitoring, Communication Protocols: TCP/IP, SNMP, SMTP, OSI, DHCP, and TCP Wrappers
Reporting Tool: Tableau & Salesforce Wave
PROFESSIONAL EXPERIENCE:
Confidential, Atlanta, GA
Big Data Engineer
Environment: Hadoop 2.7.2, Hive, Hbase, Yarn, Kafka, Apache Nifi, Java JDK 1.8, Scala 2.10, SQL Server, Spark 1.6.0, 1.4, REST web services, Wave, Maven, Gradle, RHL 6.5, Apache Solr, GitHub, micro services, Spring boot, log4j 1.2, Docker, Kubernetes, Tableau, Jenkins, AWS
Responsibilities:
- Involved in all phases of Software Development Life Cycle (SDLC) and Worked on all activities related to the development, implementation, administration and support for Hadoop.
- Leveraged the Hortonworks platform by using the Hadoop, Hive, Hbase and Spark service in the project.
- Working in creating the processor in Apache Nifi to automate the flow of data between systems.
- Good understanding in working on Apache Spark Integration with Apache Nifi to stream the data to Hadoop HDFS.
- Using the micro services architecture developed the spring boot Java code as a Docker containers.
- Developed the code by consuming the Kafka API and Apache Spark API to stream the data flow in real time.
- Developed the RDD and data frames using the Scala machine language solved the given metrics in Scala IDE.
- Coded the Scala programs for various metrics given by the business by submitting those metrics as spark jobs to generate the results and populating on the Wave dashboard.
- Designed the architecture and automated the entire process by sending the input files from different teams saving those files to HDFS in CSV and Parquet formats by reading them into spark master class process the Spark job to Yarn cluster and generate the output to HDFS.
- Worked on analyzing the real-time data coming from Confidential and write that data to Spark metrics analyze and give the business solution to the teams.
- Developed the rest full web services in spring boot using annotations and provide those rest endpoints to the team.
- Integrated end to end Tableau, Cognos with Hive connection using the ODBC driver and doing data analysis.
- Installed and configured the AWS Classic load balancers over the existing AWS Containers.
- Developed the Alerts API which will handle real time stream of data which will subscribe and write it to database, also developed the unsubscribe and retrieval process from the database.
- Performed the day to day deployments to productions and making sure all the application servers are running.
- Developed the Docker builds cycle and automated the individual applications as a Docker containers.
- Converted and parsed the xml’s data to JSON and write that JSON data to Kafka in the form of topics.
- Worked in developing multiple Gradle Java projects for Marketo leveraging Java API created the rest endpoints reading the data from Static Lists processing to SFDC.
- Worked in creating the complex SQL queries in the Scala platform to run and execute millions of records over each metrics.
- Configured the Jenkin nodes and did the Jenkins configuration to automatically build the project and download the jars to AWS instances and Hadoop box.
- Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
- Performed the day to day deployments to productions and making sure all the application servers are UP and running.
- Worked on POC for Contract project to save the data into CASSANDRA database also written the API’s to connect to Apache Solr for faster search operations.
- Involved in designing, capacity arrangement, cluster set up, performance fine-tuning, monitoring, structure planning, scaling and administration.
- Monitor Hadoop cluster connectivity and performance also setup new Hadoop users.
- Installed certificates on the new nodes where using the SSL for Yarn cluster.
- Deployed all the Docker images in the Kubernetes cluster orchestration.
Confidential, Houston, TX
Senior Software Engineer
Environment: InfluxDB, Java JDK 1.8, Spark 1.4, Scala 2.10, Hadoop, Hive, Maven, Ubuntu 14.04, Bootstrap, IgniteUI, AngularJS 1.3, Web services, Maven, Github, Wireshark, Microsoft Azure, VMware, IOT devices, Sensor Technology, JAMON, SAM Electronics.
Responsibilities:
- Install, configure, and troubleshoot components of the applications in IOT platform.
- Lead day-to-day database administration and ensure integrity, security and availability of applications environment.
- Installed a multi-node Hadoop cluster using Ambari HDP installation for our POC project.
- Researched on the ways to utilize our real-time data coming in from Sensors by saving that data into Hadoop Hive and Hbase and given demo to the client.
- Running our Spark jobs over the Hadoop MapR box.
- Integrated current NS applications by designing database architecture and server scripting; studying and establishing connectivity with network systems, search engines, and information servers.
- Developed and Build Ubuntu 14.0 servers for supporting existing NS applications.
- Build the Automation system software using JavaAPI’s to get the real time data from the IOT sensors and write to InfluxDB database.
- Experienced in collecting and developing the data from IOT sensors like AirMar sensors to get speed and temperature of ship, Geolocation Sensor to get latitude and longitude of ships and others.
- Used Wireshark to track the packet level communication to identify byte streams to develop the software. Developed application using UDP protocol for packet level communication.
- Used TCP socket communication.
- Ensure network configurations and resolve the day to day network connectivity issues.
- Performed the complete databases upgrade for InfluxDB and migrated all the data without having any data loss for this time series databases.
- Developed NS application by implementing business requirements, using the following technologies: Ubuntu 14, J2EE, XML, Unix, Linux, scripting, configuration, InfluxDB, Eclipse.
- Worked in the fast pace platform with InfluxDB similar to MongoDB, Cassandra databases.
- Automated the application by doing the process of data extraction and data imports to onshore instance.
- Build the mailProcessor software to automatically download and read the mails to dump the data to the database and the log the last events to DB using POP3, IMAP & SMTP protocols.
- Designed and Developed single page application using AngularJS and JQuery IgniteUI to show the real-time data.
- Implemented metric field level conversion plug-in capabilities for capture adapters allowing for sensor field specific adjustments prior to logging. This is used to convert NMEA GPS latitude and longitude to be converted to decimal degrees before logging to influx.
- Developed the Report Engine to automatically pull the data daily reports and mail back to the receipt back based upon the sent request.
- Data Visualization pattern design and development of the software.
Confidential, Lafayette, LA
Senior Software Consultant
Environment: RabbitMQ, TIBCO SmartSockets, Oracle (BEA) Weblogic server, EJB 3.0, ClearQuest, GitHub, JMS, AngularJS 1.3, Java 1.6, Java Swings, Struts 1.2,Web services, SOAP, Synergy, ANT(1.8.4), J-Unit 3.8, log4j 1.2, MYSQL, Red Hat Linux (RHEL 5.10, 6.4), Oracle 10g &11g, VMware, vSphere ESX
Responsibilities:
- Led Design and Documentation to replace the existing message Broker which is TIBCO SmartSockets in replacing with RABBITMQ on AMQP protocol.
- Involved in writing few python/bash scripting to automate the jobs.
- Configured and build the physical servers with RHEL 6.4 Version, and good working experience and understand on Linux platform servers, added the users to the LDAP to grant access.
- Integrate applications using web services, including SOAP and RESTful web services, and MQ Queues using JMS.
- Experience in using Angular UI for developing a responsive interface over bootstrap, followed MVC pattern on the UI using AngularJS and developed everything else as Data Rest Services.
- Worked with server side Java Script with AngularJS version 1.2 using JQuery.
- Good Understanding on current DMS application with complete end to end Prototype Developing of RabbitMQ Message Broker.
- Designed and developed Web Service Client Handler and SSO Engine projects as services to Member portal with XML as the communication media.
- Created REST FULL web services and deployed them using Spring Framework.
- Developed MDB for listening to a Queue configured on AMQP Protocol. Worked on other EJB’s including Session beans.
- Worked in the daily agile development methodology.
- Used Hibernate an ORM tool along with spring to perform all the Database actions.
- Worked with Quality Assurance Group to resolve critical bugs and answer technical questions.
- Created security clients for message broker RABBITMQ by doing secure encryption, authentication mechanisms for the DMS application.
Confidential, Austin & Dallas, TX
Java Developer / Linux Admin
Environment: Tomcat 6.0, Oracle (BEA) WebLogic server, EJB 3.0, JMS, Tortoise Hg, Java 1.6, Struts 1.2, Web services, SSO, TOAD, Apache Open sources, SOAP, Synergy, XSD, XML, ANT(1.8.4), J-Unit 3.8, log4j 1.2, MYSQL, HTML5, Red Hat Linux (RHEL 5.10, 6.4), HP-UX, SunOS, Oracle 10g &11g, Sybase, DB2 Client, TCP/IP, DNS, DHCP, NFS, Web logic, Tuxedo, Apache, POP3, SMTP, VMware, vSphere ESX.
Responsibilities:
- Implemented backend and frontend applications including interfaces with internal and external applications in the company (CRM & OMS).
- Integrate applications using web services, including SOAP and RESTful web services, and MQ Queues using JMS.
- Design and develop J2EE applications for Mundial Mexico for fraud prevention systems.
- Develop SMS and charging applications for Telcel operators using Java support third-parties for integration with Telcel charging.
- Developed the entire architecture using standard design patterns.
- Developed tools to integrate them in all kinds of devices using AJAX calls.
- Developed web applications for both the OSS and the Softswitch using Struts, EJB with OC4j and WebLogic application servers.
- Installation and configuration of Red Hat Enterprise Linux (RHEL) 5x, 6x Servers on HP, Dell Hardware and VMware virtual environment.
- Worked on Hitachi, EMC SAN. Worked on Disk Array based Environments.
- Working with firewall team to run firewall rules for server going to be built behind firewall and setup rules for server to be built in DMZ. Configure and Support FC-SAN environment and LUN provisioning.
- Developed the code with the help of python scripting to automatically backup all ESXi host servers to the scheduled time.
- Configuring and maintaining virtual server environment using VMWARE ESX/ESXi, VSphere, VCenter.
- Create, Manage and configure virtual machine(VM)s and VM templates on VMWare platform.
- Worked with backup team to backup Linux servers using Netbackup
- Worked with VMware ESXi 4.0, Oracle Linux, Oracle VM and third party HP Software Automated system.
- Configured Linux servers as clients to satellite server to use it as repository server and used it for quarterly patching.
- Configuration and administration of DNS, LDAP, NFS, NIS, NIS+ and Sendmail on Redhat Linux servers.
- Configured Crash Dumps, Linux Kernel Tunable, OS patching/upgrading on Linux Servers
- Written shell scripts for system/application in automating server task, installing and monitoring applications and data feeding file transfer and log files.
- Worked with Aquality Assurance Group in mexico to resolve defects and answer technical questions.
Confidential, Phoenix, AZ
Software Engineer
Environment: Glassfish(3.1), Java 1.5, Java1.7, Spring MVC (2.5),Web services, SSO,TOAD, Apache Open sources, SOAP, REST, StarTeam(2.0), XSD, XML, ANT(1.8.4), J-Unit 3.8, log4j 1.2, WSS4j 1.5.3, SAAJ1.2.
Responsibilities:
- Developed custom templates to display content using Marriott Rewards.
- Designed and developed the project utilizing the spring container and leveraged Spring MVC for developing the view components.
- Involved in development of a Single-Sign on mechanism for users to seamlessly login to multiple Marriott Rewards redemption websites with the help of a secure client side Cookie.
- Developed the entire web application from GUI Front end to Backend.
- Developed Web service calls along with the digital signature specifications provided by Confidential .
- Worked with multiple teams for Web services and SSO under Confidential and coordinated accordingly.
- Developed reusable and configurable Adapters for each service implementation under SSO and web services. Also Developed tools to integrate them onto mobile devices using AJAX calls.
- Developed MixedPayments for the Credit card transactions under Confidential Standards
- Developed jquery mobile slider bar on the tickets sites.
- Also developed the mobile websites for Android and Mac devices.
- Involved in designing and creating the Database Tables and Views.
- Developed the entire architecture using standard design patterns, Business Delegate, Session Facade, Value Objects.
Confidential, Boston, MA
Software Consultant
Environment: IBM RAD, IBM Web Sphere 6.1, Java1.5, Spring 2.5.5, Spring web flow, Struts 2.0, Web services, Apache Axis 1.4, Apache Open sources, SOAP, WSDL, XSD, XML, ANT, J-Unit 3.8, log4j 1.2, WSS4j 1.5.3, SAAJ1.2.
Responsibilities:
- Involved in designing and creating the Database Tables and Views.
- Hibernate is used to access data from the DB.
- Worked on Agile Methodology of SDLC process.
- Involved in writing Ant Scripts for building and deploying the war files.
- Involved in writing JUnit Scripts for testing the application.
- Worked with the IWAY integration tool. Involved in importing the database schemas.
- Configured spring to maintain the transaction on the service layer.
- Created a mapping xml file and used Dozer to map objects from the hibernate layer to the data transfer objects.
- Developed a Query Engine, which dynamically builds the query from an xml depending on the input parameters and runs it against the database.
- Maintained the MVC pattern using Struts frame work on the User Interface layer, which then connects to spring frame work in the Service layer and this talks to the hibernate at the data layer to retrieve data from the database.
- Involved in creating stored procedures and writing triggers on the database.
- Generation of unit test plan, testing and code review.
Confidential
Web Developer
Environment: Ajax, Tomcat, Java and Oracle Database, JSP, HTML, CSS, Photoshop, IBM Web Sphere 5.0, Java1.4, Oracle 10g, J-Unit 3.8, log4j 1.2.
Responsibilities:
- Developed the static webpages in HTML and JSP based on the given designs.
- Involved in designing and creating the Database Tables and Views.
- Writing well designed, testable, efficient code by using best software development practices.
- Creating website layout/user interfaces by using standard HTML/CSS practices.
- Integrating data from various back-end services and databases.
- Gather and refine specifications and requirements based on technical needs.
- Stay plugged into emerging technologies/industry trends and apply them into operations and activities
- Create and maintain software documentation.
Confidential
Jr. Developer
Environment: Windows xp/2003 server, Java and Oracle Database, Visual C++, .Net Framework 2008, TOAD for Oracle, IBM RAD, IBM Web Sphere 5.0, Java1.4, Oracle 10g, J-Unit 3.8, log4j 1.2.
Responsibilities:
- Involved in designing and creating the Database Tables and Views.
- Supervised the full software development lifecycle (analysis, design, code, test and document) from inception to production transition for web application with J2EE and JBOSS application server.
- Used ValueObject, ValueListHandler, Data Transfer Objects, Data Access Object, and Session Façade design patterns in development.
- Involved in Requirements gathering and designing the architecture for entire module.
- Created a mapping xml file and used Dozer to map objects from the hibernate layer to the data transfer objects.
- Developed a Query Engine, which dynamically builds the query from an xml depending on the input parameters and runs it against the database.
- Maintained the MVC pattern using Struts frame work on the User Interface layer, which then connects to spring frame work in the Service layer and this talks to the hibernate at the data layer to retrieve data from the database.
- Implemented the business logic using Stored Procedures with a set of SQL Queries.
- Involved in creating stored procedures and writing triggers on the database.
- Generation of unit test plan, testing and code review.