Senior Software Engineer (big Data/hadoop/automation) Resume
San Ramon, CA
SUMMARY:
- 14+ years of work experience with fortune 100 companies like Confidential, Confidential, Confidential, and Confidential . Expertise in full project life cycle development of various enterprise applications along with next - generation Big Data and Cloud Computing technologies. Strong experience with clusters loaded with Hadoop ecosystem components, client-server technologies in a distributed environment, and responsibilities like automation, DevOps, platform, and database solutions including requirement gathering. Proven history of utilizing integration skills of various business intelligence and analytic applications. Extensive experience in automating product modules on different platforms.
- Experience with tech giants like Confidential, Confidential, Confidential Enterprise, Confidential, etc.
- Proficiency in problem solving, troubleshooting, multi-tasking, and documentation.
- Experienced in turning innovative ideas into effective results.
- Well organized, disciplined work ethic, Confidential ce, and positive attitude.
- Ability to manage stress, time, and people effectively.
- Created documents were considered as Industry s standard documents.
TECHNICAL SKILLS:
Big Data/Hadoop:: Confidential Predix, Microservices, AWS, Dockers, Kibernetis, Cloud Foundry, Timeseries,Cloudera, Hortonworks, MapR, Hive, HBase, Greenplum, HDFS, YARN, Sqoop
Languages:: Scala, Java, Python, Ant, Shell Scripting, Perl, C, C++, XML, PostScripts
BI/ETL/Analytics:: Tableau, OBIEE, Informatica, MicroStrategy, SAP, BO, SAP HANA, SAS
Databases/DW:: Predix Timeseries, Hive, PostgreSQL, Teradata, Oracle, MySQL, Trafodion
Automation/IDE:: Ansible, Puppet, Protractor, Cucumber, Jenkins, Selenium, Talend, EC, ALM
UX/Web:: AngularJS, JavaScript, PHP, JSP, CSS, ASP, HTML, Maven, JBoss, Tomcat
OS:: Linux (CentOS, Ubuntu, Red Hat), HP-UX, AIX, Windows, Mac OSX, DOS
PROFESSIONAL EXPERIENCE:
Confidential, San Ramon, CA
Senior Software Engineer (Big Data/Hadoop/Automation)
Technologies: Timeseries, Dockers, Python, AWS, Cloud Foundry, Java, Cassandra, Jenkins, REST, Greenplum, Postgres, Talend, Shell Scripts, Oracle, JIRA, Rally.
Responsibilities:
- Introduced methodologies and best practices using Docker containers that enhanced product definition, release processes and customization of applications to user needs.
- Architect, design, and develop QA validation process and automation suite according to customer requirements using Java, Junit, TestNG, Python, and Shell Scripting.
- Prepare and load test data in Cassandra and Postgress using Spark and Kafka and extensively worked on creating variations of test datasets to validate multi - tenancy features.
- Developed validation framework for functional validation as well as system integration to validate various components of the product and automated the entire validation and regression process.
- Extensive experience working with Greenplum and HAWQ data warehouse for Confidential Aviation data and implementing strategies for creating multi-partitioned data warehouse.
- Worked with data scientist team and helping them with data extraction of Confidential Oil & Gas data using complex SQL queries and performed QA/validation of the extracted data.
- Design a complete automated test solution by developing complex analytical functions and provided automated scripts for data validation in Greenplum as well as in HAWQ.
- Performed DevOps practices for project repositories and used Jenkins for build promotion pipelines.
- Worked with Talend to migrate data between databases from ERPs, DWHs, and Flat files to HDFS.
- Worked with physical data modelling for data warehouses and data marts and architected database providing high availability database solutions with automated daily refresh of the data.
- Interacting with business users and compiling end user requirements for development process.
- Contributed software engineering expertise in the development of software product for 3D printing project through the agile software development practices.
Confidential, San Jose, CA
Hadoop Lead/Data Engineer
Technologies: Hadoop, Hive, Sqoop, Teradata, Oracle, Tableau, Shell Scripts, Python, Rally, Linux.
Responsibilities:
- Worked with the developers, business analysts and subject matter experts on day-to-day bases to gather requirements and produce integrated end-to-end solution for the product.
- Worked on the development of dashboard reports for the Key Performance Indicators for the top management including SOX compliance reports.
- Wrote scoop scripts to pull data from datalike and load it into Hive.
- Worked on Data Modelling and wrote Hive scripts to transform data loaded in various schemas.
- Prepared shell scripts and automation scripts to eliminate manual processes in the project.
- Worked extensively with Tableau dashboards having various reporting objects like facts, hierarchies, transformations, filters, prompts, calculated fields, sets, groups, parameters etc.
Confidential, Palo Alto, CA
Software/Big Data Engineer, Hadoop Architect, DevOps/SRE
Technologies: Java, Python, Ant, Puppet, Ansible, Shell Scripts, Hadoop, HDFS, Hive, HBase, Git, Spark, Kafka, Maven, Jenkins, YARN, Selenium, Teradata, Informatica, BO, OBIEE.
Responsibilities:
- At HP, we had developed an open source product called Trafodion (trafodion.org), to provide a transactional layer on top of Hadoop. Being in R&D team, we were focused on architecting and building a data processing solution on a large-scale Hadoop clusters and performance tuning.
- Successfully developed and automated infrastructure components on HP Helion using various Big Data technologies like Git, Jenkins, Puppet, and Maven for cluster management and provisioning.
- Installed Spark and Kafka on clusters and written Kafka Rest API to collect events from front end.
- Automated build process for SeaQuest and Trafodion using Jenkins, Zuul, Git, Electric Commander and successfully completed integration with bug tracking tools like JIRA, and ALM/QC.
- Developed JUnit test cases to migrate existing test cases written in DFM framework and developed unit and functional tests in Java and a Python using Tox and Testr.
- Worked on automating test process modules and Trafodion APIs like ODBC, Problem Management, and Instance Management using Selenium, EC, Load Runner with Java, Python, Shell Script for the purpose of scalability analysis and data validation and comparing results with Oracle and Teradata.
- Created, maintained and administered GIT repository as source code management and configured and managed Jenkins for running the automated builds and deployments to CI dev servers.
- Worked with Cloudera, Hortonworks, and MapR to deploy major Hadoop eco system components like HBase, HDFS, Zookeeper, and automating the installation and configuration process.
- Conducted process for ISV products by performing multiple benchmarks against HP’s proprietary data warehouse and comparing results with benchmark against Teradata and Oracle.
- Installed, implemented, and supported Puppet masters and agents for configuration management along with successfully implementing CI and CD using Jenkins and Maven.
- Built a tool to automate backup/restore process for Trafodion data using HBase Snapshots and MapReduce and performed query level and HBase level tuning on large scale Hadoop clusters.
- Worked with driver manager like Data Direct and performed ODBC tuning to ensured ISV products work seamlessly and efficiently with Neoview/SeaQuest data warehouse.
- Created the master/slave pattern for Jenkins to distribute the load and created new build jobs, integration testing jobs, and deployment jobs to automate the process around the build farm.
- Developed test suits using Python and shell scripts for self-healing of the clusters and provided High Availability (HA) solution to architect and deliver optimal Big Data solution.
- Automated integration of various ISV products like OBIEE, Informatica, BusinessObjects, and SAS to ensure they work seamlessly and efficiently with Trafodion/SeaQuest/Neoview.
- Created and managed all dependency packages and worked with creating the Linux repositories and RPMs containing all required packages to deliver proposed product on Hadoop clusters.
- Performed system administration, troubleshooting, and debugging tasks on CentOS to ensure Trafodion supports enterprise-class operational capability like OLTP.
Confidential, CA
Software Engineer
Technologies: Siebel Analytics (OBIEE), Oracle, Teradata, Informatica, ETL, TOAD, BO, JavaScript.
Responsibilities:
- Worked with stakeholders for requirement gathering, solicitation, and validation.
- Converted complex business requirement into functional specification and technical specification.
- Delivered the Business Intelligence solution that was processing over 1 million transactions daily.
- Worked on developing accounting reports vital to booking $3.5 Billion annual revenue on P&L.
- Designed and developed relational data model with providing role-based security in the product.
- Developed an intuitive drilldown feature for all modules of the application using JavaScript.
- Involved in ETL processes, performance tuning, load balancing, and server level tuning.
- Strongly involved in designing the architecture of the proposed product for Disney.
- Developed a data migration application using OBIEE, Oracle, and MS SQL Server.
- Strongly involved in the process of database migration from Oracle DW to Teradata DW.
- Developed stored procedures and packages to populate fact and dimension tables.
Information Technology Intern
Technologies:Perl, MS SQL 2000, DOS Scripting, JavaScript, CVS, CSS, HTML.
Responsibilities:
- Designed a data model to enhance PRMS and implemented a role-based security.
- Developed scripts using Perl to migrate Customer Support System across environments.
- Worked with MS SQL server to make the proposed application efficient and scalable.
- Worked with DMZ machines, CVS, and developed scripts to control PRMS resources.
- Converted business document to thin-HTML and developed a portal to manage it.
Confidential, San Jose, CA
Database Administrator & Web Developer
Technologies:PHP, JavaScript, Oracle, MySQL, MS Access, Apache, CSS, HTML.
Responsibilities:
- Used PHP and JavaScript to develop various web application modules for the library.
- Involved in challenging DB migration and implementing online university applications.
- Implemented a business logic layer and a database layer of the application.
- Documented the requirements to create online tutorials for SJSU students.
- Performed database administrator tasks to maintain robust database.
- Maintained and managed various application servers and the database servers.
Confidential
Software Developer, System Specialist, and Database Administrator
Technologies:Visual Studio, Oracle DB, TOAD, MS SQL, MS Access, Dreamweaver, MS Project.
Responsibilities:
- Developed a prototype for Election Commissionerate and Industries Commissionerate.
- Involved in a complete software development life cycle of the product.
- Developed an application for offline barcode printing for ONGC, the 3rd largest in India.
- Carried out detailed study of the proposed system and requirement analysis.
- Generated DFD, ERD, URS, SRS, and screen designs according to CMM standards.
- Developed various components of the proposed product including core and reporting module.
- Administered large scale database and data integration process for the premium client (ONGC).
- Developed an application to generate PDF files using PostScript and Visual Studio.
- Successfully transmitted sensitive data using secure FTP for India General Election 2002.