We provide IT Staff Augmentation Services!

Resume

3.00/5 (Submit Your Rating)

Big Data Developer Dallas, TX

SUMMARY:

  • Around 5 years of extensive experience in complete Software Development Life Cycle (SDLC) including gathering System Requirements, Design, Coding, Development, Testing, Production Support, Maintenance and Enhancement.
  • Hands on experience in installing, configuring and using Apache Hadoop ecosystems such as HDFS, MapReduce, HIVE, PIG, SQOOP, FLUME and OOZIE.
  • Hands on experience on Hortonworks and Cloudera Hadoop environments.
  • Strong understanding of Hadoop daemons and MapReduce concepts.
  • Strong understanding of Scala
  • Developed scripts in importing - exporting data into HDFS format.
  • Developed scripts in analyzing big data using Hadoop environment.
  • Developed scripts in handling Hadoop Ecosystem Projects such as Hive, Pig and Sqoop.
  • Developed in devUDFs for Hive using Java.
  • Strong experience in working with Elastic MapReduce and setting up environments on Amazon AWS EC2 instances
  • Strong understanding of NoSQL databases like HBase, MongoDB.
  • Experience in writing Microsoft SQL queries, Stored Procedures, Functions and Packages for various validation purposes using, MySQL, SQLServer, Crystal reports, TSQL, TOAD.
  • Experience in Data modeling for Data Warehouse/Data Mart development, Data Analysis for Online Transaction Processing (OLTP) and Data Warehousing (OLAP)/Business Intelligence (BI) applications and tuning SQL Server Integration Services (SSIS) packages and SSRS reports.
  • Comprehensive knowledge and experience in process improvement, Experience in extracting, transforming and loading (ETL ) data from spreadsheets, database normalization/de-normalization, data extraction, data cleansing, data manipulation.
  • Strong Programming Skills in designing and implementation of multi-tier applications using PHP, JSP, HTML, JavaScript, J Query, CSS, Bootstrap.
  • Experience in BI Development, SSIS packages from MS-Access, Excel, Oracle.
  • Designing and Deployment of Reports for the End-User requests using Web Interface & SSRS.
  • Experience in controlling the User privileges, Monitoring the Security and Audit issues as per Standards.
  • Expertise in generating reports using SQL Server reporting Services, Crystal Reports, and MS Excel Power View.
  • Excellent T-SQL Developer skills including Stored Procedures, Indexed Views, User Defined Functions, Triggers, and Distributed Transactions.
  • Implemented new solutions with emerging technologies for resolving issues such as inefficiency, performance bottleneck in distributed applications.
  • Proficient in Unit Testing and Integrated Testing.
  • Mentored team member to learn and use emerging technologies and domain knowledge.
  • Provided solutions to automate development and testing for web services-based application.
  • Strong analytical thinking and problem-solving skills backed by good communication skill.

TECHNICAL SKILLS:

Hadoop/Big Data: HDFS, MapReduce, Hive, Pig, Sqoop, Spark, Hbase, Flume, Oozie, and ZooKeeper, AWS (EMR)

Programming Languages: PHP,JAVA, C, C++, C#, XML, SQL, PL/SQL, Unix shell scripts.

Web/XML Technologies: HTML, CSS, JavaScript, JSP, DOM, XML.

Application/Web Servers: Apache Tomcat, IBM Web Sphere, BEA Web Logic.

Tools & Utilities: SQL Server Reporting Services (SSRS), ETL and SQL Server Integration Services (SSIS), Business Intelligence Development Studio (BIDS), Eclipse, Edit plus, Text pad, WinSCP.

RDBMS: MS SQL Server 2012, DB2, BI Tools SSIS, SSRS, Oracle10g,11g,12c., MySQL 4.0, T-SQL, MS Access.

Source Control: Subversion (SVN)

Operating Systems: Windows 9x/2000/XP, Linux, Unix.

PROFESSIONAL EXPERIENCE:

Confidential

Big Data Developer

Responsibilities:

  • Involved in extracting customer’s Big data from various data sources into Hadoop HDFS. This included data from databases and also log data from servers.
  • Used Sqoop to efficiently transfer data between databases and HDFS and used Flume to stream the log data from servers.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Implemented Partitioning, Bucketing in Hive for better organization of the data.
  • Developed User Defined Functions in Pig and Hive
  • Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java Map-reduce, Hive and Sqoop as well as system specific jobs.
  • Installed and configured Hive, Sqoop, Pig, Cloudera Manager and Oozie on the Hadoop Cluster.
  • Involved in implementing High Availability and automatic failover infrastructure to overcome single point of failure for Name node utilizing Zookeeper services.
  • Involved in upgrading the cluster from CDH 4.x to CDH5.x
  • Worked on Commissioning and decommissioning the Datanodes in Hadoop Cluster.
  • Experienced in access controls, security using Kerberos and Sentry.
  • Involved in Migrating the data from Test Cluster to Virtual Cluster.
  • Worked with BI teams in generating the reports on Tableau.

Environment: Sqoop, Pig, Hive, Oozie, Flume, Tableau, HDFS, Map Reduce

Confidential

SQL and Web Developer

Platform: Amanda 6, SQL Server 2012

Responsibilities:

  • Actively involved in software development life cycle starting from requirements gathering and performing Object Oriented Analysis.
  • Involved in Requirement analysis for implementing Amanda for Public Works division in Confidential .
  • Developed and optimized database structures, stored procedures, Dynamic Management views, and user-defined functions.
  • Involved in writing Amanda batch jobs, SQL storedprocedures, database functions, database triggers.
  • Created Common Codes folders for use across the team.
  • Created custom reports using Crystal Reports and Mail Merge Documents to generate the license from Amanda application
  • Written complex batch jobs and procedures for Importing and Exporting data.
  • Involved in SQL Server Configuration, Implementation and Trouble-shooting for Business work.
  • Involved in daily code drops and archives on the production databases during maintenance windows.
  • Created custom reports using SAPCrystal Reportsreports according to business requirement.
  • Developed user interface using JSP, HTML5, jQueryand performed Client-side validations using Java Script, and CSS.
  • Developed different Scripts for the Data analysts for Verification of reports.
  • Performed detailed documentation of Report design, development, testing and deployment activities.
  • Implementing all Amanda application changes into production
  • Maintaining the production system for day to day issues and resolve all the issues opened in Forge ticketing system
  • Attending daily meeting with IDPH ( Confidential ) to walk thru fixed issues and prioritize existing issues.
  • Watching daily/weekly interfaces runs smoothly between NREMT (National Registry of Emergency Medical Technicians) And IDPH, also other interfaces from Department of workforce development.

Environment: s: Business Intelligence Development Studio (BIDS), SQL Server Reporting Services (SSRS), ETL, SQL Server Integration Services (SSIS), Microsoft SQL Server 2012, Tomcat in a Linux Environment, Crystal Reports, JSP, Java Script, jQuery, HTML, CSS, Agile Methodology.

Confidential, Dallas, TX

ETL Developer

Responsibilities:

  • Involving in creating Technical Specification Document (TSD) for the project by conducting meeting with the users to meet their business requirements.
  • Extracted data from sources like SQL server, DB2 and fixed width and delimited flat files and Loading into the Oracle target database.
  • Responsible for developing ETL processes/programs with Informatica Power Center to extract data from the client's operational databases and Flat Files, Transform and filter the data, and Load it into Target database and Flat Files
  • Used various transforms in SSIS to load data from flat files, Siebel CRM and FTP to the SQL databases.
  • Using Informatica Power Center Work Flow manager to create sessions with the task developer and workflows with workflow manager to run the logic embedded in the mappings.
  • Creating mappings with Informatica PowerCenter 9.6.1 using Transformations (Expression Transformation, Router Transformation, and Stored Procedure Transformation etc.) and Look ups.
  • Analyzing session log files in session failures to resolve errors in mapping or session configuration.
  • Verified test results (viewing the log files using VI editor, querying the database using the SQL) documenting the defects.
  • Experience working on Lotus notes for the workspace environment.
  • Working with Control-M team and Involving in Production Support in resolving issues and bugs.
  • Experience in writing UNIX shell scripts to run and schedule batch jobs.
  • Extracted the data from Oracle, DB2, XML and Flat Files into Data warehouse.
  • Provided index tuning as required for performance improvement of the applications (SQL tuning).
  • Proficient in Unit Testing and Integrated Testing.

Environment: SQL Server Reporting Services (SSRS), ETL, Informatica, DB2, Microsoft SQL Server 2012, Putty Unix Environment, Water Flow Methodology .

We'd love your feedback!