We provide IT Staff Augmentation Services!

Big Data Developer Resume

5.00/5 (Submit Your Rating)

SUMMARY:

  • Software professional with 10 years of experience in FinTech, Big Data and Technology Consulting.
  • Experience in Apache Hadoop eco system components like HDFS, MapReduce, PIG and HIVE.
  • Experience in developing and configuring applications in AWS.
  • Good understanding of Hadoop architecture and hands on experience in Hadoop components such as Job Tracker, Task Tracker, Name node and Data node.
  • Analyzed large volumes of data sets using PIG scripts and HIVE scripts.
  • Advanced SQL skills, tuned slow running queries, used parallelization and hash techniques.

SKILL:

Database Technologies: Oracle, Microsoft SQL Server, MongoDB

Operating Systems: UNIX, Linux, Ubuntu

Methodology: Agile, SDLC

Languages: PL/SQL, java, JSP, servlets, JSON, C#, Python

Other Tools: Eclipse, Toad, NetBeans, Elasticsearch, Azure, AWS, Maven

Big Data Ecosystem: Hadoop, HDFS, Yarn, MapReduce, HIVE, PIG, Spark

PROFESSIONAL EXPERIENCE:

Big Data Developer

Confidential

Responsibilities:

  • Creating access points to the data through Spark, Presto and Python for business users.
  • Analyzing functionality in the existing hydra module.
  • Developing a stand - alone module to fill the missing functionality in hydra.
  • Analyzing the business logic available in Confidential for moving the files to one lake.
  • Designed the process of loading files received from partners in to data lake.
  • Assisted in pyspark application that does the manual ingestion of file in to one lake.

Environment: AWS, Nebula, Symphony, Pyspark, Spark-SQL, Python, Java, Presto, Cobol copybook, GIT, JIRA.

Big Data Developer

Confidential

Responsibilities:

  • Load flat files and json data files in to Hadoop Data Lake. Maintain the Data lake in Hadoop by building data pipe line using Scala, Hive and Spark.
  • Involved in Configuring EMR cluster in coordination with Internal AWS Environment team.
  • Assisted in developing an Oozie workflow job that triggers ELT process using EMR cluster.
  • Developed Spark application to load the data from S3 staging bucket to S3 core bucket for data analysis using Tableau.
  • Developed Spark application to load data from S3 core bucket into PostgreSQL for faster retrieval.
  • Developed Sqoop process to pull data from Hive and load it into PostgreSQL.

Environment: HDFS, Hadoop 2.x, Oozie, YARN, Zookeeper, TEZ, Cloudera, Hue, Confidential, Hive, Sqoop, Spark, Spark-SQL, PostgreSQL, Scala, JSON File Format, Parquet File Format.

Big Data Developer

Confidential

Responsibilities:

  • Analysis of GPS generated Confidential trip data to calculate Confidential ’s pricing, to finding optimal positioning of cars to maximize profits using clusters of Confidential data.
  • Clusters are used to determine “which hours of the data and which cluster had the highest number of pickups, how many pickups occurred in each cluster etc.
  • Load the data from MySQL in to Spark data frame using Sqoop.
  • Loading data into Spark DataFrame.
  • Transform features into feature vectors to be used by machine learning algorithm using Spark ML.
  • Used VectorAssembler to transform and return a new DataFrame. Create a KMeans object, set the parameters to define the number of clusters and the maximum number of iterations to determine the clusters.
  • Confidential is used to show the Confidential results in bar charts and graphs.

Environment: HDFS, Hadoop 2.x, Hive, Spark, YARN, Zookeeper, Hortonworks, Spark-SQL, Confidential, MySQL.

Application DBA

Confidential, CO

Responsibilities:

  • Maintenance and troubleshooting of oracle RAC environment.
  • Creating logical and physical models.
  • Backing up and modifying data in different databases.
  • Creating PL/SQL code for moving and modifying data. Working with users to identify business requirements.
  • Created perl scripts to load data using SQL Performance tuning of queries and application tuning.
  • Identifying bottlenecks and assist in creating project tasks and plans.
  • Create and modify users, access privileges and permissions.

Environment: -Oracle 11g, informatica, Erwin, MS Office, PRPC

Senior DBA / Analyst

Confidential, TX

Responsibilities:

  • Modifying the Maven Model Manager, the model manager is a graphical user interface tool that contains parameters for field operators to collect blood samples.
  • These parameters are modified by IT department to collect data by different topographies, like area, age, sex, income levels etc.
  • Designed and created Actuate reports for existing health modules, tuned the actuate reports for improved performance.
  • Worked with both synchronous and asynchronous reports. Writing simple to complex stored procedures to load and modify data in development, test and production environments.
  • Testing new model manager functionality and creating JIRA tickets for any identified or unresolved bugs that may appear in the system.
  • Updating the Maven data model as per user needs.
  • Working with business users to identify business requirements and convert these business requirements to technical specification documents.
  • Work on pending JIRAs tickets created by the user group and resolve them in a timely and efficient manner.

Environment: -Oracle 11g, java / J2EE, Maven, SQL server, JIRA tool set

Database Administrator

Confidential, MA

Responsibilities:

  • Worked with corporate applications, created a business intelligence pipeline to access and run ad hoc reports that showed different views of corporate data. Administration of RTS, a network monitoring tool that acts as a gateway between banks, store and internal Refund and loss prevention systems.
  • Administration and support of RMS (Refund Management system) and Aspect LP (Loss prevention) Administration and support of Remote ware server, an FTP server responsible for daily sales and settlement files from the stores and consolidating these files and pushing them to the main frame.
  • Worked with Triversity and Global store, these systems are third part packages used on POS sale machines and the controllers that manage store operations from processing payments to end of day operations.

Environment: -Retail Transaction Switch, SQL server 200/05/08, oracle 11g, Actuate

Database Administrator

Confidential, NJ

Responsibilities:

  • Used Er/Win to design structures and Meta data in the logical model and generated physical model.
  • Created partitions using Er/Win in the logical model, generated script to be run in the physical model.
  • Developed complex stored procedures to load data from external feed to staging and ODS layers.
  • Unix data pump loads to export and import data and Meta data, writing procedures and -functions, testing queries for data quality.
  • Administration of data stage jobs, creating direct data loads using Informatica.
  • Created Cognos reports and used the Cognos BI module.
  • Extensively used Oracle Data Integrator to for the web services module and batch job feeds.
  • Confidential and application performance and tuning, adhoc requests etc.
  • Successfully implemented CDW built all components for a flip schema mechanism by creating synonyms and granting the right privileges.
  • Schema creation, user ID management, roles and privileges, oracle parameter setup, oracle directories, synonyms, server monitoring
  • Performed production support identified issues and resolutions.
  • Performed unit, regression and performance testing.

Environment: oracle10g/11g, Datastage8.1, Autosys, excel, putty, ftp, Cognos, Informatica

Oracle Developer

Confidential, TX

Responsibilities:

  • Created monthly-consolidated report in sql containing summarized loan transaction data, bad loans, collection call volume, customer interaction volumes etc.
  • Created functional specification document to create Adhoc reports, Dashboard structures and web publishing reports Identified business requirements, created mapping documents, created test cases and issue tracking for Cognos reports for the business users of Defaults and downstream departments.
  • Translate requirements into formal documentation, developed test scripts and cases.
  • Design and develop workflows to automate business processes. Data collection and data analysis.

Environment: - SQL Server 2005, Autosys, excel

DBA / Oracle Developer

Confidential, NY

Responsibilities:

  • Provided ADM 24/7 support to the Contracts DW warehouse, identifying file load failures, data discrepancies, made crontab changes etc.
  • Made modifications to the physical model of the warehouse, by changing the logical model in Erwin and re-engineered the physical model in the respective schema.
  • Wrote UNIX shell scripts to import and export schemas.
  • Created data backups periodically.

Environment: - AIX-unix SQL Server 2005, oracle10g, Autosys

OracleDeveloper

Confidential, NJ

Responsibilities:

  • Created pl/sql stored procedure for consolidation effort between P&L accounts and billing systems.
  • Wrote queries to fetch data between their trading and billing systems for report generation.
  • Part of team implemented a trading system resulting in improved process times as well as utilization of system resources.
  • Wrote requirements to generate xml files from the database for the business users.
  • Performed unit testing on the in house developed trading system, validating data quality, performance.
  • Document and assess client operations Develop and present solutions and recommendations
  • Requirements gathering and prioritization of business requirement.

Environment: - SQL Developer 2005, oracle10g

Oracle Developer / Analyst

Confidential, IN

Responsibilities:

  • Wrote stored procedure to collect data using a cursor from the staging layer to the ODS layer and into the data mart to be stored as the master record.
  • Recommended freezing loan codes and mapping them to future applications thereby eliminating the need to create redundant codes thereby improving online processing time considerably.
  • Acted as onsite coordinator for an offshore team of 2 members.
  • This involved task planning and documentation, Issue resolution, development tracking and UAT before code submission to VSS.
  • Used data stage to bring data into the trading system from the billing systems for reconciliation purposes.

Environment: - Toad, oracle10g, SQL Server 2005, excel, PowerPoint

We'd love your feedback!