Senior Big Data Developer Resume
SUMMARY:
- Over Eighteen years of experience in the IT Industry, with hands on experience on Big Data Hadoop (Eco Systems and Technologies) and Software Development.
- Deploy and manage scalable Hadoop cluster environments.
- Expertise on Hadoop Eco Systems & Environment (Spark, Hive, Pig, Kafka, Sqoop, Oozie, Flume, Zookeeper, Airflow).
- Installing/Deploying in AWS EC2/EMR/S3
- Installing/Deploying in Microsoft Azure (Virtual Machines, Storage Accounts, Databases, HDInsights)
- Strong development back ground in Scala, Perl, Python, C, C++ and Core Java.
- Optimize and tune the Hadoop environments to meet performance requirements.
- Multiple years of experience on DWH and ETL tools and best practices.
- Multiple years of RDBMS porting & development experience in Oracle and MySQL.
- Domains worked on Social media, CRM, SCM, Banking, Health Care, Insurance, Business intelligence, Telecom, Consulting.
TECHNICAL SKILLS:
DBMS: Oracle, MS SQL Server, Postgres, Hbase, Cassandra
Languages: Scala, Python, C, C++, Java
Big Data (Hadoop EcoSystem): MR/MR2, Pig, Hive, Kafka, Flume, Spark, Oozie, Zookeeper, Sqoop, Airflow
Tools: HUE, Tableau, Git, Bit bucket, Phabricator, Presto, Docker, Informatica, Arcadia, Jira, Qubole, Jenkins
Hadoop Distribution: Cloudera
AWS: EC2, EMR, S3, Redshift, Athena
Microsoft Azure: (Virtual Machines, Storage Accounts, Databases, HDInsights Clusters)
IDE: Pycharm, IntelliJ, Eclipse
Scripting: Unix, PERL
OS: Mac, Unix, Windows
PROFESSIONAL EXPERIENCE:
Senior Big data Developer
Confidential
Responsibilities:
- Create hive related workflows to load every day data which stream lines into reports and graphs for the corporate insights and decision making.
- Maintain the workflow in Python/hive, for adding, scheduling and cascading jobs.
- Tune and optimize hive queries/spark applications which take time due to influx in data.
- Back End development using Scala/Spark
- Maintaining and deploying in AWS for Spark and Hive jobs (S3, EC2, EMR)
Environment: and Tools: Spark, Scala, Python, Hive, AWS, Qubole, Airflow, Presto, Git, Phabricator, Jira, UNIX.
Senior Big data Developer
Confidential
Responsibilities:
- Analyzed large amounts of data sets to determine optimal way to aggregate and report using Scala and Spark.
- Install and configure Hadoop Eco Systems, develop multiple MapReduce jobs in Scala combining data cleansing and preprocessing.
- Managing and scheduling jobs on a cluster.
- Also use tools (Tableau and Arcadia) to create presentable graphs thru data modelling
- Continuous monitoring and managing the Hadoop cluster and generate data pipelines through Microsoft Azure.
Environment: and Tools: Hadoop, Microsoft Azure, Scala, Spark, Hive, Java, Oracle, Informatica, Arcadia, Kafka, UNIX.
Senior Hadoop Developer
Confidential
Responsibilities:
- Implemented best income logic using Pig scripts and UDFs.
- Analyzed the data by performing Hive queries (HiveQL) and using Pig scripts.
- Implemented test scripts to support test driven development and continuous integration.
- Responsible to manage data coming from different sources.
- Worked on analyzing/transferring data on different business requirement using Sqoop, Pig, Hive, and Oracle.
- Responsible for building scalable distributed data solutions using Hadoop.
Environment: Hadoop (Cloudera), Python, Hive, Pig, Sqoop, Oracle, Hbase, UNIX
Senior Python/Perl developer
Confidential
Responsibilities:
- Design forms for custom requirements about initialization, validation and notification using Python.
- Maintained and improved existing Internet/Intranet applications
- Designed prototype (GUI based Design)
- Created user information solutions and backend supports,
- Integrated applications with designing database architecture and server scripting.
Environment: Perl, Python, Oracle
Senior Developer Perl/Python
Confidential
Responsibilities:
- Developed Python based API (RESTful Web Service) to track claims.
- Programmed, tested, and implemented the user login, claim reporting & tracking, and statistics reporting system.
- Created a portable fully automated test tool, for performing the sanity check on a new build.
- Automated the daily and weekly build process to allow us to build daily builds twice a day for faster turnaround time for submitted code changes.
- Played a key role in a department wide transition from Oracle forms to Python.
- Performed system testing, integration testing and API testing using SOAPUI
Environment: Perl, Python, Oracle, XML
Java Developer
Confidential
Responsibilities:
- Involved in the design team for designing the various functionality modules.
- Involved in design, development and testing of the application.
- Created Class Diagrams and sequence diagrams using UML.
- Designed templates and screens in HTML and JavaScript.
- Involved in code and performance reviews of the code.
- Performed unit testing, system testing and integration testing Provided Technical support for production environments.
Environment: and Tools: Java, Oracle, Eclipse, Toad, Clear Case, Ant
