We provide IT Staff Augmentation Services!

Big Data Architect Resume

SUMMARY

  • BigData Architect/Technology Lead with 12+ years IT experience in the areas of Analysis, Design, Development, Coding, Implementation and Testing of Big Data ingestion platforms & analytics with a background of design, development of rest APIs & web based applications using Java/J2EE technologies
  • Expertise in building Big Data based Data Lakes, Ingestion platform (Data Pipe lines) & analytics using Spark, Scala, HBase, Solr, Apache Kafka, Flume, Hadoop and Apache Hive.
  • Expertise in building customer 360 views using graph DBs such as Titan, NEO4J & Janusgraph.
  • Expertise in building Rest Services using Spring Boot.
  • Expertise in development of applications using J2EE components which include Servlets, JSP, Java Beans, EJB, JMS, JDBC.
  • Extensive Knowledge of Spring(MVC,CORE & WebFlow) and Struts framework.
  • Experience in working with SOA architecture, Web Services (SOAP, WSDL) and Rest Services using Spring boot.
  • Experience in web development using HTML, Java Script, CSS, AJAX, jQuery,Dojo, Angular JS.
  • Excellent Hands on experience in using IDE tools like Eclipse, RAD,STS & IntelliJ IDEA.
  • Strong experience in using tools like Log4J, JUNIT,SONAR,VERACODE.
  • Good Experience with Web Servers and Application Servers, which includes ISS,IBM Web Sphere and Apache Tomcat.
  • Experience in implementing J2EE design patterns.
  • Working knowledge of NoSQL database like MongoDB, HBASE and Cassandra.
  • Strong XML experience using XML 1.0, experienced in parsing using JAXB.
  • Experience in developing applications using SQL.
  • Experience with Repository tools like GIT, Bitbucket, Clear Case.
  • Experience with build tools like Ant, Maven & Groovy.
  • Experience with production support for applications.
  • Forward - looking attitude and possess strong interpersonal, communication and problem solving skills.

PROFESSIONAL EXPERIENCE

Confidential

Big Data Architect

Responsibilities:

  • Responsible for the end to end data solution, creation of the data design and enhancements to the ingestion engine to meet the ETL requirements.
  • Ingestion of various formats of data flowing in from mainframe, oracle and DB2 based legacy platforms, doing a data profiling to establish relationships, creating entity based models, code & configure various transformations while loading the data to an entity based graph DB and a nosql DB
  • To create a combination of key, mixed and composite indices on the titan DB to allow the faster retrieval of data by the consumer
  • To leverage Sparks streaming for a faster batch processing of files, to create real time streaming jobs to load data real time using Kafka & Flume.
  • Closely working with Enterprise data modelers and data governance teams, resolving performance issues for some of the complex use cases by modifying data designs & fine tuning the queries.
  • In the role of a Technology lead for all the data migration projects to the Confidential Self Servicing Platform, he mentored and groomed a group of BigData engineers.

Environment: Apache Spark, Scala, Apache Kafka, HBase, Titan/Janus Graph,Log4j, Maven, GIT, Bamboo, JIRA, Cassandra, Apache Solr.

Confidential

Big Data & E2E Tech Lead

Responsibilities:

  • In the role of a BigData technical lead, lead a team of Data analysts to reverse engineer the current state and replicate the same functionality on the new platform
  • Solution the approach, create the data design, enhance the ingestion engine for various ETL requirements, develop Sqoop scripts to be able to extract the data and load the Sqoop import into HBase & Solr DBs
  • Work with Enterprise data modelers and the data governance teams for modeling, closing design gaps between teams during integration, make design decisions on complex technical issues, to provide solutions for complex business problems.
  • To create a combination of key, mixed and composite indices on the titan DB to allow the faster retrieval of data by the consumer.
  • To leverage Sparks streaming for a faster batch processing of files, to create real time streaming jobs to load data real time using Kafka & Flume.
  • The technology stack included a Spark based Data ingestion engine, Scala, Java, HBase, Kafka, Flume, Hadoop, Janusgraph, Unix, DevOps, Agile etc.

Environment: Apache Spark, Scala, Apache Kafka, HBase, Titan/Janus Graph,Log4j, Maven, GIT, Bamboo, JIRA, Cassandra, Apache Solr.

Confidential

Big Data Lead

Responsibilities:

  • Responsible to solutioning the approach, create the data design, enhance the ingestion engine for new ETL requirements.
  • Design, development & enhancement of the data ingestion engine using Spark, Scala, Java, HBase, Kafka, Flume, Hadoop, Titan DB, Unix, DevOps, Agile .
  • To load new entities of data into a Data lake for the consumption APIs, to create real time streaming jobs using Kafka & Flume to load data to support the servicing of Group Customers through call center to create a combination of key, mixed and composite indices on the titan DB to allow the faster retrieval of data by the consumer, to leverage Sparks streaming for a faster batch processing of files, to work with Enterprise data modelers to create a Data model and to come up with data design.
  • Working with the consumption team on the mapping specs etc, to Develop Sqoop scripts to extract the data and load the Sqoop import into HBase & Solr DBs, to work closely with the integration teams in resolving design issues, to enhance, finetune the performance of the response times.

Environment: Apache Spark, Scala, Apache Kafka, HBase, Titan/Janus Graph,Log4j, Maven, GIT, Bamboo, JIRA, Cassandra, Apache Solr.

Hire Now