An aspiring Big Data Full Stack Developer who loves leveraging different tools and technologies in mining the Big Data Mountain
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Big Data Skillset:
★Database Frameworks: Hadoop Distributed File System (HDFS), MapR FS
★NoSQL Databases: HBase and MongoDB
★Hadoop EcoSystem Tools: Hive, Sqoop, Flume, Oozie, Spark & Spark Streaming
★Data Visualization: Tableau, SpotFire, DataMeer
★Analytical programming: Pig, Python (NumPy & Pandas) and R
★Analytics: Natural Language Processing (NLP), Machine Learning and Predictive Modelling
★Data Mining Tools: RapidFire 6.0, RStudio
★Programming: Java, Python, UNIX Shell Scripting
Relevant Certifications:
✪Cloudera Certified Developer for Apache Hadoop (CCDH)
✪Computing for Data Analysis using R
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Other Technical Skill Set:
★Programming Languages: SQL and PL-SQL
★Business Intelligence Tools: Oracle Data Integrator (ODI), DB Visualizer, Oracle SQL Developer
★Databases: IBM-Netezza, Oracle 10/11g, MySQL, Microsoft SQLServer and Sybase
Relevant Certifications:
✪Oracle PL/SQL Developer Certified Associate
✪Oracle Database SQL Certified Expert
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Professional Experience:
✰Data Analytics Intern at Thermo Fisher Scientific, Carlsbad, CA (3 Months)
✰ETL PL/SQL Developer at Tata Consultancy Services, Chennai, India (34 Months)
Big Data Engineer I @ #BigData #Hadoop #MapRFS #NetworkTheory #API #Visualization #RDBMS
• Build applications leveraging different Big Data tools & Technologies – Apache Hadoop, MapRFS, Hive, Sqoop, Oozie, Elastic Search etc.
• Build customer information search engines being used by marketing, financial & fraud intelligence units
• Write Map Reduce programs to build different customer networks and to calculate different Network Statistics
• Transition existing legacy Teradata systems to use distributed computational systems like Hadoop
• Build user interactive dashboards using different tools like D3.js, Tableau
• Build ETL flows using Big-Data integrated tools like Platfora & Datameer
• Build Web Data Services and APIs using Java and Python which communicate with NoSQL Databases like HBase, MapR M7 From January 2015 to Present (1 year) Greater New York City AreaGraduate Teaching Assistant @ •Teaching Assistant for MIS 331- Database Management Systems under Prof. Susan Brown
•Assist students in understanding the RDBMS in an easy way
•Grading SQL Assignments and Exams From August 2014 to December 2014 (5 months) Data Analytics Intern @ •Pre-process unstructured data and perform statistical data analysis using different R built-in packages
•Perform predictive analytics using different Machine Learning (ML) algorithms like Support Vector Machines (SVM) and K-Nearest Neighbour (KNN)
•Analyze the CRM survey data using different kinds of Sentiment Analysis Techniques – Lexical Dictionary based using Word Net and Corpus Based using Machine Learning (ML) Algorithms in R
•Develop effective Data Visualization Dashboards in Tableau 8.2
•Work on multiple projects which required statistical analysis, unstructured data analysis using R and Tableau From June 2014 to August 2014 (3 months) Greater San Diego AreaData Analyst (Part-Time) @ • Design complex SQL/PLSQL queries to extract data as per business requirements from Microsoft SQL Server 2008 as a part of Knowledge Discovery from Databases (KDD) process
• Analyze data pertaining to the child health measure entry and find trends/patterns/insights
• Co-ordinate with Database Administrators (DBA) in University Information Technology Services
• Provide Reports in the form of Excel for easy understanding
• Visualize data and trends using tools like Tableau Desktop
• Create easy-to-use client dashboard for monitoring the metrics using Tableau
• Participate in meetings with Analysts from other universities (ASU, NAU etc) From January 2014 to May 2014 (5 months) Tucson, Arizona AreaETL Developer/Systems Engineer @ Programming Experience:
• Developed different loads which handle the ELT(Extract Load and Transform) flow using tools like Oracle Data Integrator 10/11g.
• Expertise in developing daily loads that handle loading of flat files with huge volumes (~40GB) of data from Facebook to IBM Netezza within the predefined SLA.
• Hands on experience in Data warehouse concepts such as dimensional modeling, Star Schema, Data marts, dimensions and fact identification
• Creating variables, sequences, user functions, scenarios, procedures, interfaces and packages in ODI.
• Expertise in performing Data Quality Analysis (DQA) using Ataccama Data Quality Analyzer.
• Developed different kinds of purge loads as a part of Effective Database management.
• Successfully handled UAT ETL Testing and also assisted client during Post- Deployment support
• Automated End to End Integration testing using different kinds of PLSQL procedures.
• Proficiency in tuning the SQL/PLSQL code for optimum performance.
• Designed different kinds of purge loads as a part of effective database management and to increase the performance of the ETL jobs.
Managerial Experience:
• Instrumental in leading team towards Project’s Global roll out to countries: Germany, Italy, Brazil and Australia with zero post- deployment defects.
• Work closely with the Business, Support and Testing teams to understand, document, design and code ETL Processes. From September 2010 to July 2013 (2 years 11 months) Chennai Area, India
Master's degree, Management Information Systems @ University of Arizona From 2013 to 2014 B.Tech, Electrical and Electronics Engineering @ Pondicherry University From 2006 to 2010 Yashwanth Tekena is skilled in: Hadoop, Data Mining, Big Data, Databases, Oracle, SQL, PL/SQL, ETL, Data Warehousing, Microsoft SQL Server, Linux, Business Intelligence, Data Analysis, Netezza, Unix Shell Scripting