QUALIFICATIONS:
Software Engineer with Analytical mindset and Infrastructure background
Sr Software Engineer @ Build Company profiles and insights with a group of smart, hard-working and fun people at Company Profile Engineering Team From January 2015 to Present (10 months) Mountain View, CaliforniaStaff, Business Analytics @ Overall: Drive and enable the team to take fullest advantage of LinkedIn valuable and ever-expanding proprietary rich data set, and LinkedIn world-class infrastructure. Deliver data-driven tools, metrics, insights to drive LinkedIn user growth, enterprise sales and marketing efforts.
Projects:
1) Enable LinkedIn sales and marketing teams to access and act upon data insights in as easy a manner as possible;
2) Design and implement creative data infrastructure solutions to enable high performance and low cost analytical computation; ensure sound architecture of data flow and consumption of relational and unstructured data
3) Help scale LinkedIn business analytics reporting capabilities to share with LinkedIn corporate customers From February 2013 to December 2014 (1 year 11 months) Mountain ViewSr. Software Engineer @ 1. Developed Data and Analytics Platform, which consolidates data generated by various Intuit offerings and plug in data exploration/analysis/knowledge discovery tools, to enable better decision making through more rapid and higher-quality insights and increase customers’ conversion through more effective insights
2. Designed and executed test plans to evaluate and benchmark different vendors’ Hadoop platform;
3. Developed private cloud service integration APIs (using Play framework) that facilitate virtual machine provision and monitoring. From February 2012 to February 2013 (1 year 1 month) Menlo Park, CAResearch Data Engineer @ Played a key role in designing and building scalable data warehouse on cloud and analytics products out of large-scale data set. Gained solid domain knowledge in display ads ecosystem.
Project list:
1. Designed and built hourly/daily upstream/downstream log processing pipeline that serves 10B events / month;
2. Designed and built Publisher Analytics Product that processes over 30M records and updates around 11000 publishers metrics each hour;
3. Designed and built Ads Campaign Analytics Product to offer deep insights about ads performance and behavior of targeted users;
4. Worked with data scientists and business development manager to dig out more analytics out of different datasets to support company revenue generation and new products incubation;
5. Maintained and optimized Aster/Hive Data Warehouse and Hadoop/Hive production clusters.
6. R&D: built scalable and fault-tolerent infrastructure on top of which various data analytics of different natures can be performed; advancing the state of the company's data analytics pipeline;
7. R&D: Investigated cutting-edge algorithms and novel models as required to improve the accuracy and relevance of user profile and business intelligence;
8. R&D: Investigated state-of-the-art technologies of large-scale data processing open-source tools; designing and developing novel features that can be utilized to enrich and optimize existing data processing pipeline.
Skill Sets:
Hadoop MapReduce/Hive, Aster DW, Amazon Web Service, Java.
Working knowledge about NoSQL (HBase, Cassandra, Mongodb), Apache Message Queue From May 2010 to January 2012 (1 year 9 months) Software Engineer @ Involved and contribute to many phases/aspects of the software development lifecycle
1. Data processing and Integration
1.a: Use Hadoop/MapReduce to compute various performance metrics of customer data feeds;
1.b: Use Pentaho Data Integration tool (aka Kettle) to automate data processing and loading process;
1.c: Use PostgreSQL to do database programming
2: IT/Operation
2.a: Use AWS to set up and maintain enterprise platform that enables 24X7 access to data and application development;
2.b: Configure and secure Linux servers From December 2009 to May 2010 (6 months) Visiting Scholar @ Primarily work with Prof. Jens Palsberg on how to improve efficiency and make customizable compilers for medial imaging through studying domain-specific algorithms and data structures. Statistics/probability models, Octave (Matlab's Linux alternative) and Microsoft .NET are used in this study. From August 2009 to October 2009 (3 months) PhD in Computer Science @ Please check out my academic experience at http://sites.google.com/site/zhupingnj/ From July 2002 to May 2009 (6 years 11 months) Singapore
Ph.D., Computer Science @ National University of Singapore From 2002 to 2009 B.S., M. Eng., Computer Science @ Nanjing University From 1995 to 2002 Ping Zhu is skilled in: Hadoop, Hive, MapReduce, Big Data, MongoDB, Scalability, Java, Amazon EC2, Cassandra, Amazon Web Services..., Algorithms, Software Engineering, Linux, Software Development, Cloud Computing, Distributed Systems, Databases, PostgreSQL, S3, EC2, MySQL, Apache Pig, Amazon Web Services, Pig