Big Data & Hadoop consultant.
I'm not looking for permanent positions!
I focus on consultancy around and implementation of Big Data, Hadoop & Spark projects. I've been involved in the Hadoop ecosystem since 2009.
You can find a more detailed profile and information about my availability on my homepage: http://lars-francke.de/en
Specialties: OpenStreetMap, Hadoop, Spark, HBase, Hive, NoSQL, Big Data, Elasticsearch and related technologies.
Legal notice:
Angaben gemäß § 5 TMG
Lars Francke
Sülldorfer Kirchenweg 34
22587 Hamburg
Kontakt:
Telefon: +49 172 4554978
E-Mail: mail@lars-francke.de
Umsatzsteuer-ID: DE815443126
Solutions Architect (Freelancer) @ I am a “Certified Cloudera Consultant” working with Cloudera EMEA on customer engagements.
Installation and reviews of Hadoop clusters including high availability and security as well as application development or prototyping for customers. From February 2014 to Present (1 year 11 months) Committer, Apache Hive @ From September 2015 to Present (4 months) Hadoop / Big Data consultant @ Hadoop consultancy and maintenance of a cluster based on Amazon’s EC2 From June 2015 to Present (7 months) Hadoop / Big Data consultant @ Consultancy around the BRAIN project (new BI platform) with HBase, Hadoop, Spark, Realtime From May 2015 to Present (8 months) Hamburg Area, GermanyHadoop / Big Data Consultant @ * Consultancy around Hadoop, Best Practices, Kafka
* Review of an architecture based on Kafka, Flume, Hadoop
* Review of an existing cluster regarding Best practices, performance
* Cluster sizing based on predicted usage From April 2015 to Present (9 months) Big Data / Hadoop Consultant @ * Consultancy and training on all things Hadoop, Spark and Big Data
* Development of Spark applications and Hive UDFs for PoC projects
* Tableau & Spark Integration
* Installation of a Hadoop Clusters on Microsoft Azure From March 2015 to Present (10 months) Hadoop consultancy @ * Migration of a Hadoop installation which was set up using Chef to Cloudera Manager
* Upgrade of the cluster from CDH4 to CDH5 (including switching from MRv1 to YARN/MRv2 and testing the customer's workflows)
* Setup of Kerberos with Samba4 & Univention UCS for Hadoop Security
* Troubleshooting From September 2014 to Present (1 year 4 months) Hamburg Area, GermanyBig Data/Hadoop consultant @ * Documentation and consultation around making and validating informed decisions for the following topics:
* HBase vs. Accumulo, Spark, SQL-on-Hadoop
* Backup and High availability of Hadoop clusters
* PaaS, IaaS, Bare-metal deployments in public and private cloud scenarios
* Development of code for HBase backed projects From January 2015 to Present (1 year) Hadoop consultant @ Setup of a CDH 5.4 cluster and integration of the company’s Active Directory including Sentry From August 2015 to August 2015 (1 month) Brussels Area, BelgiumSpark, Hadoop, Big Data consultant @ Spark consulting From July 2015 to July 2015 (1 month) Kreisfreie Stadt Aachen Area, GermanyBig Data / Hadoop Consultant @ * Review of a proposed Hadoop based architecture to replaced a Oracle & Informatica based Data Warehouse and ETL process
* Consultancy and training on all things Hadoop, Spark, HBase
* Setup of a development Hadoop cluster From April 2015 to June 2015 (3 months) IT Consultant for BI Big Data (Freelancer) @ Designing and implementing a BI Infrastructure for an online shop using tools such as Hadoop, Kafka and Storm.
(Collins GmbH & Co. KG was formerly called Antevorte GmbH & Co. KG) From May 2013 to January 2015 (1 year 9 months) Hamburg Area, GermanyBig Data & Java consultancy @ Consultancy around Big Data solutions for tools in the real time bidding world (e.g. generating models). Development of a prototype/proof of concept in Java using Dropwizard, Aerospike, RxJava and MongoDB. Focus on pre-processing data using MongoDB and Aerospike and low latency Java web applications. From September 2014 to December 2014 (4 months) Berlin Area, GermanyBig Data Consultant @ Big Data consultancy around a scalable solution for ingesting and processing large amounts of Geospatial data. Prototyping using Amazon’s Elastic MapReduce and Cloudera Director. From November 2014 to November 2014 (1 month) Madrid Area, SpainHadoop consultancy around processing geospatial data @ * Upgrade of the cluster from CDH4 to CDH5 (including switching from MRv1 to YARN/MRv2)
* Review and optimization of the cluster
* Training of employees on YARN and concepts such as HDFS High Availability
* Development of Hive UDFs and Queries to process large amounts of GIS data using the ESRI Spatial Framework for Hadoop From September 2014 to October 2014 (2 months) Developer (Freelancer) @ Continuation of the previous work: Realtime crawling infrastructure for the GBIF network (BioCASe, DiGIR, TAPIR and DwC-A protocols) using Hadoop & other tools from the ecosystem. From March 2013 to December 2013 (10 months) Developer @ Migrated MySQL based workflows to Hadoop, Oozie, Hive, HBase and others. Introduced software quality measures like Jenkins, Nexus and Sonar. Set up and managed Hadoop clusters from scratch using Puppet, Foreman and other related tools From October 2010 to February 2013 (2 years 5 months) Copenhagen Area, DenmarkIT Consultant @ Analysis of different OpenSource technologies to be used for large scale log file analysis.
This analysis resulted in me helping to set up an initial Hadoop cluster which some former MySQL based workflows have now been ported to and are in production. From August 2010 to October 2010 (3 months) Senior Software Engineer / Architect @ Design, build and support server-side software for the StudiVZ Platform in Java and Python. Work with the CTO and other members of a very small team to define and execute the company high-scalable technology strategy and cloud vision.
Introduced Jenkins, Sonar and Nexus as well as advise on technologies like Hadoop and HBase. From March 2010 to September 2010 (7 months) IT-Architect @ From March 2005 to August 2008 (3 years 6 months)
Diplom-Wirtschaftsinformatiker, Wirtschaftsinformatik, Computer Science for Business Administration @ Fachhochschule Wedel From 2001 to 2006 Lars Francke is skilled in: Hadoop, Big Data, Java, HBase, Python, NoSQL, Software Development, MySQL, SQL, Spark, Scalability, Agile Methodologies, Open Source, Cloud Computing, Distributed Systems
Websites:
http://lars-francke.de/en