Agenda:
- Getting a general understanding of the Pentaho platform and learn some best practices around it.
- Building a Hadoop Data Lake: loading log and master data into Hadoop using metadata injection
- Building and end-to-end data refinery: processing data in Hadoop using Spark, deliver it to business users and analyse it.
- Real World example: London Air Quality. Consume live air quality sensor data through a REST API and get it ready for analysis. You will be able to take this exercise with you as a homework to play and test by yourself.
- Real world IOT use cases: Get an understanding of the Fleet Management (Penske) demo, what’s in it and how Pentaho ties into IOT platforms.
- (Optional) IOT real-time streaming examples: Using MQTT to consume IOT data and visualize it in real-time dashboards.
- Machine Learning: get an understanding of Pentaho leverages and orchestrates machine learning for different use cases.