This course focuses on how to use KNIME Analytics Platform for in-database processing and writing/loading data into a database. Get an introduction to the Apache Hadoop ecosystem and learn how to write/load data into your big data cluster running on premise or in the cloud on Amazon EMR, Azure HDInsight, Databricks Runtime or Google Dataproc. Learn about the KNIME Spark Executor, preprocessing with Spark, machine learning with Spark, and how to export data back into KNIME/your big data cluster.
This course lets you put everything you’ve learnt into practice in a hands-on session based on the use case: Eliminating missing values by predicting their values based on other attributes.
This is a two day instructor-led course. During the course there’ll be hands-on sessions based on real-world use cases.
Sessions will be held on the following dates: November 16 and 17, 2020 at 10 AM - 2 PM UTC -6 (Chicago) which is 5 PM - 9 PM UTC +1 (Berlin).
- Introduction to KNIME and the Database Extension
- Data Processing in a Traditional Database
- Working with Hadoop and Spark
- Machine Learning with Spark
You should be an advanced KNIME user and ideally have already built some workflows. This course doesn’t provide an introduction to KNIME Analytics Platform - it focuses on more managing big data with KNIME Analytics Platform.
You'll receive a zoom link in a separate email a few days before the course starts. Make sure you have a stable internet connection!
Sure! The sessions will be recorded and you’ll have access to each one for seven days from the time the session is over.
Absolutely - fire away!
Your own laptop, ideally pre-installed with the latest version of KNIME Analytics Platform, which you can download at knime.com/downloads.
Download the latest free, open source version of knime here: knime.com/download