Pentaho Data Integration: Designing a highly available scalable solution for processing filesTutorial DetailsSoftware: PDI/Kettle 4.1 (download here) and MySQL server, both installed on your PC.Knowledge: Intermediate (To follow this tutorial you should have good knowledge of the software and hence not every single step will be described)Files: Download from herePreparationSet up slave serversOn WindowsOn LinuxDefine Slave ServersMonitor Slave...
Wednesday, 23 February 2011
Tuesday, 15 February 2011
Pentaho Data Integration: Best practice solutions for working wit
Posted on 07:39 by Unknown
Pentaho Data Integration: Best practice solutions for working with huge data setsAssign enough memoryOpen pan.sh and kitchen.sh (and spoon.sh if you use the GUI) in a text editor and assign as much memory as possible.Data inputData from a database table can only be imported as fast as the database allows it. You cannot run multiple copies of a database table input step.Text files are more convenient, as you can copy them across servers (your cluster) and read them in simultaneously. Another really cool feature that Kettle has is that you can read...
Friday, 4 February 2011
Pentaho Metadata Editor: Joining two fact tables
Posted on 14:14 by Unknown
Pentaho Metadata Model: Joining two fact tablesThis tutorial was possible with the help of Will Gorman, Vice President of Engineering at PentahoThe dataMetadata ModelMy SuggestionsTutorial DetailsSoftware: PME 3.7 (download here), a database of your liking, in example MySQLKnowledge: Intermediate (To follow this tutorial you should have good knowledge of the software and hence not every single step will be described)Tutorial files can be downloaded...
Subscribe to:
Posts (Atom)