Monday, March 17, 2014

How export and import saiku files on Pentaho 5

After install saiku plugin for Pentaho 5 available on marketplace, by default isn't possible export the saiku files (".saiku").

So, in this post I'll explain the configuration you need to do for have this feature.
  1. Stop your pentaho server;
  2. Open the file <pentaho-path>/biserver-ce/pentaho-solutions/system/importExport.xml;
  3. Add "<entry key="saiku" value-ref="streamConverter"/>" to convertersMap map;
  4. Add "<entry key="saiku" value="text/xml"/>" to the map of NameBaseMimeResolver bean;
  5. Add "<value>.saiku</value>" to the list approvedExtensionList;
  6. Start pentaho server again and try download one saiku file. 
As an example I leave this link for the file, after configured.

Tuesday, February 11, 2014

OpenShift QuickStart for SuiteCRM

In the last days, I create an OpenShift quickstart for SuiteCRM. You can find the this openshift quickstart on GitHub.

First you need create a Openshift on-line account and install the RHC client tools you can find out more in this page: Getting Started page. After you install RHC client you just need run the following command:
rhc app create suitecrm php-5.3 mysql-5.1 --from-code=
 After the command above executed with success you can access the following URL: http://suitecrm-$ and I few seconds you will have your SuiteCRM created. The default credentials are username "admin", password "admin". And as you can see in few minutes you have your CRM installed running on OpenShift Cloud.

Installation Config
This quickstart use PHP 5.x and MySQL 5.x cartridges , and you can configure the automatic install these files: file1 and file2 (like SugarCRM).

About SuiteCRM: "SuiteCRM is a software fork of the popular customer relationship management (CRM) system SugarCRM. SuiteCRM has been described by its creators as SugarCRM supercharged." - Wikipedia

Wednesday, December 18, 2013

Pentaho overview using MongoDB (Oporto MUG #2)

I'm writing this post for share the presentation and resources of my talk on the second meetup Oporto MongoDB User Group. The talk consist on demonstrate the integration between Pentaho and MongoDB, on the three aspects: ETL, Reporting and Dashboarding.
Look the presentation:

The data I use is just for demonstration, I didn't care with performance, content or data quality, the goal are just demonstrate how you can integrate Pentaho with MongoDB with simple examples. And show the potential and how easy is integrating with other systems, in this case Google Maps API.

In resume, if you want integrate Pentaho with MongoDB, there're two options for do it:
  • ETL: With Pentaho Data Integration (aka kettle) you can use the "MongoDB input" and "MongoDB output" steps for get and save data. The ETL transformations can be use on Pentaho Reporting or CDE using CDA.
  • Programming: This is the best choose for me because I'm familiar with Java development. I know, using ETL steps is more intuitive but as any platform or product, you have more performance if you design exactly what you want. So, in ETL for programming you can use the "User Defined Java Class" step. In Report designer you can use the option Scriptable or a MongoDB datasource. Finally, in dashboards using CDE, that's use CDA for data access, you can use the "scriptable over scripting" datasource and the language beanshell, for example.

This is the link with resources of presentation. Note: You need be familiar with configuration those resources to works fine.

Thursday, September 26, 2013

How publish Saiku Analytics on OpenShift

Saiku Analytics is a great open source server or Pentaho plugin for explore and vizualize data!

About Saiku: "Saiku was founded in 2008 by Tom Barber and Paul Stoellberger. Originally called the Pentaho Analysis Tool, if started life as a basic GWT based wrapper around the OLAP4J library. Over the years it has evolved, and after a complete rewrite in 2010, it was reborn as Saiku.
Saiku offers a user friendly, web based analytics solution that lets users, quickly and easily analyse corporate data and create and share reports. The solution connects to a range of OLAP Servers including Mondrian, Microsoft Analysis Services, SAP BW and Oracle Hyperion and can be deployed rapidly and cost effectively to allow users to explore data in real time." - Meteorite

About OpenShift: "OpenShift is a cloud computing platform as a service product from Red Hat. A version for private cloud is named OpenShift Enterprise.
The software that runs the service is open-sourced under the name OpenShift Origin, and is available on GitHub. Developers can use Git to deploy web applications in different languages on the platform.
OpenShift also supports binary programs that are web applications, so long as they can run on Red Hat Enterprise Linux. This allows the use of arbitrary languages and frameworks. OpenShift takes care of maintaining the services underlying the application and scaling the application as needed." - Wikipedia

In this post I'll demonstrate how you can publish Saiku Analytics server on OpenShift platform using a free account that provide 1GB storage per gear. Using another words (commercial words) is put your business analytics on the sky (ok, on the cloud :) ) by free (or low cost depends of your bussiness).

After create your account on OpenShift website, you need install/configure OpenShift RHC Client Tools. I'll describe the steps for install and configure in Linux but you can follow the instructions on this link.
The steps on Linux Ubuntu are:
  1. sudo apt-get install ruby-full rubygems git-core
  2. sudo gem install rhc
  3. rhc setup (put your credentials)

The next step is create your application on OpenShift, you have three ways for do that. One using OpenShift website with your account (as I'll demonstrate), other way is using RHC client tool and the other using JBoss Developer Studio (you can find out more about those ways in this link).
The steps using website are:

  1. Click on ADD APPLICATION button;
  2. Choose which type application, in this case I choose Tomcat 6 (JBoss EWS 1.0);
  3. Write the application name, in this case I wrote "saiku".

The next step is download Saiku WAR's (UI and Backend) file to deploy on OpenShift. After that, you need do the following steps on your command line:

  1. rhc git-clone saiku
  2. cd saiku
  3. git rm -r src pom.xml
  4. cp <war-path-file>/saiku-ui-2.5.war webapps/ROOT.war
  5. cp <war-path-file>/saiku-webapp-2.5.war webapps/saiku.war
  6. git add .
  7. git commit -m 'Deploy Saiku 2.5'
  8. git push

And is done! You can check in your URL http://<application-name>-<username> You should have something like: .

Enjoy. ;)

Interested Links:

Saturday, July 27, 2013

A new Pentaho book

A new book about Pentaho is out! The name is Instant Pentaho Data Integration Kitchen.

"The book is about kitchen and how to use the PDI's command line tools efficiently to help people in their day to day operations with Pentaho Data Integration. It is a practical book and it seemed relatively easy to write.Sergio Ramazzina

I was the technical reviewer for this book and I was very happy because the publisher choose me and is another experience that I really liked.

Also, I want to wish them congratulations to the author Sergio Ramazzina to write this book and contribute to opensource community.

Monday, July 1, 2013

PDI Apple Push Notifications Plugin is available on Pentaho Marketplace

Since 14 of June, the PDI Apple push notification is available on Pentaho Marketplace. Now, PDI can send push notifications to the two most popular smartphones platforms.

Check out what Matt Carters (Chief Data Integration for Pentaho) say about plugins available on Pentaho marketplace in the Pentaho Big Data Architecture presentation of Pentaho London User Group event (at 18 minute):

Interested Links:
Pentaho Data Integration Marketplace Wiki: Link
Pentaho Data Integration Marketplace Source: Link
PDI Apple Push Notifications Plugin Wiki: Link
PDI Apple Push Notifications Plugin Source: Link
PDI Apple Push Notifications Plugin Artifacts: Link
PDI Manager Android App: Link

Thursday, June 6, 2013

Run ETL with Pentaho Data Integration using package files

The Kettle have a small functionality to run ETL, it uses Apache VFS that let you access a set of files from inside an archive and use them directly in your processes. However, you can use this for execute ETL in somewhere on the web.

Run in file system

I create this little sample (a job that executes a transformation). And I compress this two in the zip file.
So, I have this zip file on that path in my own computer: C:\Users\latinojoel\Desktop\
The command line I need to do for execute ETL, is looks like that:
~\data-integration>Kitchen.bat -file=zip:\\"C:\Users\latinojoel\Desktop\!/job.kjb" -level=Detail -param:MSG="Wow, It's works. Very funny! :-)"

Run from web resource

With Apache VFS, you can run a zip file from web too.
For example, you can access of my zip file using this URI.
The command line you need to do is that:
~\data-integration>Kitchen.bat -file="zip:!/job.kjb" -level=Detail -param:MSG="Wow, It's works. Very funny! :-)"

Is available for Pan too. A sample:
~\data-integration>Pan.bat -file="zip:!/transf.ktr" -level=Detail -param:MSG="Wow, It's works. Very funny! :-)"