Before organising a course or seminar, we listen to the real needs and objectives of each client, in order to adapt the training and get the most out of it. We tailor each course to your needs.

We are also specialists in 'in company' trainings adapted to the needs of each organisation, where the benefit for several attendees from the same company is much greater. If this is your case, contact us.

Ponemos a disposición también plataforma Cloud con todas las herramientas instaladas y configuradas, listas para la formación, incluyendo ejercicios, bases de datos, etc... para no perder tiempo en la preparación y configuración inicial. ¡Sólo preocuparos de aprender!

Ofrecemos también la posibilidad de realizar formaciones en base a ‘Casos de Uso’

Se complementa la formación tradicional de un temario/horas/profesor con la realización de casos prácticos en las semanas posteriores al curso en base a datos reales de la propia organización, de forma que se puedan ir poniendo en producción proyectos iniciales con nuestro soporte, apoyo al desarrollo y revisión con los alumnos y equipos, etc…

En los 10 últimos años, ¡hemos formado a más de 250 organizaciones y 3.000 alumnos!

Ah, y regalamos nuestras famosas camisetas de Data Ninjas a todos los asistentes. No te quedes si las tuyas

1  2  3  4  5

ETL Kettle Pentaho

ETL Kettle Pentaho


Extraction, transformation and loading (ETL) of data is the key to success in a BI system for managing the quality of the data properly.

In this course you will have some of the best practices that we recommend for the design of ETL processes such as:

Centralization of procedures so that the coherence and consistency of the exchanged data from different sources is ensured.

Avoid redundancy calculations: if there is data previously calculated in the operational databases should not return to the calculation performed in the extraction. This premise aims to achieve a double objective.

Establishment of points of "quality control" and validation.

Implement processes charging information for possible errors in the initial information.

Consider the possibility of using intermediate tables with the most atomic level of information to be treated.

In addition, we will review the most important and used ETL tool Pentaho elements: Pentaho Data Integration and Kettle.

Target audiences

Professionals information technology, IT managers, business analysts, systems analysts, Java architects, system developers, database administrators, developers and professionals in relation to the area of technology, marketing, business and financial.


Data Quality and Integration with Pentaho


Introduction Open Source Business Intelligence Platforms
  • Architecture and features of Pentaho, SpagoBI , BIRT , Mondrian , Kettle , Talend , etc ...
  • Development Tools.
ETL (Kettle)
  • Good practices for ETL process definition.
  • Functional Overview (work, Transformations, flow control)
  • Parameterisation
    • Environment Variables
    • Parameterization of connections to databases. Shared connections.
    • Parameterization of loads and load types
  • Jobs
    • Overview
    • Steps types (Mail, File Managament, etc ...)
    • Description of Steps
    • Steps examples of more useful and common
  • Transformations
    • Overview
    • Steps types (Input, Output, Tranform, etc ...)
    • Description of Steps
    • Steps examples of more useful and common
  • Practical examples


Ajustamos cada curso a sus necesidades.

Nuestra oficina en Madrid