A simple HDFS writer job

In this part, we will learn how to create a Talend job, which uses an HDFS component to write in the Hadoop distributed file system.

To do so, we'll need a Hadoop distribution, and fortunately, most of the software vendors are providing some quick-start virtual machines to be able to kick off a Big Data project.

From my side, I'm going to use a Cloudera CDH VM, which you also must have downloaded as mentioned in the previous chapter.

If you have installed and set up your VM as described in the Appendix, Installing Your Hadoop Cluster with Cloudera CDH VM you are ready to create your first job.

We will organize our studio's workspace and create a folder for each chapter by performing the following steps:

  1. In the Repository view, ...

Get Talend for Big Data now with O’Reilly online learning.

O’Reilly members experience live online training, plus books, videos, and digital content from 200+ publishers.