Before you migrate your applications to the MapR Hadoop distribution, consider testing your applications using a small subset of data.
In this phase, you will migrate your applications to the MapR cluster test environment. The goal of this phase is to get your applications running smoothly on the MapR cluster using a subset of data. Once you have confirmed that all applications and components are running as expected you can begin migrating your data.
Migrating your applications from HDFS to MapR is relatively easy. MapR Hadoop is 100% plug-and-play compatible with Apache Hadoop, so you do not need to make changes to your applications to run them on a MapR cluster.
maprfs.jar and the
java.library.path includes
libMapRClient.somaprfs:///) rather than the HDFS
(hdfs://). If your application uses fs.default.name
then it will work automatically. If you have hardcoded HDFS links into your
applications, you must redirect those links so they point to filesystem.
Setting a default path of maprfs:/// tells your
applications to use the cluster specified in the first line of
mapr-clusters.conf. You can also specify a specific cluster with
maprfs:///mapr/<cluster name>/. Generally, the best approach to migrating your applications to MapR is to import a small subset of data and test and tune your application using that data in a test environment before you import your production data.
The following procedure offers a simple roadmap for migrating and running your applications in a MapR cluster test environment.