To incorporate MATLAB map and reduce functions into a Hadoop mapreduce job, you create a deployable archive from the map and reduce functions and pass the archive as a payload argument to a job submitted to the Hadoop cluster. A deployable archive contains a:
mapper function written in MATLAB.
reducer function written in MATLAB.
MAT-file containing a datastore that describes the structure of the data and variables to be analyzed.
Hadoop settings file that identifies the map and reduce functions, the type of data being analyzed, and other configuration details.
For more information, see Workflow to Incorporate MATLAB Map and Reduce Functions into a Hadoop Job
deploytool | Compile and package functions for external deployment |
mcc | Compile MATLAB functions for deployment |
hadoopCompiler | Package MATLAB Compiler programs for deployment against Hadoop clusters as MapReduce programs |
Hadoop Compiler | Package MATLAB programs for deployment to Hadoop clusters as MapReduce programs |
Workflow to Incorporate MATLAB Map and Reduce Functions into a Hadoop Job
Instructions on how to create a deployable archive of MATLAB map and reduce functions and incoporate it into a Hadoop mapreduce job.