diff --git a/Documentation/Cookbook/rst/OTB-Applications.rst b/Documentation/Cookbook/rst/OTB-Applications.rst
index 97b7b2f407c1d70de13aaeafaa4fa7f5ba13d331..78ffd9c1728b2cbdb40b0f58928fe61104dc191f 100644
--- a/Documentation/Cookbook/rst/OTB-Applications.rst
+++ b/Documentation/Cookbook/rst/OTB-Applications.rst
@@ -16,7 +16,7 @@ library, or compose them into high level pipelines. OTB applications allow to:
 OTB applications can be launched in different ways, and accessed from different
 entry points. The framework can be extended, but Orfeo Toolbox ships with the following:
 
--  A command-line laucher, to call applications from the terminal,
+-  A command-line launcher, to call applications from the terminal,
 
 -  A graphical launcher, with an auto-generated QT interface, providing
    ergonomic parameters setting, display of documentation, and progress
@@ -145,7 +145,7 @@ Command-line examples are provided in chapter [chap:apprefdoc], page .
 Using the GUI launcher
 ~~~~~~~~~~~~~~~~~~~~~~
 
-The graphical interface for the applications provides a usefull
+The graphical interface for the applications provides a useful
 interactive user interface to set the parameters, choose files, and
 monitor the execution progress.
 
@@ -170,7 +170,7 @@ The resulting graphical application displays a window with several tabs:
 -  Parameters is where you set the parameters and execute the
    application.
 
--  Logs is where you see the informations given by the application
+-  Logs is where you see the output given by the application
    during its execution.
 
 -  Progress is where you see a progress bar of the execution (not
@@ -258,7 +258,7 @@ application, changing the algorithm at each iteration.
       # Here we configure the smoothing algorithm
       app.SetParameterString("type", type)
 
-      # Set the output filename, using the algorithm to differenciate the outputs
+      # Set the output filename, using the algorithm to differentiate the outputs
       app.SetParameterString("out", argv[2] + type + ".tif")
 
       # This will execute the application and save the output file
@@ -313,7 +313,7 @@ An example is worth a thousand words
                     -outxml saved_applications_parameters.xml
 
 Then, you can run the applications with the same parameters using the
-output xml file previously saved. For this, you have to use the inxml
+output XML file previously saved. For this, you have to use the inxml
 parameter:
 
 ::
@@ -328,7 +328,7 @@ time
     otbcli_BandMath -inxml saved_applications_parameters.xml
                     -exp "(im1b1 - im2b1)"
 
-In this cas it will use as mathematical expression “(im1b1 - im2b1)”
+In this case it will use as mathematical expression “(im1b1 - im2b1)”
 instead of “abs(im1b1 - im2b1)”.
 
 Finally, you can also launch applications directly from the command-line
@@ -340,7 +340,7 @@ the application name. Use in this case:
     otbApplicationLauncherCommandLine -inxml saved_applications_parameters.xml
 
 It will retrieve the application name and related parameters from the
-input xml file and launch in this case the BandMath applications.
+input XML file and launch in this case the BandMath applications.
 
 In-memory connection between applications
 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
@@ -362,7 +362,7 @@ In-memory connection between applications is available both at the C++
 API level and using the  python bindings to the application presented
 in the `Using the Python interface`_ section.
 
-Here is a python code sample connecting several applications together:
+Here is a Python code sample connecting several applications together:
 
 .. code-block:: python
    
@@ -421,7 +421,7 @@ writing is only triggered if:
 - The output filename is ``.tif`` or ``.vrt``
 
   
-In this case, the output image will be divived into several tiles
+In this case, the output image will be divided into several tiles
 according to the number of MPI processes specified to the ``mpirun``
 command, and all tiles will be computed in parallel.
   
@@ -459,7 +459,7 @@ Here is an example of MPI call on a cluster::
   ------------ END JOB INFO 1043196.tu-adm01 ---------
 
 One can see that the registration and pan-sharpening of the
-panchromatic and multispectral bands of a Pleiades image has bee split
+panchromatic and multi-spectral bands of a Pleiades image has bee split
 among 560 cpus and took only 56 seconds.
 
 Note that this MPI parallel invocation of applications is only
diff --git a/Documentation/Cookbook/rst/Recipes.rst b/Documentation/Cookbook/rst/Recipes.rst
index 3fa97fbe6f6dab3835a060d32976e3db44c59074..81753aab812322503b497445299f5bd0c65ad067 100644
--- a/Documentation/Cookbook/rst/Recipes.rst
+++ b/Documentation/Cookbook/rst/Recipes.rst
@@ -3,7 +3,7 @@ Recipes
 
 This chapter presents guideline to perform various remote sensing and
 image processing tasks with either , or both. Its goal is not to be
-exhaustive, but rather to help the non-developper user to get familiar
+exhaustive, but rather to help the non-developer user to get familiar
 with these two packages, so that he can use and explore them for his
 future needs.
 
diff --git a/Documentation/Cookbook/rst/recipes/pbclassif.rst b/Documentation/Cookbook/rst/recipes/pbclassif.rst
index 3a6e37eb0c272fb453ea78678db665db6b6fc50c..e3d6755b25661fc9c0d8eef93d0bcf8e30ba09f1 100644
--- a/Documentation/Cookbook/rst/recipes/pbclassif.rst
+++ b/Documentation/Cookbook/rst/recipes/pbclassif.rst
@@ -27,7 +27,7 @@ available for each class in your image. The ``PolygonClassStatistics``
 will do this job for you. This application processes a set of training
 geometries and an image and outputs statistics about available samples
 (i.e. pixel covered by the image and out of a no-data mask if
-provided), in the form of a xml file:
+provided), in the form of a XML file:
 
 -  number of samples per class
 
@@ -122,7 +122,7 @@ to determine the sampling rate, some geometries of the training set
 might not be sampled.
 
 The application will accept as input the input image and training
-geometries, as well class statistics xml file computed during previous
+geometries, as well class statistics XML file computed during previous
 step. It will output a vector file containing point geometries which
 indicate the location of the samples.
 
@@ -215,7 +215,7 @@ images:
 * **Custom mode:** The user indicates the target number of samples for
   each image.
 
-The different behaviours for each mode and strategy are described as follows.
+The different behaviors for each mode and strategy are described as follows.
 
 :math:`T_i( c )` and :math:`N_i( c )` refers resp. to the total number and needed number
 of samples in image :math:`i` for class :math:`c`. Let's call :math:`L` the total number of
@@ -223,7 +223,7 @@ image.
 
 * **Strategy = all**
   
-  - Same behaviour for all modes proportional, equal, custom : take all samples
+  - Same behavior for all modes proportional, equal, custom : take all samples
   
 * **Strategy = constant** (let's call :math:`M` the global number of samples per
   class required)
@@ -397,10 +397,10 @@ Fancy classification results
 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 
 Color mapping can be used to apply color transformations on the final
-graylevel label image. It allows to get an RGB classification map by
+gray level label image. It allows to get an RGB classification map by
 re-mapping the image values to be suitable for display purposes. One can
 use the *ColorMapping* application. This tool will replace each label
-with an 8-bits RGB color specificied in a mapping file. The mapping file
+with an 8-bits RGB color specified in a mapping file. The mapping file
 should look like this :
 
 ::
@@ -445,7 +445,7 @@ After having processed several classifications of the same input image
 but from different models or methods (SVM, KNN, Random Forest,...), it
 is possible to make a fusion of these classification maps with the
 *FusionOfClassifications* application which uses either majority voting
-or the Demspter Shafer framework to handle this fusion. The Fusion of
+or the Dempster-Shafer framework to handle this fusion. The Fusion of
 Classifications generates a single more robust and precise
 classification map which combines the information extracted from the
 input list of labeled images.
@@ -465,7 +465,7 @@ parameters :
 
 -  ``-undecidedlabel`` label for the undecided class (default value = 0)
 
-The input pixels with the nodata class label are simply ignored by the
+The input pixels with the no-data class label are simply ignored by the
 fusion process. Moreover, the output pixels for which the fusion process
 does not result in a unique class label, are set to the undecided value.
 
@@ -543,7 +543,7 @@ each of them should be confronted with a ground truth. For this purpose,
 the masses of belief of the class labels resulting from a classifier are
 estimated from its confusion matrix, which is itself exported as a
 \*.CSV file with the help of the *ComputeConfusionMatrix* application.
-Thus, using the Dempster Shafer method to fuse classification maps needs
+Thus, using the Dempster-Shafer method to fuse classification maps needs
 an additional input list of such \*.CSV files corresponding to their
 respective confusion matrices.
 
@@ -569,9 +569,9 @@ based on the confidence level in each classifier.
 .. figure:: ../Art/MonteverdiImages/classification_chain_inputimage.jpg
 .. figure:: ../Art/MonteverdiImages/QB_1_ortho_DS_V_P_C123456_CM.png
 
-Figure 5: From left to right: Original image, and fancy colored classified image obtained by a Dempster Shafer fusion of the 6 classification maps represented in Fig. 4.13 (water: blue, roads: gray, vegetation: green, buildings with red roofs: red, undecided: white). 
+Figure 5: From left to right: Original image, and fancy colored classified image obtained by a Dempster-Shafer fusion of the 6 classification maps represented in Fig. 4.13 (water: blue, roads: gray, vegetation: green, buildings with red roofs: red, undecided: white). 
 
-Recommandations to properly use the fusion of classification maps
+Recommendations to properly use the fusion of classification maps
 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 
 In order to properly use the *FusionOfClassifications* application, some
@@ -586,7 +586,7 @@ the interpretation of the ``OutputFusedClassificationImage``.
 Majority voting based classification map regularization
 -------------------------------------------------------
 
-Resulting classification maps can be regularized in order to smoothen
+Resulting classification maps can be regularized in order to smooth
 irregular classes. Such a regularization process improves classification
 results by making more homogeneous areas which are easier to handle.
 
@@ -663,7 +663,7 @@ images, which means that the value of each pixel corresponds to the
 class label it belongs to. The ``InputLabeledImage`` is commonly an
 image generated with a classification algorithm such as the SVM
 classification. Remark: both ``InputLabeledImage`` and
-``OutputLabeledImage`` are not necessarily of the same datatype.
+``OutputLabeledImage`` are not necessarily of the same type.
 Secondly, if ip.suvbool == true, the Undecided label value must be
 different from existing labels in the input labeled image in order to
 avoid any ambiguity in the interpretation of the regularized
@@ -801,7 +801,7 @@ classification and regression mode.
 
 -  K-Nearest Neighbors
 
-The behaviour of application is very similar to . From the input data
+The behavior of application is very similar to . From the input data
 set, a portion of the samples is used for training, whereas the other
 part is used for validation. The user may also set the model to train
 and its parameters. Once the training is done, the model is stored in an