diff --git a/Documentation/Cookbook/rst/OTB-Applications.rst b/Documentation/Cookbook/rst/OTB-Applications.rst index 97b7b2f407c1d70de13aaeafaa4fa7f5ba13d331..78ffd9c1728b2cbdb40b0f58928fe61104dc191f 100644 --- a/Documentation/Cookbook/rst/OTB-Applications.rst +++ b/Documentation/Cookbook/rst/OTB-Applications.rst @@ -16,7 +16,7 @@ library, or compose them into high level pipelines. OTB applications allow to: OTB applications can be launched in different ways, and accessed from different entry points. The framework can be extended, but Orfeo Toolbox ships with the following: -- A command-line laucher, to call applications from the terminal, +- A command-line launcher, to call applications from the terminal, - A graphical launcher, with an auto-generated QT interface, providing ergonomic parameters setting, display of documentation, and progress @@ -145,7 +145,7 @@ Command-line examples are provided in chapter [chap:apprefdoc], page . Using the GUI launcher ~~~~~~~~~~~~~~~~~~~~~~ -The graphical interface for the applications provides a usefull +The graphical interface for the applications provides a useful interactive user interface to set the parameters, choose files, and monitor the execution progress. @@ -170,7 +170,7 @@ The resulting graphical application displays a window with several tabs: - Parameters is where you set the parameters and execute the application. -- Logs is where you see the informations given by the application +- Logs is where you see the output given by the application during its execution. - Progress is where you see a progress bar of the execution (not @@ -258,7 +258,7 @@ application, changing the algorithm at each iteration. # Here we configure the smoothing algorithm app.SetParameterString("type", type) - # Set the output filename, using the algorithm to differenciate the outputs + # Set the output filename, using the algorithm to differentiate the outputs app.SetParameterString("out", argv[2] + type + ".tif") # This will execute the application and save the output file @@ -313,7 +313,7 @@ An example is worth a thousand words -outxml saved_applications_parameters.xml Then, you can run the applications with the same parameters using the -output xml file previously saved. For this, you have to use the inxml +output XML file previously saved. For this, you have to use the inxml parameter: :: @@ -328,7 +328,7 @@ time otbcli_BandMath -inxml saved_applications_parameters.xml -exp "(im1b1 - im2b1)" -In this cas it will use as mathematical expression “(im1b1 - im2b1)†+In this case it will use as mathematical expression “(im1b1 - im2b1)†instead of “abs(im1b1 - im2b1)â€. Finally, you can also launch applications directly from the command-line @@ -340,7 +340,7 @@ the application name. Use in this case: otbApplicationLauncherCommandLine -inxml saved_applications_parameters.xml It will retrieve the application name and related parameters from the -input xml file and launch in this case the BandMath applications. +input XML file and launch in this case the BandMath applications. In-memory connection between applications ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ @@ -362,7 +362,7 @@ In-memory connection between applications is available both at the C++ API level and using the python bindings to the application presented in the `Using the Python interface`_ section. -Here is a python code sample connecting several applications together: +Here is a Python code sample connecting several applications together: .. code-block:: python @@ -421,7 +421,7 @@ writing is only triggered if: - The output filename is ``.tif`` or ``.vrt`` -In this case, the output image will be divived into several tiles +In this case, the output image will be divided into several tiles according to the number of MPI processes specified to the ``mpirun`` command, and all tiles will be computed in parallel. @@ -459,7 +459,7 @@ Here is an example of MPI call on a cluster:: ------------ END JOB INFO 1043196.tu-adm01 --------- One can see that the registration and pan-sharpening of the -panchromatic and multispectral bands of a Pleiades image has bee split +panchromatic and multi-spectral bands of a Pleiades image has bee split among 560 cpus and took only 56 seconds. Note that this MPI parallel invocation of applications is only diff --git a/Documentation/Cookbook/rst/Recipes.rst b/Documentation/Cookbook/rst/Recipes.rst index 3fa97fbe6f6dab3835a060d32976e3db44c59074..81753aab812322503b497445299f5bd0c65ad067 100644 --- a/Documentation/Cookbook/rst/Recipes.rst +++ b/Documentation/Cookbook/rst/Recipes.rst @@ -3,7 +3,7 @@ Recipes This chapter presents guideline to perform various remote sensing and image processing tasks with either , or both. Its goal is not to be -exhaustive, but rather to help the non-developper user to get familiar +exhaustive, but rather to help the non-developer user to get familiar with these two packages, so that he can use and explore them for his future needs. diff --git a/Documentation/Cookbook/rst/recipes/pbclassif.rst b/Documentation/Cookbook/rst/recipes/pbclassif.rst index 3a6e37eb0c272fb453ea78678db665db6b6fc50c..e3d6755b25661fc9c0d8eef93d0bcf8e30ba09f1 100644 --- a/Documentation/Cookbook/rst/recipes/pbclassif.rst +++ b/Documentation/Cookbook/rst/recipes/pbclassif.rst @@ -27,7 +27,7 @@ available for each class in your image. The ``PolygonClassStatistics`` will do this job for you. This application processes a set of training geometries and an image and outputs statistics about available samples (i.e. pixel covered by the image and out of a no-data mask if -provided), in the form of a xml file: +provided), in the form of a XML file: - number of samples per class @@ -122,7 +122,7 @@ to determine the sampling rate, some geometries of the training set might not be sampled. The application will accept as input the input image and training -geometries, as well class statistics xml file computed during previous +geometries, as well class statistics XML file computed during previous step. It will output a vector file containing point geometries which indicate the location of the samples. @@ -215,7 +215,7 @@ images: * **Custom mode:** The user indicates the target number of samples for each image. -The different behaviours for each mode and strategy are described as follows. +The different behaviors for each mode and strategy are described as follows. :math:`T_i( c )` and :math:`N_i( c )` refers resp. to the total number and needed number of samples in image :math:`i` for class :math:`c`. Let's call :math:`L` the total number of @@ -223,7 +223,7 @@ image. * **Strategy = all** - - Same behaviour for all modes proportional, equal, custom : take all samples + - Same behavior for all modes proportional, equal, custom : take all samples * **Strategy = constant** (let's call :math:`M` the global number of samples per class required) @@ -397,10 +397,10 @@ Fancy classification results ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Color mapping can be used to apply color transformations on the final -graylevel label image. It allows to get an RGB classification map by +gray level label image. It allows to get an RGB classification map by re-mapping the image values to be suitable for display purposes. One can use the *ColorMapping* application. This tool will replace each label -with an 8-bits RGB color specificied in a mapping file. The mapping file +with an 8-bits RGB color specified in a mapping file. The mapping file should look like this : :: @@ -445,7 +445,7 @@ After having processed several classifications of the same input image but from different models or methods (SVM, KNN, Random Forest,...), it is possible to make a fusion of these classification maps with the *FusionOfClassifications* application which uses either majority voting -or the Demspter Shafer framework to handle this fusion. The Fusion of +or the Dempster-Shafer framework to handle this fusion. The Fusion of Classifications generates a single more robust and precise classification map which combines the information extracted from the input list of labeled images. @@ -465,7 +465,7 @@ parameters : - ``-undecidedlabel`` label for the undecided class (default value = 0) -The input pixels with the nodata class label are simply ignored by the +The input pixels with the no-data class label are simply ignored by the fusion process. Moreover, the output pixels for which the fusion process does not result in a unique class label, are set to the undecided value. @@ -543,7 +543,7 @@ each of them should be confronted with a ground truth. For this purpose, the masses of belief of the class labels resulting from a classifier are estimated from its confusion matrix, which is itself exported as a \*.CSV file with the help of the *ComputeConfusionMatrix* application. -Thus, using the Dempster Shafer method to fuse classification maps needs +Thus, using the Dempster-Shafer method to fuse classification maps needs an additional input list of such \*.CSV files corresponding to their respective confusion matrices. @@ -569,9 +569,9 @@ based on the confidence level in each classifier. .. figure:: ../Art/MonteverdiImages/classification_chain_inputimage.jpg .. figure:: ../Art/MonteverdiImages/QB_1_ortho_DS_V_P_C123456_CM.png -Figure 5: From left to right: Original image, and fancy colored classified image obtained by a Dempster Shafer fusion of the 6 classification maps represented in Fig. 4.13 (water: blue, roads: gray, vegetation: green, buildings with red roofs: red, undecided: white). +Figure 5: From left to right: Original image, and fancy colored classified image obtained by a Dempster-Shafer fusion of the 6 classification maps represented in Fig. 4.13 (water: blue, roads: gray, vegetation: green, buildings with red roofs: red, undecided: white). -Recommandations to properly use the fusion of classification maps +Recommendations to properly use the fusion of classification maps ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In order to properly use the *FusionOfClassifications* application, some @@ -586,7 +586,7 @@ the interpretation of the ``OutputFusedClassificationImage``. Majority voting based classification map regularization ------------------------------------------------------- -Resulting classification maps can be regularized in order to smoothen +Resulting classification maps can be regularized in order to smooth irregular classes. Such a regularization process improves classification results by making more homogeneous areas which are easier to handle. @@ -663,7 +663,7 @@ images, which means that the value of each pixel corresponds to the class label it belongs to. The ``InputLabeledImage`` is commonly an image generated with a classification algorithm such as the SVM classification. Remark: both ``InputLabeledImage`` and -``OutputLabeledImage`` are not necessarily of the same datatype. +``OutputLabeledImage`` are not necessarily of the same type. Secondly, if ip.suvbool == true, the Undecided label value must be different from existing labels in the input labeled image in order to avoid any ambiguity in the interpretation of the regularized @@ -801,7 +801,7 @@ classification and regression mode. - K-Nearest Neighbors -The behaviour of application is very similar to . From the input data +The behavior of application is very similar to . From the input data set, a portion of the samples is used for training, whereas the other part is used for validation. The user may also set the model to train and its parameters. Once the training is done, the model is stored in an