Skip to content

Fix some typos (mostly found by codespell)#8

Open
stweil wants to merge 2 commits intoDoreenruirui:masterfrom
stweil:master
Open

Fix some typos (mostly found by codespell)#8
stweil wants to merge 2 commits intoDoreenruirui:masterfrom
stweil:master

Conversation

@stweil
Copy link

@stweil stweil commented Sep 19, 2019

Signed-off-by: Stefan Weil sw@weilnetz.de

</center>

**Okralact** is both a set of specifications and a prototype implementation for harmonizing the input data, parameterization and provenance tracking of training different OCR engines. It is a client/server architecture application. The interactions between the client nodes and the server are implementeqd using **Flask**, a lightweight web application framework for Python. All the training or evaluation jobs submitted to the server are handled in the background by task queues implemented wth **Redis Queue** (**RQ**).
**Okralact** is both a set of specifications and a prototype implementation for harmonizing the input data, parameterization and provenance tracking of training different OCR engines. It is a client/server architecture application. The interactions between the client nodes and the server are implemented using **Flask**, a lightweight web application framework for Python. All the training or evaluation jobs submitted to the server are handled in the background by task queues implemented with **Redis Queue** (**RQ**).
Copy link
Author

@stweil stweil Sep 19, 2019

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed implementeqd, wth.

```

Adds either an LSTM or GRU recurrent layer to the network using eiter the x (width) or y (height) dimension as the time axis. Input features are the channel dimension and the non-time-axis dimension (height/width) is treated as another batch dimension. For example, a Lfx25 layer on an 1, 16, 906, 32 input will execute 16 independent forward passes on 906x32 tensors resulting in an output of shape 1, 16, 906, 25. If this isn’t desired either run a summarizing layer in the other direction, e.g. Lfys20 for an input 1, 1, 906, 20, or prepend a reshape layer S1(1x16)1,3 combining the height and channel dimension for an 1, 1, 906, 512 input to the recurrent layer.
Adds either an LSTM or GRU recurrent layer to the network using either the x (width) or y (height) dimension as the time axis. Input features are the channel dimension and the non-time-axis dimension (height/width) is treated as another batch dimension. For example, a Lfx25 layer on an 1, 16, 906, 32 input will execute 16 independent forward passes on 906x32 tensors resulting in an output of shape 1, 16, 906, 25. If this isn’t desired either run a summarizing layer in the other direction, e.g. Lfys20 for an input 1, 1, 906, 20, or prepend a reshape layer S1(1x16)1,3 combining the height and channel dimension for an 1, 1, 906, 512 input to the recurrent layer.
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed eiter.

| Linear | | same with Kraken | output size <br>01c\<s\> | | ❌ |
| Modify Top Layers | | —append_index:[int,-1]cut the head off the network at the given index and append —net_spec network in place of cut off part | -a, --append INTEGER: remove layers before argument and then appends spec. Only work when loading an existing model | | ? |
| Loading Existing Model | | --continue_from[string, none):path to previous checkpoint from which to continue training or fine tune(training checkpoint or a recognition mdoel)<br>--stop_training[false): convert the training checkpoint in --continue_from to a recognition model<br>—convert_to_int[bool, false]: when using stop_training, convert to 8-bit integer for greater speed, with slightly less accuracy | -i, --load PATH: Load existing file to continue training | | --weights WEIGHTS, string<br>Load network weights from the given file |
| Loading Existing Model | | --continue_from[string, none):path to previous checkpoint from which to continue training or fine tune(training checkpoint or a recognition model)<br>--stop_training[false): convert the training checkpoint in --continue_from to a recognition model<br>—convert_to_int[bool, false]: when using stop_training, convert to 8-bit integer for greater speed, with slightly less accuracy | -i, --load PATH: Load existing file to continue training | | --weights WEIGHTS, string<br>Load network weights from the given file |
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed mdoel.

| Preload Data into Memory | | ❌ | --preload / --no-preload: Enables/disables preloading of the training set into memory for accelerated training. The default setting preloads data sets with less than 2500 lines, explicitly adding `--preload` will preload arbitrary sized sets. `--no-preload` disables preloading in all circumstances. | ❌ | —train_data_on_the_fly: Instead of preloading all data during the training, load the data on the fly. This is slower, but might be required for limited RAM or large datasets. —validation_data_on_the_fly:Instead of preloading all data during the training, load the data on the fly. This is slower, but might be required for limited RAM or large datasets |
| Number of openMP threads | | ❌ | --threads INTEGER[1]: Number of OpenMP threads and workers when running on CPU. | ❌ | --num_threads NUM_THREADS:The number of threads to use for all operations. —num_inter_threads,int, [0]"Tensorflow's session inter threads param") --num_intra_threads, int, [0], Tensorflow's session intra threads param |
| Special | | --max_image_MB, int[6000], maximum amount of memory to use for caching images<br>--perfect_sample_delay, int[0]: When the network gets good, only backprop a perfect sample after this many imperfect samples have been seen since the last perfect sample was allowed through.<br>--sequential_training:[bool, false], true for sequential training. Default to process all training data in round-robin fashion.<br>—traineddata[string,none]path to the starter trained data file that contains the unicharset, recorder and optional language model<br>—debug_interval[int,0]:If non-zero, show visual debugging every this many iterations. | -d, --device TEXT [cpu]:Select device to use (cpu, cuda:0, cuda:1, …)<br> | --start START[-1]:manually set the number of already learned lines, which influences the naming and stoping condition, default: -1 which will then be overriden by the value saved in the network:question: | —no_skip_invalid_gt, Do no skip invalid gt, instead raise an exception.<br>--gradient_clipping_mode GRADIENT_CLIPPING_MODE, Clipping mode of gradients. Defaults to AUTO, possible values are AUTO, NONE, CONSTANT. --gradient_clipping_const GRADIENT_CLIPPING_CONST:Clipping constant of gradients in CONSTANT mode.<br>--gt_extension GT_EXTENSION: Default extension of the gt files (expected to exist in same dir)<br> |
| Special | | --max_image_MB, int[6000], maximum amount of memory to use for caching images<br>--perfect_sample_delay, int[0]: When the network gets good, only backprop a perfect sample after this many imperfect samples have been seen since the last perfect sample was allowed through.<br>--sequential_training:[bool, false], true for sequential training. Default to process all training data in round-robin fashion.<br>—traineddata[string,none]path to the starter trained data file that contains the unicharset, recorder and optional language model<br>—debug_interval[int,0]:If non-zero, show visual debugging every this many iterations. | -d, --device TEXT [cpu]:Select device to use (cpu, cuda:0, cuda:1, …)<br> | --start START[-1]:manually set the number of already learned lines, which influences the naming and stopping condition, default: -1 which will then be overridden by the value saved in the network:question: | —no_skip_invalid_gt, Do no skip invalid gt, instead raise an exception.<br>--gradient_clipping_mode GRADIENT_CLIPPING_MODE, Clipping mode of gradients. Defaults to AUTO, possible values are AUTO, NONE, CONSTANT. --gradient_clipping_const GRADIENT_CLIPPING_CONST:Clipping constant of gradients in CONSTANT mode.<br>--gt_extension GT_EXTENSION: Default extension of the gt files (expected to exist in same dir)<br> |
Copy link
Author

@stweil stweil Sep 19, 2019

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed stoping, overriden.

@stweil
Copy link
Author

stweil commented Oct 22, 2019

Ping?

stweil added 2 commits August 20, 2020 09:11
Signed-off-by: Stefan Weil <sw@weilnetz.de>
Signed-off-by: Stefan Weil <sw@weilnetz.de>
@stweil
Copy link
Author

stweil commented Aug 20, 2020

@Doreenruirui, please merge.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant