You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I checked my configs, charset containing Chinese in particular, and ran train.py. The output model can't recognize Chinese characters at all.
I ran another friend's output model. It works in Chinese, but when I used its dataset and retrained, recognization failed again.
Appreciated for any advice.
The text was updated successfully, but these errors were encountered:
When running test.py, I found that labels read by the model from test set(lmdb format) remove all Chinese characters which actually exist in test set as unicode format.
Now I am trying to follow #9 and disable unicode-normalization and retrain.
Following #9 and disabling unicode-normalization in the main.yaml really works! Still, I don't know how to deal with that test.py will get rid of Chinese characters when reading labels groundtruth. But it doesn't matter. Let test.py run for open dataset and rewrite read.py to test my own dataset.
I checked my configs, charset containing Chinese in particular, and ran train.py. The output model can't recognize Chinese characters at all.
I ran another friend's output model. It works in Chinese, but when I used its dataset and retrained, recognization failed again.
Appreciated for any advice.
The text was updated successfully, but these errors were encountered: