You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Temporary solution in b462061 by creating a new instance of the tokenizer on each file. This is not too costly because tokenization models are very small but it cannot be counted as a correction.
And in fact, now the tagger crashes on second file.
kleag
changed the title
Deeplima tokenizer blocks on second file
If using the same tokenizer instance, deeplima tokenizer blocks on second file
May 15, 2024
Describe the bug
When deeplima is asked to analyze several files, the first one is analyzed correctly, but then the program stalls.
To Reproduce
Steps to reproduce the behavior:
deeplima --tok-model ~/.local/share/lima/resources/RnnTokenizer/ud/tokenizer-eng-UD_English-EWT.pt test-eng*.txt
segmentation_impl.cpp
here:lima/deeplima/libs/tasks/segmentation/inference/segmentation_impl.cpp
Line 156 in 7a33899
Expected behavior
All files should be analyzed successfully.
The text was updated successfully, but these errors were encountered: