You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We have been using a fast TokenBuffer API to speed up for various tokenizers in WordTokenizers.jl.
Referring to #141#140, I think it might be beneficial to extend the TokenBuffer API for Documents and Corpus that TextAnalysis.jl offers (excluding NGramDocument and TokenDocument).
This can then be used to improve the performance for preprocessing.jl.
Edit: This could also serve as a solution for #74#76
The text was updated successfully, but these errors were encountered:
We have been using a fast TokenBuffer API to speed up for various tokenizers in WordTokenizers.jl.
Referring to #141 #140, I think it might be beneficial to extend the TokenBuffer API for Documents and Corpus that TextAnalysis.jl offers (excluding NGramDocument and TokenDocument).
This can then be used to improve the performance for preprocessing.jl.
Edit: This could also serve as a solution for #74 #76
The text was updated successfully, but these errors were encountered: