You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm new to using sparknlp and have questions regarding the sparknlp.annotator.WordSegmenterModel. Applying this on english texts seems to be impossible as the pretrained models are only in Chinese/Japanese/Korean https://nlp.johnsnowlabs.com/models?task=Word+Segmentation
My texts could contain words like "hellogoodday" and "howareyou", and I need to detect the separate words as tokens. Does anyone know how to get the english language model for this annotator, or is there some other annotator that I should be using? If not, is it feasible to raise this as a new feature request?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi there,
I'm new to using sparknlp and have questions regarding the sparknlp.annotator.WordSegmenterModel. Applying this on english texts seems to be impossible as the pretrained models are only in Chinese/Japanese/Korean https://nlp.johnsnowlabs.com/models?task=Word+Segmentation
My texts could contain words like "hellogoodday" and "howareyou", and I need to detect the separate words as tokens. Does anyone know how to get the english language model for this annotator, or is there some other annotator that I should be using? If not, is it feasible to raise this as a new feature request?
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions