About language normalization #17
Replies: 5 comments 11 replies
-
Yes, I think that's a good idea. We can create a dictionary with {alpha2: service_specific_code} value and use this single dictionary in both _language_normalize and _language_denormalize. And I think we also need to make a list of languages supported by the service and check source/destinaion language by this list at BaseTranslator level. But what will we do with the languages which are not in official ISO639? For example Kazakh-Latin, Emoji, Uzbek-Cyrilic. Just leave these languages in the ISO639 list? |
Beta Was this translation helpful? Give feedback.
-
@ZhymabekRoman Just merged the |
Beta Was this translation helpful? Give feedback.
-
Ahh, yes, look at GitHub Action, for some reason 2 instances of Pylint have been idle for more than 4 hours, it will exhaust your entire free GitHub Action plan. |
Beta Was this translation helpful? Give feedback.
-
@ZhymabekRoman I finally repaired I'm just going to fix the issue with |
Beta Was this translation helpful? Give feedback.
-
I do think that the current state is ready for v2 (unless I forgot something?) What do you think about releasing it @ZhymabekRoman ? |
Beta Was this translation helpful? Give feedback.
-
@ZhymabekRoman Do you really think we should have Google, Bing, Reverso, etc. specified in the dataset we have? (calling by_google for example).
We maybe should juste have the language normalization and denormalization functions which holds the changes needed to make to the language name (with maybe a dictionary) and have the service specific language code accessible through those functions only.
Beta Was this translation helpful? Give feedback.
All reactions