Will progress in MT require some breakthrough in grammar conceptualization? Rule-based MT requires a detailed typology of grammatical categories. But it seems that the present state of grammar conceptualization is insufficient for MT purposes. Some enhanced typology is in order. In particular the common typology relating to:
- partitive article, i.e. the disambiguation of French ‘de’
is in need of a more accurate conceptualization.
What is it to make a rule-based translation software for a given language pair?
It amounts to making part of Brain Emulation, dedicated to translating one language into another i.e. emulating the brain of a bilingual individual. Arguably, ‘human cognition emulation’ is best suited here than ‘human brain emulation’, since this kind of emulation does not bear on neurons or synapses.
It shoud be noted that this applies specifically to rule-based translation. But this casts light on the fact that the present kind of translation should be better termed ‘human translation emulation’, since it aims at emulating human reasoning in the process of translating one language into another, by notably taking into account:
- grammatical disambiguation (how we humans disambiguate two inherently ambiguous terms)
- semantical disambiguation
- grammatical type mismatch’s reversal (‘la mer bleue’ = u mare turchinu)
- part-of-speech tagging