Standardization
Optimizing the production, maintenance and extension of computational lexicons is one of the crucial aspects impacting NLP. The main problem is the interoperability: various lexicons are frequently incompatible. The most frequent situation is: how to merge two lexicons, or fragments of lexicons? A secondary problem is that a lexicon is usually specifically tailored to a specific NLP program and has difficulties being used within other NLP programs or applications.
To this respect, the various data models of Computational lexicons are studied by ISO/TC37 since 2003 within the project lexical markup framework leading to an ISO standard in 2008.
Read more about this topic: Computational Lexicology