Sept. 30, 2022, 11:53 a.m. | /u/Certain-Ad827

Natural Language Processing www.reddit.com

I have a list of words that contain about 100000 word and there frequencies, I want to know the frequencies of CEFR levels to classify.

I found an [opinion](https://languagelearning.stackexchange.com/questions/3061/what-are-estimates-of-vocabulary-size-for-each-cefr-level) that suggests to divide the words with index, so the most common 500 words is A1 and next 1000 is A2 and so on, and I feel that this is a very naive way to classify words.

languagetechnology words

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst

@ SEAKR Engineering | Englewood, CO, United States

Data Analyst II

@ Postman | Bengaluru, India

Data Architect

@ FORSEVEN | Warwick, GB

Director, Data Science

@ Visa | Washington, DC, United States

Senior Manager, Data Science - Emerging ML

@ Capital One | McLean, VA