__________ is the process of splitting whole data into smaller chunks in NLP.
Answers
Answered by
9
Explanation:
ᴛᴏᴋᴇɴɪᴢᴀᴛɪᴏɴ is the process of splitting whole data into smaller chunks in NLP.
Answered by
1
Answer: Tokenization
Explanation:
- Tokenization is the process of splitting a phrase, word or sentences in further smaller units. The smaller units in which it has been splitted are called tokens.
- Tokens could be anything. It could be words, numbers or punctuation marks.
- The end point of a particular word and the beginning of the next word is known as the word boundary.
- The importance of this method is that it enables the easy understanding of the meaning of the text and its interpretation.
- Tokens are also considered as the first steps of stemming and lemmatization.
- Creating the tokens is entire based on the objective of the study and sometimes it is also referred as the targeted study of the articles as it enable the specific study of the article.
To know more about Tokenization refer the links given below:
https://brainly.in/question/2038451?referrer=searchResults
https://brainly.in/question/34010129?referrer=searchResults
#SPJ5
Similar questions
French,
2 months ago
English,
2 months ago
Geography,
4 months ago
English,
10 months ago
Social Sciences,
10 months ago