Tokenization in Natural Language Processing


Episode Artwork
1.0x
0% played 00:00 00:00
Sep 14 2020 2 mins  

In this episode we discuss about tokenization in Natural Language Processing. As discussed in previous episode, tokenisation is an important step in data cleaning and it entails dividing a large piece of text into smaller chunks. In this episode we discuss some of the basic tokenizers available from nltk.tokenize in nltk.

If you liked this episode, do follow and do connect with me on twitter @sarvesh0829

follow my blog at www.stacklearn.org.

If you sell something locally, do it using BagUp app available at play store, It would help a lot.

--- Send in a voice message: https://podcasters.spotify.com/pod/show/sarvesh-bhatnagar/message