Before returning home, package nltk_data of Python natural language processing into 360 cloud disks and share it with friends, saving everyone as much time as I do.
Download and decompress the package at one time. The official nltk. Download () always fails to be downloaded. Countless times. I wasted a lot of time.
Package download (recommended ):
Http://l3.yunpan.cn/lk/QvLSuskVd6vCU? SID = 1, 305
Download the package and put it in the python/nltk_data directory.
Record the error message so that you can search for it.
The error was:
Nltk. Download ()
Cocould not find maxent_treebank_pos_tagger/English. xxx
Searched index:
C: \\ python27 \\ nltk_data \
C: \ nltk_data
D: \ nltk_data
E: \ nltk_data
Then, I used nltk. Download () to download it. The error always occurred, so I had
Download http://nltk.org/nltk_data/ official
You can also download it from nltk.code.google.com,
However, after downloading the data file, you can directly store it in the nltk_data directory. However, the program cannot find the dataset.
So I tried to download one with nltk. Download () and observe its file location, so that I found that there were several original document directories.
Use the Tree Command to structure its directory as follows:
Folder path list
The volume serial number is 00000200 b2f8: ed9d
├ ── Chunkers# This level is the folder under nltk_data
│ └ ── Maxent_ne_chunker# This level is the data file under the corresponding folder
─ ── Zhanga# This level is the folder under nltk_data
│ ─ ── ABC# This level is the data file under the corresponding folder
│ ─ ── Alpino
│ ├ ── Basque_grammars
│ ├ ── Biocreative_ppi
│ ├ ── Book_grammars
│ ├ ── Brown
│ ├ ── Brown_tei
│ ─ ── Cess_cat
│ ├ ── Cess_esp
│ ├ ── Chat80
│ ─ ── City_database
│ ─ ── Cmudict
│ ─ ── Comtrans
│ ─ ── Conll2000
│ ─ ── Conll2002
│ ─ ── Conll2007
│ Dependency_treebank
│ ─ ── Europarl_raw
│
│ ├ ── Floresta
│ ├ ── Gazetteers
│ ├-Genesis
│ ─ ── Gutenberg
│ ─ ── Hmm_treebank_pos_tagger
│ ├ ── IEER
│ ├ ── Inaugural
│ ─ ── Indian
│ ├ ── Jeita
│ ├ ── Kimmo
│ ─ ── Knbc
│
│ ─ ── Langid
│ ├ ── Large_grammars
│ Machado
│
│ ├ ── Mac_morpho
│ ├ ── Maxent_ne_chunker
│ ├ ── Maxent_treebank_pos_tagger
│ ─ ── Movie_reviews
│
│ ├ ── Names
│ ─ ── Nombank.1.0
│
│ ─ ── Nps_chat
│ ─ ── Oanc_masc
│
│ ─ ── Paradigms
│ ─ ── Pe08
│ ── PIL
│ ├ ── Pl196x
│ ├ ── Ppattach
│ ├ ── Problem_reports
│ ─ ── Propbank
│
│ ─-PTB
│ ├ ── Punkt
│ ─ ── QC
│ ├ ── Reuters
│
│ ─ ── Rslp
│ ─ ── RTE
│ Sampled-sample_grammars
│ Eclip── semcor
│
│ Sampled-senseval
│ ├ ── Shakespeare
│ ├ ── Sinica_treebank
│ ─ ── Smultron
│ ── Spanish_grammars
│ ├ ── State_union
│ ─ ── Stopwords
│ ── Swadesh
│ ├ ── Switchboard
│ ─ ── Tagsets
│ Commit-timit
│
│ ─ ── Toolbox
│
│ Zhu── treebank
│
│ ─ ── UDHR
│ ─ ── Udhr2
│ ── Unicode_samples
│ ─ ── Verbnet
│ ─ ── Webtext
│ ├ ── WordNet
│ ─ ── Wordnet_ic
│ Words ── words
│ Polic── ycoe
Grammars
│ ├ ── Basque_grammars
│ ├ ── Book_grammars
│ ├ ── Large_grammars
│ Sampled-sample_grammars
│ ── Spanish_grammars
├ ── Help
│ ─ ── Tagsets
Memory-stemmers
│ ─ ── Rslp
├ ── Taggers
│ ─ ── Hmm_treebank_pos_tagger
│ ├ ── Maxent_ne_chunker
│ └ ── Maxent_treebank_pos_tagger
└ ── Tokenizers
└ ── Punkt