Recently, I have improved the search program in the blog garden. The search function of the blog Park uses the mongoe.net search engine. When the search function was added to the blog Park, mongoe.net did not support Chinese word segmentation. Later, I got help from http://www.cnblogs.com/yuhen/to answer this question. (A problem has occurred in the search program of the blog site recently. google is used for the moment ).
Currently, word segmentation is supported in Alibaba e.net. I downloaded the Language Processing Package (
Lucene. Net NLS PackUsing Lucene. Net. Analysis. Cn. ChineseAnalyzer to create an index, but every time an index is created, a deadlock occurs. After reading this article on sf.net, CJK Analysis maybe cause dead lock, I think it may be Lucene. net. analysis. the code in Cn has a Bug. Check Lucene. net. analysis. the code in Cn really has a problem. In line 148th of ChineseTokenizer, change dataLen =-1 to dataLen = 0, and the deadlock problem is solved. Later, I found this article. I changed the two bugs and put them on the blog garden to download to the publisher. Although only two lines of code have been modified, I want to put it here, which is helpful for some first-time users.
Bin: http://files.cnblogs.com/dudu/Lucene.Net.Analysis.Cn.rar
Src: http://files.cnblogs.com/dudu/Lucene.Net.NLS.rar