搜索资源列表
dict
- 这是我的中文信息处理的小作业,是一个自动分词系统。- This is my Chinese information processing 灏忎綔 industry, is an automatic participle system.
关于分词规范的探讨(语言文字应用
- 分词规范的探诫坟-Word standardized detection commandment graves
SegExample
- 对一个文件可以分词的系统,挺有用的哦-right can be a Word document system, quite the oh
cutdic
- 用于中文分词的切分词典,还有词性标注,以这个词典数据库为基础建立文本分类,文本检索或文本过滤可以节省很多时间.很全,强烈推荐!-for the Chinese word segmentation Dictionary, and tagging to the dictionary database based text classification, text retrieval or text filtering can save a
WordSeg
- 利用最大匹配法进行汉语句子的分词 最大匹配算法是最常用的分词算法,简单实用正确率可达到80%以上-the maximum matching method for the Chinese Sentence Word maximum matching algorithm is the most commonly used word segmentation algorithm, simple and practical accuracy r
ProbWordSeg
- 最大概率分词法,这种分词算法能够较好的解决汉语分词中的歧义问题,但分词效率比最大匹配分词算法要低-greatest probability points accidence, Segmentation algorithm can be used to solve the Chinese word segmentation of Ambiguity, but Word efficient than the largest matching
Codes and Application
- 这是一个中科院中文词法分析器,能对输入的文本进行分词,词性标注,未登陆词识别等功能,正确率相当高,很有用哦-This is an Academia Sinica Chinese morphology analyzer, can carry on the participle to the input text, the lexical category sign note, has not landed function and so
Qiyi
- 最大概率法分词的数据结构与算法,用这样的方法分词可以提高分词中歧义词的辨别率,非常经典啊-greatest probability- term data structure and algorithm, this approach can increase Word word-of ambiguous words in the identification rate, very classic ah
lucene-1.4.3
- java分词技术,只实现英文分词,但是该分词算法很经典(来源于apache)-java-term technology, achieving only English Word, but the Word algorithm classic (from apache)
3_31_1
- 中文分词的代码,是用vc++编的.文件名为3_31_1.zip-Chinese word code is used vc series. Documents called 3_31_1. Zip
yangpengwei
- 中文分词程序,是用java语言编的.文件名为yangpengwei.zip-Chinese word segmentation procedure is the use of java language series. Documents called yangpengwei.zip
zya0132_freebbssearch
- 基于中文分词的bbs源程序。具有很好的全站信息检索功能。-based on the Chinese word of bulletin board source. With good station information retrieval functions.
mlct_public
- 这是一个基于Java的分词、N-gram统计、分段 、分句等功能的程序,支持多种语言-This is a Java-based segmentation, N-gram statistics, the sub-clause of the function procedures, multilingual support
0000000000
- 中文分词: 输入一个汉字的文本文件,可将里面的汉字分解成一个个符合人们心理的词-Chinese word : a Chinese character input text files can be inside of a Chinese character is divided into 000 people with mental word
SplitCNWord
- 一个中文分词的实现及演示程序,可用与中文和英文词组的分割.-a Chinese word achieve and demonstrate the procedure can be used with the Chinese and English phrase segmentation.
FreeICTCLAS
- 中科院的分词动态连接库,不用我说了哈,在国内是大名顶顶哈-CAS-term dynamic link library, I do not have said Kazakhstan, in Kazakhstan after another is Daming
Xerdoc XDMF
- 一个比较好用的中文分词模块,是车东等人开发的,比较有名,java语言实现-a better quality of the Chinese word module, East cars were developed, more famous, java language!
语料库
- 一份很重要的语料库,为你的分词程序是一个很好用的资料库文件-a very important corpus, as your segmentation procedure is a very good use of the database file
lzsearch
- 用javascrip编写的分词系统 可以解决现在许多网站中文搜索支持不好的问题 无解压密码 -javascrip prepared with the sub-term system can solve many Web sites now support Chinese search the problems without extracting passwords
wordseg
- 中文文本分类中对文本进行分词处理,以降温本表示称词的集合,然后用词向量来进行分类。-Chinese text classification of text-word processing, said to cool the pool said the words, then word vector to classify.