搜索资源列表
jieba
- jieba分词软件,是python下的开源分词软件,里面有使用例子,简单易用-jieba segmentation software, is under the open source python segmentation software, there are examples of the use, easy to use
jieba-analysis-master
- 结巴分词,在Lucene中用来分词的,该分词器具有自动提取关键字的功能-Stuttered word, the word is used to Lucene, the word has the function of automatic extraction of keywords
jieba-0.31
- jeiba分词、中文分词、python、开源-jeiba word, Chinese word segmentation, python, open source
jieba-analysis-master
- 结巴分词(java版)只保留的原项目针对搜索引擎分词的功能(cut_for_index、cut_for_search),词性标注,关键词提取没有实现(今后如用到,可以考虑实现)。-Stammer participle (Java version) to retain only the original project for search engine participle (cut for index, cut for search)
OntologyLearning
- 本体学习相关研究 (1)ConcepLearing 从文本中学习本体概念,使用中科院NLPIR进行分词(也有jieba分词版本的),然后在根据统计(互信息/卡方值)和Topic Model(PLSA)的方法进行实验对比。 -Ontology learning research (1) ConcepLearing learning ontology concept the text, use the CAS NLPIR wo
jieba分词
- jieba 的java分词包,一般都是python的包,这个可用于java的jieba分词(Jieba Java word segmentation package, generally Python package, this can be used for the Java Jieba participle)
jieba for Python
- jieba分词功能在python中的实现方法(The Method of jieba for word-split in python)
jieba
- 将句子分成很小的独立词,来提取信息,对照数据字典得到有用的关键信息,进行智能筛选题目或回答问题。(The sentence is divided into very small independent words to extract information, and the data dictionary is used to obtain useful key information.)
jieba
- python结巴分词模块.能对句子进行分词.(python module that make sentences into words)
jieba-0.39
- Python非常强大的中文分词包,用于自然语言处理。(Python is a very powerful Chinese word wrapper for natural language processing.)
jieba分词
- 分词出差的武器的呼气和对区华东区希望成为(xwijidwdjdowslkmxkszmwksww)
cppjieba-master
- 结巴分词,提供针对中文的分词方法,使用C++语言编写(jieba provides a word segmentation method for Chinese, which is written in the C++ language)
ChPreprocess
- 使用jieba包从excel表中读取数据,进行中文分词,预料分析(Using Jieba package for Chinese analysis, expected analysis)
jieba.NET.0.38.2
- jieba 分词,用于诶文档进行分词处理,可以作为文本分析的材料。(jieba fenci,use yin doc wordsplit,and it can use in the case word analist . and i fuxxk this site)
Downloads
- jieba分词和ansj分词使用的java包(The package used by the ansj participle)
CPP
- 分词,结巴分词C++版本。自己学习用,或者可以加入自己的项目(participle,C++ version of participle.Learn to use, or join your own project)
jieba-jieba3k
- MATLAB 结巴分词的工具包,用于很多中文分词的模式识别代码程序,利用已有函数工具包提高工作效率,内有安装说明(MATLAB jieba toolkit, used for many Chinese word segmentation pattern recognition code programs, using existing function toolkits to improve work efficiency, with
jieba
- 精确模式,试图将句子最精确地切开,适合文本分析; 全模式,把句子中所有的可以成词的词语都扫描出来, 速度非常快,但是不能解决歧义; 搜索引擎模式,在精确模式的基础上,对长词再次切分,提高召回率,适合用于搜索引擎分词。(Accurate mode, trying to cut the sentence up to the most accurate, suitable for text analysis. The whole m
jieba-0.38
- jieba 分词,用在Python中,对中文文本进行分词(Jieba participle, used in Python to segment Chinese text;)
chatbot
- 聊天机器人 原理: 严谨的说叫 ”基于深度学习的开放域生成对话模型“,框架为Keras(Tensorflow的高层包装),方案为主流的RNN(循环神经网络)的变种LSTM(长短期记忆网络)+seq2seq(序列到序列模型),外加算法Attention Mechanism(注意力机制),分词工具为jieba,UI为Tkinter,基于”青云“语料(10万+闲聊对话)训练。 运行环境:python3.6以上,Tensorflow,pan