搜索资源列表
jieba
- jieba分词软件,是python下的开源分词软件,里面有使用例子,简单易用-jieba segmentation software, is under the open source python segmentation software, there are examples of the use, easy to use
jieba-0.31
- jeiba分词、中文分词、python、开源-jeiba word, Chinese word segmentation, python, open source
OntologyLearning
- 本体学习相关研究 (1)ConcepLearing 从文本中学习本体概念,使用中科院NLPIR进行分词(也有jieba分词版本的),然后在根据统计(互信息/卡方值)和Topic Model(PLSA)的方法进行实验对比。 -Ontology learning research (1) ConcepLearing learning ontology concept the text, use the CAS NLPIR word segmentation (also jieba Wo
jieba分词
- jieba 的java分词包,一般都是python的包,这个可用于java的jieba分词(Jieba Java word segmentation package, generally Python package, this can be used for the Java Jieba participle)
jieba for Python
- jieba分词功能在python中的实现方法(The Method of jieba for word-split in python)
jieba
- 将句子分成很小的独立词,来提取信息,对照数据字典得到有用的关键信息,进行智能筛选题目或回答问题。(The sentence is divided into very small independent words to extract information, and the data dictionary is used to obtain useful key information.)
jieba_plus
- 解决jieba分词中部分bug,包括全角字母和数字等,更新中(solve part of the bugs in Jieba segmentation, update)
jieba
- python结巴分词模块.能对句子进行分词.(python module that make sentences into words)
jieba-0.39
- Python非常强大的中文分词包,用于自然语言处理。(Python is a very powerful Chinese word wrapper for natural language processing.)
fenci
- 对文件夹中的所有文件进行批量、分词处理并保存(Batch word segmentation for all files in a folder)
jieba分词
- 分词出差的武器的呼气和对区华东区希望成为(xwijidwdjdowslkmxkszmwksww)
cppjieba-master
- 结巴分词,提供针对中文的分词方法,使用C++语言编写(jieba provides a word segmentation method for Chinese, which is written in the C++ language)
ChPreprocess
- 使用jieba包从excel表中读取数据,进行中文分词,预料分析(Using Jieba package for Chinese analysis, expected analysis)
Downloads
- jieba分词和ansj分词使用的java包(The package used by the ansj participle)
CPP
- 分词,结巴分词C++版本。自己学习用,或者可以加入自己的项目(participle,C++ version of participle.Learn to use, or join your own project)
jieba-jieba3k
- MATLAB 结巴分词的工具包,用于很多中文分词的模式识别代码程序,利用已有函数工具包提高工作效率,内有安装说明(MATLAB jieba toolkit, used for many Chinese word segmentation pattern recognition code programs, using existing function toolkits to improve work efficiency, with installation instructions)
jieba
- 精确模式,试图将句子最精确地切开,适合文本分析; 全模式,把句子中所有的可以成词的词语都扫描出来, 速度非常快,但是不能解决歧义; 搜索引擎模式,在精确模式的基础上,对长词再次切分,提高召回率,适合用于搜索引擎分词。(Accurate mode, trying to cut the sentence up to the most accurate, suitable for text analysis. The whole mode can scan all the words tha
jieba-0.38
- jieba 分词,用在Python中,对中文文本进行分词(Jieba participle, used in Python to segment Chinese text;)
jieba分词.net源码
- 该项目是jieba分词组件的.net版本源码实现,生成的库可以使用,分词也较好
chatbot
- 聊天机器人 原理: 严谨的说叫 ”基于深度学习的开放域生成对话模型“,框架为Keras(Tensorflow的高层包装),方案为主流的RNN(循环神经网络)的变种LSTM(长短期记忆网络)+seq2seq(序列到序列模型),外加算法Attention Mechanism(注意力机制),分词工具为jieba,UI为Tkinter,基于”青云“语料(10万+闲聊对话)训练。 运行环境:python3.6以上,Tensorflow,pandas,numpy,jieba。(Chat Robot