搜索资源列表
jieba-analysis-master
- 结巴分词,在Lucene中用来分词的,该分词器具有自动提取关键字的功能-Stuttered word, the word is used to Lucene, the word has the function of automatic extraction of keywords
jieba-analysis-master
- 结巴分词(java版)只保留的原项目针对搜索引擎分词的功能(cut_for_index、cut_for_search),词性标注,关键词提取没有实现(今后如用到,可以考虑实现)。-Stammer participle (Java version) to retain only the original project for search engine participle (cut for index, cut for search), part of speech tagging, keyw
jieba
- python结巴分词模块.能对句子进行分词.(python module that make sentences into words)
zh_lstm
- lstm做情感分类,中文,用到豆瓣影评,结巴分词,lstm模型,环境python3做编码处理。(lstm for sentiment analyse)
cppjieba-master
- 结巴分词,提供针对中文的分词方法,使用C++语言编写(jieba provides a word segmentation method for Chinese, which is written in the C++ language)
CPP
- 分词,结巴分词C++版本。自己学习用,或者可以加入自己的项目(participle,C++ version of participle.Learn to use, or join your own project)
jiebacut.py
- 通过结巴分词处理中文分词问题【对文本进行分词以及词频统计处理】。(The problem of Chinese participle is dealt with by the branch word segmentation.)
jieba-jieba3k
- MATLAB 结巴分词的工具包,用于很多中文分词的模式识别代码程序,利用已有函数工具包提高工作效率,内有安装说明(MATLAB jieba toolkit, used for many Chinese word segmentation pattern recognition code programs, using existing function toolkits to improve work efficiency, with installation instructions)
text_classification_AI100-master
- 实现LSTM情感分析,中文,用到结巴分词,LSTM模型,环境python3.0(Achieve LSTM sentiment analysis, Chinese, use stuttering participle, LSTM model, environment python3.0)
lda
- 建立文本主题模型,中文结巴分词,一致性分数计算(lda model ,jieba,co score)
垃圾短信分类
- 基于文本内容的垃圾短信识别,对数据进行了数据清洗,分词等,进行 了模型训练及评价(Based on the text content of spam short message recognition, data cleaning, segmentation, model training and evaluation are carried out)