搜索资源列表
pymmseg-cpp-win32-1.0.1.tar.gz
- 基于python的中文分词程序,易用性高,可以作为接口直接在python程序中使用,Python-based Chinese word segmentation process, ease of use high, can be used as interface directly in the python program to use
segChnWord
- 中文分词评测系统,用于评测中文分词的质量,给出准确率等-Chinese word segmentation evaluation system for evaluating the quality of Chinese word segmentation, given the accuracy of such
ICTCLAS50_Windows_32_C
- python中文分词\ CDict.py-Chinese word python \ python Chinese word \ CDict.py
tmsvm_src_v1.1.0
- SVM中文分词分类器的python实现,在LINUX环境下运行不需太多配置和代码修改。-SVM for Chinese word classification python implementation running under LINUX environment without too much configuration and code changes.
python-ictclas
- 运用python编写的中文分词,运行try文件并将路径改变就可以实现中文分词 中文存储在txt文档中-Use python to write the Chinese word, you can run try to change file and path to achieve the Chinese word Chinese storage txt document
tmsvm_for_win_1.2.0
- python svm 中文分词、特征提取、文本分类-python svm
python-cutword
- 用python写的中文分词,分词效果很好,有接口可以直接调用接口。-cut word process using python,it s good for cutting word.It has interface to be used easily.
自然语言处理的最好的资源
- python,java代码,可以完成分词,训练词向量,等功能,效果突出。
jieba-0.31
- jeiba分词、中文分词、python、开源-jeiba word, Chinese word segmentation, python, open source
CWSS17.1.1.4
- 基于隐马尔科夫模型的中文分词系统,上交ieee专业大一作业,界面一般,主要用于学习,在此分享,注:开发环境python3.5(Based on Hidden Markov model of Chinese word segmentation system, on the IEEE professional freshman job, interface is common, mainly used for learning, in this share, note: development en
jieba-0.39
- Python非常强大的中文分词包,用于自然语言处理。(Python is a very powerful Chinese word wrapper for natural language processing.)
work
- 贝叶斯下文本分类器,python语言,包括中文分词和英文分词(Bias's lower text classifier, python language, including Chinese participle and English participle)
ChPreprocess
- 使用jieba包从excel表中读取数据,进行中文分词,预料分析(Using Jieba package for Chinese analysis, expected analysis)
chinese_seg_update
- 中文分词,采用逆向最大匹配方法实现,利用字典作为索引(The Chinese word segmentation is realized by the reverse maximum matching method, and the dictionary is used as the index.)
jieba-jieba3k
- MATLAB 结巴分词的工具包,用于很多中文分词的模式识别代码程序,利用已有函数工具包提高工作效率,内有安装说明(MATLAB jieba toolkit, used for many Chinese word segmentation pattern recognition code programs, using existing function toolkits to improve work efficiency, with installation instructions)
jieba-0.38
- jieba 分词,用在Python中,对中文文本进行分词(Jieba participle, used in Python to segment Chinese text;)
pyltp
- 调用哈工大的语言云平台lpt进行中文分词处理,支持自定义词典(Word segregation is the process that cutting the string of natural language into words.you can use this to syncopate words)
Python逆向最大匹配实现-代码-文档-截图等
- 对于文本进行自然语言处理,中文分词。北邮计算机院NLP课的作业。(This is a homework for BUPT student, implementing the word segment of natural language processing.)
HMM-master
- 隐马尔科夫模型进行中文分词 模型训练 python HMM_train.py RenMinData.txt_utf8 RenMinData.RenMinData_utf8 为人民日报已经人工分词的预料。 生成三个文件 * prob_start.py 为模型的初始概率 * prob_trans.py 为模型状态转移概率 * prob_emit.py 为发射概率 测试模型效果 python HMM.py reference 维特比算法:(python
Python中文文本预处理
- 包括删除标点、分词、删除停用词、计算相似度、文本聚类等功能