搜索资源列表
分词程序代码
- 分词程序的源代码-Word program's source code
MFC查词典、分词、词频统计程序
- MFC编程,功能是查词典(用户可自己导入文本),分词,统计词频,还可以保存结果!我们MFC课的期末作业,强烈推荐!-MFC programming function is to check dictionary (users can import their own version), participle, statistical, frequency, the results can be saved! We MFC class at the end operations, strongly
Free ICTCLAS 中科院的分词软件ICTCLAS
- 中科院的分词软件ICTCLAS,自己已经把他用到程序里了感觉效果很好,分享给大家-Chinese Academy of Sciences of the sub-word software ICTCLAS, he has used his program works well in a sense, we share
pymmseg.用python写的分词程序
- 用python写的分词程序,实现的是最大匹配方法,简单易用,Using python to write the sub-term process of implementation is the largest matching method, easy-to-use
pymmseg-cpp-win32-1.0.1.tar.gz
- 基于python的中文分词程序,易用性高,可以作为接口直接在python程序中使用,Python-based Chinese word segmentation process, ease of use high, can be used as interface directly in the python program to use
Chinese-Segmentation.rar
- 自己编写的中文分词源程序,用vc++编写,附有完整的文档,以及标准的分词数据库,I have written the source code of the Chinese word segmentation, using vc++ to prepare, with complete documentation, as well as sub-standard speech database
windows_c_32.rar
- 中国科学院的最新版本的中文分析程序,可以进行分词、词性标注等,The latest version of the Chinese Academy of Sciences of the Chinese language analysis procedures, can be sub-word-of-speech tagging, etc.
mmseg
- 基于双数组trie的分词程序,分词速度20MB/S,能够支持GBK、UTF8编码-Double array trie-based sub-word procedure word speed 20MB/S, can support GBK, UTF8 encoding
NICTCLAS_Release
- 中科院分词程序 开源但是词库非开源 中科院分词程序 开源但是词库非开源-wordspilt
SW_I_WordSegment
- SW-I中文分词算法,MFC程序,在visual studio 2008中调试通过。默认词库为mdb,由于较大未包含在源文件中,请自行下载mdb格式的词典。-SW-I Chinese word segmentation algorithm, MFC procedures, visual studio 2008 in debug through. Default thesaurus for the mdb, as a result of the larger not included in the
中文分词算法
- 本程序使用给出的字典进行学习并对训练语料进行分词处理,采用C语言编写,高效易懂!
splitword
- 自己写的小分词程序,中文分词测试版,仅供参考,谢谢!-Writing their own small sub-term process, English sub-test version of the word, for reference purposes only, thank you!
text
- python写的gbk分词分句程序 可以使用sogou或者谷歌输入法的词库进行分词-python written procedures gbk participle clause can use Google sogou or input method for segmentation of the thesaurus
Separate_word_program
- 中文分词程序,能够在短时间内对中文内容进行分解split.txt中是要分解的内容而word.TXT则是分词表,分词表越完整最后结果越正确。-Chinese word segmentation process, can in a short time for Chinese content split.txt in decomposition is to decompose the content is word.TXT participle form, word form the final o
MmFenCi
- 基于MM的分词算法,有兴趣者可以把程序中没有完成的部分继续。-MM sub-word based algorithm, are interested in can not complete the program part to continue.
WordSeg
- 简单分词程序 读入一个pdf 输出一个分好词的txt-Reading of simple segmentation procedure into a pdf output of a good word txt
2004050215271615762
- 本文是关于中文分词程序源代码,仅供大家学习和研究使用。-This article is about the Chinese word segmentation program source code, for everyone to learn and study.
word_split
- 这个一个基于逆向最大匹配的分词程序,语料规模比较小。-The maximum matching based on the reverse of the sub-term process, relatively small-scale corpus.
imdict-chinese-analyzer
- imdict-chinese-analyzer 是 imdict智能词典 的智能中文分词模块,算法基于隐马尔科夫模型(Hidden Markov Model, HMM),是中国科学院计算技术研究所的ictclas中文分词程序的重新实现(基于Java),可以直接为lucene搜索引擎提供简体中文分词支持。-imdict-chinese-analyzer is a smart imdict Chinese Dictionary smart module segmentation algorithm
IKAnalyzer3.1.1_userguide
- java分词程序,能够精确分词,包含词库等-java word program, word accurately, including the thesaurus, etc.