搜索资源列表
Win32Cut
- 分词程序,Win32窗口界面程序,含设计文档,具有打开文档,显示分词结果,保存结果等功能,欢迎讨论。- The participle procedure, the Win32 window contact surface procedure, contains the design documents, has opens the documents, demonstrated the participle result, pres
wordppl
- 本程序采用正向 逆向最大匹配才实现汉字分词-the procedures being used in reverse to get the maximum matching Chinese Word
framework
- 基于动态规划的中文分词程序,用vc写的,便于扩展。-based on dynamic programming of the Chinese word segmentation procedures using vc write, easy expansion.
chentian.nutch
- 实现了基于词库的nutch中文分词,主要修改了其中的.jj文件等-realized based on the thesaurus nutch Chinese word, the main change of them. Jj documents
chentian.fenci
- 实现了基于词库的nutch中文分词,这一部分是其中的dll文件-realized based on the thesaurus nutch Chinese word, this part is one of the dll file
firtex_beta102_src
- FirteX介绍 功能: 支持增量索引,差量索引,多字段索引,提供了3种前向索引方式; 支持纯文本,HTML,PDF等文件格式; 提供快速中文分词; 从底层到高层,提供了多种索引访问接口,灵活自由地使用索引文件; 提供丰富的检索语法,支持多字段检索,日期范围检索,检索结果自定义排序等。 性能: 在Pentium 4 2.8G 2GRAM的机器上超过200Mb每分钟的索引速度 在近7G的索引文件(10
PWSWNRCODE
- 最大概率法分词。这种技术的分词效率极高。大家共享了。-greatest probability method segmentation. This segmentation of the very efficient. Share of.
ChineseTokenizer20060426
- 分词文件,是一个比较简单词库.简单的说就是个强大的搜索-word document is a relatively simple thesaurus. Simply means that the powerful search
Demo-3GWS
- 这是我在作研究是所使用的分词系统软件,现在提交给大家,供大家研究使用.-This is my research is the use of segmentation system software, we now introduce to you, for your study.
Wordsegmentation2
- NLP技术实现,对语料库进行自动统计生成分词词典,对训练集进行分词,列出所有的分词可能并计算每种可能的概率。请使用者自行加入语料库和测试集。-NLP technology to automatically Corpus Health Statistics ingredients dictionary, the training set for segmentation, list all the sub-term may calcula
2006111
- 支持英文、数字、中文(简体)混合分词 常用的数量和人名的匹配 超过22万词的词库整理 实现正向最大匹配算法 智能数字,日期,人名识别。-support English, figures, Chinese (simplified) mixed at the number of commonly used words and the names of more than 220,000 matching word thesaurus fin
SQLET_split
- SQLET分词算法,一个C程序,供大家参考。-SQLET segmentation algorithm, a C program, for your reference.
HLsegment
- 海量分词研究版的java语言接口,并附dll动态链接库.-Massive Segmentation Research version of the java language interface, with dynamic link library dll.
fencicode
- 挺不错的东西,分词库要自己去添加,我没有乱写啊。-quite good things, thesaurus minutes to add to his, ah, I do not write without basis.
CSW_CN_dll_5.0.0
- CSW中文分词组件_5.0.0 CSW中文分词组件_5-CSW Chinese word segmentation components _5.0.0 CSW Chinese word segmentation components _5
SegmentRAM
- 1.基于luncene倒排索引格式的高性能索引数据访问接口。 2. 实现若干中文分词的的分词算法。-1. Based on luncene inverted index format of high-performance data access interface Index. 2. A number of Chinese word segmentation algorithm of.
3GWS-jni
- 又一个free中文分词 不过需要向作者索要注册码~这不是什么问题吧-also a free Chinese word but it needs to license key-authors claim this is not what it
lunce-splitword
- 一个运用于lunce中的中文分词算法源码,与大家共享-a lunce applied to the Chinese word segmentation algorithm source code and share
sousyinqing
- 搜索引擎技术的研究论文,本文阐述了搜索引擎的基本原理,着重分析了中文分词的设计与实现。-search engine technology research papers, the paper deals with the search engine's basic principles focused on analysis of the Chinese word Design and Implementation.
guide
- 中科院的分词程序说明,缺少此文档的可以看看看。-CAS-word descr iption of the processes, the lack of this document can watch to see.