為了《中國哲學書電子化計劃》輸入用
-
Updated
Jun 1, 2024 - C#
為了《中國哲學書電子化計劃》輸入用
Web Adjuster + Annotator Generator
Specifically built for the research proposal: Estimating sector attention index with deep learning methods : example of Chinese stock market, Jan. 4, 2024.
Chinese-English Dictionary Public-domain Additions for Names Etc (CedPane)
SymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
Python port of SymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
zhparser is a PostgreSQL extension for full-text search of Chinese language
Pytorch-NLU,一个中文文本分类、序列标注工具包,支持中文长文本、短文本的多类、多标签分类任务,支持中文命名实体识别、词性标注、分词、抽取式文本摘要等序列标注任务。 Ptorch NLU, a Chinese text classification and sequence annotation toolkit, supports multi class and multi label classification tasks of Chinese long text and short text, and supports sequence annotation tasks such as Chinese named entity recognition, part of spee
ik-analyzer for rust; chinese tokenizer for tantivy
Tokenizer support Lucene5/6/7/8/9+ version, LTS
基于 jieba-rs 的中文分词插件
Jcseg is a light weight NLP framework developed with Java. Provide CJK and English segmentation based on MMSEG algorithm, With also keywords extraction, key sentence extraction, summary extraction implemented based on TEXTRANK algorithm. Jcseg had a build-in http server and search modules for lucene,solr,elasticsearch,opensearch
Using Flask export jieba, SnowNLP, pkuseg as http API web service.
Jiebago 的性能优化版, 支持从 io.Reader 加载字典
A copy-cat implementation of jieba as a learning exercise.
"結巴"中文分詞:做最好的 PHP 中文分詞、中文斷詞組件。 / "Jieba" (Chinese for "to stutter") Chinese text segmentation: built to be the best PHP Chinese word segmentation module.
Mandarin Chinese text segmentation and mobile dictionary Android app (中文分词)
一些容易被分词工具被分错的句子。
开源中文分词工具包,中文分词Web API,Lucene中文分词,中英文混合分词
Add a description, image, and links to the chinese-text-segmentation topic page so that developers can more easily learn about it.
To associate your repository with the chinese-text-segmentation topic, visit your repo's landing page and select "manage topics."