Yasuhiro Horimoto 2019-01-04 10:35:39 +0900 (Fri, 04 Jan 2019) Revision: 28078c85361aefe7771fb1bce3ce23c7aa6426f0 https://github.com/groonga/groonga/commit/28078c85361aefe7771fb1bce3ce23c7aa6426f0 Message: doc: Separate from tokenizers page Added files: doc/source/reference/tokenizers/token_bigram_split_symbol.rst Modified files: doc/locale/ja/LC_MESSAGES/reference.po Modified: doc/locale/ja/LC_MESSAGES/reference.po (+34 -0) =================================================================== --- doc/locale/ja/LC_MESSAGES/reference.po 2019-01-04 09:58:48 +0900 (86e738871) +++ doc/locale/ja/LC_MESSAGES/reference.po 2019-01-04 10:35:39 +0900 (6b9b44c79) @@ -27409,6 +27409,40 @@ msgid "``TokenBigram`` hasn't parameter::" msgstr "``TokenBigram`` には、引数がありません。" msgid "" +"``TokenBigramSplitSymbol`` is similar to :ref:`token-bigram`. The difference " +"between them is symbol handling." +msgstr "" +"``TokenBigramSplitSymbol`` は :ref:`token-bigram` と似ています。違いは記号の" +"扱いです。" + +msgid "``TokenBigramSplitSymbol`` hasn't parameter::" +msgstr "``TokenBigramSplitSymbol`` には、引数がありません。" + +msgid "``TokenBigramSplitSymbol`` tokenizes symbols by bigram tokenize method:" +msgstr "" +"``TokenBigramSplitSymbol`` は記号のトークナイズ方法にバイグラムを使います。" + +#, fuzzy +msgid "" +"``TokenBigramSplitSymbolAlpha`` is similar to :ref:`token-bigram`. The " +"difference between them is symbol and alphabet handling." +msgstr "" +"``TokenBigramIgnoreBlankSplitSymbolAlpha`` は :ref:`token-bigram` と似ていま" +"す。違いは次の通りです。" + +#, fuzzy +msgid "``TokenBigramSplitSymbolAlpha`` hasn't parameter::" +msgstr "``TokenBigram`` には、引数がありません。" + +#, fuzzy +msgid "" +"``TokenBigramSplitSymbolAlpha`` tokenizes symbols and alphabets by bigram " +"tokenize method:" +msgstr "" +"``TokenBigramIgnoreBlankSplitSymbolAlpha`` は記号とアルファベットをバイグラム" +"でトークナイズします。" + +msgid "" "``TokenDelimit`` extracts token by splitting one or more space characters " "(``U+0020``). For example, ``Hello World`` is tokenized to ``Hello`` and " "``World``." Added: doc/source/reference/tokenizers/token_bigram_split_symbol.rst (+31 -0) 100644 =================================================================== --- /dev/null +++ doc/source/reference/tokenizers/token_bigram_split_symbol.rst 2019-01-04 10:35:39 +0900 (229068ae0) @@ -0,0 +1,31 @@ +.. -*- rst -*- + +.. highlightlang:: none + +.. groonga-command +.. database: tokenizers + +``TokenBigramSplitSymbol`` +========================== + +Summary +------- + +``TokenBigramSplitSymbol`` is similar to :ref:`token-bigram`. The +difference between them is symbol handling. + +Syntax +------ + +``TokenBigramSplitSymbol`` hasn't parameter:: + + TokenBigramSplitSymbol + +Usage +----- + +``TokenBigramSplitSymbol`` tokenizes symbols by bigram tokenize method: + +.. groonga-command +.. include:: ../../example/reference/tokenizers/token-bigram-split-symbol-with-normalizer.log +.. tokenize TokenBigramSplitSymbol "100cents!!!" NormalizerAuto -------------- next part -------------- An HTML attachment was scrubbed... URL: <https://lists.osdn.me/mailman/archives/groonga-commit/attachments/20190104/2f0cea9e/attachment-0001.html>