[Groonga-commit] groonga/groonga at 28078c8 [master] doc: Separate from tokenizers page

Back to archive index
Yasuhiro Horimoto null+****@clear*****
Fri Jan 4 10:35:39 JST 2019


Yasuhiro Horimoto	2019-01-04 10:35:39 +0900 (Fri, 04 Jan 2019)

  Revision: 28078c85361aefe7771fb1bce3ce23c7aa6426f0
  https://github.com/groonga/groonga/commit/28078c85361aefe7771fb1bce3ce23c7aa6426f0

  Message:
    doc: Separate from tokenizers page

  Added files:
    doc/source/reference/tokenizers/token_bigram_split_symbol.rst
  Modified files:
    doc/locale/ja/LC_MESSAGES/reference.po

  Modified: doc/locale/ja/LC_MESSAGES/reference.po (+34 -0)
===================================================================
--- doc/locale/ja/LC_MESSAGES/reference.po    2019-01-04 09:58:48 +0900 (86e738871)
+++ doc/locale/ja/LC_MESSAGES/reference.po    2019-01-04 10:35:39 +0900 (6b9b44c79)
@@ -27409,6 +27409,40 @@ msgid "``TokenBigram`` hasn't parameter::"
 msgstr "``TokenBigram`` には、引数がありません。"
 
 msgid ""
+"``TokenBigramSplitSymbol`` is similar to :ref:`token-bigram`. The difference "
+"between them is symbol handling."
+msgstr ""
+"``TokenBigramSplitSymbol`` は :ref:`token-bigram` と似ています。違いは記号の"
+"扱いです。"
+
+msgid "``TokenBigramSplitSymbol`` hasn't parameter::"
+msgstr "``TokenBigramSplitSymbol`` には、引数がありません。"
+
+msgid "``TokenBigramSplitSymbol`` tokenizes symbols by bigram tokenize method:"
+msgstr ""
+"``TokenBigramSplitSymbol`` は記号のトークナイズ方法にバイグラムを使います。"
+
+#, fuzzy
+msgid ""
+"``TokenBigramSplitSymbolAlpha`` is similar to :ref:`token-bigram`. The "
+"difference between them is symbol and alphabet handling."
+msgstr ""
+"``TokenBigramIgnoreBlankSplitSymbolAlpha`` は :ref:`token-bigram` と似ていま"
+"す。違いは次の通りです。"
+
+#, fuzzy
+msgid "``TokenBigramSplitSymbolAlpha`` hasn't parameter::"
+msgstr "``TokenBigram`` には、引数がありません。"
+
+#, fuzzy
+msgid ""
+"``TokenBigramSplitSymbolAlpha`` tokenizes symbols and alphabets by bigram "
+"tokenize method:"
+msgstr ""
+"``TokenBigramIgnoreBlankSplitSymbolAlpha`` は記号とアルファベットをバイグラム"
+"でトークナイズします。"
+
+msgid ""
 "``TokenDelimit`` extracts token by splitting one or more space characters "
 "(``U+0020``). For example, ``Hello World`` is tokenized to ``Hello`` and "
 "``World``."

  Added: doc/source/reference/tokenizers/token_bigram_split_symbol.rst (+31 -0) 100644
===================================================================
--- /dev/null
+++ doc/source/reference/tokenizers/token_bigram_split_symbol.rst    2019-01-04 10:35:39 +0900 (229068ae0)
@@ -0,0 +1,31 @@
+.. -*- rst -*-
+
+.. highlightlang:: none
+
+.. groonga-command
+.. database: tokenizers
+
+``TokenBigramSplitSymbol``
+==========================
+
+Summary
+-------
+
+``TokenBigramSplitSymbol`` is similar to :ref:`token-bigram`. The
+difference between them is symbol handling.
+
+Syntax
+------
+
+``TokenBigramSplitSymbol`` hasn't parameter::
+
+  TokenBigramSplitSymbol
+
+Usage
+-----
+
+``TokenBigramSplitSymbol`` tokenizes symbols by bigram tokenize method:
+
+.. groonga-command
+.. include:: ../../example/reference/tokenizers/token-bigram-split-symbol-with-normalizer.log
+.. tokenize TokenBigramSplitSymbol "100cents!!!" NormalizerAuto
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.osdn.me/mailman/archives/groonga-commit/attachments/20190104/2f0cea9e/attachment-0001.html>


More information about the Groonga-commit mailing list
Back to archive index