[Groonga-commit] groonga/groonga at 34aacd1 [master] TokenRegexp: fix wrong end mark character...

Back to archive index

Kouhei Sutou null+****@clear*****
Thu Mar 12 21:11:50 JST 2015


Kouhei Sutou	2015-03-12 21:11:50 +0900 (Thu, 12 Mar 2015)

  New Revision: 34aacd121be08147e7f1a90fdb545cafac921ec2
  https://github.com/groonga/groonga/commit/34aacd121be08147e7f1a90fdb545cafac921ec2

  Message:
    TokenRegexp: fix wrong end mark character...

  Modified files:
    lib/tokenizers.c
    test/command/suite/tokenizers/regexp/add/four.expected
    test/command/suite/tokenizers/regexp/add/one.expected
    test/command/suite/tokenizers/regexp/add/three.expected
    test/command/suite/tokenizers/regexp/add/two.expected
    test/command/suite/tokenizers/regexp/get/begin/one.expected
    test/command/suite/tokenizers/regexp/get/begin/three.expected
    test/command/suite/tokenizers/regexp/get/begin/two.expected
    test/command/suite/tokenizers/regexp/get/end/four.expected
    test/command/suite/tokenizers/regexp/get/end/one.expected
    test/command/suite/tokenizers/regexp/get/end/three.expected
    test/command/suite/tokenizers/regexp/get/end/two.expected

  Modified: lib/tokenizers.c (+4 -4)
===================================================================
--- lib/tokenizers.c    2015-03-12 20:36:16 +0900 (796c02d)
+++ lib/tokenizers.c    2015-03-12 21:11:50 +0900 (d5b3db9)
@@ -586,8 +586,8 @@ regexp_next(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data)
       status |= GRN_TOKEN_LAST | GRN_TOKEN_REACH_END;
       grn_tokenizer_token_push(ctx,
                                &(tokenizer->token),
-                               GRN_TOKENIZER_BEGIN_MARK_UTF8,
-                               GRN_TOKENIZER_BEGIN_MARK_UTF8_LEN,
+                               GRN_TOKENIZER_END_MARK_UTF8,
+                               GRN_TOKENIZER_END_MARK_UTF8_LEN,
                                status);
       return NULL;
     }
@@ -606,8 +606,8 @@ regexp_next(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data)
       status |= GRN_TOKEN_LAST | GRN_TOKEN_REACH_END;
       grn_tokenizer_token_push(ctx,
                                &(tokenizer->token),
-                               GRN_TOKENIZER_BEGIN_MARK_UTF8,
-                               GRN_TOKENIZER_BEGIN_MARK_UTF8_LEN,
+                               GRN_TOKENIZER_END_MARK_UTF8,
+                               GRN_TOKENIZER_END_MARK_UTF8_LEN,
                                status);
       return NULL;
     }

  Modified: test/command/suite/tokenizers/regexp/add/four.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/add/four.expected    2015-03-12 20:36:16 +0900 (ae80e56)
+++ test/command/suite/tokenizers/regexp/add/four.expected    2015-03-12 21:11:50 +0900 (9942df3)
@@ -27,7 +27,7 @@ tokenize TokenRegexp "abcd"
       "position": 4
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 5
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/add/one.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/add/one.expected    2015-03-12 20:36:16 +0900 (6c2e513)
+++ test/command/suite/tokenizers/regexp/add/one.expected    2015-03-12 21:11:50 +0900 (f53d7bb)
@@ -15,7 +15,7 @@ tokenize TokenRegexp "x"
       "position": 1
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 2
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/add/three.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/add/three.expected    2015-03-12 20:36:16 +0900 (e865c8c)
+++ test/command/suite/tokenizers/regexp/add/three.expected    2015-03-12 21:11:50 +0900 (cbcc334)
@@ -23,7 +23,7 @@ tokenize TokenRegexp "xyz"
       "position": 3
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 4
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/add/two.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/add/two.expected    2015-03-12 20:36:16 +0900 (27eef72)
+++ test/command/suite/tokenizers/regexp/add/two.expected    2015-03-12 21:11:50 +0900 (4907a52)
@@ -19,7 +19,7 @@ tokenize TokenRegexp "xy"
       "position": 2
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 3
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/begin/one.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/get/begin/one.expected    2015-03-12 20:36:16 +0900 (254c754)
+++ test/command/suite/tokenizers/regexp/get/begin/one.expected    2015-03-12 21:11:50 +0900 (8c9747a)
@@ -17,7 +17,7 @@ table_tokenize Lexicon "x" --mode ADD
       "position": 1
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 2
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/begin/three.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/get/begin/three.expected    2015-03-12 20:36:16 +0900 (49bc507)
+++ test/command/suite/tokenizers/regexp/get/begin/three.expected    2015-03-12 21:11:50 +0900 (878c093)
@@ -25,7 +25,7 @@ table_tokenize Lexicon "xyz" --mode ADD
       "position": 3
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 4
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/begin/two.expected (+1 -1)
===================================================================
--- test/command/suite/tokenizers/regexp/get/begin/two.expected    2015-03-12 20:36:16 +0900 (2808762)
+++ test/command/suite/tokenizers/regexp/get/begin/two.expected    2015-03-12 21:11:50 +0900 (8e82fb0)
@@ -21,7 +21,7 @@ table_tokenize Lexicon "xy" --mode ADD
       "position": 2
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 3
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/end/four.expected (+2 -2)
===================================================================
--- test/command/suite/tokenizers/regexp/get/end/four.expected    2015-03-12 20:36:16 +0900 (af115be)
+++ test/command/suite/tokenizers/regexp/get/end/four.expected    2015-03-12 21:11:50 +0900 (b621183)
@@ -29,7 +29,7 @@ table_tokenize Lexicon "abcd" --mode ADD
       "position": 4
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 5
     }
   ]
@@ -55,7 +55,7 @@ table_tokenize Lexicon "abcd\\z" --mode GET
       "position": 2
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 4
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/end/one.expected (+2 -2)
===================================================================
--- test/command/suite/tokenizers/regexp/get/end/one.expected    2015-03-12 20:36:16 +0900 (d692b4a)
+++ test/command/suite/tokenizers/regexp/get/end/one.expected    2015-03-12 21:11:50 +0900 (acaf793)
@@ -17,10 +17,10 @@ table_tokenize Lexicon "x" --mode ADD
       "position": 1
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 2
     }
   ]
 ]
 table_tokenize Lexicon "x\\z" --mode GET
-[[0,0.0,0.0],[{"value":"x","position":0},{"value":"￯","position":1}]]
+[[0,0.0,0.0],[{"value":"x","position":0},{"value":"￰","position":1}]]

  Modified: test/command/suite/tokenizers/regexp/get/end/three.expected (+2 -2)
===================================================================
--- test/command/suite/tokenizers/regexp/get/end/three.expected    2015-03-12 20:36:16 +0900 (7759db6)
+++ test/command/suite/tokenizers/regexp/get/end/three.expected    2015-03-12 21:11:50 +0900 (d082e85)
@@ -25,7 +25,7 @@ table_tokenize Lexicon "xyz" --mode ADD
       "position": 3
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 4
     }
   ]
@@ -47,7 +47,7 @@ table_tokenize Lexicon "xyz\\z" --mode GET
       "position": 1
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 3
     }
   ]

  Modified: test/command/suite/tokenizers/regexp/get/end/two.expected (+2 -2)
===================================================================
--- test/command/suite/tokenizers/regexp/get/end/two.expected    2015-03-12 20:36:16 +0900 (ccd7ce2)
+++ test/command/suite/tokenizers/regexp/get/end/two.expected    2015-03-12 21:11:50 +0900 (40ed20b)
@@ -21,10 +21,10 @@ table_tokenize Lexicon "xy" --mode ADD
       "position": 2
     },
     {
-      "value": "￯",
+      "value": "￰",
       "position": 3
     }
   ]
 ]
 table_tokenize Lexicon "xy\\z" --mode GET
-[[0,0.0,0.0],[{"value":"xy","position":0},{"value":"￯","position":2}]]
+[[0,0.0,0.0],[{"value":"xy","position":0},{"value":"￰","position":2}]]
-------------- next part --------------
HTML����������������������������...
Download 



More information about the Groonga-commit mailing list
Back to archive index