Kouhei Sutou
null+****@clear*****
Thu Mar 12 21:11:50 JST 2015
Kouhei Sutou 2015-03-12 21:11:50 +0900 (Thu, 12 Mar 2015) New Revision: 34aacd121be08147e7f1a90fdb545cafac921ec2 https://github.com/groonga/groonga/commit/34aacd121be08147e7f1a90fdb545cafac921ec2 Message: TokenRegexp: fix wrong end mark character... Modified files: lib/tokenizers.c test/command/suite/tokenizers/regexp/add/four.expected test/command/suite/tokenizers/regexp/add/one.expected test/command/suite/tokenizers/regexp/add/three.expected test/command/suite/tokenizers/regexp/add/two.expected test/command/suite/tokenizers/regexp/get/begin/one.expected test/command/suite/tokenizers/regexp/get/begin/three.expected test/command/suite/tokenizers/regexp/get/begin/two.expected test/command/suite/tokenizers/regexp/get/end/four.expected test/command/suite/tokenizers/regexp/get/end/one.expected test/command/suite/tokenizers/regexp/get/end/three.expected test/command/suite/tokenizers/regexp/get/end/two.expected Modified: lib/tokenizers.c (+4 -4) =================================================================== --- lib/tokenizers.c 2015-03-12 20:36:16 +0900 (796c02d) +++ lib/tokenizers.c 2015-03-12 21:11:50 +0900 (d5b3db9) @@ -586,8 +586,8 @@ regexp_next(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data) status |= GRN_TOKEN_LAST | GRN_TOKEN_REACH_END; grn_tokenizer_token_push(ctx, &(tokenizer->token), - GRN_TOKENIZER_BEGIN_MARK_UTF8, - GRN_TOKENIZER_BEGIN_MARK_UTF8_LEN, + GRN_TOKENIZER_END_MARK_UTF8, + GRN_TOKENIZER_END_MARK_UTF8_LEN, status); return NULL; } @@ -606,8 +606,8 @@ regexp_next(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data) status |= GRN_TOKEN_LAST | GRN_TOKEN_REACH_END; grn_tokenizer_token_push(ctx, &(tokenizer->token), - GRN_TOKENIZER_BEGIN_MARK_UTF8, - GRN_TOKENIZER_BEGIN_MARK_UTF8_LEN, + GRN_TOKENIZER_END_MARK_UTF8, + GRN_TOKENIZER_END_MARK_UTF8_LEN, status); return NULL; } Modified: test/command/suite/tokenizers/regexp/add/four.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/add/four.expected 2015-03-12 20:36:16 +0900 (ae80e56) +++ test/command/suite/tokenizers/regexp/add/four.expected 2015-03-12 21:11:50 +0900 (9942df3) @@ -27,7 +27,7 @@ tokenize TokenRegexp "abcd" "position": 4 }, { - "value": "", + "value": "", "position": 5 } ] Modified: test/command/suite/tokenizers/regexp/add/one.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/add/one.expected 2015-03-12 20:36:16 +0900 (6c2e513) +++ test/command/suite/tokenizers/regexp/add/one.expected 2015-03-12 21:11:50 +0900 (f53d7bb) @@ -15,7 +15,7 @@ tokenize TokenRegexp "x" "position": 1 }, { - "value": "", + "value": "", "position": 2 } ] Modified: test/command/suite/tokenizers/regexp/add/three.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/add/three.expected 2015-03-12 20:36:16 +0900 (e865c8c) +++ test/command/suite/tokenizers/regexp/add/three.expected 2015-03-12 21:11:50 +0900 (cbcc334) @@ -23,7 +23,7 @@ tokenize TokenRegexp "xyz" "position": 3 }, { - "value": "", + "value": "", "position": 4 } ] Modified: test/command/suite/tokenizers/regexp/add/two.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/add/two.expected 2015-03-12 20:36:16 +0900 (27eef72) +++ test/command/suite/tokenizers/regexp/add/two.expected 2015-03-12 21:11:50 +0900 (4907a52) @@ -19,7 +19,7 @@ tokenize TokenRegexp "xy" "position": 2 }, { - "value": "", + "value": "", "position": 3 } ] Modified: test/command/suite/tokenizers/regexp/get/begin/one.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/get/begin/one.expected 2015-03-12 20:36:16 +0900 (254c754) +++ test/command/suite/tokenizers/regexp/get/begin/one.expected 2015-03-12 21:11:50 +0900 (8c9747a) @@ -17,7 +17,7 @@ table_tokenize Lexicon "x" --mode ADD "position": 1 }, { - "value": "", + "value": "", "position": 2 } ] Modified: test/command/suite/tokenizers/regexp/get/begin/three.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/get/begin/three.expected 2015-03-12 20:36:16 +0900 (49bc507) +++ test/command/suite/tokenizers/regexp/get/begin/three.expected 2015-03-12 21:11:50 +0900 (878c093) @@ -25,7 +25,7 @@ table_tokenize Lexicon "xyz" --mode ADD "position": 3 }, { - "value": "", + "value": "", "position": 4 } ] Modified: test/command/suite/tokenizers/regexp/get/begin/two.expected (+1 -1) =================================================================== --- test/command/suite/tokenizers/regexp/get/begin/two.expected 2015-03-12 20:36:16 +0900 (2808762) +++ test/command/suite/tokenizers/regexp/get/begin/two.expected 2015-03-12 21:11:50 +0900 (8e82fb0) @@ -21,7 +21,7 @@ table_tokenize Lexicon "xy" --mode ADD "position": 2 }, { - "value": "", + "value": "", "position": 3 } ] Modified: test/command/suite/tokenizers/regexp/get/end/four.expected (+2 -2) =================================================================== --- test/command/suite/tokenizers/regexp/get/end/four.expected 2015-03-12 20:36:16 +0900 (af115be) +++ test/command/suite/tokenizers/regexp/get/end/four.expected 2015-03-12 21:11:50 +0900 (b621183) @@ -29,7 +29,7 @@ table_tokenize Lexicon "abcd" --mode ADD "position": 4 }, { - "value": "", + "value": "", "position": 5 } ] @@ -55,7 +55,7 @@ table_tokenize Lexicon "abcd\\z" --mode GET "position": 2 }, { - "value": "", + "value": "", "position": 4 } ] Modified: test/command/suite/tokenizers/regexp/get/end/one.expected (+2 -2) =================================================================== --- test/command/suite/tokenizers/regexp/get/end/one.expected 2015-03-12 20:36:16 +0900 (d692b4a) +++ test/command/suite/tokenizers/regexp/get/end/one.expected 2015-03-12 21:11:50 +0900 (acaf793) @@ -17,10 +17,10 @@ table_tokenize Lexicon "x" --mode ADD "position": 1 }, { - "value": "", + "value": "", "position": 2 } ] ] table_tokenize Lexicon "x\\z" --mode GET -[[0,0.0,0.0],[{"value":"x","position":0},{"value":"","position":1}]] +[[0,0.0,0.0],[{"value":"x","position":0},{"value":"","position":1}]] Modified: test/command/suite/tokenizers/regexp/get/end/three.expected (+2 -2) =================================================================== --- test/command/suite/tokenizers/regexp/get/end/three.expected 2015-03-12 20:36:16 +0900 (7759db6) +++ test/command/suite/tokenizers/regexp/get/end/three.expected 2015-03-12 21:11:50 +0900 (d082e85) @@ -25,7 +25,7 @@ table_tokenize Lexicon "xyz" --mode ADD "position": 3 }, { - "value": "", + "value": "", "position": 4 } ] @@ -47,7 +47,7 @@ table_tokenize Lexicon "xyz\\z" --mode GET "position": 1 }, { - "value": "", + "value": "", "position": 3 } ] Modified: test/command/suite/tokenizers/regexp/get/end/two.expected (+2 -2) =================================================================== --- test/command/suite/tokenizers/regexp/get/end/two.expected 2015-03-12 20:36:16 +0900 (ccd7ce2) +++ test/command/suite/tokenizers/regexp/get/end/two.expected 2015-03-12 21:11:50 +0900 (40ed20b) @@ -21,10 +21,10 @@ table_tokenize Lexicon "xy" --mode ADD "position": 2 }, { - "value": "", + "value": "", "position": 3 } ] ] table_tokenize Lexicon "xy\\z" --mode GET -[[0,0.0,0.0],[{"value":"xy","position":0},{"value":"","position":2}]] +[[0,0.0,0.0],[{"value":"xy","position":0},{"value":"","position":2}]] -------------- next part -------------- HTML����������������������������...Download