diff options
author | tobo <tobo@yandex-team.ru> | 2022-02-10 16:47:27 +0300 |
---|---|---|
committer | Daniil Cherednik <dcherednik@yandex-team.ru> | 2022-02-10 16:47:27 +0300 |
commit | 7fe839092527589b38f014d854c51565b3c1adfa (patch) | |
tree | 309e97022d3530044b712b8f71318c78faf7856e /util/stream/tokenizer_ut.cpp | |
parent | d0d68c395c10da4cb56a1c845504570a04d7893e (diff) | |
download | ydb-7fe839092527589b38f014d854c51565b3c1adfa.tar.gz |
Restoring authorship annotation for <tobo@yandex-team.ru>. Commit 1 of 2.
Diffstat (limited to 'util/stream/tokenizer_ut.cpp')
-rw-r--r-- | util/stream/tokenizer_ut.cpp | 18 |
1 files changed, 9 insertions, 9 deletions
diff --git a/util/stream/tokenizer_ut.cpp b/util/stream/tokenizer_ut.cpp index afc566da86..37b30d8e5f 100644 --- a/util/stream/tokenizer_ut.cpp +++ b/util/stream/tokenizer_ut.cpp @@ -40,7 +40,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(LastTokenendDoesntSatisfyPredicateTest) { const char data[] = "abc\ndef\nxxxxxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; + const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input}; @@ -58,7 +58,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(FirstTokenIsEmptyTest) { const char data[] = "\ndef\nxxxxxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf(), TStringBuf("def"), TStringBuf("xxxxxx")}; + const TStringBuf tokens[] = {TStringBuf(), TStringBuf("def"), TStringBuf("xxxxxx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input}; @@ -91,7 +91,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(SimpleTest) { const char data[] = "qwerty\n1234567890\n"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("qwerty"), TStringBuf("1234567890")}; + const TStringBuf tokens[] = {TStringBuf("qwerty"), TStringBuf("1234567890")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input}; @@ -115,7 +115,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { const char data[] = "abc|def|xxxxxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; + const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TIsVerticalBar>{&input}; @@ -139,8 +139,8 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { const char data[] = "abc|def|xxxxxx,abc|def|xxxxxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx"), - TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; + const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx"), + TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TIsVerticalBar>{&input}; @@ -199,7 +199,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(FirstTokenHasSizeOfTheBufferTest) { const char data[] = "xxxxx\nxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("xxxxx"), TStringBuf("xx")}; + const TStringBuf tokens[] = {TStringBuf("xxxxx"), TStringBuf("xx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input, TEol{}, tokens[0].size()}; @@ -231,7 +231,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(BufferSizeInitialSizeSmallerThanTokenTest) { const char data[] = "xxxxx\nxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("xxxxx"), TStringBuf("xx")}; + const TStringBuf tokens[] = {TStringBuf("xxxxx"), TStringBuf("xx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input, TEol{}, 1}; @@ -248,7 +248,7 @@ Y_UNIT_TEST_SUITE(TStreamTokenizerTests) { Y_UNIT_TEST(RangeBasedForTest) { const char data[] = "abc\ndef\nxxxxxx"; const auto dataSize = Y_ARRAY_SIZE(data) - 1; - const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; + const TStringBuf tokens[] = {TStringBuf("abc"), TStringBuf("def"), TStringBuf("xxxxxx")}; const auto tokensSize = Y_ARRAY_SIZE(tokens); auto&& input = TMemoryInput{data, dataSize}; auto&& tokenizer = TStreamTokenizer<TEol>{&input}; |