aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorAndrey Serebryanskiy <serebryanskiy@ydb.tech>2025-06-02 15:26:30 +0300
committerGitHub <noreply@github.com>2025-06-02 15:26:30 +0300
commit399438989847cb85bfbd3b3fff289329fe79cb9a (patch)
tree1665f724f0a385394ced64310cb56219adec5d9f
parenta435e8cf166dfaaa493ebd81987713728e16a079 (diff)
downloadydb-399438989847cb85bfbd3b3fff289329fe79cb9a.tar.gz
[Kafka proxy] ProduceActor: fix asan test fails (#19150)
-rw-r--r--ydb/core/kafka_proxy/ut/ut_produce_actor.cpp11
1 files changed, 6 insertions, 5 deletions
diff --git a/ydb/core/kafka_proxy/ut/ut_produce_actor.cpp b/ydb/core/kafka_proxy/ut/ut_produce_actor.cpp
index 001bd519e47..7ab376de43d 100644
--- a/ydb/core/kafka_proxy/ut/ut_produce_actor.cpp
+++ b/ydb/core/kafka_proxy/ut/ut_produce_actor.cpp
@@ -48,6 +48,9 @@ namespace {
TActorId ActorId;
const TString Database = "/Root/PQ";
const TString TopicName = "topic"; // as specified in pq_ut_common
+ const NKikimrConfig::TKafkaProxyConfig KafkaConfig = {};
+ const TString KeyToProduce = "record-key";
+ const TString ValueToProduce = "record-value";
TString TransactionalId = "123";
void SetUp(NUnitTest::TTestContext&) override {
@@ -59,7 +62,7 @@ namespace {
Ctx->Runtime->DisableBreakOnStopCondition();
Ctx->Runtime->SetLogPriority(NKikimrServices::KAFKA_PROXY, NLog::PRI_TRACE);
Ctx->Runtime->SetLogPriority(NKikimrServices::PQ_WRITE_PROXY, NLog::PRI_TRACE);
- TContext::TPtr kafkaContext = std::make_shared<TContext>(NKikimrConfig::TKafkaProxyConfig());
+ TContext::TPtr kafkaContext = std::make_shared<TContext>(KafkaConfig);
kafkaContext->DatabasePath = "/Root/PQ";
ActorId = Ctx->Runtime->Register(CreateKafkaProduceActor(kafkaContext));
auto dummySchemeCacheId = Ctx->Runtime->Register(new TDummySchemeCacheActor(Ctx->TabletId));
@@ -83,15 +86,13 @@ namespace {
records->ProducerId = producerId;
records->ProducerEpoch = producerEpoch;
- TString key = "record-key";
- TString value = "record-value";
TKafkaRecordBatch batch;
records->BaseOffset = 3;
records->BaseSequence = 5;
records->Magic = 2; // Current supported
records->Records.resize(1);
- records->Records[0].Key = TKafkaRawBytes(key.data(), key.size());
- records->Records[0].Value = TKafkaRawBytes(value.data(), value.size());
+ records->Records[0].Key = TKafkaRawBytes(KeyToProduce.data(), KeyToProduce.size());
+ records->Records[0].Value = TKafkaRawBytes(ValueToProduce.data(), ValueToProduce.size());
partitionData.Records = records;
topicData.PartitionData.push_back(partitionData);