fix: hide nodelay Kafka client socket option

also fix a few descriptions
This commit is contained in:
Zaiming (Stone) Shi 2023-03-29 11:39:18 +02:00
parent 11c120f87c
commit 624c168dde
2 changed files with 26 additions and 21 deletions

View File

@ -216,6 +216,7 @@ emqx_ee_bridge_kafka {
zh: "Socket 收包缓存大小" zh: "Socket 收包缓存大小"
} }
} }
# hidden
socket_nodelay { socket_nodelay {
desc { desc {
en: "When set to 'true', TCP buffer is sent as soon as possible. " en: "When set to 'true', TCP buffer is sent as soon as possible. "
@ -571,26 +572,27 @@ emqx_ee_bridge_kafka {
zh: "指定从哪个 Kafka 主题消费消息。" zh: "指定从哪个 Kafka 主题消费消息。"
} }
label { label {
en: "Kafka topic" en: "Kafka Topic"
zh: "Kafka 主题 " zh: "Kafka 主题"
} }
} }
consumer_max_batch_bytes { consumer_max_batch_bytes {
desc { desc {
en: "Maximum bytes to fetch in a batch of messages." en: "Set how many bytes to pull from Kafka in each fetch request. "
"Please note that if the configured value is smaller than the message size in Kafka, it may negatively impact the fetch performance." "Please note that if the configured value is smaller than the message size in Kafka, it may negatively impact the fetch performance."
zh: "在一批消息中要取的最大字节数。" zh: "设置每次从 Kafka 拉取数据的字节数。"
"如果该配置小于 Kafka 中消息到大小,则可能会影响消费性能。" "如果该配置小于 Kafka 中消息到大小,则可能会影响消费性能。"
} }
label { label {
en: "Max Bytes" en: "Fetch Bytes"
zh: "最大字节数" zh: "拉取字节数"
} }
} }
# hidden
consumer_max_rejoin_attempts { consumer_max_rejoin_attempts {
desc { desc {
en: "Maximum number of times allowed for a member to re-join the group. If the consumer group can not reach balance after this configured number of attempts, the consumer group member will restart after a delay." en: "Maximum number of times allowed for a member to re-join the group. If the consumer group can not reach balance after this configured number of attempts, the consumer group member will restart after a delay."
zh: "允许一个成员重新加入小组的最大次数。如果超过改配置次数后仍不能成功加入消费组,则会在延迟一段时间后再重试。" zh: "允许一个消费组成员重新加入小组的最大次数。如果超过改配置次数后仍不能成功加入消费组,则会在延迟一段时间后再重试。"
} }
label { label {
en: "Max Rejoin Attempts" en: "Max Rejoin Attempts"
@ -599,10 +601,9 @@ emqx_ee_bridge_kafka {
} }
consumer_offset_reset_policy { consumer_offset_reset_policy {
desc { desc {
en: "Defines from which offset a consumer should start fetching when there" en: "Defines from which offset a consumer should start fetching when there "
" is no commit history or when the commit history becomes invalid." "is no commit history or when the commit history becomes invalid."
zh: "当没有主题分区没有偏移量的历史记录,或则历史记录失效后," zh: "当没有偏移量的历史记录,或则历史记录失效后,消费者应该使用哪个偏移量重新开始消费。"
"消费者应该使用哪个偏移量重新开始消费"
} }
label { label {
en: "Offset Reset Policy" en: "Offset Reset Policy"
@ -616,13 +617,13 @@ emqx_ee_bridge_kafka {
} }
label { label {
en: "Offset Commit Interval" en: "Offset Commit Interval"
zh: "偏移承诺间隔" zh: "偏移提交间隔"
} }
} }
consumer_topic_mapping { consumer_topic_mapping {
desc { desc {
en: "Defines the mapping between Kafka topics and MQTT topics. Must contain at least one item." en: "Defines the mapping between Kafka topics and MQTT topics. Must contain at least one item."
zh: "指定 Kafka 主题和 MQTT 主题之间的映射。 必须至少包含一。" zh: "指定 Kafka 主题和 MQTT 主题之间的映射关系。 必须至少包含一项。"
} }
label { label {
en: "Topic Mapping" en: "Topic Mapping"
@ -632,9 +633,9 @@ emqx_ee_bridge_kafka {
consumer_key_encoding_mode { consumer_key_encoding_mode {
desc { desc {
en: "Defines how the key from the Kafka message is" en: "Defines how the key from the Kafka message is"
" dealt with before being forwarded via MQTT.\n" " encoded before being forwarded via MQTT.\n"
"<code>none</code> Uses the key from the Kafka message unchanged." "<code>none</code> Uses the key from the Kafka message unchanged."
" Note: in this case, then the key must be a valid UTF-8 string.\n" " Note: in this case, the key must be a valid UTF-8 string.\n"
"<code>base64</code> Uses base-64 encoding on the received key." "<code>base64</code> Uses base-64 encoding on the received key."
zh: "定义了在通过MQTT转发之前如何处理Kafka消息的 Key。" zh: "定义了在通过MQTT转发之前如何处理Kafka消息的 Key。"
"<code>none</code> 使用Kafka消息中的 Key 原始值,不进行编码。" "<code>none</code> 使用Kafka消息中的 Key 原始值,不进行编码。"
@ -649,9 +650,9 @@ emqx_ee_bridge_kafka {
consumer_value_encoding_mode { consumer_value_encoding_mode {
desc { desc {
en: "Defines how the value from the Kafka message is" en: "Defines how the value from the Kafka message is"
" dealt with before being forwarded via MQTT.\n" " encoded before being forwarded via MQTT.\n"
"<code>none</code> Uses the value from the Kafka message unchanged." "<code>none</code> Uses the value from the Kafka message unchanged."
" Note: in this case, then the value must be a valid UTF-8 string.\n" " Note: in this case, the value must be a valid UTF-8 string.\n"
"<code>base64</code> Uses base-64 encoding on the received value." "<code>base64</code> Uses base-64 encoding on the received value."
zh: "定义了在通过MQTT转发之前如何处理Kafka消息的 Value。" zh: "定义了在通过MQTT转发之前如何处理Kafka消息的 Value。"
"<code>none</code> 使用Kafka消息中的 Value 原始值,不进行编码。" "<code>none</code> 使用Kafka消息中的 Value 原始值,不进行编码。"

View File

@ -221,17 +221,21 @@ fields(socket_opts) ->
{sndbuf, {sndbuf,
mk( mk(
emqx_schema:bytesize(), emqx_schema:bytesize(),
#{default => <<"1024KB">>, desc => ?DESC(socket_send_buffer)} #{default => <<"1MB">>, desc => ?DESC(socket_send_buffer)}
)}, )},
{recbuf, {recbuf,
mk( mk(
emqx_schema:bytesize(), emqx_schema:bytesize(),
#{default => <<"1024KB">>, desc => ?DESC(socket_receive_buffer)} #{default => <<"1MB">>, desc => ?DESC(socket_receive_buffer)}
)}, )},
{nodelay, {nodelay,
mk( mk(
boolean(), boolean(),
#{default => true, desc => ?DESC(socket_nodelay)} #{
default => true,
hidden => true,
desc => ?DESC(socket_nodelay)
}
)} )}
]; ];
fields(producer_opts) -> fields(producer_opts) ->