Merge pull request #8947 from zmstone/0913-sync-master-to-ee50
0913 sync master to ee50
This commit is contained in:
commit
477d4b0b03
|
@ -185,9 +185,19 @@ jobs:
|
|||
img_suffix="elixir-${{ matrix.arch }}"
|
||||
img_labels="org.opencontainers.image.elixir.version=${{ matrix.elixir }}\n${img_labels}"
|
||||
fi
|
||||
|
||||
if [ ${{ matrix.profile }} = "emqx" ]; then
|
||||
img_labels="org.opencontainers.image.edition=Opensource\n${img_labels}"
|
||||
fi
|
||||
|
||||
if [ ${{ matrix.profile }} = "emqx-enterprise" ]; then
|
||||
img_labels="org.opencontainers.image.edition=Enterprise\n${img_labels}"
|
||||
fi
|
||||
|
||||
if [[ ${{ matrix.os[0] }} =~ "alpine" ]]; then
|
||||
img_suffix="${img_suffix}-alpine"
|
||||
fi
|
||||
|
||||
echo "::set-output name=emqx_name::${emqx_name}"
|
||||
echo "::set-output name=img_suffix::${img_suffix}"
|
||||
echo "::set-output name=img_labels::${img_labels}"
|
||||
|
@ -303,10 +313,19 @@ jobs:
|
|||
img_suffix="elixir-${{ matrix.arch }}"
|
||||
img_labels="org.opencontainers.image.elixir.version=${{ matrix.elixir }}\n$img_labels"
|
||||
fi
|
||||
|
||||
if [ ${{ matrix.profile }} = "emqx" ]; then
|
||||
img_labels="org.opencontainers.image.edition=Opensource\n${img_labels}"
|
||||
fi
|
||||
|
||||
if [ ${{ matrix.profile }} = "emqx-enterprise" ]; then
|
||||
img_labels="org.opencontainers.image.edition=Enterprise\n${img_labels}"
|
||||
fi
|
||||
|
||||
if [[ ${{ matrix.os[0] }} =~ "alpine" ]]; then
|
||||
img_suffix="${img_suffix}-alpine"
|
||||
fi
|
||||
echo "::set-output name=img::${img}"
|
||||
|
||||
echo "::set-output name=emqx_name::${emqx_name}"
|
||||
echo "::set-output name=img_suffix::${img_suffix}"
|
||||
echo "::set-output name=img_labels::${img_labels}"
|
||||
|
|
|
@ -80,7 +80,9 @@ jobs:
|
|||
- uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: "${{ matrix.profile }}_schema_dump"
|
||||
path: _build/*/lib/emqx_dashboard/priv/www/static/schema.json
|
||||
path: |
|
||||
scripts/spellcheck
|
||||
_build/${{ matrix.profile }}/lib/emqx_dashboard/priv/www/static/schema.json
|
||||
|
||||
windows:
|
||||
runs-on: windows-2019
|
||||
|
@ -205,7 +207,6 @@ jobs:
|
|||
- emqx
|
||||
- emqx-enterprise
|
||||
runs-on: aws-amd64
|
||||
container: "ghcr.io/emqx/emqx-schema-validate:0.3.5"
|
||||
steps:
|
||||
- uses: actions/download-artifact@v2
|
||||
name: Download schema dump
|
||||
|
@ -214,9 +215,7 @@ jobs:
|
|||
path: /tmp/
|
||||
- name: Run spellcheck
|
||||
run: |
|
||||
cd /LanguageTool
|
||||
bash start.sh > /dev/null &
|
||||
./emqx_schema_validate /tmp/${{ matrix.profile }}/lib/emqx_dashboard/priv/www/static/schema.json
|
||||
bash /tmp/scripts/spellcheck/spellcheck.sh /tmp/_build/${{ matrix.profile }}/lib/emqx_dashboard/priv/www/static/schema.json
|
||||
|
||||
allgood_packaging:
|
||||
runs-on: ubuntu-latest
|
||||
|
|
|
@ -6,12 +6,18 @@
|
|||
* Fix JWT plugin don't support non-integer timestamp claims. [#8867](https://github.com/emqx/emqx/pull/8867)
|
||||
* Avoid publishing will message when client fails to auhtenticate. [#8887](https://github.com/emqx/emqx/pull/8887)
|
||||
* Speed up dispatching of shared subscription messages in a cluster [#8893](https://github.com/emqx/emqx/pull/8893)
|
||||
* Fix the extra / prefix when CoAP gateway parsing client topics. [#8658](https://github.com/emqx/emqx/pull/8658)
|
||||
* Speed up updating the configuration, When some nodes in the cluster are down. [#8857](https://github.com/emqx/emqx/pull/8857)
|
||||
* Fix delayed publish inaccurate caused by os time change. [#8926](https://github.com/emqx/emqx/pull/8926)
|
||||
* Fix that EMQX can't start when the retainer is disabled [#8911](https://github.com/emqx/emqx/pull/8911)
|
||||
|
||||
## Enhancements
|
||||
|
||||
* Print a warning message when boot with the default (insecure) Erlang cookie. [#8905](https://github.com/emqx/emqx/pull/8905)
|
||||
* Change the `/gateway` API path to plural form. [#8823](https://github.com/emqx/emqx/pull/8823)
|
||||
* Remove `node.etc_dir` from emqx.conf, because it is never used.
|
||||
Also allow user to customize the logging directory [#8892](https://github.com/emqx/emqx/pull/8892)
|
||||
* Added a new API `POST /listeners` for creating listener. [#8876](https://github.com/emqx/emqx/pull/8876)
|
||||
|
||||
# 5.0.7
|
||||
|
||||
|
|
87
README-RU.md
87
README-RU.md
|
@ -88,9 +88,9 @@ docker run -d --name emqx-ee -p 1883:1883 -p 8081:8081 -p 8083:8083 -p 8084:8084
|
|||
|
||||
## Сборка из исходного кода
|
||||
|
||||
Начиная с релиза 3.0, для сборки требуется Erlang/OTP R21 или выше.
|
||||
Ветка `master` предназначена для последней версии 5, переключитесь на ветку `main-v4.3` для версии 4.3 и `main-v4.4` для версии 4.4.
|
||||
|
||||
Инструкция для сборки версии 4.3 и выше:
|
||||
EMQX требует OTP 22 или 23 для версии 4.3 и OTP 24 для версий 4.4 и 5.0.
|
||||
|
||||
```bash
|
||||
git clone https://github.com/emqx/emqx.git
|
||||
|
@ -99,7 +99,7 @@ make
|
|||
_build/emqx/rel/emqx/bin/emqx console
|
||||
```
|
||||
|
||||
Более ранние релизы могут быть собраны с помощью другого репозитория:
|
||||
Версии до 4.2 (включительно) нужно собирать из другого репозитория:
|
||||
|
||||
```bash
|
||||
git clone https://github.com/emqx/emqx-rel.git
|
||||
|
@ -108,79 +108,24 @@ make
|
|||
_build/emqx/rel/emqx/bin/emqx console
|
||||
```
|
||||
|
||||
## Первый запуск
|
||||
### Сборка на Apple silicon (M1, M2)
|
||||
|
||||
Если emqx был собран из исходников: `cd _build/emqx/rel/emqx`.
|
||||
Или перейдите в директорию, куда emqx был установлен из бинарного пакета.
|
||||
Пакетный менеджер Homebrew, когда установлен на Apple silicon, [стал использовать другую домашнюю папку по умолчанию](https://github.com/Homebrew/brew/issues/9177), `/opt/homebrew` вместо `/usr/local`. В результате некоторые библиотеки перестали собираться автоматически.
|
||||
|
||||
Касательно EMQX, сборка Erlang из исходного кода не найдёт библиотеку `unixodbc`, установленную с homebrew, без дополнительных действий:
|
||||
|
||||
```bash
|
||||
# Запуск:
|
||||
./bin/emqx start
|
||||
|
||||
# Проверка статуса:
|
||||
./bin/emqx_ctl status
|
||||
|
||||
# Остановка:
|
||||
./bin/emqx stop
|
||||
brew install unixodbc kerl
|
||||
sudo ln -s $(realpath $(brew --prefix unixodbc)) /usr/local/odbc
|
||||
export CC="/usr/bin/gcc -I$(brew --prefix unixodbc)/include"
|
||||
export LDFLAGS="-L$(brew --prefix unixodbc)/lib"
|
||||
kerl build 24.3
|
||||
mkdir ~/.kerl/installations
|
||||
kerl install 24.3 ~/.kerl/installations/24.3
|
||||
. ~/.kerl/installations/24.3/activate
|
||||
```
|
||||
|
||||
Веб-интерфейс брокера будет доступен по ссылке: http://localhost:18083
|
||||
|
||||
## Тесты
|
||||
|
||||
### Полное тестирование
|
||||
|
||||
```
|
||||
make eunit ct
|
||||
```
|
||||
|
||||
### Запуск части тестов
|
||||
|
||||
Пример:
|
||||
|
||||
```bash
|
||||
make apps/emqx_retainer-ct
|
||||
```
|
||||
|
||||
### Dialyzer
|
||||
##### Статический анализ всех приложений
|
||||
```
|
||||
make dialyzer
|
||||
```
|
||||
|
||||
##### Статический анализ части приложений (список через запятую)
|
||||
```
|
||||
DIALYZER_ANALYSE_APP=emqx_lwm2m,emqx_authz make dialyzer
|
||||
```
|
||||
|
||||
## Сообщество
|
||||
|
||||
### FAQ
|
||||
|
||||
Наиболее частые проблемы разобраны в [EMQX FAQ](https://www.emqx.io/docs/en/latest/faq/faq.html).
|
||||
|
||||
|
||||
### Вопросы
|
||||
|
||||
Задать вопрос или поделиться идеей можно в [GitHub Discussions](https://github.com/emqx/emqx/discussions).
|
||||
|
||||
### Предложения
|
||||
|
||||
Более масштабные предложения можно присылать в виде pull request в репозиторий [EIP](https://github.com/emqx/eip).
|
||||
|
||||
### Разработка плагинов
|
||||
|
||||
Инструкция по разработке собственных плагинов доступна по ссылке: [PLUGIN.md](./PLUGIN.md)
|
||||
|
||||
## Спецификации стандарта MQTT
|
||||
|
||||
Следующие ссылки содержат спецификации стандартов:
|
||||
|
||||
[MQTT Version 3.1.1](https://docs.oasis-open.org/mqtt/mqtt/v3.1.1/os/mqtt-v3.1.1-os.html)
|
||||
|
||||
[MQTT Version 5.0](https://docs.oasis-open.org/mqtt/mqtt/v5.0/cs02/mqtt-v5.0-cs02.html)
|
||||
|
||||
[MQTT SN](https://www.oasis-open.org/committees/download.php/66091/MQTT-SN_spec_v1.2.pdf)
|
||||
Дальше можно собирать emqx как обычно, с помощью `make`.
|
||||
|
||||
## Лицензия
|
||||
|
||||
|
|
|
@ -361,7 +361,12 @@ handle_response(Headers, Body) ->
|
|||
_ ->
|
||||
ignore
|
||||
end;
|
||||
{error, _Reason} ->
|
||||
{error, Reason} ->
|
||||
?TRACE_AUTHN_PROVIDER(
|
||||
error,
|
||||
"parse_http_response_failed",
|
||||
#{content_type => ContentType, body => Body, reason => Reason}
|
||||
),
|
||||
ignore
|
||||
end.
|
||||
|
||||
|
|
|
@ -399,7 +399,7 @@ do_verify(JWT, [JWK | More], VerifyClaims) ->
|
|||
end.
|
||||
|
||||
verify_claims(Claims, VerifyClaims0) ->
|
||||
Now = os:system_time(seconds),
|
||||
Now = erlang:system_time(seconds),
|
||||
VerifyClaims =
|
||||
[
|
||||
{<<"exp">>, fun(ExpireTime) ->
|
||||
|
|
|
@ -149,8 +149,8 @@ authenticate(
|
|||
of
|
||||
ok ->
|
||||
{ok, emqx_authn_utils:is_superuser(Selected)};
|
||||
{error, Reason} ->
|
||||
{error, Reason}
|
||||
{error, _Reason} ->
|
||||
ignore
|
||||
end;
|
||||
{error, Reason} ->
|
||||
?TRACE_AUTHN_PROVIDER(error, "redis_query_failed", #{
|
||||
|
|
|
@ -173,6 +173,9 @@ test_user_auth(#{
|
|||
{create_authenticator, ?GLOBAL, AuthConfig}
|
||||
),
|
||||
|
||||
{ok, [#{provider := emqx_authn_redis, state := State}]} =
|
||||
emqx_authentication:list_authenticators(?GLOBAL),
|
||||
|
||||
Credentials = Credentials0#{
|
||||
listener => 'tcp:default',
|
||||
protocol => mqtt
|
||||
|
@ -180,6 +183,15 @@ test_user_auth(#{
|
|||
|
||||
?assertEqual(Result, emqx_access_control:authenticate(Credentials)),
|
||||
|
||||
AuthnResult =
|
||||
case Result of
|
||||
{error, _} ->
|
||||
ignore;
|
||||
Any ->
|
||||
Any
|
||||
end,
|
||||
?assertEqual(AuthnResult, emqx_authn_redis:authenticate(Credentials, State)),
|
||||
|
||||
emqx_authn_test_lib:delete_authenticators(
|
||||
[authentication],
|
||||
?GLOBAL
|
||||
|
@ -466,7 +478,7 @@ user_seeds() ->
|
|||
<<"cmd">> => <<"HMGET mqtt_user:${username} password_hash salt is_superuser">>,
|
||||
<<"password_hash_algorithm">> => #{<<"name">> => <<"bcrypt">>}
|
||||
},
|
||||
result => {error, bad_username_or_password}
|
||||
result => {error, not_authorized}
|
||||
},
|
||||
|
||||
#{
|
||||
|
|
|
@ -10,7 +10,7 @@
|
|||
|
||||
node {
|
||||
name = "emqx@127.0.0.1"
|
||||
cookie = emqxsecretcookie
|
||||
cookie = "{{ emqx_default_erlang_cookie }}"
|
||||
data_dir = "{{ platform_data_dir }}"
|
||||
}
|
||||
|
||||
|
|
|
@ -72,6 +72,7 @@
|
|||
-define(TIMEOUT, timer:minutes(1)).
|
||||
-define(APPLY_KIND_REPLICATE, replicate).
|
||||
-define(APPLY_KIND_INITIATE, initiate).
|
||||
-define(IS_STATUS(_A_), (_A_ =:= peers_lagging orelse _A_ =:= stopped_nodes)).
|
||||
|
||||
-type tnx_id() :: pos_integer().
|
||||
|
||||
|
@ -123,13 +124,13 @@ start_link(Node, Name, RetryMs) ->
|
|||
%% the result is expected to be `ok | {ok, _}' to indicate success,
|
||||
%% and `{error, _}' to indicate failure.
|
||||
%%
|
||||
%% The excpetion of the MFA evaluation is captured and translated
|
||||
%% The exception of the MFA evaluation is captured and translated
|
||||
%% into an `{error, _}' tuple.
|
||||
%% This call tries to wait for all peer nodes to be in-sync before
|
||||
%% returning the result.
|
||||
%%
|
||||
%% In case of partial success, an `error' level log is emitted
|
||||
%% but the initial localy apply result is returned.
|
||||
%% but the initial local apply result is returned.
|
||||
-spec multicall(module(), atom(), list()) -> term().
|
||||
multicall(M, F, A) ->
|
||||
multicall(M, F, A, all, timer:minutes(2)).
|
||||
|
@ -141,11 +142,12 @@ multicall(M, F, A, RequiredSyncs, Timeout) when RequiredSyncs =:= all orelse Req
|
|||
Result;
|
||||
{init_failure, Error} ->
|
||||
Error;
|
||||
{peers_lagging, TnxId, Res, Nodes} ->
|
||||
{Status, TnxId, Res, Nodes} when ?IS_STATUS(Status) ->
|
||||
%% The init MFA return ok, but some other nodes failed.
|
||||
?SLOG(error, #{
|
||||
msg => "cluster_rpc_peers_lagging",
|
||||
lagging_nodes => Nodes,
|
||||
status => Status,
|
||||
nodes => Nodes,
|
||||
tnx_id => TnxId
|
||||
}),
|
||||
Res
|
||||
|
@ -193,9 +195,9 @@ do_multicall(M, F, A, RequiredSyncs, Timeout) ->
|
|||
InitRes;
|
||||
{init_failure, Error0} ->
|
||||
{init_failure, Error0};
|
||||
{peers_lagging, Nodes} ->
|
||||
{Status, Nodes} when ?IS_STATUS(Status) ->
|
||||
{ok, TnxId0, MFARes} = InitRes,
|
||||
{peers_lagging, TnxId0, MFARes, Nodes}
|
||||
{Status, TnxId0, MFARes, Nodes}
|
||||
end.
|
||||
|
||||
-spec query(pos_integer()) -> {'atomic', map()} | {'aborted', Reason :: term()}.
|
||||
|
@ -509,14 +511,18 @@ do_alarm(Fun, Res, #{tnx_id := Id} = Meta) ->
|
|||
emqx_alarm:Fun(cluster_rpc_apply_failed, Meta#{result => ?TO_BIN(Res)}, AlarmMsg).
|
||||
|
||||
wait_for_all_nodes_commit(TnxId, Delay, Remain) ->
|
||||
case lagging_node(TnxId) of
|
||||
Lagging = lagging_nodes(TnxId),
|
||||
Stopped = stopped_nodes(),
|
||||
case Lagging -- Stopped of
|
||||
[] when Stopped =:= [] ->
|
||||
ok;
|
||||
[] ->
|
||||
{stopped_nodes, Stopped};
|
||||
[_ | _] when Remain > 0 ->
|
||||
ok = timer:sleep(Delay),
|
||||
wait_for_all_nodes_commit(TnxId, Delay, Remain - Delay);
|
||||
[] ->
|
||||
ok;
|
||||
Nodes ->
|
||||
{peers_lagging, Nodes}
|
||||
[_ | _] ->
|
||||
{peers_lagging, Lagging}
|
||||
end.
|
||||
|
||||
wait_for_nodes_commit(RequiredSyncs, TnxId, Delay, Remain) ->
|
||||
|
@ -527,14 +533,18 @@ wait_for_nodes_commit(RequiredSyncs, TnxId, Delay, Remain) ->
|
|||
false when Remain > 0 ->
|
||||
wait_for_nodes_commit(RequiredSyncs, TnxId, Delay, Remain - Delay);
|
||||
false ->
|
||||
case lagging_node(TnxId) of
|
||||
%% All commit but The succeedNum > length(nodes()).
|
||||
[] -> ok;
|
||||
Nodes -> {peers_lagging, Nodes}
|
||||
case lagging_nodes(TnxId) of
|
||||
[] ->
|
||||
ok;
|
||||
Lagging ->
|
||||
case stopped_nodes() of
|
||||
[] -> {peers_lagging, Lagging};
|
||||
Stopped -> {stopped_nodes, Stopped}
|
||||
end
|
||||
end
|
||||
end.
|
||||
|
||||
lagging_node(TnxId) ->
|
||||
lagging_nodes(TnxId) ->
|
||||
{atomic, Nodes} = transaction(fun ?MODULE:commit_status_trans/2, ['<', TnxId]),
|
||||
Nodes.
|
||||
|
||||
|
@ -548,6 +558,9 @@ commit_status_trans(Operator, TnxId) ->
|
|||
Result = '$2',
|
||||
mnesia:select(?CLUSTER_COMMIT, [{MatchHead, [Guard], [Result]}]).
|
||||
|
||||
stopped_nodes() ->
|
||||
ekka_cluster:info(stopped_nodes).
|
||||
|
||||
get_retry_ms() ->
|
||||
emqx_conf:get([node, cluster_call, retry_interval], timer:minutes(1)).
|
||||
|
||||
|
|
|
@ -399,7 +399,7 @@ fields("node") ->
|
|||
string(),
|
||||
#{
|
||||
mapping => "vm_args.-setcookie",
|
||||
default => "emqxsecretcookie",
|
||||
required => true,
|
||||
'readOnly' => true,
|
||||
sensitive => true,
|
||||
desc => ?DESC(node_cookie)
|
||||
|
|
|
@ -785,7 +785,7 @@ to_bin(List) when is_list(List) ->
|
|||
to_bin(Boolean) when is_boolean(Boolean) -> Boolean;
|
||||
to_bin(Atom) when is_atom(Atom) -> atom_to_binary(Atom, utf8);
|
||||
to_bin({Type, Args}) ->
|
||||
unicode:characters_to_binary(io_lib:format("~p(~p)", [Type, Args]));
|
||||
unicode:characters_to_binary(io_lib:format("~ts(~p)", [Type, Args]));
|
||||
to_bin(X) ->
|
||||
X.
|
||||
|
||||
|
|
|
@ -153,7 +153,7 @@ init(
|
|||
mountpoint => Mountpoint
|
||||
}
|
||||
),
|
||||
|
||||
%% FIXME: it should coap.hearbeat instead of idle_timeout?
|
||||
Heartbeat = ?GET_IDLE_TIME(Config),
|
||||
#channel{
|
||||
ctx = Ctx,
|
||||
|
@ -447,6 +447,7 @@ enrich_conninfo(
|
|||
conninfo = ConnInfo
|
||||
}
|
||||
) ->
|
||||
%% FIXME: generate a random clientid if absent
|
||||
case Queries of
|
||||
#{<<"clientid">> := ClientId} ->
|
||||
Interval = maps:get(interval, emqx_keepalive:info(KeepAlive)),
|
||||
|
@ -467,6 +468,9 @@ enrich_clientinfo(
|
|||
{Queries, Msg},
|
||||
Channel = #channel{clientinfo = ClientInfo0}
|
||||
) ->
|
||||
%% FIXME:
|
||||
%% 1. generate a random clientid if absent;
|
||||
%% 2. assgin username, password to `undefined` if absent
|
||||
case Queries of
|
||||
#{
|
||||
<<"username">> := UserName,
|
||||
|
@ -542,6 +546,7 @@ process_connect(
|
|||
)
|
||||
of
|
||||
{ok, _Sess} ->
|
||||
%% FIXME: Token in cluster wide?
|
||||
RandVal = rand:uniform(?TOKEN_MAXIMUM),
|
||||
Token = erlang:list_to_binary(erlang:integer_to_list(RandVal)),
|
||||
NResult = Result#{events => [{event, connected}]},
|
||||
|
|
|
@ -69,17 +69,7 @@ handle_method(_, _, Msg, _, _) ->
|
|||
check_topic([]) ->
|
||||
error;
|
||||
check_topic(Path) ->
|
||||
Sep = <<"/">>,
|
||||
{ok,
|
||||
emqx_http_lib:uri_decode(
|
||||
lists:foldl(
|
||||
fun(Part, Acc) ->
|
||||
<<Acc/binary, Sep/binary, Part/binary>>
|
||||
end,
|
||||
<<>>,
|
||||
Path
|
||||
)
|
||||
)}.
|
||||
{ok, emqx_http_lib:uri_decode(iolist_to_binary(lists:join(<<"/">>, Path)))}.
|
||||
|
||||
get_sub_opts(#coap_message{options = Opts} = Msg) ->
|
||||
SubOpts = maps:fold(fun parse_sub_opts/3, #{}, Opts),
|
||||
|
@ -124,25 +114,30 @@ get_publish_qos(Msg) ->
|
|||
end.
|
||||
|
||||
apply_publish_opts(Msg, MQTTMsg) ->
|
||||
maps:fold(
|
||||
fun
|
||||
(<<"retain">>, V, Acc) ->
|
||||
Val = erlang:binary_to_atom(V),
|
||||
emqx_message:set_flag(retain, Val, Acc);
|
||||
(<<"expiry">>, V, Acc) ->
|
||||
Val = erlang:binary_to_integer(V),
|
||||
Props = emqx_message:get_header(properties, Acc),
|
||||
emqx_message:set_header(
|
||||
properties,
|
||||
Props#{'Message-Expiry-Interval' => Val},
|
||||
Acc
|
||||
);
|
||||
(_, _, Acc) ->
|
||||
Acc
|
||||
end,
|
||||
MQTTMsg,
|
||||
emqx_coap_message:get_option(uri_query, Msg)
|
||||
).
|
||||
case emqx_coap_message:get_option(uri_query, Msg) of
|
||||
undefined ->
|
||||
MQTTMsg;
|
||||
Qs ->
|
||||
maps:fold(
|
||||
fun
|
||||
(<<"retain">>, V, Acc) ->
|
||||
Val = erlang:binary_to_atom(V),
|
||||
emqx_message:set_flag(retain, Val, Acc);
|
||||
(<<"expiry">>, V, Acc) ->
|
||||
Val = erlang:binary_to_integer(V),
|
||||
Props = emqx_message:get_header(properties, Acc),
|
||||
emqx_message:set_header(
|
||||
properties,
|
||||
Props#{'Message-Expiry-Interval' => Val},
|
||||
Acc
|
||||
);
|
||||
(_, _, Acc) ->
|
||||
Acc
|
||||
end,
|
||||
MQTTMsg,
|
||||
Qs
|
||||
)
|
||||
end.
|
||||
|
||||
subscribe(#coap_message{token = <<>>} = Msg, _, _, _) ->
|
||||
reply({error, bad_request}, <<"observe without token">>, Msg);
|
||||
|
|
|
@ -143,12 +143,15 @@ t_connection_with_authn_failed(_) ->
|
|||
ok.
|
||||
|
||||
t_publish(_) ->
|
||||
Action = fun(Channel, Token) ->
|
||||
Topic = <<"/abc">>,
|
||||
%% can publish to a normal topic
|
||||
Topics = [
|
||||
<<"abc">>,
|
||||
%% can publish to a `/` leading topic
|
||||
<<"/abc">>
|
||||
],
|
||||
Action = fun(Topic, Channel, Token) ->
|
||||
Payload = <<"123">>,
|
||||
|
||||
TopicStr = binary_to_list(Topic),
|
||||
URI = ?PS_PREFIX ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = pubsub_uri(binary_to_list(Topic), Token),
|
||||
|
||||
%% Sub topic first
|
||||
emqx:subscribe(Topic),
|
||||
|
@ -164,24 +167,28 @@ t_publish(_) ->
|
|||
?assert(false)
|
||||
end
|
||||
end,
|
||||
with_connection(Action).
|
||||
with_connection(Topics, Action).
|
||||
|
||||
t_subscribe(_) ->
|
||||
Topic = <<"/abc">>,
|
||||
Fun = fun(Channel, Token) ->
|
||||
TopicStr = binary_to_list(Topic),
|
||||
%% can subscribe to a normal topic
|
||||
Topics = [
|
||||
<<"abc">>,
|
||||
%% can subscribe to a `/` leading topic
|
||||
<<"/abc">>
|
||||
],
|
||||
Fun = fun(Topic, Channel, Token) ->
|
||||
Payload = <<"123">>,
|
||||
|
||||
URI = ?PS_PREFIX ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = pubsub_uri(binary_to_list(Topic), Token),
|
||||
Req = make_req(get, Payload, [{observe, 0}]),
|
||||
{ok, content, _} = do_request(Channel, URI, Req),
|
||||
?LOGT("observer topic:~ts~n", [Topic]),
|
||||
|
||||
%% ensure subscribe succeed
|
||||
timer:sleep(100),
|
||||
[SubPid] = emqx:subscribers(Topic),
|
||||
?assert(is_pid(SubPid)),
|
||||
|
||||
%% Publish a message
|
||||
%% publish a message
|
||||
emqx:publish(emqx_message:make(Topic, Payload)),
|
||||
{ok, content, Notify} = with_response(Channel),
|
||||
?LOGT("observer get Notif=~p", [Notify]),
|
||||
|
@ -191,18 +198,27 @@ t_subscribe(_) ->
|
|||
?assertEqual(Payload, PayloadRecv)
|
||||
end,
|
||||
|
||||
with_connection(Fun),
|
||||
timer:sleep(100),
|
||||
with_connection(Topics, Fun),
|
||||
|
||||
?assertEqual([], emqx:subscribers(Topic)).
|
||||
%% subscription removed if coap client disconnected
|
||||
timer:sleep(100),
|
||||
lists:foreach(
|
||||
fun(Topic) ->
|
||||
?assertEqual([], emqx:subscribers(Topic))
|
||||
end,
|
||||
Topics
|
||||
).
|
||||
|
||||
t_un_subscribe(_) ->
|
||||
Topic = <<"/abc">>,
|
||||
Fun = fun(Channel, Token) ->
|
||||
TopicStr = binary_to_list(Topic),
|
||||
%% can unsubscribe to a normal topic
|
||||
Topics = [
|
||||
<<"abc">>,
|
||||
%% can unsubscribe to a `/` leading topic
|
||||
<<"/abc">>
|
||||
],
|
||||
Fun = fun(Topic, Channel, Token) ->
|
||||
Payload = <<"123">>,
|
||||
|
||||
URI = ?PS_PREFIX ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = pubsub_uri(binary_to_list(Topic), Token),
|
||||
|
||||
Req = make_req(get, Payload, [{observe, 0}]),
|
||||
{ok, content, _} = do_request(Channel, URI, Req),
|
||||
|
@ -219,16 +235,15 @@ t_un_subscribe(_) ->
|
|||
?assertEqual([], emqx:subscribers(Topic))
|
||||
end,
|
||||
|
||||
with_connection(Fun).
|
||||
with_connection(Topics, Fun).
|
||||
|
||||
t_observe_wildcard(_) ->
|
||||
Fun = fun(Channel, Token) ->
|
||||
%% resolve_url can't process wildcard with #
|
||||
Topic = <<"/abc/+">>,
|
||||
TopicStr = binary_to_list(Topic),
|
||||
Topic = <<"abc/+">>,
|
||||
Payload = <<"123">>,
|
||||
|
||||
URI = ?PS_PREFIX ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = pubsub_uri(binary_to_list(Topic), Token),
|
||||
Req = make_req(get, Payload, [{observe, 0}]),
|
||||
{ok, content, _} = do_request(Channel, URI, Req),
|
||||
?LOGT("observer topic:~ts~n", [Topic]),
|
||||
|
@ -238,7 +253,7 @@ t_observe_wildcard(_) ->
|
|||
?assert(is_pid(SubPid)),
|
||||
|
||||
%% Publish a message
|
||||
PubTopic = <<"/abc/def">>,
|
||||
PubTopic = <<"abc/def">>,
|
||||
emqx:publish(emqx_message:make(PubTopic, Payload)),
|
||||
{ok, content, Notify} = with_response(Channel),
|
||||
|
||||
|
@ -320,7 +335,7 @@ t_clients_get_subscription_api(_) ->
|
|||
|
||||
{200, [Subs]} = request(get, Path),
|
||||
|
||||
?assertEqual(<<"/coap/observe">>, maps:get(topic, Subs)),
|
||||
?assertEqual(<<"coap/observe">>, maps:get(topic, Subs)),
|
||||
|
||||
observe(Channel, Token, false),
|
||||
|
||||
|
@ -386,6 +401,9 @@ observe(Channel, Token, false) ->
|
|||
{ok, nocontent, _Data} = do_request(Channel, URI, Req),
|
||||
ok.
|
||||
|
||||
pubsub_uri(Topic, Token) when is_list(Topic), is_list(Token) ->
|
||||
?PS_PREFIX ++ "/" ++ Topic ++ "?clientid=client1&token=" ++ Token.
|
||||
|
||||
make_req(Method) ->
|
||||
make_req(Method, <<>>).
|
||||
|
||||
|
@ -442,6 +460,16 @@ with_connection(Action) ->
|
|||
end,
|
||||
do(Fun).
|
||||
|
||||
with_connection(Checks, Action) ->
|
||||
Fun = fun(Channel) ->
|
||||
Token = connection(Channel),
|
||||
timer:sleep(100),
|
||||
lists:foreach(fun(E) -> Action(E, Channel, Token) end, Checks),
|
||||
disconnection(Channel, Token),
|
||||
timer:sleep(100)
|
||||
end,
|
||||
do(Fun).
|
||||
|
||||
receive_deliver(Wait) ->
|
||||
receive
|
||||
{deliver, _, Msg} ->
|
||||
|
|
|
@ -98,7 +98,7 @@ t_case_coap_publish(_) ->
|
|||
Prefix = Mod:ps_prefix(),
|
||||
Fun = fun(Channel, Token, Topic, Checker) ->
|
||||
TopicStr = binary_to_list(Topic),
|
||||
URI = Prefix ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = Prefix ++ "/" ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
|
||||
Req = Mod:make_req(post, <<>>),
|
||||
Checker(Mod:do_request(Channel, URI, Req))
|
||||
|
@ -114,7 +114,7 @@ t_case_coap_subscribe(_) ->
|
|||
Prefix = Mod:ps_prefix(),
|
||||
Fun = fun(Channel, Token, Topic, Checker) ->
|
||||
TopicStr = binary_to_list(Topic),
|
||||
URI = Prefix ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
URI = Prefix ++ "/" ++ TopicStr ++ "?clientid=client1&token=" ++ Token,
|
||||
|
||||
Req = Mod:make_req(get, <<>>, [{observe, 0}]),
|
||||
Checker(Mod:do_request(Channel, URI, Req))
|
||||
|
|
|
@ -96,6 +96,16 @@ schema("/listeners") ->
|
|||
listener_id_status_example()
|
||||
)
|
||||
}
|
||||
},
|
||||
post => #{
|
||||
tags => [<<"listeners">>],
|
||||
desc => <<"Create the specified listener on all nodes.">>,
|
||||
parameters => [],
|
||||
'requestBody' => create_listener_schema(#{bind => true}),
|
||||
responses => #{
|
||||
200 => listener_schema(#{bind => true}),
|
||||
400 => error_codes(['BAD_LISTENER_ID', 'BAD_REQUEST'])
|
||||
}
|
||||
}
|
||||
};
|
||||
schema("/listeners/:id") ->
|
||||
|
@ -129,7 +139,8 @@ schema("/listeners/:id") ->
|
|||
responses => #{
|
||||
200 => listener_schema(#{bind => true}),
|
||||
400 => error_codes(['BAD_LISTENER_ID', 'BAD_REQUEST'])
|
||||
}
|
||||
},
|
||||
deprecated => true
|
||||
},
|
||||
delete => #{
|
||||
tags => [<<"listeners">>],
|
||||
|
@ -251,10 +262,10 @@ fields(node_status) ->
|
|||
})},
|
||||
{status, ?HOCON(?R_REF(status))}
|
||||
];
|
||||
fields({Type, with_name}) ->
|
||||
listener_struct_with_name(Type);
|
||||
fields(Type) ->
|
||||
Listeners = listeners_info(#{bind => true}) ++ listeners_info(#{bind => false}),
|
||||
[Schema] = [S || #{ref := ?R_REF(_, T), schema := S} <- Listeners, T =:= Type],
|
||||
Schema.
|
||||
listener_struct(Type).
|
||||
|
||||
listener_schema(Opts) ->
|
||||
emqx_dashboard_swagger:schema_with_example(
|
||||
|
@ -262,6 +273,17 @@ listener_schema(Opts) ->
|
|||
tcp_schema_example()
|
||||
).
|
||||
|
||||
create_listener_schema(Opts) ->
|
||||
Schemas = [
|
||||
?R_REF(Mod, {Type, with_name})
|
||||
|| #{ref := ?R_REF(Mod, Type)} <- listeners_info(Opts)
|
||||
],
|
||||
Example = maps:remove(id, tcp_schema_example()),
|
||||
emqx_dashboard_swagger:schema_with_example(
|
||||
?UNION(Schemas),
|
||||
Example#{name => <<"demo">>}
|
||||
).
|
||||
|
||||
listeners_type() ->
|
||||
lists:map(
|
||||
fun({Type, _}) -> list_to_existing_atom(Type) end,
|
||||
|
@ -339,7 +361,9 @@ list_listeners(get, #{query_string := Query}) ->
|
|||
{ok, Type} -> listener_type_filter(atom_to_binary(Type), Listeners);
|
||||
error -> Listeners
|
||||
end,
|
||||
{200, listener_status_by_id(NodeL)}.
|
||||
{200, listener_status_by_id(NodeL)};
|
||||
list_listeners(post, #{body := Body}) ->
|
||||
create_listener(Body).
|
||||
|
||||
crud_listeners_by_id(get, #{bindings := #{id := Id0}}) ->
|
||||
Listeners =
|
||||
|
@ -382,23 +406,8 @@ crud_listeners_by_id(put, #{bindings := #{id := Id}, body := Body0}) ->
|
|||
_ ->
|
||||
{400, #{code => 'BAD_LISTENER_ID', message => ?LISTENER_ID_INCONSISTENT}}
|
||||
end;
|
||||
crud_listeners_by_id(post, #{bindings := #{id := Id}, body := Body0}) ->
|
||||
case parse_listener_conf(Body0) of
|
||||
{Id, Type, Name, Conf} ->
|
||||
Path = [listeners, Type, Name],
|
||||
case create(Path, Conf) of
|
||||
{ok, #{raw_config := _RawConf}} ->
|
||||
crud_listeners_by_id(get, #{bindings => #{id => Id}});
|
||||
{error, already_exist} ->
|
||||
{400, #{code => 'BAD_LISTENER_ID', message => <<"Already Exist">>}};
|
||||
{error, Reason} ->
|
||||
{400, #{code => 'BAD_REQUEST', message => err_msg(Reason)}}
|
||||
end;
|
||||
{error, Reason} ->
|
||||
{400, #{code => 'BAD_REQUEST', message => err_msg(Reason)}};
|
||||
_ ->
|
||||
{400, #{code => 'BAD_LISTENER_ID', message => ?LISTENER_ID_INCONSISTENT}}
|
||||
end;
|
||||
crud_listeners_by_id(post, #{body := Body}) ->
|
||||
create_listener(Body);
|
||||
crud_listeners_by_id(delete, #{bindings := #{id := Id}}) ->
|
||||
{ok, #{type := Type, name := Name}} = emqx_listeners:parse_listener_id(Id),
|
||||
case ensure_remove([listeners, Type, Name]) of
|
||||
|
@ -408,13 +417,24 @@ crud_listeners_by_id(delete, #{bindings := #{id := Id}}) ->
|
|||
|
||||
parse_listener_conf(Conf0) ->
|
||||
Conf1 = maps:without([<<"running">>, <<"current_connections">>], Conf0),
|
||||
{IdBin, Conf2} = maps:take(<<"id">>, Conf1),
|
||||
{TypeBin, Conf3} = maps:take(<<"type">>, Conf2),
|
||||
{ok, #{type := Type, name := Name}} = emqx_listeners:parse_listener_id(IdBin),
|
||||
{TypeBin, Conf2} = maps:take(<<"type">>, Conf1),
|
||||
TypeAtom = binary_to_existing_atom(TypeBin),
|
||||
case Type =:= TypeAtom of
|
||||
true -> {binary_to_existing_atom(IdBin), TypeAtom, Name, Conf3};
|
||||
false -> {error, listener_type_inconsistent}
|
||||
|
||||
case maps:take(<<"id">>, Conf2) of
|
||||
{IdBin, Conf3} ->
|
||||
{ok, #{type := Type, name := Name}} = emqx_listeners:parse_listener_id(IdBin),
|
||||
case Type =:= TypeAtom of
|
||||
true -> {binary_to_existing_atom(IdBin), TypeAtom, Name, Conf3};
|
||||
false -> {error, listener_type_inconsistent}
|
||||
end;
|
||||
_ ->
|
||||
case maps:take(<<"name">>, Conf2) of
|
||||
{Name, Conf3} ->
|
||||
IdBin = <<TypeBin/binary, $:, Name/binary>>,
|
||||
{binary_to_atom(IdBin), TypeAtom, Name, Conf3};
|
||||
_ ->
|
||||
{error, listener_config_invalid}
|
||||
end
|
||||
end.
|
||||
|
||||
stop_listeners_by_id(Method, Body = #{bindings := Bindings}) ->
|
||||
|
@ -787,3 +807,37 @@ tcp_schema_example() ->
|
|||
type => tcp,
|
||||
zone => default
|
||||
}.
|
||||
|
||||
create_listener(Body) ->
|
||||
case parse_listener_conf(Body) of
|
||||
{Id, Type, Name, Conf} ->
|
||||
Path = [listeners, Type, Name],
|
||||
case create(Path, Conf) of
|
||||
{ok, #{raw_config := _RawConf}} ->
|
||||
crud_listeners_by_id(get, #{bindings => #{id => Id}});
|
||||
{error, already_exist} ->
|
||||
{400, #{code => 'BAD_LISTENER_ID', message => <<"Already Exist">>}};
|
||||
{error, Reason} ->
|
||||
{400, #{code => 'BAD_REQUEST', message => err_msg(Reason)}}
|
||||
end;
|
||||
{error, Reason} ->
|
||||
{400, #{code => 'BAD_REQUEST', message => err_msg(Reason)}}
|
||||
end.
|
||||
|
||||
listener_struct(Type) ->
|
||||
Listeners = listeners_info(#{bind => true}) ++ listeners_info(#{bind => false}),
|
||||
[Schema] = [S || #{ref := ?R_REF(_, T), schema := S} <- Listeners, T =:= Type],
|
||||
Schema.
|
||||
|
||||
listener_struct_with_name(Type) ->
|
||||
BaseSchema = listener_struct(Type),
|
||||
lists:keyreplace(
|
||||
id,
|
||||
1,
|
||||
BaseSchema,
|
||||
{name,
|
||||
?HOCON(binary(), #{
|
||||
desc => "Listener name",
|
||||
required => true
|
||||
})}
|
||||
).
|
||||
|
|
|
@ -37,6 +37,35 @@ t_list_listeners(_) ->
|
|||
Res = request(get, Path, [], []),
|
||||
#{<<"listeners">> := Expect} = emqx_mgmt_api_listeners:do_list_listeners(),
|
||||
?assertEqual(length(Expect), length(Res)),
|
||||
|
||||
%% POST /listeners
|
||||
ListenerId = <<"tcp:default">>,
|
||||
NewListenerId = <<"tcp:new">>,
|
||||
|
||||
OriginPath = emqx_mgmt_api_test_util:api_path(["listeners", ListenerId]),
|
||||
NewPath = emqx_mgmt_api_test_util:api_path(["listeners", NewListenerId]),
|
||||
|
||||
OriginListener = request(get, OriginPath, [], []),
|
||||
|
||||
%% create with full options
|
||||
?assertEqual({error, not_found}, is_running(NewListenerId)),
|
||||
?assertMatch({error, {"HTTP/1.1", 404, _}}, request(get, NewPath, [], [])),
|
||||
|
||||
OriginListener2 = maps:remove(<<"id">>, OriginListener),
|
||||
NewConf = OriginListener2#{
|
||||
<<"name">> => <<"new">>,
|
||||
<<"bind">> => <<"0.0.0.0:2883">>
|
||||
},
|
||||
Create = request(post, Path, [], NewConf),
|
||||
?assertEqual(lists:sort(maps:keys(OriginListener)), lists:sort(maps:keys(Create))),
|
||||
Get1 = request(get, NewPath, [], []),
|
||||
?assertMatch(Create, Get1),
|
||||
?assert(is_running(NewListenerId)),
|
||||
|
||||
%% delete
|
||||
?assertEqual([], delete(NewPath)),
|
||||
?assertEqual({error, not_found}, is_running(NewListenerId)),
|
||||
?assertMatch({error, {"HTTP/1.1", 404, _}}, request(get, NewPath, [], [])),
|
||||
ok.
|
||||
|
||||
t_tcp_crud_listeners_by_id(_) ->
|
||||
|
|
|
@ -296,7 +296,7 @@ handle_cast(Msg, State) ->
|
|||
|
||||
%% Do Publish...
|
||||
handle_info({timeout, TRef, do_publish}, State = #{publish_timer := TRef}) ->
|
||||
DeletedKeys = do_publish(mnesia:dirty_first(?TAB), os:system_time(seconds)),
|
||||
DeletedKeys = do_publish(mnesia:dirty_first(?TAB), erlang:system_time(seconds)),
|
||||
lists:foreach(fun(Key) -> mria:dirty_delete(?TAB, Key) end, DeletedKeys),
|
||||
{noreply, ensure_publish_timer(State#{publish_timer := undefined, publish_at := 0})};
|
||||
handle_info(stats, State = #{stats_fun := StatsFun}) ->
|
||||
|
@ -347,12 +347,12 @@ ensure_publish_timer(State) ->
|
|||
ensure_publish_timer('$end_of_table', State) ->
|
||||
State#{publish_timer := undefined, publish_at := 0};
|
||||
ensure_publish_timer({Ts, _Id}, State = #{publish_timer := undefined}) ->
|
||||
ensure_publish_timer(Ts, os:system_time(seconds), State);
|
||||
ensure_publish_timer(Ts, erlang:system_time(seconds), State);
|
||||
ensure_publish_timer({Ts, _Id}, State = #{publish_timer := TRef, publish_at := PubAt}) when
|
||||
Ts < PubAt
|
||||
->
|
||||
ok = emqx_misc:cancel_timer(TRef),
|
||||
ensure_publish_timer(Ts, os:system_time(seconds), State);
|
||||
ensure_publish_timer(Ts, erlang:system_time(seconds), State);
|
||||
ensure_publish_timer(_Key, State) ->
|
||||
State.
|
||||
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
%% -*- mode: erlang -*-
|
||||
{application, emqx_modules, [
|
||||
{description, "EMQX Modules"},
|
||||
{vsn, "5.0.3"},
|
||||
{vsn, "5.0.4"},
|
||||
{modules, []},
|
||||
{applications, [kernel, stdlib, emqx]},
|
||||
{mod, {emqx_modules_app, []}},
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
{application, emqx_retainer, [
|
||||
{description, "EMQX Retainer"},
|
||||
% strict semver, bump manually!
|
||||
{vsn, "5.0.4"},
|
||||
{vsn, "5.0.5"},
|
||||
{modules, []},
|
||||
{registered, [emqx_retainer_sup]},
|
||||
{applications, [kernel, stdlib, emqx]},
|
||||
|
|
|
@ -348,16 +348,12 @@ enable_retainer(
|
|||
#{context_id := ContextId} = State,
|
||||
#{
|
||||
msg_clear_interval := ClearInterval,
|
||||
backend := BackendCfg,
|
||||
flow_control := FlowControl
|
||||
backend := BackendCfg
|
||||
}
|
||||
) ->
|
||||
NewContextId = ContextId + 1,
|
||||
Context = create_resource(new_context(NewContextId), BackendCfg),
|
||||
load(Context),
|
||||
emqx_limiter_server:add_bucket(
|
||||
?APP, internal, maps:get(batch_deliver_limiter, FlowControl, undefined)
|
||||
),
|
||||
State#{
|
||||
enable := true,
|
||||
context_id := NewContextId,
|
||||
|
@ -373,7 +369,6 @@ disable_retainer(
|
|||
} = State
|
||||
) ->
|
||||
unload(),
|
||||
emqx_limiter_server:del_bucket(?APP, internal),
|
||||
ok = close_resource(Context),
|
||||
State#{
|
||||
enable := false,
|
||||
|
|
|
@ -18,6 +18,8 @@
|
|||
|
||||
-behaviour(application).
|
||||
|
||||
-include("emqx_retainer.hrl").
|
||||
|
||||
-export([
|
||||
start/2,
|
||||
stop/1
|
||||
|
@ -25,8 +27,19 @@
|
|||
|
||||
start(_Type, _Args) ->
|
||||
ok = emqx_retainer_mnesia_cli:load(),
|
||||
init_bucket(),
|
||||
emqx_retainer_sup:start_link().
|
||||
|
||||
stop(_State) ->
|
||||
ok = emqx_retainer_mnesia_cli:unload(),
|
||||
delete_bucket(),
|
||||
ok.
|
||||
|
||||
init_bucket() ->
|
||||
#{flow_control := FlowControl} = emqx:get_config([retainer]),
|
||||
emqx_limiter_server:add_bucket(
|
||||
?APP, internal, maps:get(batch_deliver_limiter, FlowControl, undefined)
|
||||
).
|
||||
|
||||
delete_bucket() ->
|
||||
emqx_limiter_server:del_bucket(?APP, internal).
|
||||
|
|
|
@ -31,14 +31,16 @@ all() ->
|
|||
[
|
||||
{group, mnesia_without_indices},
|
||||
{group, mnesia_with_indices},
|
||||
{group, mnesia_reindex}
|
||||
{group, mnesia_reindex},
|
||||
{group, test_disable_then_start}
|
||||
].
|
||||
|
||||
groups() ->
|
||||
[
|
||||
{mnesia_without_indices, [sequence], common_tests()},
|
||||
{mnesia_with_indices, [sequence], common_tests()},
|
||||
{mnesia_reindex, [sequence], [t_reindex]}
|
||||
{mnesia_reindex, [sequence], [t_reindex]},
|
||||
{test_disable_then_start, [sequence], [test_disable_then_start]}
|
||||
].
|
||||
|
||||
common_tests() ->
|
||||
|
@ -624,6 +626,19 @@ t_get_basic_usage_info(_Config) ->
|
|||
?assertEqual(#{retained_messages => 5}, emqx_retainer:get_basic_usage_info()),
|
||||
ok.
|
||||
|
||||
%% test whether the app can start normally after disabling emqx_retainer
|
||||
%% fix: https://github.com/emqx/emqx/pull/8911
|
||||
test_disable_then_start(_Config) ->
|
||||
emqx_retainer:update_config(#{<<"enable">> => false}),
|
||||
?assertNotEqual([], gproc_pool:active_workers(emqx_retainer_dispatcher)),
|
||||
ok = application:stop(emqx_retainer),
|
||||
timer:sleep(100),
|
||||
?assertEqual([], gproc_pool:active_workers(emqx_retainer_dispatcher)),
|
||||
ok = application:ensure_started(emqx_retainer),
|
||||
timer:sleep(100),
|
||||
?assertNotEqual([], gproc_pool:active_workers(emqx_retainer_dispatcher)),
|
||||
ok.
|
||||
|
||||
%%--------------------------------------------------------------------
|
||||
%% Helper functions
|
||||
%%--------------------------------------------------------------------
|
||||
|
|
17
bin/emqx
17
bin/emqx
|
@ -7,7 +7,7 @@ set -euo pipefail
|
|||
DEBUG="${DEBUG:-0}"
|
||||
[ "$DEBUG" -eq 1 ] && set -x
|
||||
|
||||
RUNNER_ROOT_DIR="$(cd "$(dirname "$(readlink "$0" || echo "$0")")"/..; pwd -P)"
|
||||
RUNNER_ROOT_DIR="$(cd "$(dirname "$(realpath "$0" || echo "$0")")"/..; pwd -P)"
|
||||
|
||||
# shellcheck disable=SC1090,SC1091
|
||||
. "$RUNNER_ROOT_DIR"/releases/emqx_vars
|
||||
|
@ -600,7 +600,7 @@ is_down() {
|
|||
if ps -p "$PID" | grep -q 'defunct'; then
|
||||
# zombie state, print parent pid
|
||||
parent="$(ps -o ppid= -p "$PID" | tr -d ' ')"
|
||||
echo "WARN: $PID is marked <defunct>, parent:"
|
||||
echo "WARNING: $PID is marked <defunct>, parent:"
|
||||
ps -p "$parent"
|
||||
return 0
|
||||
fi
|
||||
|
@ -748,8 +748,9 @@ export ESCRIPT_NAME="$SHORT_NAME"
|
|||
|
||||
PIPE_DIR="${PIPE_DIR:-/$DATA_DIR/${WHOAMI}_erl_pipes/$NAME/}"
|
||||
|
||||
## make EMQX_NODE_COOKIE right
|
||||
## Resolve Erlang cookie.
|
||||
if [ -n "${EMQX_NODE_COOKIE:-}" ]; then
|
||||
## To be backward compatible, read EMQX_NODE_COOKIE
|
||||
export EMQX_NODE__COOKIE="${EMQX_NODE_COOKIE}"
|
||||
unset EMQX_NODE_COOKIE
|
||||
fi
|
||||
|
@ -762,9 +763,13 @@ if [ -z "$COOKIE" ]; then
|
|||
COOKIE="$(grep -E '^-setcookie' "${vm_args_file}" | awk '{print $2}')"
|
||||
fi
|
||||
fi
|
||||
|
||||
if [ -z "$COOKIE" ]; then
|
||||
die "Please set node.cookie in $EMQX_ETC_DIR/emqx.conf or override from environment variable EMQX_NODE__COOKIE"
|
||||
[ -z "$COOKIE" ] && COOKIE="$EMQX_DEFAULT_ERLANG_COOKIE"
|
||||
if [ $IS_BOOT_COMMAND = 'yes' ] && [ "$COOKIE" = "$EMQX_DEFAULT_ERLANG_COOKIE" ]; then
|
||||
echoerr "!!!!!!"
|
||||
echoerr "WARNING: Default (insecure) Erlang cookie is in use."
|
||||
echoerr "WARNING: Configure node.cookie in $EMQX_ETC_DIR/emqx.conf or override from environment variable EMQX_NODE__COOKIE"
|
||||
echoerr "NOTE: Use the same config value for all nodes in the cluster."
|
||||
echoerr "!!!!!!"
|
||||
fi
|
||||
|
||||
## check if OTP version has mnesia_hook feature; if not, fallback to
|
||||
|
|
1
build
1
build
|
@ -163,6 +163,7 @@ make_relup() {
|
|||
local name_pattern
|
||||
name_pattern="${PROFILE}-$(./pkg-vsn.sh "$PROFILE" --vsn_matcher --long)"
|
||||
local releases=()
|
||||
mkdir -p _upgrade_base
|
||||
while read -r tgzfile ; do
|
||||
local base_vsn
|
||||
base_vsn="$(echo "$tgzfile" | grep -oE "[0-9]+\.[0-9]+\.[0-9]+(-(alpha|beta|rc)\.[0-9])?(-[0-9a-f]{8})?" | head -1)"
|
||||
|
|
6
mix.exs
6
mix.exs
|
@ -562,6 +562,7 @@ defmodule EMQXUmbrella.MixProject do
|
|||
|
||||
defp template_vars(release, release_type, :bin = _package_type, edition_type) do
|
||||
[
|
||||
emqx_default_erlang_cookie: default_cookie(),
|
||||
platform_data_dir: "data",
|
||||
platform_etc_dir: "etc",
|
||||
platform_log_dir: "log",
|
||||
|
@ -584,6 +585,7 @@ defmodule EMQXUmbrella.MixProject do
|
|||
|
||||
defp template_vars(release, release_type, :pkg = _package_type, edition_type) do
|
||||
[
|
||||
emqx_default_erlang_cookie: default_cookie(),
|
||||
platform_data_dir: "/var/lib/emqx",
|
||||
platform_etc_dir: "/etc/emqx",
|
||||
platform_log_dir: "/var/log/emqx",
|
||||
|
@ -604,6 +606,10 @@ defmodule EMQXUmbrella.MixProject do
|
|||
] ++ build_info()
|
||||
end
|
||||
|
||||
defp default_cookie() do
|
||||
"emqx50elixir"
|
||||
end
|
||||
|
||||
defp emqx_description(release_type, edition_type) do
|
||||
case {release_type, edition_type} do
|
||||
{:cloud, :enterprise} ->
|
||||
|
|
|
@ -298,14 +298,13 @@ relform() ->
|
|||
emqx_description(cloud, ee) -> "EMQX Enterprise";
|
||||
emqx_description(cloud, ce) -> "EMQX".
|
||||
|
||||
overlay_vars(RelType, PkgType, Edition) ->
|
||||
overlay_vars_rel(RelType) ++
|
||||
overlay_vars(cloud, PkgType, Edition) ->
|
||||
[
|
||||
{emqx_default_erlang_cookie, "emqxsecretcookie"}
|
||||
] ++
|
||||
overlay_vars_pkg(PkgType) ++
|
||||
overlay_vars_edition(Edition).
|
||||
|
||||
overlay_vars_rel(cloud) ->
|
||||
[{vm_args_file, "vm.args"}].
|
||||
|
||||
overlay_vars_edition(ce) ->
|
||||
[
|
||||
{emqx_schema_mod, emqx_conf_schema},
|
||||
|
@ -485,11 +484,16 @@ emqx_etc_overlay_per_edition(ee) ->
|
|||
].
|
||||
|
||||
get_vsn(Profile) ->
|
||||
%% to make it compatible to Linux and Windows,
|
||||
%% we must use bash to execute the bash file
|
||||
%% because "./" will not be recognized as an internal or external command
|
||||
os_cmd("pkg-vsn.sh " ++ atom_to_list(Profile)).
|
||||
case os:getenv("PKG_VSN") of
|
||||
false ->
|
||||
os_cmd("pkg-vsn.sh " ++ atom_to_list(Profile));
|
||||
Vsn ->
|
||||
Vsn
|
||||
end.
|
||||
|
||||
%% to make it compatible to Linux and Windows,
|
||||
%% we must use bash to execute the bash file
|
||||
%% because "./" will not be recognized as an internal or external command
|
||||
os_cmd(Cmd) ->
|
||||
Output = os:cmd("bash " ++ Cmd),
|
||||
re:replace(Output, "\n", "", [{return, list}]).
|
||||
|
|
|
@ -9,19 +9,17 @@ ERL_OPTS="{{ erl_opts }}"
|
|||
RUNNER_BIN_DIR="{{ runner_bin_dir }}"
|
||||
RUNNER_LIB_DIR="{{ runner_lib_dir }}"
|
||||
IS_ELIXIR="${IS_ELIXIR:-{{ is_elixir }}}"
|
||||
|
||||
## Allow users to pre-set `RUNNER_LOG_DIR` because it only affects boot commands like `start` and `console`,
|
||||
## but not other commands such as `ping` and `ctl`.
|
||||
RUNNER_LOG_DIR="${RUNNER_LOG_DIR:-{{ runner_log_dir }}}"
|
||||
|
||||
EMQX_ETC_DIR="{{ emqx_etc_dir }}"
|
||||
RUNNER_USER="{{ runner_user }}"
|
||||
SCHEMA_MOD="{{ emqx_schema_mod }}"
|
||||
IS_ENTERPRISE="{{ is_enterprise }}"
|
||||
|
||||
## Do not change EMQX_DEFAULT_ERLANG_COOKIE.
|
||||
## Configure EMQX_NODE_COOKIE instead
|
||||
EMQX_DEFAULT_ERLANG_COOKIE='{{ emqx_default_erlang_cookie }}'
|
||||
REL_NAME="emqx"
|
||||
export EMQX_DESCRIPTION='{{ emqx_description }}'
|
||||
|
||||
## computed vars
|
||||
REL_NAME="emqx"
|
||||
|
||||
## updated vars here
|
||||
|
|
|
@ -42,7 +42,7 @@ curl -L --silent --show-error \
|
|||
--output "${RELEASE_ASSET_FILE}" \
|
||||
"$DIRECT_DOWNLOAD_URL"
|
||||
|
||||
unzip -q "$RELEASE_ASSET_FILE" -d "$DASHBOARD_PATH"
|
||||
unzip -o -q "$RELEASE_ASSET_FILE" -d "$DASHBOARD_PATH"
|
||||
rm -rf "$DASHBOARD_PATH/www"
|
||||
mv "$DASHBOARD_PATH/dist" "$DASHBOARD_PATH/www"
|
||||
rm -f "$RELEASE_ASSET_FILE"
|
||||
|
|
|
@ -1,21 +0,0 @@
|
|||
#!/bin/bash
|
||||
set -uo pipefail
|
||||
|
||||
if [ -z "${1:-}" ]; then
|
||||
SCHEMA="_build/emqx/lib/emqx_dashboard/priv/www/static/schema.json"
|
||||
else
|
||||
SCHEMA="$1"
|
||||
fi
|
||||
|
||||
docker run -d --name langtool "ghcr.io/emqx/emqx-schema-validate:0.3.5"
|
||||
|
||||
docker exec -i langtool ./emqx_schema_validate - < "${SCHEMA}"
|
||||
success="$?"
|
||||
|
||||
docker kill langtool || true
|
||||
docker rm langtool || true
|
||||
|
||||
echo "If this script finds a false positive (e.g. when it things that a protocol name is a typo),
|
||||
make a PR here: https://github.com/emqx/emqx-schema-validate/blob/master/dict/en_spelling_additions.txt"
|
||||
|
||||
exit "$success"
|
|
@ -0,0 +1,265 @@
|
|||
ACL
|
||||
AES
|
||||
APIs
|
||||
BPAPI
|
||||
BSON
|
||||
Backplane
|
||||
CA
|
||||
CAs
|
||||
CHACHA
|
||||
CLI
|
||||
CMD
|
||||
CN
|
||||
CONNACK
|
||||
CoAP
|
||||
Cygwin
|
||||
DES
|
||||
DN
|
||||
DNS
|
||||
DTLS
|
||||
DevOps
|
||||
Dialyzer
|
||||
Diffie
|
||||
EIP
|
||||
EMQX
|
||||
EPMD
|
||||
ERL
|
||||
ETS
|
||||
FIXME
|
||||
GCM
|
||||
HMAC
|
||||
HOCON
|
||||
HTTPS
|
||||
JSON
|
||||
JWK
|
||||
JWKs
|
||||
JWT
|
||||
Kubernetes
|
||||
LwM
|
||||
MQTT
|
||||
Makefile
|
||||
MitM
|
||||
Multicast
|
||||
NIF
|
||||
OTP
|
||||
PEM
|
||||
PINGREQ
|
||||
PSK
|
||||
PSK
|
||||
PSKs
|
||||
PUBREL
|
||||
QoS
|
||||
RESTful
|
||||
ROADMAP
|
||||
RSA
|
||||
Req
|
||||
Riak
|
||||
SHA
|
||||
SMS
|
||||
Struct
|
||||
TCP
|
||||
TLS
|
||||
TTL
|
||||
UDP
|
||||
URI
|
||||
XMLs
|
||||
acceptors
|
||||
ack
|
||||
acked
|
||||
addr
|
||||
api
|
||||
apiserver
|
||||
arg
|
||||
args
|
||||
async
|
||||
attr
|
||||
auth
|
||||
authenticator
|
||||
authenticators
|
||||
authn
|
||||
authz
|
||||
autoclean
|
||||
autoheal
|
||||
backend
|
||||
backends
|
||||
backoff
|
||||
backplane
|
||||
backtrace
|
||||
badarg
|
||||
badkey
|
||||
bcrypt
|
||||
behaviour
|
||||
bhvr
|
||||
boolean
|
||||
bytesize
|
||||
cacert
|
||||
cacertfile
|
||||
certfile
|
||||
ci
|
||||
clientid
|
||||
clientinfo
|
||||
cmake
|
||||
coap
|
||||
conf
|
||||
config
|
||||
configs
|
||||
confirmable
|
||||
conn
|
||||
connectionless
|
||||
cors
|
||||
cpu
|
||||
ctx
|
||||
customizable
|
||||
datagram
|
||||
datagrams
|
||||
desc
|
||||
dir
|
||||
dns
|
||||
downlink
|
||||
downlink
|
||||
dtls
|
||||
ekka
|
||||
emqx
|
||||
enablement
|
||||
enqueue
|
||||
enqueued
|
||||
env
|
||||
eof
|
||||
epmd
|
||||
erl
|
||||
erts
|
||||
escript
|
||||
etcd
|
||||
eval
|
||||
exe
|
||||
executables
|
||||
exhook
|
||||
exproto
|
||||
extensibility
|
||||
formatter
|
||||
gRPC
|
||||
github
|
||||
goto
|
||||
grpcbox
|
||||
hocon
|
||||
hoconsc
|
||||
hostname
|
||||
hrl
|
||||
http
|
||||
https
|
||||
iface
|
||||
img
|
||||
impl
|
||||
inet
|
||||
inflight
|
||||
ini
|
||||
init
|
||||
ip
|
||||
ipv
|
||||
jenkins
|
||||
jq
|
||||
kb
|
||||
keepalive
|
||||
libcoap
|
||||
lifecycle
|
||||
localhost
|
||||
lwm
|
||||
mnesia
|
||||
mountpoint
|
||||
mqueue
|
||||
mria
|
||||
msg
|
||||
multicalls
|
||||
multicasts
|
||||
namespace
|
||||
natively
|
||||
nodelay
|
||||
nodetool
|
||||
nullable
|
||||
num
|
||||
os
|
||||
params
|
||||
peerhost
|
||||
peername
|
||||
perf
|
||||
powershell
|
||||
procmem
|
||||
procs
|
||||
progname
|
||||
prometheus
|
||||
proto
|
||||
ps
|
||||
psk
|
||||
pubsub
|
||||
pushgateway
|
||||
qlen
|
||||
qmode
|
||||
qos
|
||||
quic
|
||||
ratelimit
|
||||
rebar
|
||||
recbuf
|
||||
relup
|
||||
replayq
|
||||
replicant
|
||||
repo
|
||||
reuseaddr
|
||||
rh
|
||||
rlog
|
||||
rootdir
|
||||
rpc
|
||||
runtime
|
||||
sc
|
||||
scalable
|
||||
seg
|
||||
setcookie
|
||||
sharded
|
||||
shareload
|
||||
sn
|
||||
sndbuf
|
||||
sockname
|
||||
sql
|
||||
src
|
||||
ssl
|
||||
statsd
|
||||
structs
|
||||
subprotocol
|
||||
subprotocols
|
||||
superset
|
||||
sys
|
||||
sysmem
|
||||
sysmon
|
||||
tcp
|
||||
ticktime
|
||||
tlog
|
||||
tls
|
||||
tlsv
|
||||
travis
|
||||
trie
|
||||
ttl
|
||||
typerefl
|
||||
udp
|
||||
uid
|
||||
un-acked
|
||||
unsub
|
||||
uplink
|
||||
url
|
||||
utc
|
||||
util
|
||||
ver
|
||||
vm
|
||||
vsn
|
||||
wakaama
|
||||
websocket
|
||||
ws
|
||||
wss
|
||||
xml
|
||||
HStream
|
||||
HStreamDB
|
||||
hstream
|
||||
hstreamDB
|
||||
hstream
|
||||
hstreamdb
|
||||
SASL
|
||||
GSSAPI
|
||||
keytab
|
|
@ -0,0 +1,30 @@
|
|||
#!/usr/bin/env bash
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
# ensure dir
|
||||
cd -P -- "$(dirname -- "${BASH_SOURCE[0]}")/../.."
|
||||
PROJ_ROOT="$(pwd)"
|
||||
|
||||
if [ -z "${1:-}" ]; then
|
||||
SCHEMA="${PROJ_ROOT}/_build/emqx/lib/emqx_dashboard/priv/www/static/schema.json"
|
||||
else
|
||||
SCHEMA="$(realpath "$1")"
|
||||
fi
|
||||
|
||||
set +e
|
||||
docker run --rm -i --name spellcheck \
|
||||
-v "${PROJ_ROOT}"/scripts/spellcheck/dicts:/dicts \
|
||||
-v "$SCHEMA":/schema.json \
|
||||
ghcr.io/emqx/emqx-schema-validate:0.4.0 /schema.json
|
||||
|
||||
result="$?"
|
||||
|
||||
if [ "$result" -eq 0 ]; then
|
||||
echo "Spellcheck OK"
|
||||
exit 0
|
||||
fi
|
||||
|
||||
echo "If this script finds a false positive (e.g. when it thinks that a protocol name is a typo),"
|
||||
echo "Add the word to dictionary in scripts/spellcheck/dicts"
|
||||
exit $result
|
Loading…
Reference in New Issue