refactor(config): split config and schema
This commit is contained in:
parent
4bbd398550
commit
522d8e0a4a
|
@ -1,702 +1,3 @@
|
||||||
## master-88df1713
|
|
||||||
|
|
||||||
## NOTE: The configurations in this file will be overridden by
|
|
||||||
## `<path-to-emqx-installation>/data/emqx_overrides.conf`
|
|
||||||
|
|
||||||
##==================================================================
|
|
||||||
## Node
|
|
||||||
##==================================================================
|
|
||||||
node {
|
|
||||||
## Node name.
|
|
||||||
## See: http://erlang.org/doc/reference_manual/distributed.html
|
|
||||||
##
|
|
||||||
## @doc node.name
|
|
||||||
## ValueType: NodeName
|
|
||||||
## Default: emqx@127.0.0.1
|
|
||||||
name: "emqx@127.0.0.1"
|
|
||||||
|
|
||||||
## Cookie for distributed node communication.
|
|
||||||
##
|
|
||||||
## @doc node.cookie
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqxsecretcookie
|
|
||||||
cookie: emqxsecretcookie
|
|
||||||
|
|
||||||
## Data dir for the node
|
|
||||||
##
|
|
||||||
## @doc node.data_dir
|
|
||||||
## ValueType: Folder
|
|
||||||
## Default: "{{ platform_data_dir }}/"
|
|
||||||
data_dir: "{{ platform_data_dir }}/"
|
|
||||||
|
|
||||||
## Dir of crash dump file.
|
|
||||||
##
|
|
||||||
## @doc node.crash_dump_dir
|
|
||||||
## ValueType: Folder
|
|
||||||
## Default: "{{ platform_log_dir }}/"
|
|
||||||
crash_dump_dir: "{{ platform_log_dir }}/"
|
|
||||||
|
|
||||||
## Global GC Interval.
|
|
||||||
##
|
|
||||||
## @doc node.global_gc_interval
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 15m
|
|
||||||
global_gc_interval: 15m
|
|
||||||
|
|
||||||
## Sets the net_kernel tick time in seconds.
|
|
||||||
## Notice that all communicating nodes are to have the same
|
|
||||||
## TickTime value specified.
|
|
||||||
##
|
|
||||||
## See: http://www.erlang.org/doc/man/kernel_app.html#net_ticktime
|
|
||||||
##
|
|
||||||
## @doc node.dist_net_ticktime
|
|
||||||
## ValueType: Number
|
|
||||||
## Default: 2m
|
|
||||||
dist_net_ticktime: 2m
|
|
||||||
|
|
||||||
## Sets the port range for the listener socket of a distributed
|
|
||||||
## Erlang node.
|
|
||||||
## Note that if there are firewalls between clustered nodes, this
|
|
||||||
## port segment for nodes’ communication should be allowed.
|
|
||||||
##
|
|
||||||
## See: http://www.erlang.org/doc/man/kernel_app.html
|
|
||||||
##
|
|
||||||
## @doc node.dist_listen_min
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [1024,65535]
|
|
||||||
## Default: 6369
|
|
||||||
dist_listen_min: 6369
|
|
||||||
|
|
||||||
## Sets the port range for the listener socket of a distributed
|
|
||||||
## Erlang node.
|
|
||||||
## Note that if there are firewalls between clustered nodes, this
|
|
||||||
## port segment for nodes’ communication should be allowed.
|
|
||||||
##
|
|
||||||
## See: http://www.erlang.org/doc/man/kernel_app.html
|
|
||||||
##
|
|
||||||
## @doc node.dist_listen_max
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [1024,65535]
|
|
||||||
## Default: 6369
|
|
||||||
dist_listen_max: 6369
|
|
||||||
|
|
||||||
## Sets the maximum depth of call stack back-traces in the exit
|
|
||||||
## reason element of 'EXIT' tuples.
|
|
||||||
## The flag also limits the stacktrace depth returned by
|
|
||||||
## process_info item current_stacktrace.
|
|
||||||
##
|
|
||||||
## @doc node.backtrace_depth
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [0,1024]
|
|
||||||
## Default: 23
|
|
||||||
backtrace_depth: 23
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
##==================================================================
|
|
||||||
## Cluster
|
|
||||||
##==================================================================
|
|
||||||
cluster {
|
|
||||||
## Cluster name.
|
|
||||||
##
|
|
||||||
## @doc cluster.name
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqxcl
|
|
||||||
name: emqxcl
|
|
||||||
|
|
||||||
## Enable cluster autoheal from network partition.
|
|
||||||
##
|
|
||||||
## @doc cluster.autoheal
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: true
|
|
||||||
autoheal: true
|
|
||||||
|
|
||||||
## Autoclean down node. A down node will be removed from the cluster
|
|
||||||
## if this value > 0.
|
|
||||||
##
|
|
||||||
## @doc cluster.autoclean
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 5m
|
|
||||||
autoclean: 5m
|
|
||||||
|
|
||||||
## Node discovery strategy to join the cluster.
|
|
||||||
##
|
|
||||||
## @doc cluster.discovery_strategy
|
|
||||||
## ValueType: manual | static | mcast | dns | etcd | k8s
|
|
||||||
## - manual: Manual join command
|
|
||||||
## - static: Static node list
|
|
||||||
## - mcast: IP Multicast
|
|
||||||
## - dns: DNS A Record
|
|
||||||
## - etcd: etcd
|
|
||||||
## - k8s: Kubernetes
|
|
||||||
##
|
|
||||||
## Default: manual
|
|
||||||
discovery_strategy: manual
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Cluster using static node list
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
static {
|
|
||||||
## Node list of the cluster
|
|
||||||
##
|
|
||||||
## @doc cluster.static.seeds
|
|
||||||
## ValueType: Array<NodeName>
|
|
||||||
## Default: []
|
|
||||||
seeds: ["emqx1@127.0.0.1", "emqx2@127.0.0.1"]
|
|
||||||
}
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Cluster using IP Multicast
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
mcast {
|
|
||||||
## IP Multicast Address.
|
|
||||||
##
|
|
||||||
## @doc cluster.mcast.addr
|
|
||||||
## ValueType: IPAddress
|
|
||||||
## Default: "239.192.0.1"
|
|
||||||
addr: "239.192.0.1"
|
|
||||||
|
|
||||||
## Multicast Ports.
|
|
||||||
##
|
|
||||||
## @doc cluster.mcast.ports
|
|
||||||
## ValueType: Array<Port>
|
|
||||||
## Default: [4369, 4370]
|
|
||||||
ports: [4369, 4370]
|
|
||||||
|
|
||||||
## Multicast Iface.
|
|
||||||
##
|
|
||||||
## @doc cluster.mcast.iface
|
|
||||||
## ValueType: IPAddress
|
|
||||||
## Default: "0.0.0.0"
|
|
||||||
iface: "0.0.0.0"
|
|
||||||
|
|
||||||
## Multicast Ttl.
|
|
||||||
##
|
|
||||||
## @doc cluster.mcast.ttl
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [0,255]
|
|
||||||
## Default: 255
|
|
||||||
ttl: 255
|
|
||||||
|
|
||||||
## Multicast loop.
|
|
||||||
##
|
|
||||||
## @doc cluster.mcast.loop
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: true
|
|
||||||
loop: true
|
|
||||||
}
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Cluster using DNS A records
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
dns {
|
|
||||||
## DNS name.
|
|
||||||
##
|
|
||||||
## @doc cluster.dns.name
|
|
||||||
## ValueType: String
|
|
||||||
## Default: localhost
|
|
||||||
name: localhost
|
|
||||||
|
|
||||||
## The App name is used to build 'node.name' with IP address.
|
|
||||||
##
|
|
||||||
## @doc cluster.dns.app
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqx
|
|
||||||
app: emqx
|
|
||||||
}
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Cluster using etcd
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
etcd {
|
|
||||||
## Etcd server list, seperated by ','.
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.server
|
|
||||||
## ValueType: URL
|
|
||||||
## Required: true
|
|
||||||
server: "http://127.0.0.1:2379"
|
|
||||||
|
|
||||||
## The prefix helps build nodes path in etcd. Each node in the cluster
|
|
||||||
## will create a path in etcd: v2/keys/<prefix>/<name>/<node.name>
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.prefix
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqxcl
|
|
||||||
prefix: emqxcl
|
|
||||||
|
|
||||||
## The TTL for node's path in etcd.
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.node_ttl
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 1m
|
|
||||||
node_ttl: 1m
|
|
||||||
|
|
||||||
## Path to the file containing the user's private PEM-encoded key.
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.ssl.keyfile
|
|
||||||
## ValueType: File
|
|
||||||
## Default: "{{ platform_etc_dir }}/certs/key.pem"
|
|
||||||
ssl.keyfile: "{{ platform_etc_dir }}/certs/key.pem"
|
|
||||||
|
|
||||||
## Path to a file containing the user certificate.
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.ssl.certfile
|
|
||||||
## ValueType: File
|
|
||||||
## Default: "{{ platform_etc_dir }}/certs/cert.pem"
|
|
||||||
ssl.certfile: "{{ platform_etc_dir }}/certs/cert.pem"
|
|
||||||
|
|
||||||
## Path to the file containing PEM-encoded CA certificates. The CA certificates
|
|
||||||
## are used during server authentication and when building the client certificate chain.
|
|
||||||
##
|
|
||||||
## @doc cluster.etcd.ssl.cacertfile
|
|
||||||
## ValueType: File
|
|
||||||
## Default: "{{ platform_etc_dir }}/certs/cacert.pem"
|
|
||||||
ssl.cacertfile: "{{ platform_etc_dir }}/certs/cacert.pem"
|
|
||||||
}
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Cluster using Kubernetes
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
k8s {
|
|
||||||
## Kubernetes API server list, seperated by ','.
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.apiserver
|
|
||||||
## ValueType: URL
|
|
||||||
## Required: true
|
|
||||||
apiserver: "http://10.110.111.204:8080"
|
|
||||||
|
|
||||||
## The service name helps lookup EMQ nodes in the cluster.
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.service_name
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqx
|
|
||||||
service_name: emqx
|
|
||||||
|
|
||||||
## The address type is used to extract host from k8s service.
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.address_type
|
|
||||||
## ValueType: ip | dns | hostname
|
|
||||||
## Default: ip
|
|
||||||
address_type: ip
|
|
||||||
|
|
||||||
## The app name helps build 'node.name'.
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.app_name
|
|
||||||
## ValueType: String
|
|
||||||
## Default: emqx
|
|
||||||
app_name: emqx
|
|
||||||
|
|
||||||
## The suffix added to dns and hostname get from k8s service
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.suffix
|
|
||||||
## ValueType: String
|
|
||||||
## Default: "pod.local"
|
|
||||||
suffix: "pod.local"
|
|
||||||
|
|
||||||
## Kubernetes Namespace
|
|
||||||
##
|
|
||||||
## @doc cluster.k8s.namespace
|
|
||||||
## ValueType: String
|
|
||||||
## Default: default
|
|
||||||
namespace: default
|
|
||||||
}
|
|
||||||
|
|
||||||
db_backend: mnesia
|
|
||||||
|
|
||||||
rlog: {
|
|
||||||
# role: core
|
|
||||||
# core_nodes: []
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
##==================================================================
|
|
||||||
## Log
|
|
||||||
##==================================================================
|
|
||||||
log {
|
|
||||||
## The primary log level
|
|
||||||
##
|
|
||||||
## - all the log messages with levels lower than this level will
|
|
||||||
## be dropped.
|
|
||||||
## - all the log messages with levels higher than this level will
|
|
||||||
## go into the log handlers. The handlers then decide to log it
|
|
||||||
## out or drop it according to the level setting of the handler.
|
|
||||||
##
|
|
||||||
## Note: Only the messages with severity level higher than or
|
|
||||||
## equal to this level will be logged.
|
|
||||||
##
|
|
||||||
## @doc log.primary_level
|
|
||||||
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
|
||||||
## Default: warning
|
|
||||||
primary_level: warning
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## The console log handler send log messages to emqx console
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## Log to single line
|
|
||||||
## @doc log.console_handler.enable
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: false
|
|
||||||
console_handler.enable: false
|
|
||||||
|
|
||||||
## The log level of this handler
|
|
||||||
## All the log messages with levels lower than this level will
|
|
||||||
## be dropped.
|
|
||||||
##
|
|
||||||
## @doc log.console_handler.level
|
|
||||||
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
|
||||||
## Default: warning
|
|
||||||
console_handler.level: warning
|
|
||||||
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## The file log handlers send log messages to files
|
|
||||||
##----------------------------------------------------------------
|
|
||||||
## file_handlers.<name>
|
|
||||||
file_handlers.emqx_log: {
|
|
||||||
## The log level filter of this handler
|
|
||||||
## All the log messages with levels lower than this level will
|
|
||||||
## be dropped.
|
|
||||||
##
|
|
||||||
## @doc log.file_handlers.<name>.level
|
|
||||||
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
|
||||||
## Default: warning
|
|
||||||
level: warning
|
|
||||||
|
|
||||||
## The log file for specified level.
|
|
||||||
##
|
|
||||||
## If `rotation` is disabled, this is the file of the log files.
|
|
||||||
##
|
|
||||||
## If `rotation` is enabled, this is the base name of the files.
|
|
||||||
## Each file in a rotated log is named <base_name>.N, where N is an integer.
|
|
||||||
##
|
|
||||||
## Note: Log files for a specific log level will only contain all the logs
|
|
||||||
## that higher than or equal to that level
|
|
||||||
##
|
|
||||||
## @doc log.file_handlers.<name>.file
|
|
||||||
## ValueType: File
|
|
||||||
## Required: true
|
|
||||||
file: "{{ platform_log_dir }}/emqx.log"
|
|
||||||
|
|
||||||
## Enables the log rotation.
|
|
||||||
## With this enabled, new log files will be created when the current
|
|
||||||
## log file is full, max to `rotation_count` files will be created.
|
|
||||||
##
|
|
||||||
## @doc log.file_handlers.<name>.rotation.enable
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: true
|
|
||||||
rotation.enable: true
|
|
||||||
|
|
||||||
## Maximum rotation count of log files.
|
|
||||||
##
|
|
||||||
## @doc log.file_handlers.<name>.rotation.count
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [1, 2048]
|
|
||||||
## Default: 10
|
|
||||||
rotation.count: 10
|
|
||||||
|
|
||||||
## Maximum size of each log file.
|
|
||||||
##
|
|
||||||
## If the max_size reached and `rotation` is disabled, the handler
|
|
||||||
## will stop sending log messages, if the `rotation` is enabled,
|
|
||||||
## the file rotates.
|
|
||||||
##
|
|
||||||
## @doc log.file_handlers.<name>.max_size
|
|
||||||
## ValueType: Size | infinity
|
|
||||||
## Default: 10MB
|
|
||||||
max_size: 10MB
|
|
||||||
}
|
|
||||||
|
|
||||||
## file_handlers.<name>
|
|
||||||
##
|
|
||||||
## You could also create multiple file handlers for different
|
|
||||||
## log level for example:
|
|
||||||
file_handlers.emqx_error_log: {
|
|
||||||
level: error
|
|
||||||
file: "{{ platform_log_dir }}/error.log"
|
|
||||||
}
|
|
||||||
|
|
||||||
## Timezone offset to display in logs
|
|
||||||
##
|
|
||||||
## @doc log.time_offset
|
|
||||||
## ValueType: system | utc | String
|
|
||||||
## - "system" use system zone
|
|
||||||
## - "utc" for Universal Coordinated Time (UTC)
|
|
||||||
## - "+hh:mm" or "-hh:mm" for a specified offset
|
|
||||||
## Default: system
|
|
||||||
time_offset: system
|
|
||||||
|
|
||||||
## Limits the total number of characters printed for each log event.
|
|
||||||
##
|
|
||||||
## @doc log.chars_limit
|
|
||||||
## ValueType: Integer | infinity
|
|
||||||
## Range: [0, infinity)
|
|
||||||
## Default: infinity
|
|
||||||
chars_limit: infinity
|
|
||||||
|
|
||||||
## Maximum depth for Erlang term log formatting
|
|
||||||
## and Erlang process message queue inspection.
|
|
||||||
##
|
|
||||||
## @doc log.max_depth
|
|
||||||
## ValueType: Integer | infinity
|
|
||||||
## Default: 80
|
|
||||||
max_depth: 80
|
|
||||||
|
|
||||||
## Log formatter
|
|
||||||
## @doc log.formatter
|
|
||||||
## ValueType: text | json
|
|
||||||
## Default: text
|
|
||||||
formatter: text
|
|
||||||
|
|
||||||
## Log to single line
|
|
||||||
## @doc log.single_line
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: true
|
|
||||||
single_line: true
|
|
||||||
|
|
||||||
## The max allowed queue length before switching to sync mode.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. If the message queue grows
|
|
||||||
## larger than this value the handler switches from anync to sync mode.
|
|
||||||
##
|
|
||||||
## @doc log.sync_mode_qlen
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [0, ${log.drop_mode_qlen}]
|
|
||||||
## Default: 100
|
|
||||||
sync_mode_qlen: 100
|
|
||||||
|
|
||||||
## The max allowed queue length before switching to drop mode.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. When the message queue grows
|
|
||||||
## larger than this threshold, the handler switches to a mode in which
|
|
||||||
## it drops all new events that senders want to log.
|
|
||||||
##
|
|
||||||
## @doc log.drop_mode_qlen
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [${log.sync_mode_qlen}, ${log.flush_qlen}]
|
|
||||||
## Default: 3000
|
|
||||||
drop_mode_qlen: 3000
|
|
||||||
|
|
||||||
## The max allowed queue length before switching to flush mode.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. If the length of the message queue
|
|
||||||
## grows larger than this threshold, a flush (delete) operation takes place.
|
|
||||||
## To flush events, the handler discards the messages in the message queue
|
|
||||||
## by receiving them in a loop without logging.
|
|
||||||
##
|
|
||||||
## @doc log.flush_qlen
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [${log.drop_mode_qlen}, infinity)
|
|
||||||
## Default: 8000
|
|
||||||
flush_qlen: 8000
|
|
||||||
|
|
||||||
## Kill the log handler when it gets overloaded.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. It is possible that a handler,
|
|
||||||
## even if it can successfully manage peaks of high load without crashing,
|
|
||||||
## can build up a large message queue, or use a large amount of memory.
|
|
||||||
## We could kill the log handler in these cases and restart it after a
|
|
||||||
## few seconds.
|
|
||||||
##
|
|
||||||
## @doc log.overload_kill.enable
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: true
|
|
||||||
overload_kill.enable: true
|
|
||||||
|
|
||||||
## The max allowed queue length before killing the log hanlder.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. This is the maximum allowed queue
|
|
||||||
## length. If the message queue grows larger than this, the handler
|
|
||||||
## process is terminated.
|
|
||||||
##
|
|
||||||
## @doc log.overload_kill.qlen
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [0, 1048576]
|
|
||||||
## Default: 20000
|
|
||||||
overload_kill.qlen: 20000
|
|
||||||
|
|
||||||
## The max allowed memory size before killing the log hanlder.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. This is the maximum memory size
|
|
||||||
## that the handler process is allowed to use. If the handler grows
|
|
||||||
## larger than this, the process is terminated.
|
|
||||||
##
|
|
||||||
## @doc log.overload_kill.mem_size
|
|
||||||
## ValueType: Size
|
|
||||||
## Default: 30MB
|
|
||||||
overload_kill.mem_size: 30MB
|
|
||||||
|
|
||||||
## Restart the log hanlder after some seconds.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. If the handler is terminated,
|
|
||||||
## it restarts automatically after a delay specified in seconds.
|
|
||||||
##
|
|
||||||
## @doc log.overload_kill.restart_after
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 5s
|
|
||||||
overload_kill.restart_after: 5s
|
|
||||||
|
|
||||||
## Controlling Bursts of Log Requests.
|
|
||||||
##
|
|
||||||
## Log overload protection parameter. Large bursts of log events - many
|
|
||||||
## events received by the handler under a short period of time - can
|
|
||||||
## potentially cause problems. By specifying the maximum number of events
|
|
||||||
## to be handled within a certain time frame, the handler can avoid
|
|
||||||
## choking the log with massive amounts of printouts.
|
|
||||||
##
|
|
||||||
## Note that there would be no warning if any messages were
|
|
||||||
## dropped because of burst control.
|
|
||||||
##
|
|
||||||
## @doc log.burst_limit.enable
|
|
||||||
## ValueType: Boolean
|
|
||||||
## Default: false
|
|
||||||
burst_limit.enable: false
|
|
||||||
|
|
||||||
## This config controls the maximum number of events to handle within
|
|
||||||
## a time frame. After the limit is reached, successive events are
|
|
||||||
## dropped until the end of the time frame defined by `window_time`.
|
|
||||||
##
|
|
||||||
## @doc log.burst_limit.max_count
|
|
||||||
## ValueType: Integer
|
|
||||||
## Default: 10000
|
|
||||||
burst_limit.max_count: 10000
|
|
||||||
|
|
||||||
## See the previous description of burst_limit_max_count.
|
|
||||||
##
|
|
||||||
## @doc log.burst_limit.window_time
|
|
||||||
## ValueType: duration
|
|
||||||
## Default: 1s
|
|
||||||
burst_limit.window_time: 1s
|
|
||||||
}
|
|
||||||
|
|
||||||
##==================================================================
|
|
||||||
## RPC
|
|
||||||
##==================================================================
|
|
||||||
rpc {
|
|
||||||
## RPC Mode.
|
|
||||||
##
|
|
||||||
## @doc rpc.mode
|
|
||||||
## ValueType: sync | async
|
|
||||||
## Default: async
|
|
||||||
mode: async
|
|
||||||
|
|
||||||
## Max batch size of async RPC requests.
|
|
||||||
##
|
|
||||||
## NOTE: RPC batch won't work when rpc.mode = sync
|
|
||||||
## Zero value disables rpc batching.
|
|
||||||
##
|
|
||||||
## @doc rpc.async_batch_size
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [0, 1048576]
|
|
||||||
## Default: 0
|
|
||||||
async_batch_size: 256
|
|
||||||
|
|
||||||
## RPC port discovery
|
|
||||||
##
|
|
||||||
## The strategy for discovering the RPC listening port of
|
|
||||||
## other nodes.
|
|
||||||
##
|
|
||||||
## @doc cluster.discovery_strategy
|
|
||||||
## ValueType: manual | stateless
|
|
||||||
## - manual: discover ports by `tcp_server_port`.
|
|
||||||
## - stateless: discover ports in a stateless manner.
|
|
||||||
## If node name is `emqx<N>@127.0.0.1`, where the `<N>` is
|
|
||||||
## an integer, then the listening port will be `5370 + <N>`
|
|
||||||
##
|
|
||||||
## Default: `stateless`.
|
|
||||||
port_discovery: stateless
|
|
||||||
|
|
||||||
## TCP server port for RPC.
|
|
||||||
##
|
|
||||||
## Only takes effect when `rpc.port_discovery` = `manual`.
|
|
||||||
##
|
|
||||||
## @doc rpc.tcp_server_port
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [1024-65535]
|
|
||||||
## Defaults: 5369
|
|
||||||
tcp_server_port: 5369
|
|
||||||
|
|
||||||
## Number of outgoing RPC connections.
|
|
||||||
##
|
|
||||||
## Set this to 1 to keep the message order sent from the same
|
|
||||||
## client.
|
|
||||||
##
|
|
||||||
## @doc rpc.tcp_client_num
|
|
||||||
## ValueType: Integer
|
|
||||||
## Range: [1, 256]
|
|
||||||
## Defaults: 1
|
|
||||||
tcp_client_num: 1
|
|
||||||
|
|
||||||
## RCP Client connect timeout.
|
|
||||||
##
|
|
||||||
## @doc rpc.connect_timeout
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 5s
|
|
||||||
connect_timeout: 5s
|
|
||||||
|
|
||||||
## TCP send timeout of RPC client and server.
|
|
||||||
##
|
|
||||||
## @doc rpc.send_timeout
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 5s
|
|
||||||
send_timeout: 5s
|
|
||||||
|
|
||||||
## Authentication timeout
|
|
||||||
##
|
|
||||||
## @doc rpc.authentication_timeout
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 5s
|
|
||||||
authentication_timeout: 5s
|
|
||||||
|
|
||||||
## Default receive timeout for call() functions
|
|
||||||
##
|
|
||||||
## @doc rpc.call_receive_timeout
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 15s
|
|
||||||
call_receive_timeout: 15s
|
|
||||||
|
|
||||||
## Socket idle keepalive.
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_keepalive_idle
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 900s
|
|
||||||
socket_keepalive_idle: 900s
|
|
||||||
|
|
||||||
## TCP Keepalive probes interval.
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_keepalive_interval
|
|
||||||
## ValueType: Duration
|
|
||||||
## Default: 75s
|
|
||||||
socket_keepalive_interval: 75s
|
|
||||||
|
|
||||||
## Probes lost to close the connection
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_keepalive_count
|
|
||||||
## ValueType: Integer
|
|
||||||
## Default: 9
|
|
||||||
socket_keepalive_count: 9
|
|
||||||
|
|
||||||
## Size of TCP send buffer.
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_sndbuf
|
|
||||||
## ValueType: Size
|
|
||||||
## Default: 1MB
|
|
||||||
socket_sndbuf: 1MB
|
|
||||||
|
|
||||||
## Size of TCP receive buffer.
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_recbuf
|
|
||||||
## ValueType: Size
|
|
||||||
## Default: 1MB
|
|
||||||
socket_recbuf: 1MB
|
|
||||||
|
|
||||||
## Size of user-level software socket buffer.
|
|
||||||
##
|
|
||||||
## @doc rpc.socket_buffer
|
|
||||||
## ValueType: Size
|
|
||||||
## Default: 1MB
|
|
||||||
socket_buffer: 1MB
|
|
||||||
}
|
|
||||||
|
|
||||||
##==================================================================
|
##==================================================================
|
||||||
## Broker
|
## Broker
|
||||||
##==================================================================
|
##==================================================================
|
||||||
|
|
|
@ -24,7 +24,6 @@
|
||||||
|
|
||||||
-include_lib("typerefl/include/types.hrl").
|
-include_lib("typerefl/include/types.hrl").
|
||||||
|
|
||||||
-type log_level() :: debug | info | notice | warning | error | critical | alert | emergency | all.
|
|
||||||
-type duration() :: integer().
|
-type duration() :: integer().
|
||||||
-type duration_s() :: integer().
|
-type duration_s() :: integer().
|
||||||
-type duration_ms() :: integer().
|
-type duration_ms() :: integer().
|
||||||
|
@ -60,179 +59,18 @@
|
||||||
|
|
||||||
-behaviour(hocon_schema).
|
-behaviour(hocon_schema).
|
||||||
|
|
||||||
-reflect_type([ log_level/0, duration/0, duration_s/0, duration_ms/0,
|
-reflect_type([ duration/0, duration_s/0, duration_ms/0,
|
||||||
bytesize/0, wordsize/0, percent/0, file/0,
|
bytesize/0, wordsize/0, percent/0, file/0,
|
||||||
comma_separated_list/0, bar_separated_list/0, ip_port/0,
|
comma_separated_list/0, bar_separated_list/0, ip_port/0,
|
||||||
cipher/0,
|
cipher/0,
|
||||||
comma_separated_atoms/0]).
|
comma_separated_atoms/0]).
|
||||||
|
|
||||||
-export([structs/0, fields/1, translations/0, translation/1]).
|
-export([structs/0, fields/1]).
|
||||||
-export([t/1, t/3, t/4, ref/1]).
|
-export([t/1, t/3, t/4, ref/1]).
|
||||||
-export([conf_get/2, conf_get/3, keys/2, filter/1]).
|
-export([conf_get/2, conf_get/3, keys/2, filter/1]).
|
||||||
-export([ssl/1]).
|
-export([ssl/1]).
|
||||||
|
|
||||||
%% will be used by emqx_ct_helper to find the dependent apps
|
structs() -> ["zones", "listeners", "broker", "plugins", "sysmon", "alarm"].
|
||||||
-export([includes/0, extra_schema_fields/1]).
|
|
||||||
|
|
||||||
structs() -> ["cluster", "node", "rpc", "log",
|
|
||||||
"zones", "listeners", "broker",
|
|
||||||
"plugins", "sysmon", "alarm"]
|
|
||||||
++ ?MODULE:includes().
|
|
||||||
|
|
||||||
-ifndef(EMQX_EXT_SCHEMAS).
|
|
||||||
includes() -> [].
|
|
||||||
-else.
|
|
||||||
includes() ->
|
|
||||||
[FieldName || {FieldName, _SchemaMod} <- ?EMQX_EXT_SCHEMAS].
|
|
||||||
-endif.
|
|
||||||
|
|
||||||
fields("cluster") ->
|
|
||||||
[ {"name", t(atom(), "ekka.cluster_name", emqxcl)}
|
|
||||||
, {"discovery_strategy", t(union([manual, static, mcast, dns, etcd, k8s]),
|
|
||||||
undefined, manual)}
|
|
||||||
, {"autoclean", t(duration(), "ekka.cluster_autoclean", "5m")}
|
|
||||||
, {"autoheal", t(boolean(), "ekka.cluster_autoheal", true)}
|
|
||||||
, {"static", ref("static")}
|
|
||||||
, {"mcast", ref("mcast")}
|
|
||||||
, {"proto_dist", t(union([inet_tcp, inet6_tcp, inet_tls]), "ekka.proto_dist", inet_tcp)}
|
|
||||||
, {"dns", ref("dns")}
|
|
||||||
, {"etcd", ref("etcd")}
|
|
||||||
, {"k8s", ref("k8s")}
|
|
||||||
, {"db_backend", t(union([mnesia, rlog]), "ekka.db_backend", mnesia)}
|
|
||||||
, {"rlog", ref("rlog")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("static") ->
|
|
||||||
[ {"seeds", t(hoconsc:array(string()), undefined, [])}];
|
|
||||||
|
|
||||||
fields("mcast") ->
|
|
||||||
[ {"addr", t(string(), undefined, "239.192.0.1")}
|
|
||||||
, {"ports", t(hoconsc:array(integer()), undefined, [4369, 4370])}
|
|
||||||
, {"iface", t(string(), undefined, "0.0.0.0")}
|
|
||||||
, {"ttl", t(range(0, 255), undefined, 255)}
|
|
||||||
, {"loop", t(boolean(), undefined, true)}
|
|
||||||
, {"sndbuf", t(bytesize(), undefined, "16KB")}
|
|
||||||
, {"recbuf", t(bytesize(), undefined, "16KB")}
|
|
||||||
, {"buffer", t(bytesize(), undefined, "32KB")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("dns") ->
|
|
||||||
[ {"name", t(string(), undefined, "localhost")}
|
|
||||||
, {"app", t(string(), undefined, "emqx")}];
|
|
||||||
|
|
||||||
fields("etcd") ->
|
|
||||||
[ {"server", t(comma_separated_list())}
|
|
||||||
, {"prefix", t(string(), undefined, "emqxcl")}
|
|
||||||
, {"node_ttl", t(duration(), undefined, "1m")}
|
|
||||||
, {"ssl", ref("etcd_ssl")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("etcd_ssl") ->
|
|
||||||
ssl(#{});
|
|
||||||
|
|
||||||
fields("k8s") ->
|
|
||||||
[ {"apiserver", t(string())}
|
|
||||||
, {"service_name", t(string(), undefined, "emqx")}
|
|
||||||
, {"address_type", t(union([ip, dns, hostname]))}
|
|
||||||
, {"app_name", t(string(), undefined, "emqx")}
|
|
||||||
, {"namespace", t(string(), undefined, "default")}
|
|
||||||
, {"suffix", t(string(), undefined, "pod.local")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("rlog") ->
|
|
||||||
[ {"role", t(union([core, replicant]), "ekka.node_role", core)}
|
|
||||||
, {"core_nodes", t(comma_separated_atoms(), "ekka.core_nodes", [])}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("node") ->
|
|
||||||
[ {"name", hoconsc:t(string(), #{default => "emqx@127.0.0.1",
|
|
||||||
override_env => "EMQX_NODE_NAME"
|
|
||||||
})}
|
|
||||||
, {"cookie", hoconsc:t(string(), #{mapping => "vm_args.-setcookie",
|
|
||||||
default => "emqxsecretcookie",
|
|
||||||
sensitive => true,
|
|
||||||
override_env => "EMQX_NODE_COOKIE"
|
|
||||||
})}
|
|
||||||
, {"data_dir", t(string())}
|
|
||||||
, {"config_files", t(comma_separated_list())}
|
|
||||||
, {"global_gc_interval", t(duration(), undefined, "15m")}
|
|
||||||
, {"crash_dump_dir", t(file(), "vm_args.-env ERL_CRASH_DUMP", undefined)}
|
|
||||||
, {"dist_net_ticktime", t(duration(), "vm_args.-kernel net_ticktime", "2m")}
|
|
||||||
, {"dist_listen_min", t(range(1024, 65535), "kernel.inet_dist_listen_min", 6369)}
|
|
||||||
, {"dist_listen_max", t(range(1024, 65535), "kernel.inet_dist_listen_max", 6369)}
|
|
||||||
, {"backtrace_depth", t(integer(), "emqx_machine.backtrace_depth", 23)}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("rpc") ->
|
|
||||||
[ {"mode", t(union(sync, async), undefined, async)}
|
|
||||||
, {"async_batch_size", t(integer(), "gen_rpc.max_batch_size", 256)}
|
|
||||||
, {"port_discovery",t(union(manual, stateless), "gen_rpc.port_discovery", stateless)}
|
|
||||||
, {"tcp_server_port", t(integer(), "gen_rpc.tcp_server_port", 5369)}
|
|
||||||
, {"tcp_client_num", t(range(1, 256), undefined, 1)}
|
|
||||||
, {"connect_timeout", t(duration(), "gen_rpc.connect_timeout", "5s")}
|
|
||||||
, {"send_timeout", t(duration(), "gen_rpc.send_timeout", "5s")}
|
|
||||||
, {"authentication_timeout", t(duration(), "gen_rpc.authentication_timeout", "5s")}
|
|
||||||
, {"call_receive_timeout", t(duration(), "gen_rpc.call_receive_timeout", "15s")}
|
|
||||||
, {"socket_keepalive_idle", t(duration_s(), "gen_rpc.socket_keepalive_idle", "7200s")}
|
|
||||||
, {"socket_keepalive_interval", t(duration_s(), "gen_rpc.socket_keepalive_interval", "75s")}
|
|
||||||
, {"socket_keepalive_count", t(integer(), "gen_rpc.socket_keepalive_count", 9)}
|
|
||||||
, {"socket_sndbuf", t(bytesize(), "gen_rpc.socket_sndbuf", "1MB")}
|
|
||||||
, {"socket_recbuf", t(bytesize(), "gen_rpc.socket_recbuf", "1MB")}
|
|
||||||
, {"socket_buffer", t(bytesize(), "gen_rpc.socket_buffer", "1MB")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("log") ->
|
|
||||||
[ {"primary_level", t(log_level(), undefined, warning)}
|
|
||||||
, {"console_handler", ref("console_handler")}
|
|
||||||
, {"file_handlers", ref("file_handlers")}
|
|
||||||
, {"time_offset", t(string(), undefined, "system")}
|
|
||||||
, {"chars_limit", maybe_infinity(range(1, inf))}
|
|
||||||
, {"supervisor_reports", t(union([error, progress]), undefined, error)}
|
|
||||||
, {"max_depth", t(union([infinity, integer()]),
|
|
||||||
"kernel.error_logger_format_depth", 80)}
|
|
||||||
, {"formatter", t(union([text, json]), undefined, text)}
|
|
||||||
, {"single_line", t(boolean(), undefined, true)}
|
|
||||||
, {"sync_mode_qlen", t(integer(), undefined, 100)}
|
|
||||||
, {"drop_mode_qlen", t(integer(), undefined, 3000)}
|
|
||||||
, {"flush_qlen", t(integer(), undefined, 8000)}
|
|
||||||
, {"overload_kill", ref("log_overload_kill")}
|
|
||||||
, {"burst_limit", ref("log_burst_limit")}
|
|
||||||
, {"error_logger", t(atom(), "kernel.error_logger", silent)}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("console_handler") ->
|
|
||||||
[ {"enable", t(boolean(), undefined, false)}
|
|
||||||
, {"level", t(log_level(), undefined, warning)}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("file_handlers") ->
|
|
||||||
[ {"$name", ref("log_file_handler")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("log_file_handler") ->
|
|
||||||
[ {"level", t(log_level(), undefined, warning)}
|
|
||||||
, {"file", t(file(), undefined, undefined)}
|
|
||||||
, {"rotation", ref("log_rotation")}
|
|
||||||
, {"max_size", maybe_infinity(bytesize(), "10MB")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("log_rotation") ->
|
|
||||||
[ {"enable", t(boolean(), undefined, true)}
|
|
||||||
, {"count", t(range(1, 2048), undefined, 10)}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("log_overload_kill") ->
|
|
||||||
[ {"enable", t(boolean(), undefined, true)}
|
|
||||||
, {"mem_size", t(bytesize(), undefined, "30MB")}
|
|
||||||
, {"qlen", t(integer(), undefined, 20000)}
|
|
||||||
, {"restart_after", t(union(duration(), infinity), undefined, "5s")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("log_burst_limit") ->
|
|
||||||
[ {"enable", t(boolean(), undefined, true)}
|
|
||||||
, {"max_count", t(integer(), undefined, 10000)}
|
|
||||||
, {"window_time", t(duration(), undefined, "1s")}
|
|
||||||
];
|
|
||||||
|
|
||||||
fields("stats") ->
|
fields("stats") ->
|
||||||
[ {"enable", t(boolean(), undefined, true)}
|
[ {"enable", t(boolean(), undefined, true)}
|
||||||
|
@ -480,20 +318,7 @@ fields("alarm") ->
|
||||||
[ {"actions", t(hoconsc:array(atom()), undefined, [log, publish])}
|
[ {"actions", t(hoconsc:array(atom()), undefined, [log, publish])}
|
||||||
, {"size_limit", t(integer(), undefined, 1000)}
|
, {"size_limit", t(integer(), undefined, 1000)}
|
||||||
, {"validity_period", t(duration(), undefined, "24h")}
|
, {"validity_period", t(duration(), undefined, "24h")}
|
||||||
];
|
].
|
||||||
|
|
||||||
fields(FieldName) ->
|
|
||||||
?MODULE:extra_schema_fields(FieldName).
|
|
||||||
|
|
||||||
-ifndef(EMQX_EXT_SCHEMAS).
|
|
||||||
%% Function extra_schema_fields/1 only terminates with explicit exception
|
|
||||||
-dialyzer([{nowarn_function, [extra_schema_fields/1]}]).
|
|
||||||
extra_schema_fields(FieldName) -> error({unknown_field, FieldName}).
|
|
||||||
-else.
|
|
||||||
extra_schema_fields(FieldName) ->
|
|
||||||
{_, Mod} = lists:keyfind(FieldName, 1, ?EMQX_EXT_SCHEMAS),
|
|
||||||
Mod:fields(FieldName).
|
|
||||||
-endif.
|
|
||||||
|
|
||||||
mqtt_listener() ->
|
mqtt_listener() ->
|
||||||
base_listener() ++
|
base_listener() ++
|
||||||
|
@ -509,117 +334,6 @@ base_listener() ->
|
||||||
, {"rate_limit", ref("rate_limit")}
|
, {"rate_limit", ref("rate_limit")}
|
||||||
].
|
].
|
||||||
|
|
||||||
translations() -> ["ekka", "kernel", "emqx"].
|
|
||||||
|
|
||||||
translation("ekka") ->
|
|
||||||
[ {"cluster_discovery", fun tr_cluster__discovery/1}];
|
|
||||||
|
|
||||||
translation("kernel") ->
|
|
||||||
[ {"logger_level", fun tr_logger_level/1}
|
|
||||||
, {"logger", fun tr_logger/1}];
|
|
||||||
|
|
||||||
translation("emqx") ->
|
|
||||||
[ {"config_files", fun tr_config_files/1}
|
|
||||||
].
|
|
||||||
|
|
||||||
tr_config_files(Conf) ->
|
|
||||||
case conf_get("emqx.config_files", Conf) of
|
|
||||||
[_ | _] = Files ->
|
|
||||||
Files;
|
|
||||||
_ ->
|
|
||||||
case os:getenv("RUNNER_ETC_DIR") of
|
|
||||||
false ->
|
|
||||||
[filename:join([code:lib_dir(emqx), "etc", "emqx.conf"])];
|
|
||||||
Dir ->
|
|
||||||
[filename:join([Dir, "emqx.conf"])]
|
|
||||||
end
|
|
||||||
end.
|
|
||||||
|
|
||||||
tr_cluster__discovery(Conf) ->
|
|
||||||
Strategy = conf_get("cluster.discovery_strategy", Conf),
|
|
||||||
{Strategy, filter(options(Strategy, Conf))}.
|
|
||||||
|
|
||||||
tr_logger_level(Conf) -> conf_get("log.primary_level", Conf).
|
|
||||||
|
|
||||||
tr_logger(Conf) ->
|
|
||||||
CharsLimit = case conf_get("log.chars_limit", Conf) of
|
|
||||||
infinity -> unlimited;
|
|
||||||
V -> V
|
|
||||||
end,
|
|
||||||
SingleLine = conf_get("log.single_line", Conf),
|
|
||||||
FmtName = conf_get("log.formatter", Conf),
|
|
||||||
Formatter = formatter(FmtName, CharsLimit, SingleLine),
|
|
||||||
BasicConf = #{
|
|
||||||
sync_mode_qlen => conf_get("log.sync_mode_qlen", Conf),
|
|
||||||
drop_mode_qlen => conf_get("log.drop_mode_qlen", Conf),
|
|
||||||
flush_qlen => conf_get("log.flush_qlen", Conf),
|
|
||||||
overload_kill_enable => conf_get("log.overload_kill.enable", Conf),
|
|
||||||
overload_kill_qlen => conf_get("log.overload_kill.qlen", Conf),
|
|
||||||
overload_kill_mem_size => conf_get("log.overload_kill.mem_size", Conf),
|
|
||||||
overload_kill_restart_after => conf_get("log.overload_kill.restart_after", Conf),
|
|
||||||
burst_limit_enable => conf_get("log.burst_limit.enable", Conf),
|
|
||||||
burst_limit_max_count => conf_get("log.burst_limit.max_count", Conf),
|
|
||||||
burst_limit_window_time => conf_get("log.burst_limit.window_time", Conf)
|
|
||||||
},
|
|
||||||
Filters = case conf_get("log.supervisor_reports", Conf) of
|
|
||||||
error -> [{drop_progress_reports, {fun logger_filters:progress/2, stop}}];
|
|
||||||
progress -> []
|
|
||||||
end,
|
|
||||||
%% For the default logger that outputs to console
|
|
||||||
ConsoleHandler =
|
|
||||||
case conf_get("log.console_handler.enable", Conf) of
|
|
||||||
true ->
|
|
||||||
[{handler, console, logger_std_h, #{
|
|
||||||
level => conf_get("log.console_handler.level", Conf),
|
|
||||||
config => BasicConf#{type => standard_io},
|
|
||||||
formatter => Formatter,
|
|
||||||
filters => Filters
|
|
||||||
}}];
|
|
||||||
false -> []
|
|
||||||
end,
|
|
||||||
%% For the file logger
|
|
||||||
FileHandlers =
|
|
||||||
[{handler, binary_to_atom(HandlerName, latin1), logger_disk_log_h, #{
|
|
||||||
level => conf_get("level", SubConf),
|
|
||||||
config => BasicConf#{
|
|
||||||
type => case conf_get("rotation.enable", SubConf) of
|
|
||||||
true -> wrap;
|
|
||||||
_ -> halt
|
|
||||||
end,
|
|
||||||
file => conf_get("file", SubConf),
|
|
||||||
max_no_files => conf_get("rotation.count", SubConf),
|
|
||||||
max_no_bytes => conf_get("max_size", SubConf)
|
|
||||||
},
|
|
||||||
formatter => Formatter,
|
|
||||||
filters => Filters,
|
|
||||||
filesync_repeat_interval => no_repeat
|
|
||||||
}}
|
|
||||||
|| {HandlerName, SubConf} <- maps:to_list(conf_get("log.file_handlers", Conf, #{}))],
|
|
||||||
|
|
||||||
[{handler, default, undefined}] ++ ConsoleHandler ++ FileHandlers.
|
|
||||||
|
|
||||||
%% helpers
|
|
||||||
formatter(json, CharsLimit, SingleLine) ->
|
|
||||||
{emqx_logger_jsonfmt,
|
|
||||||
#{chars_limit => CharsLimit,
|
|
||||||
single_line => SingleLine
|
|
||||||
}};
|
|
||||||
formatter(text, CharsLimit, SingleLine) ->
|
|
||||||
{emqx_logger_textfmt,
|
|
||||||
#{template =>
|
|
||||||
[time," [",level,"] ",
|
|
||||||
{clientid,
|
|
||||||
[{peername,
|
|
||||||
[clientid,"@",peername," "],
|
|
||||||
[clientid, " "]}],
|
|
||||||
[{peername,
|
|
||||||
[peername," "],
|
|
||||||
[]}]},
|
|
||||||
msg,"\n"],
|
|
||||||
chars_limit => CharsLimit,
|
|
||||||
single_line => SingleLine
|
|
||||||
}}.
|
|
||||||
|
|
||||||
%% utils
|
%% utils
|
||||||
-spec(conf_get(string() | [string()], hocon:config()) -> term()).
|
-spec(conf_get(string() | [string()], hocon:config()) -> term()).
|
||||||
conf_get(Key, Conf) ->
|
conf_get(Key, Conf) ->
|
||||||
|
@ -740,8 +454,7 @@ t(Type, Mapping, Default, OverrideEnv, Validator) ->
|
||||||
, validator => Validator
|
, validator => Validator
|
||||||
}).
|
}).
|
||||||
|
|
||||||
ref(Field) ->
|
ref(Field) -> hoconsc:t(hoconsc:ref(?MODULE, Field)).
|
||||||
fun (type) -> Field; (_) -> undefined end.
|
|
||||||
|
|
||||||
maybe_disabled(T) ->
|
maybe_disabled(T) ->
|
||||||
maybe_sth(disabled, T, disabled).
|
maybe_sth(disabled, T, disabled).
|
||||||
|
@ -817,37 +530,6 @@ to_erl_cipher_suite(Str) ->
|
||||||
Cipher -> Cipher
|
Cipher -> Cipher
|
||||||
end.
|
end.
|
||||||
|
|
||||||
options(static, Conf) ->
|
|
||||||
[{seeds, [to_atom(S) || S <- conf_get("cluster.static.seeds", Conf, [])]}];
|
|
||||||
options(mcast, Conf) ->
|
|
||||||
{ok, Addr} = inet:parse_address(conf_get("cluster.mcast.addr", Conf)),
|
|
||||||
{ok, Iface} = inet:parse_address(conf_get("cluster.mcast.iface", Conf)),
|
|
||||||
Ports = conf_get("cluster.mcast.ports", Conf),
|
|
||||||
[{addr, Addr}, {ports, Ports}, {iface, Iface},
|
|
||||||
{ttl, conf_get("cluster.mcast.ttl", Conf, 1)},
|
|
||||||
{loop, conf_get("cluster.mcast.loop", Conf, true)}];
|
|
||||||
options(dns, Conf) ->
|
|
||||||
[{name, conf_get("cluster.dns.name", Conf)},
|
|
||||||
{app, conf_get("cluster.dns.app", Conf)}];
|
|
||||||
options(etcd, Conf) ->
|
|
||||||
Namespace = "cluster.etcd.ssl",
|
|
||||||
SslOpts = fun(C) ->
|
|
||||||
Options = keys(Namespace, C),
|
|
||||||
lists:map(fun(Key) -> {to_atom(Key), conf_get([Namespace, Key], Conf)} end, Options) end,
|
|
||||||
[{server, conf_get("cluster.etcd.server", Conf)},
|
|
||||||
{prefix, conf_get("cluster.etcd.prefix", Conf, "emqxcl")},
|
|
||||||
{node_ttl, conf_get("cluster.etcd.node_ttl", Conf, 60)},
|
|
||||||
{ssl_options, filter(SslOpts(Conf))}];
|
|
||||||
options(k8s, Conf) ->
|
|
||||||
[{apiserver, conf_get("cluster.k8s.apiserver", Conf)},
|
|
||||||
{service_name, conf_get("cluster.k8s.service_name", Conf)},
|
|
||||||
{address_type, conf_get("cluster.k8s.address_type", Conf, ip)},
|
|
||||||
{app_name, conf_get("cluster.k8s.app_name", Conf)},
|
|
||||||
{namespace, conf_get("cluster.k8s.namespace", Conf)},
|
|
||||||
{suffix, conf_get("cluster.k8s.suffix", Conf, "")}];
|
|
||||||
options(manual, _Conf) ->
|
|
||||||
[].
|
|
||||||
|
|
||||||
to_atom(Atom) when is_atom(Atom) ->
|
to_atom(Atom) when is_atom(Atom) ->
|
||||||
Atom;
|
Atom;
|
||||||
to_atom(Str) when is_list(Str) ->
|
to_atom(Str) when is_list(Str) ->
|
||||||
|
|
|
@ -0,0 +1,696 @@
|
||||||
|
## NOTE: The configurations in this file will be overridden by
|
||||||
|
## `<path-to-emqx-installation>/data/emqx_overrides.conf`
|
||||||
|
|
||||||
|
##==================================================================
|
||||||
|
## Node
|
||||||
|
##==================================================================
|
||||||
|
node {
|
||||||
|
## Node name.
|
||||||
|
## See: http://erlang.org/doc/reference_manual/distributed.html
|
||||||
|
##
|
||||||
|
## @doc node.name
|
||||||
|
## ValueType: NodeName
|
||||||
|
## Default: emqx@127.0.0.1
|
||||||
|
name: "emqx@127.0.0.1"
|
||||||
|
|
||||||
|
## Cookie for distributed node communication.
|
||||||
|
##
|
||||||
|
## @doc node.cookie
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqxsecretcookie
|
||||||
|
cookie: emqxsecretcookie
|
||||||
|
|
||||||
|
## Data dir for the node
|
||||||
|
##
|
||||||
|
## @doc node.data_dir
|
||||||
|
## ValueType: Folder
|
||||||
|
## Default: "{{ platform_data_dir }}/"
|
||||||
|
data_dir: "{{ platform_data_dir }}/"
|
||||||
|
|
||||||
|
## Dir of crash dump file.
|
||||||
|
##
|
||||||
|
## @doc node.crash_dump_dir
|
||||||
|
## ValueType: Folder
|
||||||
|
## Default: "{{ platform_log_dir }}/"
|
||||||
|
crash_dump_dir: "{{ platform_log_dir }}/"
|
||||||
|
|
||||||
|
## Global GC Interval.
|
||||||
|
##
|
||||||
|
## @doc node.global_gc_interval
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 15m
|
||||||
|
global_gc_interval: 15m
|
||||||
|
|
||||||
|
## Sets the net_kernel tick time in seconds.
|
||||||
|
## Notice that all communicating nodes are to have the same
|
||||||
|
## TickTime value specified.
|
||||||
|
##
|
||||||
|
## See: http://www.erlang.org/doc/man/kernel_app.html#net_ticktime
|
||||||
|
##
|
||||||
|
## @doc node.dist_net_ticktime
|
||||||
|
## ValueType: Number
|
||||||
|
## Default: 2m
|
||||||
|
dist_net_ticktime: 2m
|
||||||
|
|
||||||
|
## Sets the port range for the listener socket of a distributed
|
||||||
|
## Erlang node.
|
||||||
|
## Note that if there are firewalls between clustered nodes, this
|
||||||
|
## port segment for nodes’ communication should be allowed.
|
||||||
|
##
|
||||||
|
## See: http://www.erlang.org/doc/man/kernel_app.html
|
||||||
|
##
|
||||||
|
## @doc node.dist_listen_min
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [1024,65535]
|
||||||
|
## Default: 6369
|
||||||
|
dist_listen_min: 6369
|
||||||
|
|
||||||
|
## Sets the port range for the listener socket of a distributed
|
||||||
|
## Erlang node.
|
||||||
|
## Note that if there are firewalls between clustered nodes, this
|
||||||
|
## port segment for nodes’ communication should be allowed.
|
||||||
|
##
|
||||||
|
## See: http://www.erlang.org/doc/man/kernel_app.html
|
||||||
|
##
|
||||||
|
## @doc node.dist_listen_max
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [1024,65535]
|
||||||
|
## Default: 6369
|
||||||
|
dist_listen_max: 6369
|
||||||
|
|
||||||
|
## Sets the maximum depth of call stack back-traces in the exit
|
||||||
|
## reason element of 'EXIT' tuples.
|
||||||
|
## The flag also limits the stacktrace depth returned by
|
||||||
|
## process_info item current_stacktrace.
|
||||||
|
##
|
||||||
|
## @doc node.backtrace_depth
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [0,1024]
|
||||||
|
## Default: 23
|
||||||
|
backtrace_depth: 23
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
##==================================================================
|
||||||
|
## Cluster
|
||||||
|
##==================================================================
|
||||||
|
cluster {
|
||||||
|
## Cluster name.
|
||||||
|
##
|
||||||
|
## @doc cluster.name
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqxcl
|
||||||
|
name: emqxcl
|
||||||
|
|
||||||
|
## Enable cluster autoheal from network partition.
|
||||||
|
##
|
||||||
|
## @doc cluster.autoheal
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: true
|
||||||
|
autoheal: true
|
||||||
|
|
||||||
|
## Autoclean down node. A down node will be removed from the cluster
|
||||||
|
## if this value > 0.
|
||||||
|
##
|
||||||
|
## @doc cluster.autoclean
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 5m
|
||||||
|
autoclean: 5m
|
||||||
|
|
||||||
|
## Node discovery strategy to join the cluster.
|
||||||
|
##
|
||||||
|
## @doc cluster.discovery_strategy
|
||||||
|
## ValueType: manual | static | mcast | dns | etcd | k8s
|
||||||
|
## - manual: Manual join command
|
||||||
|
## - static: Static node list
|
||||||
|
## - mcast: IP Multicast
|
||||||
|
## - dns: DNS A Record
|
||||||
|
## - etcd: etcd
|
||||||
|
## - k8s: Kubernetes
|
||||||
|
##
|
||||||
|
## Default: manual
|
||||||
|
discovery_strategy: manual
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Cluster using static node list
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
static {
|
||||||
|
## Node list of the cluster
|
||||||
|
##
|
||||||
|
## @doc cluster.static.seeds
|
||||||
|
## ValueType: Array<NodeName>
|
||||||
|
## Default: []
|
||||||
|
seeds: ["emqx1@127.0.0.1", "emqx2@127.0.0.1"]
|
||||||
|
}
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Cluster using IP Multicast
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
mcast {
|
||||||
|
## IP Multicast Address.
|
||||||
|
##
|
||||||
|
## @doc cluster.mcast.addr
|
||||||
|
## ValueType: IPAddress
|
||||||
|
## Default: "239.192.0.1"
|
||||||
|
addr: "239.192.0.1"
|
||||||
|
|
||||||
|
## Multicast Ports.
|
||||||
|
##
|
||||||
|
## @doc cluster.mcast.ports
|
||||||
|
## ValueType: Array<Port>
|
||||||
|
## Default: [4369, 4370]
|
||||||
|
ports: [4369, 4370]
|
||||||
|
|
||||||
|
## Multicast Iface.
|
||||||
|
##
|
||||||
|
## @doc cluster.mcast.iface
|
||||||
|
## ValueType: IPAddress
|
||||||
|
## Default: "0.0.0.0"
|
||||||
|
iface: "0.0.0.0"
|
||||||
|
|
||||||
|
## Multicast Ttl.
|
||||||
|
##
|
||||||
|
## @doc cluster.mcast.ttl
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [0,255]
|
||||||
|
## Default: 255
|
||||||
|
ttl: 255
|
||||||
|
|
||||||
|
## Multicast loop.
|
||||||
|
##
|
||||||
|
## @doc cluster.mcast.loop
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: true
|
||||||
|
loop: true
|
||||||
|
}
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Cluster using DNS A records
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
dns {
|
||||||
|
## DNS name.
|
||||||
|
##
|
||||||
|
## @doc cluster.dns.name
|
||||||
|
## ValueType: String
|
||||||
|
## Default: localhost
|
||||||
|
name: localhost
|
||||||
|
|
||||||
|
## The App name is used to build 'node.name' with IP address.
|
||||||
|
##
|
||||||
|
## @doc cluster.dns.app
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqx
|
||||||
|
app: emqx
|
||||||
|
}
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Cluster using etcd
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
etcd {
|
||||||
|
## Etcd server list, seperated by ','.
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.server
|
||||||
|
## ValueType: URL
|
||||||
|
## Required: true
|
||||||
|
server: "http://127.0.0.1:2379"
|
||||||
|
|
||||||
|
## The prefix helps build nodes path in etcd. Each node in the cluster
|
||||||
|
## will create a path in etcd: v2/keys/<prefix>/<name>/<node.name>
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.prefix
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqxcl
|
||||||
|
prefix: emqxcl
|
||||||
|
|
||||||
|
## The TTL for node's path in etcd.
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.node_ttl
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 1m
|
||||||
|
node_ttl: 1m
|
||||||
|
|
||||||
|
## Path to the file containing the user's private PEM-encoded key.
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.ssl.keyfile
|
||||||
|
## ValueType: File
|
||||||
|
## Default: "{{ platform_etc_dir }}/certs/key.pem"
|
||||||
|
ssl.keyfile: "{{ platform_etc_dir }}/certs/key.pem"
|
||||||
|
|
||||||
|
## Path to a file containing the user certificate.
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.ssl.certfile
|
||||||
|
## ValueType: File
|
||||||
|
## Default: "{{ platform_etc_dir }}/certs/cert.pem"
|
||||||
|
ssl.certfile: "{{ platform_etc_dir }}/certs/cert.pem"
|
||||||
|
|
||||||
|
## Path to the file containing PEM-encoded CA certificates. The CA certificates
|
||||||
|
## are used during server authentication and when building the client certificate chain.
|
||||||
|
##
|
||||||
|
## @doc cluster.etcd.ssl.cacertfile
|
||||||
|
## ValueType: File
|
||||||
|
## Default: "{{ platform_etc_dir }}/certs/cacert.pem"
|
||||||
|
ssl.cacertfile: "{{ platform_etc_dir }}/certs/cacert.pem"
|
||||||
|
}
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Cluster using Kubernetes
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
k8s {
|
||||||
|
## Kubernetes API server list, seperated by ','.
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.apiserver
|
||||||
|
## ValueType: URL
|
||||||
|
## Required: true
|
||||||
|
apiserver: "http://10.110.111.204:8080"
|
||||||
|
|
||||||
|
## The service name helps lookup EMQ nodes in the cluster.
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.service_name
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqx
|
||||||
|
service_name: emqx
|
||||||
|
|
||||||
|
## The address type is used to extract host from k8s service.
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.address_type
|
||||||
|
## ValueType: ip | dns | hostname
|
||||||
|
## Default: ip
|
||||||
|
address_type: ip
|
||||||
|
|
||||||
|
## The app name helps build 'node.name'.
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.app_name
|
||||||
|
## ValueType: String
|
||||||
|
## Default: emqx
|
||||||
|
app_name: emqx
|
||||||
|
|
||||||
|
## The suffix added to dns and hostname get from k8s service
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.suffix
|
||||||
|
## ValueType: String
|
||||||
|
## Default: "pod.local"
|
||||||
|
suffix: "pod.local"
|
||||||
|
|
||||||
|
## Kubernetes Namespace
|
||||||
|
##
|
||||||
|
## @doc cluster.k8s.namespace
|
||||||
|
## ValueType: String
|
||||||
|
## Default: default
|
||||||
|
namespace: default
|
||||||
|
}
|
||||||
|
|
||||||
|
db_backend: mnesia
|
||||||
|
|
||||||
|
rlog: {
|
||||||
|
# role: core
|
||||||
|
# core_nodes: []
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
##==================================================================
|
||||||
|
## Log
|
||||||
|
##==================================================================
|
||||||
|
log {
|
||||||
|
## The primary log level
|
||||||
|
##
|
||||||
|
## - all the log messages with levels lower than this level will
|
||||||
|
## be dropped.
|
||||||
|
## - all the log messages with levels higher than this level will
|
||||||
|
## go into the log handlers. The handlers then decide to log it
|
||||||
|
## out or drop it according to the level setting of the handler.
|
||||||
|
##
|
||||||
|
## Note: Only the messages with severity level higher than or
|
||||||
|
## equal to this level will be logged.
|
||||||
|
##
|
||||||
|
## @doc log.primary_level
|
||||||
|
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
||||||
|
## Default: warning
|
||||||
|
primary_level: warning
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## The console log handler send log messages to emqx console
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## Log to single line
|
||||||
|
## @doc log.console_handler.enable
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: false
|
||||||
|
console_handler.enable: false
|
||||||
|
|
||||||
|
## The log level of this handler
|
||||||
|
## All the log messages with levels lower than this level will
|
||||||
|
## be dropped.
|
||||||
|
##
|
||||||
|
## @doc log.console_handler.level
|
||||||
|
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
||||||
|
## Default: warning
|
||||||
|
console_handler.level: warning
|
||||||
|
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## The file log handlers send log messages to files
|
||||||
|
##----------------------------------------------------------------
|
||||||
|
## file_handlers.<name>
|
||||||
|
file_handlers.emqx_log: {
|
||||||
|
## The log level filter of this handler
|
||||||
|
## All the log messages with levels lower than this level will
|
||||||
|
## be dropped.
|
||||||
|
##
|
||||||
|
## @doc log.file_handlers.<name>.level
|
||||||
|
## ValueType: debug | info | notice | warning | error | critical | alert | emergency
|
||||||
|
## Default: warning
|
||||||
|
level: warning
|
||||||
|
|
||||||
|
## The log file for specified level.
|
||||||
|
##
|
||||||
|
## If `rotation` is disabled, this is the file of the log files.
|
||||||
|
##
|
||||||
|
## If `rotation` is enabled, this is the base name of the files.
|
||||||
|
## Each file in a rotated log is named <base_name>.N, where N is an integer.
|
||||||
|
##
|
||||||
|
## Note: Log files for a specific log level will only contain all the logs
|
||||||
|
## that higher than or equal to that level
|
||||||
|
##
|
||||||
|
## @doc log.file_handlers.<name>.file
|
||||||
|
## ValueType: File
|
||||||
|
## Required: true
|
||||||
|
file: "{{ platform_log_dir }}/emqx.log"
|
||||||
|
|
||||||
|
## Enables the log rotation.
|
||||||
|
## With this enabled, new log files will be created when the current
|
||||||
|
## log file is full, max to `rotation_count` files will be created.
|
||||||
|
##
|
||||||
|
## @doc log.file_handlers.<name>.rotation.enable
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: true
|
||||||
|
rotation.enable: true
|
||||||
|
|
||||||
|
## Maximum rotation count of log files.
|
||||||
|
##
|
||||||
|
## @doc log.file_handlers.<name>.rotation.count
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [1, 2048]
|
||||||
|
## Default: 10
|
||||||
|
rotation.count: 10
|
||||||
|
|
||||||
|
## Maximum size of each log file.
|
||||||
|
##
|
||||||
|
## If the max_size reached and `rotation` is disabled, the handler
|
||||||
|
## will stop sending log messages, if the `rotation` is enabled,
|
||||||
|
## the file rotates.
|
||||||
|
##
|
||||||
|
## @doc log.file_handlers.<name>.max_size
|
||||||
|
## ValueType: Size | infinity
|
||||||
|
## Default: 10MB
|
||||||
|
max_size: 10MB
|
||||||
|
}
|
||||||
|
|
||||||
|
## file_handlers.<name>
|
||||||
|
##
|
||||||
|
## You could also create multiple file handlers for different
|
||||||
|
## log level for example:
|
||||||
|
file_handlers.emqx_error_log: {
|
||||||
|
level: error
|
||||||
|
file: "{{ platform_log_dir }}/error.log"
|
||||||
|
}
|
||||||
|
|
||||||
|
## Timezone offset to display in logs
|
||||||
|
##
|
||||||
|
## @doc log.time_offset
|
||||||
|
## ValueType: system | utc | String
|
||||||
|
## - "system" use system zone
|
||||||
|
## - "utc" for Universal Coordinated Time (UTC)
|
||||||
|
## - "+hh:mm" or "-hh:mm" for a specified offset
|
||||||
|
## Default: system
|
||||||
|
time_offset: system
|
||||||
|
|
||||||
|
## Limits the total number of characters printed for each log event.
|
||||||
|
##
|
||||||
|
## @doc log.chars_limit
|
||||||
|
## ValueType: Integer | infinity
|
||||||
|
## Range: [0, infinity)
|
||||||
|
## Default: infinity
|
||||||
|
chars_limit: infinity
|
||||||
|
|
||||||
|
## Maximum depth for Erlang term log formatting
|
||||||
|
## and Erlang process message queue inspection.
|
||||||
|
##
|
||||||
|
## @doc log.max_depth
|
||||||
|
## ValueType: Integer | infinity
|
||||||
|
## Default: 80
|
||||||
|
max_depth: 80
|
||||||
|
|
||||||
|
## Log formatter
|
||||||
|
## @doc log.formatter
|
||||||
|
## ValueType: text | json
|
||||||
|
## Default: text
|
||||||
|
formatter: text
|
||||||
|
|
||||||
|
## Log to single line
|
||||||
|
## @doc log.single_line
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: true
|
||||||
|
single_line: true
|
||||||
|
|
||||||
|
## The max allowed queue length before switching to sync mode.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. If the message queue grows
|
||||||
|
## larger than this value the handler switches from anync to sync mode.
|
||||||
|
##
|
||||||
|
## @doc log.sync_mode_qlen
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [0, ${log.drop_mode_qlen}]
|
||||||
|
## Default: 100
|
||||||
|
sync_mode_qlen: 100
|
||||||
|
|
||||||
|
## The max allowed queue length before switching to drop mode.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. When the message queue grows
|
||||||
|
## larger than this threshold, the handler switches to a mode in which
|
||||||
|
## it drops all new events that senders want to log.
|
||||||
|
##
|
||||||
|
## @doc log.drop_mode_qlen
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [${log.sync_mode_qlen}, ${log.flush_qlen}]
|
||||||
|
## Default: 3000
|
||||||
|
drop_mode_qlen: 3000
|
||||||
|
|
||||||
|
## The max allowed queue length before switching to flush mode.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. If the length of the message queue
|
||||||
|
## grows larger than this threshold, a flush (delete) operation takes place.
|
||||||
|
## To flush events, the handler discards the messages in the message queue
|
||||||
|
## by receiving them in a loop without logging.
|
||||||
|
##
|
||||||
|
## @doc log.flush_qlen
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [${log.drop_mode_qlen}, infinity)
|
||||||
|
## Default: 8000
|
||||||
|
flush_qlen: 8000
|
||||||
|
|
||||||
|
## Kill the log handler when it gets overloaded.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. It is possible that a handler,
|
||||||
|
## even if it can successfully manage peaks of high load without crashing,
|
||||||
|
## can build up a large message queue, or use a large amount of memory.
|
||||||
|
## We could kill the log handler in these cases and restart it after a
|
||||||
|
## few seconds.
|
||||||
|
##
|
||||||
|
## @doc log.overload_kill.enable
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: true
|
||||||
|
overload_kill.enable: true
|
||||||
|
|
||||||
|
## The max allowed queue length before killing the log hanlder.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. This is the maximum allowed queue
|
||||||
|
## length. If the message queue grows larger than this, the handler
|
||||||
|
## process is terminated.
|
||||||
|
##
|
||||||
|
## @doc log.overload_kill.qlen
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [0, 1048576]
|
||||||
|
## Default: 20000
|
||||||
|
overload_kill.qlen: 20000
|
||||||
|
|
||||||
|
## The max allowed memory size before killing the log hanlder.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. This is the maximum memory size
|
||||||
|
## that the handler process is allowed to use. If the handler grows
|
||||||
|
## larger than this, the process is terminated.
|
||||||
|
##
|
||||||
|
## @doc log.overload_kill.mem_size
|
||||||
|
## ValueType: Size
|
||||||
|
## Default: 30MB
|
||||||
|
overload_kill.mem_size: 30MB
|
||||||
|
|
||||||
|
## Restart the log hanlder after some seconds.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. If the handler is terminated,
|
||||||
|
## it restarts automatically after a delay specified in seconds.
|
||||||
|
##
|
||||||
|
## @doc log.overload_kill.restart_after
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 5s
|
||||||
|
overload_kill.restart_after: 5s
|
||||||
|
|
||||||
|
## Controlling Bursts of Log Requests.
|
||||||
|
##
|
||||||
|
## Log overload protection parameter. Large bursts of log events - many
|
||||||
|
## events received by the handler under a short period of time - can
|
||||||
|
## potentially cause problems. By specifying the maximum number of events
|
||||||
|
## to be handled within a certain time frame, the handler can avoid
|
||||||
|
## choking the log with massive amounts of printouts.
|
||||||
|
##
|
||||||
|
## Note that there would be no warning if any messages were
|
||||||
|
## dropped because of burst control.
|
||||||
|
##
|
||||||
|
## @doc log.burst_limit.enable
|
||||||
|
## ValueType: Boolean
|
||||||
|
## Default: false
|
||||||
|
burst_limit.enable: false
|
||||||
|
|
||||||
|
## This config controls the maximum number of events to handle within
|
||||||
|
## a time frame. After the limit is reached, successive events are
|
||||||
|
## dropped until the end of the time frame defined by `window_time`.
|
||||||
|
##
|
||||||
|
## @doc log.burst_limit.max_count
|
||||||
|
## ValueType: Integer
|
||||||
|
## Default: 10000
|
||||||
|
burst_limit.max_count: 10000
|
||||||
|
|
||||||
|
## See the previous description of burst_limit_max_count.
|
||||||
|
##
|
||||||
|
## @doc log.burst_limit.window_time
|
||||||
|
## ValueType: duration
|
||||||
|
## Default: 1s
|
||||||
|
burst_limit.window_time: 1s
|
||||||
|
}
|
||||||
|
|
||||||
|
##==================================================================
|
||||||
|
## RPC
|
||||||
|
##==================================================================
|
||||||
|
rpc {
|
||||||
|
## RPC Mode.
|
||||||
|
##
|
||||||
|
## @doc rpc.mode
|
||||||
|
## ValueType: sync | async
|
||||||
|
## Default: async
|
||||||
|
mode: async
|
||||||
|
|
||||||
|
## Max batch size of async RPC requests.
|
||||||
|
##
|
||||||
|
## NOTE: RPC batch won't work when rpc.mode = sync
|
||||||
|
## Zero value disables rpc batching.
|
||||||
|
##
|
||||||
|
## @doc rpc.async_batch_size
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [0, 1048576]
|
||||||
|
## Default: 0
|
||||||
|
async_batch_size: 256
|
||||||
|
|
||||||
|
## RPC port discovery
|
||||||
|
##
|
||||||
|
## The strategy for discovering the RPC listening port of
|
||||||
|
## other nodes.
|
||||||
|
##
|
||||||
|
## @doc cluster.discovery_strategy
|
||||||
|
## ValueType: manual | stateless
|
||||||
|
## - manual: discover ports by `tcp_server_port`.
|
||||||
|
## - stateless: discover ports in a stateless manner.
|
||||||
|
## If node name is `emqx<N>@127.0.0.1`, where the `<N>` is
|
||||||
|
## an integer, then the listening port will be `5370 + <N>`
|
||||||
|
##
|
||||||
|
## Default: `stateless`.
|
||||||
|
port_discovery: stateless
|
||||||
|
|
||||||
|
## TCP server port for RPC.
|
||||||
|
##
|
||||||
|
## Only takes effect when `rpc.port_discovery` = `manual`.
|
||||||
|
##
|
||||||
|
## @doc rpc.tcp_server_port
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [1024-65535]
|
||||||
|
## Defaults: 5369
|
||||||
|
tcp_server_port: 5369
|
||||||
|
|
||||||
|
## Number of outgoing RPC connections.
|
||||||
|
##
|
||||||
|
## Set this to 1 to keep the message order sent from the same
|
||||||
|
## client.
|
||||||
|
##
|
||||||
|
## @doc rpc.tcp_client_num
|
||||||
|
## ValueType: Integer
|
||||||
|
## Range: [1, 256]
|
||||||
|
## Defaults: 1
|
||||||
|
tcp_client_num: 1
|
||||||
|
|
||||||
|
## RCP Client connect timeout.
|
||||||
|
##
|
||||||
|
## @doc rpc.connect_timeout
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 5s
|
||||||
|
connect_timeout: 5s
|
||||||
|
|
||||||
|
## TCP send timeout of RPC client and server.
|
||||||
|
##
|
||||||
|
## @doc rpc.send_timeout
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 5s
|
||||||
|
send_timeout: 5s
|
||||||
|
|
||||||
|
## Authentication timeout
|
||||||
|
##
|
||||||
|
## @doc rpc.authentication_timeout
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 5s
|
||||||
|
authentication_timeout: 5s
|
||||||
|
|
||||||
|
## Default receive timeout for call() functions
|
||||||
|
##
|
||||||
|
## @doc rpc.call_receive_timeout
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 15s
|
||||||
|
call_receive_timeout: 15s
|
||||||
|
|
||||||
|
## Socket idle keepalive.
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_keepalive_idle
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 900s
|
||||||
|
socket_keepalive_idle: 900s
|
||||||
|
|
||||||
|
## TCP Keepalive probes interval.
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_keepalive_interval
|
||||||
|
## ValueType: Duration
|
||||||
|
## Default: 75s
|
||||||
|
socket_keepalive_interval: 75s
|
||||||
|
|
||||||
|
## Probes lost to close the connection
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_keepalive_count
|
||||||
|
## ValueType: Integer
|
||||||
|
## Default: 9
|
||||||
|
socket_keepalive_count: 9
|
||||||
|
|
||||||
|
## Size of TCP send buffer.
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_sndbuf
|
||||||
|
## ValueType: Size
|
||||||
|
## Default: 1MB
|
||||||
|
socket_sndbuf: 1MB
|
||||||
|
|
||||||
|
## Size of TCP receive buffer.
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_recbuf
|
||||||
|
## ValueType: Size
|
||||||
|
## Default: 1MB
|
||||||
|
socket_recbuf: 1MB
|
||||||
|
|
||||||
|
## Size of user-level software socket buffer.
|
||||||
|
##
|
||||||
|
## @doc rpc.socket_buffer
|
||||||
|
## ValueType: Size
|
||||||
|
## Default: 1MB
|
||||||
|
socket_buffer: 1MB
|
||||||
|
}
|
|
@ -30,6 +30,8 @@ start(_Type, _Args) ->
|
||||||
ok = print_otp_version_warning(),
|
ok = print_otp_version_warning(),
|
||||||
_ = load_modules(),
|
_ = load_modules(),
|
||||||
|
|
||||||
|
ok = load_config_files(),
|
||||||
|
|
||||||
{ok, _} = application:ensure_all_started(emqx),
|
{ok, _} = application:ensure_all_started(emqx),
|
||||||
|
|
||||||
_ = emqx_plugins:load(),
|
_ = emqx_plugins:load(),
|
||||||
|
@ -75,3 +77,12 @@ load_modules() ->
|
||||||
start_modules() ->
|
start_modules() ->
|
||||||
ok.
|
ok.
|
||||||
-endif.
|
-endif.
|
||||||
|
|
||||||
|
load_config_files() ->
|
||||||
|
%% the app env 'config_files' for 'emqx` app should be set
|
||||||
|
%% in app.time.config by boot script before starting Erlang VM
|
||||||
|
ConfFiles = application:get_env(emqx, config_files, []),
|
||||||
|
%% emqx_machine_schema is a superset of emqx_schema
|
||||||
|
ok = emqx_config:init_load(emqx_machine_schema, ConfFiles),
|
||||||
|
%% to avoid config being loaded again when emqx app starts.
|
||||||
|
ok = emqx_app:set_init_config_load_done().
|
||||||
|
|
|
@ -0,0 +1,421 @@
|
||||||
|
%%--------------------------------------------------------------------
|
||||||
|
%% Copyright (c) 2021 EMQ Technologies Co., Ltd. All Rights Reserved.
|
||||||
|
%%
|
||||||
|
%% Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
%% you may not use this file except in compliance with the License.
|
||||||
|
%% You may obtain a copy of the License at
|
||||||
|
%%
|
||||||
|
%% http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
%%
|
||||||
|
%% Unless required by applicable law or agreed to in writing, software
|
||||||
|
%% distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
%% WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
%% See the License for the specific language governing permissions and
|
||||||
|
%% limitations under the License.
|
||||||
|
%%--------------------------------------------------------------------
|
||||||
|
|
||||||
|
-module(emqx_machine_schema).
|
||||||
|
|
||||||
|
-dialyzer(no_return).
|
||||||
|
-dialyzer(no_match).
|
||||||
|
-dialyzer(no_contracts).
|
||||||
|
-dialyzer(no_unused).
|
||||||
|
-dialyzer(no_fail_call).
|
||||||
|
|
||||||
|
-include_lib("typerefl/include/types.hrl").
|
||||||
|
|
||||||
|
-type log_level() :: debug | info | notice | warning | error | critical | alert | emergency | all.
|
||||||
|
-type file() :: string().
|
||||||
|
-type cipher() :: map().
|
||||||
|
|
||||||
|
-behaviour(hocon_schema).
|
||||||
|
|
||||||
|
-reflect_type([ log_level/0,
|
||||||
|
file/0,
|
||||||
|
cipher/0]).
|
||||||
|
|
||||||
|
-export([structs/0, fields/1, translations/0, translation/1]).
|
||||||
|
-export([t/1, t/3, t/4, ref/1]).
|
||||||
|
-export([conf_get/2, conf_get/3, keys/2, filter/1]).
|
||||||
|
|
||||||
|
%% Static apps which merge their configs into the merged emqx.conf
|
||||||
|
%% The list can not be made a dynamic read at run-time as it is used
|
||||||
|
%% by nodetool to generate app.<time>.config before EMQ X is started
|
||||||
|
-define(MERGED_CONFIGS,
|
||||||
|
[ emqx_schema
|
||||||
|
, emqx_data_bridge_schema
|
||||||
|
, emqx_retainer_schema
|
||||||
|
, emqx_statsd_schema
|
||||||
|
, emqx_authn_schema
|
||||||
|
, emqx_authz_schema
|
||||||
|
, emqx_bridge_mqtt_schema
|
||||||
|
, emqx_modules_schema
|
||||||
|
, emqx_management_schema
|
||||||
|
, emqx_dashboard_schema
|
||||||
|
, emqx_gateway_schema
|
||||||
|
, emqx_prometheus_schema
|
||||||
|
]).
|
||||||
|
|
||||||
|
%% TODO: add a test case to ensure the list elements are unique
|
||||||
|
structs() ->
|
||||||
|
["cluster", "node", "rpc", "log"]
|
||||||
|
++ lists:flatmap(fun(Mod) -> Mod:structs() end, ?MERGED_CONFIGS).
|
||||||
|
|
||||||
|
fields("cluster") ->
|
||||||
|
[ {"name", t(atom(), "ekka.cluster_name", emqxcl)}
|
||||||
|
, {"discovery_strategy", t(union([manual, static, mcast, dns, etcd, k8s]),
|
||||||
|
undefined, manual)}
|
||||||
|
, {"autoclean", t(emqx_schema:duration(), "ekka.cluster_autoclean", "5m")}
|
||||||
|
, {"autoheal", t(boolean(), "ekka.cluster_autoheal", true)}
|
||||||
|
, {"static", ref("static")}
|
||||||
|
, {"mcast", ref("mcast")}
|
||||||
|
, {"proto_dist", t(union([inet_tcp, inet6_tcp, inet_tls]), "ekka.proto_dist", inet_tcp)}
|
||||||
|
, {"dns", ref("dns")}
|
||||||
|
, {"etcd", ref("etcd")}
|
||||||
|
, {"k8s", ref("k8s")}
|
||||||
|
, {"db_backend", t(union([mnesia, rlog]), "ekka.db_backend", mnesia)}
|
||||||
|
, {"rlog", ref("rlog")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("static") ->
|
||||||
|
[ {"seeds", t(hoconsc:array(string()), undefined, [])}];
|
||||||
|
|
||||||
|
fields("mcast") ->
|
||||||
|
[ {"addr", t(string(), undefined, "239.192.0.1")}
|
||||||
|
, {"ports", t(hoconsc:array(integer()), undefined, [4369, 4370])}
|
||||||
|
, {"iface", t(string(), undefined, "0.0.0.0")}
|
||||||
|
, {"ttl", t(range(0, 255), undefined, 255)}
|
||||||
|
, {"loop", t(boolean(), undefined, true)}
|
||||||
|
, {"sndbuf", t(emqx_schema:bytesize(), undefined, "16KB")}
|
||||||
|
, {"recbuf", t(emqx_schema:bytesize(), undefined, "16KB")}
|
||||||
|
, {"buffer", t(emqx_schema:bytesize(), undefined, "32KB")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("dns") ->
|
||||||
|
[ {"name", t(string(), undefined, "localhost")}
|
||||||
|
, {"app", t(string(), undefined, "emqx")}];
|
||||||
|
|
||||||
|
fields("etcd") ->
|
||||||
|
[ {"server", t(emqx_schema:comma_separated_list())}
|
||||||
|
, {"prefix", t(string(), undefined, "emqxcl")}
|
||||||
|
, {"node_ttl", t(emqx_schema:duration(), undefined, "1m")}
|
||||||
|
, {"ssl", ref("etcd_ssl")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("etcd_ssl") ->
|
||||||
|
emqx_schema:ssl(#{});
|
||||||
|
|
||||||
|
fields("k8s") ->
|
||||||
|
[ {"apiserver", t(string())}
|
||||||
|
, {"service_name", t(string(), undefined, "emqx")}
|
||||||
|
, {"address_type", t(union([ip, dns, hostname]))}
|
||||||
|
, {"app_name", t(string(), undefined, "emqx")}
|
||||||
|
, {"namespace", t(string(), undefined, "default")}
|
||||||
|
, {"suffix", t(string(), undefined, "pod.local")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("rlog") ->
|
||||||
|
[ {"role", t(union([core, replicant]), "ekka.node_role", core)}
|
||||||
|
, {"core_nodes", t(emqx_schema:comma_separated_atoms(), "ekka.core_nodes", [])}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("node") ->
|
||||||
|
[ {"name", hoconsc:t(string(), #{default => "emqx@127.0.0.1",
|
||||||
|
override_env => "EMQX_NODE_NAME"
|
||||||
|
})}
|
||||||
|
, {"cookie", hoconsc:t(string(), #{mapping => "vm_args.-setcookie",
|
||||||
|
default => "emqxsecretcookie",
|
||||||
|
sensitive => true,
|
||||||
|
override_env => "EMQX_NODE_COOKIE"
|
||||||
|
})}
|
||||||
|
, {"data_dir", t(string(), undefined, undefined)}
|
||||||
|
, {"config_files", t(list(string()), "emqx.config_files",
|
||||||
|
[ filename:join([os:getenv("RUNNER_ETC_DIR"), "emqx.conf"])
|
||||||
|
])}
|
||||||
|
, {"global_gc_interval", t(emqx_schema:duration(), undefined, "15m")}
|
||||||
|
, {"crash_dump_dir", t(file(), "vm_args.-env ERL_CRASH_DUMP", undefined)}
|
||||||
|
, {"dist_net_ticktime", t(emqx_schema:duration(), "vm_args.-kernel net_ticktime", "2m")}
|
||||||
|
, {"dist_listen_min", t(range(1024, 65535), "kernel.inet_dist_listen_min", 6369)}
|
||||||
|
, {"dist_listen_max", t(range(1024, 65535), "kernel.inet_dist_listen_max", 6369)}
|
||||||
|
, {"backtrace_depth", t(integer(), "emqx_machine.backtrace_depth", 23)}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("rpc") ->
|
||||||
|
[ {"mode", t(union(sync, async), undefined, async)}
|
||||||
|
, {"async_batch_size", t(integer(), "gen_rpc.max_batch_size", 256)}
|
||||||
|
, {"port_discovery",t(union(manual, stateless), "gen_rpc.port_discovery", stateless)}
|
||||||
|
, {"tcp_server_port", t(integer(), "gen_rpc.tcp_server_port", 5369)}
|
||||||
|
, {"tcp_client_num", t(range(1, 256), undefined, 1)}
|
||||||
|
, {"connect_timeout", t(emqx_schema:duration(), "gen_rpc.connect_timeout", "5s")}
|
||||||
|
, {"send_timeout", t(emqx_schema:duration(), "gen_rpc.send_timeout", "5s")}
|
||||||
|
, {"authentication_timeout", t(emqx_schema:duration(), "gen_rpc.authentication_timeout", "5s")}
|
||||||
|
, {"call_receive_timeout", t(emqx_schema:duration(), "gen_rpc.call_receive_timeout", "15s")}
|
||||||
|
, {"socket_keepalive_idle", t(emqx_schema:duration_s(), "gen_rpc.socket_keepalive_idle", "7200s")}
|
||||||
|
, {"socket_keepalive_interval", t(emqx_schema:duration_s(), "gen_rpc.socket_keepalive_interval", "75s")}
|
||||||
|
, {"socket_keepalive_count", t(integer(), "gen_rpc.socket_keepalive_count", 9)}
|
||||||
|
, {"socket_sndbuf", t(emqx_schema:bytesize(), "gen_rpc.socket_sndbuf", "1MB")}
|
||||||
|
, {"socket_recbuf", t(emqx_schema:bytesize(), "gen_rpc.socket_recbuf", "1MB")}
|
||||||
|
, {"socket_buffer", t(emqx_schema:bytesize(), "gen_rpc.socket_buffer", "1MB")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("log") ->
|
||||||
|
[ {"primary_level", t(log_level(), undefined, warning)}
|
||||||
|
, {"console_handler", ref("console_handler")}
|
||||||
|
, {"file_handlers", ref("file_handlers")}
|
||||||
|
, {"time_offset", t(string(), undefined, "system")}
|
||||||
|
, {"chars_limit", maybe_infinity(range(1, inf))}
|
||||||
|
, {"supervisor_reports", t(union([error, progress]), undefined, error)}
|
||||||
|
, {"max_depth", t(union([infinity, integer()]),
|
||||||
|
"kernel.error_logger_format_depth", 80)}
|
||||||
|
, {"formatter", t(union([text, json]), undefined, text)}
|
||||||
|
, {"single_line", t(boolean(), undefined, true)}
|
||||||
|
, {"sync_mode_qlen", t(integer(), undefined, 100)}
|
||||||
|
, {"drop_mode_qlen", t(integer(), undefined, 3000)}
|
||||||
|
, {"flush_qlen", t(integer(), undefined, 8000)}
|
||||||
|
, {"overload_kill", ref("log_overload_kill")}
|
||||||
|
, {"burst_limit", ref("log_burst_limit")}
|
||||||
|
, {"error_logger", t(atom(), "kernel.error_logger", silent)}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("console_handler") ->
|
||||||
|
[ {"enable", t(boolean(), undefined, false)}
|
||||||
|
, {"level", t(log_level(), undefined, warning)}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("file_handlers") ->
|
||||||
|
[ {"$name", ref("log_file_handler")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("log_file_handler") ->
|
||||||
|
[ {"level", t(log_level(), undefined, warning)}
|
||||||
|
, {"file", t(file(), undefined, undefined)}
|
||||||
|
, {"rotation", ref("log_rotation")}
|
||||||
|
, {"max_size", maybe_infinity(emqx_schema:bytesize(), "10MB")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("log_rotation") ->
|
||||||
|
[ {"enable", t(boolean(), undefined, true)}
|
||||||
|
, {"count", t(range(1, 2048), undefined, 10)}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("log_overload_kill") ->
|
||||||
|
[ {"enable", t(boolean(), undefined, true)}
|
||||||
|
, {"mem_size", t(emqx_schema:bytesize(), undefined, "30MB")}
|
||||||
|
, {"qlen", t(integer(), undefined, 20000)}
|
||||||
|
, {"restart_after", t(union(emqx_schema:duration(), infinity), undefined, "5s")}
|
||||||
|
];
|
||||||
|
|
||||||
|
fields("log_burst_limit") ->
|
||||||
|
[ {"enable", t(boolean(), undefined, true)}
|
||||||
|
, {"max_count", t(integer(), undefined, 10000)}
|
||||||
|
, {"window_time", t(emqx_schema:duration(), undefined, "1s")}
|
||||||
|
];
|
||||||
|
fields(Name) ->
|
||||||
|
find_field(Name, ?MERGED_CONFIGS).
|
||||||
|
|
||||||
|
find_field(Name, []) ->
|
||||||
|
error({unknown_config_struct_field, Name});
|
||||||
|
find_field(Name, [SchemaModule | Rest]) ->
|
||||||
|
case lists:member(Name, SchemaModule:structs()) of
|
||||||
|
true -> SchemaModule:fields(Name);
|
||||||
|
false -> find_field(Name, Rest)
|
||||||
|
end.
|
||||||
|
|
||||||
|
translations() -> ["ekka", "kernel", "emqx"].
|
||||||
|
|
||||||
|
translation("ekka") ->
|
||||||
|
[ {"cluster_discovery", fun tr_cluster__discovery/1}];
|
||||||
|
translation("kernel") ->
|
||||||
|
[ {"logger_level", fun tr_logger_level/1}
|
||||||
|
, {"logger", fun tr_logger/1}];
|
||||||
|
translation("emqx") ->
|
||||||
|
[ {"config_files", fun tr_config_files/1}
|
||||||
|
].
|
||||||
|
|
||||||
|
tr_config_files(Conf) ->
|
||||||
|
case conf_get("emqx.config_files", Conf) of
|
||||||
|
[_ | _] = Files ->
|
||||||
|
Files;
|
||||||
|
_ ->
|
||||||
|
case os:getenv("RUNNER_ETC_DIR") of
|
||||||
|
false ->
|
||||||
|
[filename:join([code:lib_dir(emqx), "etc", "emqx.conf"])];
|
||||||
|
Dir ->
|
||||||
|
[filename:join([Dir, "emqx.conf"])]
|
||||||
|
end
|
||||||
|
end.
|
||||||
|
|
||||||
|
tr_cluster__discovery(Conf) ->
|
||||||
|
Strategy = conf_get("cluster.discovery_strategy", Conf),
|
||||||
|
{Strategy, filter(options(Strategy, Conf))}.
|
||||||
|
|
||||||
|
tr_logger_level(Conf) -> conf_get("log.primary_level", Conf).
|
||||||
|
|
||||||
|
tr_logger(Conf) ->
|
||||||
|
CharsLimit = case conf_get("log.chars_limit", Conf) of
|
||||||
|
infinity -> unlimited;
|
||||||
|
V -> V
|
||||||
|
end,
|
||||||
|
SingleLine = conf_get("log.single_line", Conf),
|
||||||
|
FmtName = conf_get("log.formatter", Conf),
|
||||||
|
Formatter = formatter(FmtName, CharsLimit, SingleLine),
|
||||||
|
BasicConf = #{
|
||||||
|
sync_mode_qlen => conf_get("log.sync_mode_qlen", Conf),
|
||||||
|
drop_mode_qlen => conf_get("log.drop_mode_qlen", Conf),
|
||||||
|
flush_qlen => conf_get("log.flush_qlen", Conf),
|
||||||
|
overload_kill_enable => conf_get("log.overload_kill.enable", Conf),
|
||||||
|
overload_kill_qlen => conf_get("log.overload_kill.qlen", Conf),
|
||||||
|
overload_kill_mem_size => conf_get("log.overload_kill.mem_size", Conf),
|
||||||
|
overload_kill_restart_after => conf_get("log.overload_kill.restart_after", Conf),
|
||||||
|
burst_limit_enable => conf_get("log.burst_limit.enable", Conf),
|
||||||
|
burst_limit_max_count => conf_get("log.burst_limit.max_count", Conf),
|
||||||
|
burst_limit_window_time => conf_get("log.burst_limit.window_time", Conf)
|
||||||
|
},
|
||||||
|
Filters = case conf_get("log.supervisor_reports", Conf) of
|
||||||
|
error -> [{drop_progress_reports, {fun logger_filters:progress/2, stop}}];
|
||||||
|
progress -> []
|
||||||
|
end,
|
||||||
|
%% For the default logger that outputs to console
|
||||||
|
ConsoleHandler =
|
||||||
|
case conf_get("log.console_handler.enable", Conf) of
|
||||||
|
true ->
|
||||||
|
[{handler, console, logger_std_h, #{
|
||||||
|
level => conf_get("log.console_handler.level", Conf),
|
||||||
|
config => BasicConf#{type => standard_io},
|
||||||
|
formatter => Formatter,
|
||||||
|
filters => Filters
|
||||||
|
}}];
|
||||||
|
false -> []
|
||||||
|
end,
|
||||||
|
%% For the file logger
|
||||||
|
FileHandlers =
|
||||||
|
[{handler, binary_to_atom(HandlerName, latin1), logger_disk_log_h, #{
|
||||||
|
level => conf_get("level", SubConf),
|
||||||
|
config => BasicConf#{
|
||||||
|
type => case conf_get("rotation.enable", SubConf) of
|
||||||
|
true -> wrap;
|
||||||
|
_ -> halt
|
||||||
|
end,
|
||||||
|
file => conf_get("file", SubConf),
|
||||||
|
max_no_files => conf_get("rotation.count", SubConf),
|
||||||
|
max_no_bytes => conf_get("max_size", SubConf)
|
||||||
|
},
|
||||||
|
formatter => Formatter,
|
||||||
|
filters => Filters,
|
||||||
|
filesync_repeat_interval => no_repeat
|
||||||
|
}}
|
||||||
|
|| {HandlerName, SubConf} <- maps:to_list(conf_get("log.file_handlers", Conf, #{}))],
|
||||||
|
|
||||||
|
[{handler, default, undefined}] ++ ConsoleHandler ++ FileHandlers.
|
||||||
|
|
||||||
|
%% helpers
|
||||||
|
formatter(json, CharsLimit, SingleLine) ->
|
||||||
|
{emqx_logger_jsonfmt,
|
||||||
|
#{chars_limit => CharsLimit,
|
||||||
|
single_line => SingleLine
|
||||||
|
}};
|
||||||
|
formatter(text, CharsLimit, SingleLine) ->
|
||||||
|
{emqx_logger_textfmt,
|
||||||
|
#{template =>
|
||||||
|
[time," [",level,"] ",
|
||||||
|
{clientid,
|
||||||
|
[{peername,
|
||||||
|
[clientid,"@",peername," "],
|
||||||
|
[clientid, " "]}],
|
||||||
|
[{peername,
|
||||||
|
[peername," "],
|
||||||
|
[]}]},
|
||||||
|
msg,"\n"],
|
||||||
|
chars_limit => CharsLimit,
|
||||||
|
single_line => SingleLine
|
||||||
|
}}.
|
||||||
|
|
||||||
|
%% utils
|
||||||
|
-spec(conf_get(string() | [string()], hocon:config()) -> term()).
|
||||||
|
conf_get(Key, Conf) ->
|
||||||
|
V = hocon_schema:get_value(Key, Conf),
|
||||||
|
case is_binary(V) of
|
||||||
|
true ->
|
||||||
|
binary_to_list(V);
|
||||||
|
false ->
|
||||||
|
V
|
||||||
|
end.
|
||||||
|
|
||||||
|
conf_get(Key, Conf, Default) ->
|
||||||
|
V = hocon_schema:get_value(Key, Conf, Default),
|
||||||
|
case is_binary(V) of
|
||||||
|
true ->
|
||||||
|
binary_to_list(V);
|
||||||
|
false ->
|
||||||
|
V
|
||||||
|
end.
|
||||||
|
|
||||||
|
filter(Opts) ->
|
||||||
|
[{K, V} || {K, V} <- Opts, V =/= undefined].
|
||||||
|
|
||||||
|
%% @private return a list of keys in a parent field
|
||||||
|
-spec(keys(string(), hocon:config()) -> [string()]).
|
||||||
|
keys(Parent, Conf) ->
|
||||||
|
[binary_to_list(B) || B <- maps:keys(conf_get(Parent, Conf, #{}))].
|
||||||
|
|
||||||
|
%% types
|
||||||
|
|
||||||
|
t(Type) -> hoconsc:t(Type).
|
||||||
|
|
||||||
|
t(Type, Mapping, Default) ->
|
||||||
|
hoconsc:t(Type, #{mapping => Mapping, default => Default}).
|
||||||
|
|
||||||
|
t(Type, Mapping, Default, OverrideEnv) ->
|
||||||
|
hoconsc:t(Type, #{ mapping => Mapping
|
||||||
|
, default => Default
|
||||||
|
, override_env => OverrideEnv
|
||||||
|
}).
|
||||||
|
|
||||||
|
ref(Field) -> hoconsc:t(hoconsc:ref(Field)).
|
||||||
|
|
||||||
|
maybe_infinity(T) ->
|
||||||
|
maybe_sth(infinity, T, infinity).
|
||||||
|
|
||||||
|
maybe_infinity(T, Default) ->
|
||||||
|
maybe_sth(infinity, T, Default).
|
||||||
|
|
||||||
|
maybe_sth(What, Type, Default) ->
|
||||||
|
t(union([What, Type]), undefined, Default).
|
||||||
|
|
||||||
|
options(static, Conf) ->
|
||||||
|
[{seeds, [to_atom(S) || S <- conf_get("cluster.static.seeds", Conf, [])]}];
|
||||||
|
options(mcast, Conf) ->
|
||||||
|
{ok, Addr} = inet:parse_address(conf_get("cluster.mcast.addr", Conf)),
|
||||||
|
{ok, Iface} = inet:parse_address(conf_get("cluster.mcast.iface", Conf)),
|
||||||
|
Ports = conf_get("cluster.mcast.ports", Conf),
|
||||||
|
[{addr, Addr}, {ports, Ports}, {iface, Iface},
|
||||||
|
{ttl, conf_get("cluster.mcast.ttl", Conf, 1)},
|
||||||
|
{loop, conf_get("cluster.mcast.loop", Conf, true)}];
|
||||||
|
options(dns, Conf) ->
|
||||||
|
[{name, conf_get("cluster.dns.name", Conf)},
|
||||||
|
{app, conf_get("cluster.dns.app", Conf)}];
|
||||||
|
options(etcd, Conf) ->
|
||||||
|
Namespace = "cluster.etcd.ssl",
|
||||||
|
SslOpts = fun(C) ->
|
||||||
|
Options = keys(Namespace, C),
|
||||||
|
lists:map(fun(Key) -> {to_atom(Key), conf_get([Namespace, Key], Conf)} end, Options) end,
|
||||||
|
[{server, conf_get("cluster.etcd.server", Conf)},
|
||||||
|
{prefix, conf_get("cluster.etcd.prefix", Conf, "emqxcl")},
|
||||||
|
{node_ttl, conf_get("cluster.etcd.node_ttl", Conf, 60)},
|
||||||
|
{ssl_options, filter(SslOpts(Conf))}];
|
||||||
|
options(k8s, Conf) ->
|
||||||
|
[{apiserver, conf_get("cluster.k8s.apiserver", Conf)},
|
||||||
|
{service_name, conf_get("cluster.k8s.service_name", Conf)},
|
||||||
|
{address_type, conf_get("cluster.k8s.address_type", Conf, ip)},
|
||||||
|
{app_name, conf_get("cluster.k8s.app_name", Conf)},
|
||||||
|
{namespace, conf_get("cluster.k8s.namespace", Conf)},
|
||||||
|
{suffix, conf_get("cluster.k8s.suffix", Conf, "")}];
|
||||||
|
options(manual, _Conf) ->
|
||||||
|
[].
|
||||||
|
|
||||||
|
to_atom(Atom) when is_atom(Atom) ->
|
||||||
|
Atom;
|
||||||
|
to_atom(Str) when is_list(Str) ->
|
||||||
|
list_to_atom(Str);
|
||||||
|
to_atom(Bin) when is_binary(Bin) ->
|
||||||
|
binary_to_atom(Bin, utf8).
|
9
bin/emqx
9
bin/emqx
|
@ -16,6 +16,7 @@ ROOT_DIR="$(cd "$(dirname "$(readlink "$0" || echo "$0")")"/..; pwd -P)"
|
||||||
RUNNER_SCRIPT="$RUNNER_BIN_DIR/$REL_NAME"
|
RUNNER_SCRIPT="$RUNNER_BIN_DIR/$REL_NAME"
|
||||||
CODE_LOADING_MODE="${CODE_LOADING_MODE:-embedded}"
|
CODE_LOADING_MODE="${CODE_LOADING_MODE:-embedded}"
|
||||||
REL_DIR="$RUNNER_ROOT_DIR/releases/$REL_VSN"
|
REL_DIR="$RUNNER_ROOT_DIR/releases/$REL_VSN"
|
||||||
|
SCHEMA_MOD=emqx_machine_schema
|
||||||
|
|
||||||
WHOAMI=$(whoami)
|
WHOAMI=$(whoami)
|
||||||
|
|
||||||
|
@ -240,7 +241,7 @@ generate_config() {
|
||||||
## ths command populates two files: app.<time>.config and vm.<time>.args
|
## ths command populates two files: app.<time>.config and vm.<time>.args
|
||||||
## disable SC2086 to allow EMQX_LICENSE_CONF_OPTION to split
|
## disable SC2086 to allow EMQX_LICENSE_CONF_OPTION to split
|
||||||
# shellcheck disable=SC2086
|
# shellcheck disable=SC2086
|
||||||
call_hocon -v -t "$NOW_TIME" -s emqx_schema -c "$RUNNER_ETC_DIR"/emqx.conf $EMQX_LICENSE_CONF_OPTION -d "$RUNNER_DATA_DIR"/configs generate
|
call_hocon -v -t "$NOW_TIME" -s $SCHEMA_MOD -c "$RUNNER_ETC_DIR"/emqx.conf $EMQX_LICENSE_CONF_OPTION -d "$RUNNER_DATA_DIR"/configs generate
|
||||||
|
|
||||||
## filenames are per-hocon convention
|
## filenames are per-hocon convention
|
||||||
local CONF_FILE="$CONFIGS_DIR/app.$NOW_TIME.config"
|
local CONF_FILE="$CONFIGS_DIR/app.$NOW_TIME.config"
|
||||||
|
@ -331,7 +332,7 @@ if [ -z "$NAME" ]; then
|
||||||
NAME="$(grep -E '^-s?name' "$LATEST_VM_ARGS" | awk '{print $2}')"
|
NAME="$(grep -E '^-s?name' "$LATEST_VM_ARGS" | awk '{print $2}')"
|
||||||
else
|
else
|
||||||
# for boot commands, inspect emqx.conf for node name
|
# for boot commands, inspect emqx.conf for node name
|
||||||
NAME="$(call_hocon -s emqx_schema -c "$RUNNER_ETC_DIR"/emqx.conf get node.name | tr -d \")"
|
NAME="$(call_hocon -s $SCHEMA_MOD -c "$RUNNER_ETC_DIR"/emqx.conf get node.name | tr -d \")"
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
@ -353,7 +354,7 @@ PIPE_DIR="${PIPE_DIR:-/$RUNNER_DATA_DIR/${WHOAMI}_erl_pipes/$NAME/}"
|
||||||
COOKIE="${EMQX_NODE_COOKIE:-}"
|
COOKIE="${EMQX_NODE_COOKIE:-}"
|
||||||
if [ -z "$COOKIE" ]; then
|
if [ -z "$COOKIE" ]; then
|
||||||
if [ "$IS_BOOT_COMMAND" = 'yes' ]; then
|
if [ "$IS_BOOT_COMMAND" = 'yes' ]; then
|
||||||
COOKIE="$(call_hocon -s emqx_schema -c "$RUNNER_ETC_DIR"/emqx.conf get node.cookie | tr -d \")"
|
COOKIE="$(call_hocon -s $SCHEMA_MOD -c "$RUNNER_ETC_DIR"/emqx.conf get node.cookie | tr -d \")"
|
||||||
else
|
else
|
||||||
# shellcheck disable=SC2012,SC2086
|
# shellcheck disable=SC2012,SC2086
|
||||||
LATEST_VM_ARGS="$(ls -t $CONFIGS_DIR/vm.*.args | head -1)"
|
LATEST_VM_ARGS="$(ls -t $CONFIGS_DIR/vm.*.args | head -1)"
|
||||||
|
@ -370,7 +371,7 @@ if [ -z "$COOKIE" ]; then
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Support for IPv6 Dist. See: https://github.com/emqtt/emqttd/issues/1460
|
# Support for IPv6 Dist. See: https://github.com/emqtt/emqttd/issues/1460
|
||||||
PROTO_DIST="$(call_hocon -s emqx_schema -c "$RUNNER_ETC_DIR"/emqx.conf get cluster.proto_dist | tr -d \")"
|
PROTO_DIST="$(call_hocon -s $SCHEMA_MOD -c "$RUNNER_ETC_DIR"/emqx.conf get cluster.proto_dist | tr -d \")"
|
||||||
if [ -z "$PROTO_DIST" ]; then
|
if [ -z "$PROTO_DIST" ]; then
|
||||||
PROTO_DIST_ARG=""
|
PROTO_DIST_ARG=""
|
||||||
else
|
else
|
||||||
|
|
|
@ -1,22 +0,0 @@
|
||||||
%% -*-: erlang -*-
|
|
||||||
%% This file lists all the hocon schemas other than the `emqx_schema`.
|
|
||||||
%% Each element is a two-element tuple where the first element is the field name,
|
|
||||||
%% and the second element is the schema module name.
|
|
||||||
%%
|
|
||||||
[ {"emqx_data_bridge", emqx_data_bridge_schema}
|
|
||||||
, {"emqx_retainer", emqx_retainer_schema}
|
|
||||||
, {"authentication", emqx_authn_schema}
|
|
||||||
, {"authorization", emqx_authz_schema}
|
|
||||||
, {"emqx_bridge_mqtt", emqx_bridge_mqtt_schema}
|
|
||||||
, {"emqx_management", emqx_management_schema}
|
|
||||||
, {"emqx_dashboard", emqx_dashboard_schema}
|
|
||||||
, {"gateway", emqx_gateway_schema}
|
|
||||||
, {"prometheus", emqx_prometheus_schema}
|
|
||||||
, {"statsd", emqx_statsd_schema}
|
|
||||||
, {"delayed", emqx_modules_schema}
|
|
||||||
, {"recon", emqx_modules_schema}
|
|
||||||
, {"telemetry", emqx_modules_schema}
|
|
||||||
, {"event_message", emqx_modules_schema}
|
|
||||||
, {"rewrite", emqx_modules_schema}
|
|
||||||
, {"topic_metrics", emqx_modules_schema}
|
|
||||||
].
|
|
|
@ -79,10 +79,6 @@ is_cover_enabled() ->
|
||||||
is_enterprise() ->
|
is_enterprise() ->
|
||||||
filelib:is_regular("EMQX_ENTERPRISE").
|
filelib:is_regular("EMQX_ENTERPRISE").
|
||||||
|
|
||||||
emqx_ext_schemas() ->
|
|
||||||
{ok, Schemas} = file:script("extension_schemas.config"),
|
|
||||||
Schemas.
|
|
||||||
|
|
||||||
is_quicer_supported() ->
|
is_quicer_supported() ->
|
||||||
not (false =/= os:getenv("BUILD_WITHOUT_QUIC") orelse
|
not (false =/= os:getenv("BUILD_WITHOUT_QUIC") orelse
|
||||||
is_win32() orelse is_centos_6()
|
is_win32() orelse is_centos_6()
|
||||||
|
@ -144,7 +140,6 @@ common_compile_opts() ->
|
||||||
] ++
|
] ++
|
||||||
[{d, 'EMQX_DEP_APPS', AppNames -- [emqx]}] ++
|
[{d, 'EMQX_DEP_APPS', AppNames -- [emqx]}] ++
|
||||||
[{d, 'EMQX_ENTERPRISE'} || is_enterprise()] ++
|
[{d, 'EMQX_ENTERPRISE'} || is_enterprise()] ++
|
||||||
[{d, 'EMQX_EXT_SCHEMAS', emqx_ext_schemas()}] ++
|
|
||||||
[{d, 'EMQX_BENCHMARK'} || os:getenv("EMQX_BENCHMARK") =:= "1" ].
|
[{d, 'EMQX_BENCHMARK'} || os:getenv("EMQX_BENCHMARK") =:= "1" ].
|
||||||
|
|
||||||
prod_compile_opts() ->
|
prod_compile_opts() ->
|
||||||
|
@ -383,7 +378,7 @@ emqx_etc_overlay(edge) ->
|
||||||
].
|
].
|
||||||
|
|
||||||
emqx_etc_overlay_common() ->
|
emqx_etc_overlay_common() ->
|
||||||
[ {"{{base_dir}}/lib/emqx/etc/emqx.conf.all", "etc/emqx.conf"}
|
[ {"{{base_dir}}/lib/emqx_machine/etc/emqx.conf.all", "etc/emqx.conf"}
|
||||||
, {"{{base_dir}}/lib/emqx/etc/ssl_dist.conf", "etc/ssl_dist.conf"}
|
, {"{{base_dir}}/lib/emqx/etc/ssl_dist.conf", "etc/ssl_dist.conf"}
|
||||||
].
|
].
|
||||||
|
|
||||||
|
|
|
@ -11,9 +11,9 @@
|
||||||
-mode(compile).
|
-mode(compile).
|
||||||
|
|
||||||
main(_) ->
|
main(_) ->
|
||||||
BaseConf = "apps/emqx/etc/emqx.conf",
|
BaseConf = "apps/emqx_machine/etc/emqx_machine.conf",
|
||||||
{ok, Bin} = file:read_file(BaseConf),
|
{ok, Bin} = file:read_file(BaseConf),
|
||||||
Apps = filelib:wildcard("emqx_*", "apps/"),
|
Apps = filelib:wildcard("*", "apps/"),
|
||||||
Conf = lists:foldl(fun(App, Acc) ->
|
Conf = lists:foldl(fun(App, Acc) ->
|
||||||
Filename = filename:join([apps, App, "etc", App]) ++ ".conf",
|
Filename = filename:join([apps, App, "etc", App]) ++ ".conf",
|
||||||
case filelib:is_regular(Filename) of
|
case filelib:is_regular(Filename) of
|
||||||
|
@ -23,4 +23,4 @@ main(_) ->
|
||||||
false -> Acc
|
false -> Acc
|
||||||
end
|
end
|
||||||
end, Bin, Apps),
|
end, Bin, Apps),
|
||||||
ok = file:write_file("apps/emqx/etc/emqx.conf.all", Conf).
|
ok = file:write_file("apps/emqx_machine/etc/emqx.conf.all", Conf).
|
||||||
|
|
Loading…
Reference in New Issue