mirror of
https://github.com/prometheus/node_exporter.git
synced 2025-03-05 21:00:12 -08:00
fix lint
Signed-off-by: dongjiang1989 <dongjiang1989@126.com>
This commit is contained in:
parent
8d8b4a415e
commit
42c96a05a1
|
@ -2283,14 +2283,14 @@ node_mountstats_nfs_transport_pending_queue_total{export="192.168.1.1:/srv/test"
|
|||
# TYPE node_mountstats_nfs_transport_pullup_copy_total counter
|
||||
node_mountstats_nfs_transport_pullup_copy_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_pullup_copy_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_reply_total Total number of rdma reply accessed when receiving a reply.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_reply_total gauge
|
||||
node_mountstats_nfs_transport_rdma_reply_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_reply_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_request_total Total number of rdma request accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_request_total gauge
|
||||
node_mountstats_nfs_transport_rdma_request_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_request_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_reply_sum Total number of rdma reply accessed when receiving a reply.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_reply_sum gauge
|
||||
node_mountstats_nfs_transport_rdma_reply_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_reply_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_request_sum Total number of rdma request accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_request_sum gauge
|
||||
node_mountstats_nfs_transport_rdma_request_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_request_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_read_chunk_total Total number of read chunk accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_read_chunk_total counter
|
||||
node_mountstats_nfs_transport_read_chunk_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
|
|
|
@ -2305,14 +2305,14 @@ node_mountstats_nfs_transport_pending_queue_total{export="192.168.1.1:/srv/test"
|
|||
# TYPE node_mountstats_nfs_transport_pullup_copy_total counter
|
||||
node_mountstats_nfs_transport_pullup_copy_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_pullup_copy_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_reply_total Total number of rdma reply accessed when receiving a reply.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_reply_total gauge
|
||||
node_mountstats_nfs_transport_rdma_reply_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_reply_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_request_total Total number of rdma request accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_request_total gauge
|
||||
node_mountstats_nfs_transport_rdma_request_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_request_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_reply_sum Total number of rdma reply accessed when receiving a reply.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_reply_sum gauge
|
||||
node_mountstats_nfs_transport_rdma_reply_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_reply_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_rdma_request_sum Total number of rdma request accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_rdma_request_sum gauge
|
||||
node_mountstats_nfs_transport_rdma_request_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
node_mountstats_nfs_transport_rdma_request_sum{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="udp"} 0
|
||||
# HELP node_mountstats_nfs_transport_read_chunk_total Total number of read chunk accessed when sending a call.
|
||||
# TYPE node_mountstats_nfs_transport_read_chunk_total counter
|
||||
node_mountstats_nfs_transport_read_chunk_total{export="192.168.1.1:/srv/test",mountaddr="192.168.1.1",protocol="tcp"} 0
|
||||
|
|
|
@ -308,7 +308,7 @@ func NewMountStatsCollector(logger log.Logger) (Collector, error) {
|
|||
),
|
||||
|
||||
NFSTransportRdmaRequestTotal: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(namespace, subsystem, "transport_rdma_request_total"),
|
||||
prometheus.BuildFQName(namespace, subsystem, "transport_rdma_request_sum"),
|
||||
"Total number of rdma request accessed when sending a call.",
|
||||
labels,
|
||||
nil,
|
||||
|
@ -371,7 +371,7 @@ func NewMountStatsCollector(logger log.Logger) (Collector, error) {
|
|||
),
|
||||
|
||||
NFSTransportRdmaReplyTotal: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(namespace, subsystem, "transport_rdma_reply_total"),
|
||||
prometheus.BuildFQName(namespace, subsystem, "transport_rdma_reply_sum"),
|
||||
"Total number of rdma reply accessed when receiving a reply.",
|
||||
labels,
|
||||
nil,
|
||||
|
|
Loading…
Reference in a new issue