Skip to content

Commit 41d6864

Browse files
author
TeamCityBot
committed
Update metrics-list.csv for v24.3
1 parent 3159501 commit 41d6864

File tree

1 file changed

+67
-6
lines changed

1 file changed

+67
-6
lines changed

src/current/_data/v24.3/metrics/metrics-list.csv

Lines changed: 67 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -762,6 +762,8 @@ have a good estimate for this information for all of its followers, and since
762762
followers are expected to be behind (when they are not required as part of a
763763
quorum) *and* the aggregate thus scales like the count of such followers, it is
764764
difficult to meaningfully interpret this metric.",Log Entries,GAUGE,COUNT,AVG,NONE
765+
STORAGE,raftlog.size.max,Approximate size of the largest Raft log on the store.,Bytes,GAUGE,BYTES,AVG,NONE
766+
STORAGE,raftlog.size.total,Approximate size of all Raft logs on the store.,Bytes,GAUGE,BYTES,AVG,NONE
765767
STORAGE,raftlog.truncated,Number of Raft log entries truncated,Log Entries,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
766768
STORAGE,range.adds,Number of range additions,Range Ops,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
767769
STORAGE,range.merges,Number of range merges,Range Ops,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -1252,7 +1254,7 @@ APPLICATION,changefeed.frontier_updates,Number of change frontier updates across
12521254
APPLICATION,changefeed.internal_retry_message_count,Number of messages for which an attempt to retry them within an aggregator node was made,Messages,GAUGE,COUNT,AVG,NONE
12531255
APPLICATION,changefeed.kafka_throttling_hist_nanos,Time spent in throttling due to exceeding kafka quota,Nanoseconds,HISTOGRAM,NANOSECONDS,AVG,NONE
12541256
APPLICATION,changefeed.lagging_ranges,The number of ranges considered to be lagging behind,Ranges,GAUGE,COUNT,AVG,NONE
1255-
APPLICATION,changefeed.max_behind_nanos,(Deprecated in favor of checkpoint_progress) The most any changefeed's persisted checkpoint is behind the present,Nanoseconds,GAUGE,NANOSECONDS,AVG,NONE
1257+
APPLICATION,changefeed.max_behind_nanos,The most any changefeed's persisted checkpoint is behind the present,Nanoseconds,GAUGE,NANOSECONDS,AVG,NONE
12561258
APPLICATION,changefeed.message_size_hist,Message size histogram,Bytes,HISTOGRAM,BYTES,AVG,NONE
12571259
APPLICATION,changefeed.messages.messages_pushback_nanos,Total time spent throttled for messages quota,Nanoseconds,COUNTER,NANOSECONDS,AVG,NON_NEGATIVE_DERIVATIVE
12581260
APPLICATION,changefeed.network.bytes_in,The number of bytes received from the network by changefeeds,Bytes,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -1355,6 +1357,7 @@ APPLICATION,distsender.rangefeed.retry.replica_removed,Number of ranges that enc
13551357
APPLICATION,distsender.rangefeed.retry.send,Number of ranges that encountered retryable send error,Ranges,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
13561358
APPLICATION,distsender.rangefeed.retry.slow_consumer,Number of ranges that encountered retryable SLOW_CONSUMER error,Ranges,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
13571359
APPLICATION,distsender.rangefeed.retry.store_not_found,Number of ranges that encountered retryable store not found error,Ranges,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
1360+
APPLICATION,distsender.rangefeed.retry.unknown,Number of ranges that encountered retryable unknown error,Ranges,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
13581361
APPLICATION,distsender.rangefeed.total_ranges,"Number of ranges executing rangefeed
13591362

13601363
This counts the number of ranges with an active rangefeed.
@@ -2210,6 +2213,7 @@ APPLICATION,jobs.row_level_ttl.fail_or_cancel_completed,Number of row_level_ttl
22102213
APPLICATION,jobs.row_level_ttl.fail_or_cancel_failed,Number of row_level_ttl jobs which failed with a non-retriable error on their failure or cancelation process,jobs,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22112214
APPLICATION,jobs.row_level_ttl.fail_or_cancel_retry_error,Number of row_level_ttl jobs which failed with a retriable error on their failure or cancelation process,jobs,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22122215
APPLICATION,jobs.row_level_ttl.num_active_spans,Number of active spans the TTL job is deleting from.,num_active_spans,GAUGE,COUNT,AVG,NONE
2216+
APPLICATION,jobs.row_level_ttl.num_delete_batch_retries,Number of times the row level TTL job had to reduce the delete batch size and retry.,num_retries,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22132217
APPLICATION,jobs.row_level_ttl.protected_age_sec,The age of the oldest PTS record protected by row_level_ttl jobs,seconds,GAUGE,SECONDS,AVG,NONE
22142218
APPLICATION,jobs.row_level_ttl.protected_record_count,Number of protected timestamp records held by row_level_ttl jobs,records,GAUGE,COUNT,AVG,NONE
22152219
APPLICATION,jobs.row_level_ttl.resume_completed,Number of row_level_ttl jobs which successfully resumed to completion,jobs,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -2288,16 +2292,16 @@ APPLICATION,kv.protectedts.reconciliation.records_processed,number of records pr
22882292
APPLICATION,kv.protectedts.reconciliation.records_removed,number of records removed during reconciliation runs on this node,Count,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22892293
APPLICATION,logical_replication.batch_hist_nanos,Time spent flushing a batch,Nanoseconds,HISTOGRAM,NANOSECONDS,AVG,NONE
22902294
APPLICATION,logical_replication.catchup_ranges,Source side ranges undergoing catch up scans (inaccurate with multiple LDR jobs),Ranges,GAUGE,COUNT,AVG,NONE
2291-
APPLICATION,logical_replication.catchup_ranges_by_label,Source side ranges undergoing catch up scans,Ranges,GAUGE,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2295+
APPLICATION,logical_replication.catchup_ranges_by_label,Source side ranges undergoing catch up scans,Ranges,GAUGE,COUNT,AVG,NONE
22922296
APPLICATION,logical_replication.checkpoint_events_ingested,Checkpoint events ingested by all replication jobs,Events,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22932297
APPLICATION,logical_replication.commit_latency,"Event commit latency: a difference between event MVCC timestamp and the time it was flushed into disk. If we batch events, then the difference between the oldest event in the batch and flush is recorded",Nanoseconds,HISTOGRAM,NANOSECONDS,AVG,NONE
22942298
APPLICATION,logical_replication.events_dlqed,Row update events sent to DLQ,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22952299
APPLICATION,logical_replication.events_dlqed_age,Row update events sent to DLQ due to reaching the maximum time allowed in the retry queue,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2296-
APPLICATION,logical_replication.events_dlqed_by_label,Row update events sent to DLQ by label,Failures,GAUGE,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2300+
APPLICATION,logical_replication.events_dlqed_by_label,Row update events sent to DLQ by label,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22972301
APPLICATION,logical_replication.events_dlqed_errtype,Row update events sent to DLQ due to an error not considered retryable,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22982302
APPLICATION,logical_replication.events_dlqed_space,Row update events sent to DLQ due to capacity of the retry queue,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
22992303
APPLICATION,logical_replication.events_ingested,Events ingested by all replication jobs,Events,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2300-
APPLICATION,logical_replication.events_ingested_by_label,Events ingested by all replication jobs by label,Events,GAUGE,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2304+
APPLICATION,logical_replication.events_ingested_by_label,Events ingested by all replication jobs by label,Events,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
23012305
APPLICATION,logical_replication.events_initial_failure,Failed attempts to apply an incoming row update,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
23022306
APPLICATION,logical_replication.events_initial_success,Successful applications of an incoming row update,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
23032307
APPLICATION,logical_replication.events_retry_failure,Failed re-attempts to apply a row update,Failures,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -2306,12 +2310,12 @@ APPLICATION,logical_replication.kv.update_too_old,Total number of updates that w
23062310
APPLICATION,logical_replication.kv.value_refreshes,Total number of batches that refreshed the previous value,Events,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
23072311
APPLICATION,logical_replication.logical_bytes,Logical bytes (sum of keys + values) received by all replication jobs,Bytes,COUNTER,BYTES,AVG,NON_NEGATIVE_DERIVATIVE
23082312
APPLICATION,logical_replication.replan_count,Total number of dist sql replanning events,Events,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2309-
APPLICATION,logical_replication.replicated_time_by_label,Replicated time of the logical replication stream by label,Seconds,GAUGE,SECONDS,AVG,NON_NEGATIVE_DERIVATIVE
2313+
APPLICATION,logical_replication.replicated_time_by_label,Replicated time of the logical replication stream by label,Seconds,GAUGE,SECONDS,AVG,NONE
23102314
APPLICATION,logical_replication.replicated_time_seconds,The replicated time of the logical replication stream in seconds since the unix epoch.,Seconds,GAUGE,SECONDS,AVG,NONE
23112315
APPLICATION,logical_replication.retry_queue_bytes,The replicated time of the logical replication stream in seconds since the unix epoch.,Bytes,GAUGE,BYTES,AVG,NONE
23122316
APPLICATION,logical_replication.retry_queue_events,The replicated time of the logical replication stream in seconds since the unix epoch.,Events,GAUGE,COUNT,AVG,NONE
23132317
APPLICATION,logical_replication.scanning_ranges,Source side ranges undergoing an initial scan (inaccurate with multiple LDR jobs),Ranges,GAUGE,COUNT,AVG,NONE
2314-
APPLICATION,logical_replication.scanning_ranges_by_label,Source side ranges undergoing an initial scan,Ranges,GAUGE,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2318+
APPLICATION,logical_replication.scanning_ranges_by_label,Source side ranges undergoing an initial scan,Ranges,GAUGE,COUNT,AVG,NONE
23152319
APPLICATION,obs.tablemetadata.update_job.duration,Time spent running the update table metadata job.,Duration,HISTOGRAM,NANOSECONDS,AVG,NONE
23162320
APPLICATION,obs.tablemetadata.update_job.errors,The total number of errors that have been emitted from the update table metadata job.,Errors,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
23172321
APPLICATION,obs.tablemetadata.update_job.runs,The total number of runs of the update table metadata job.,Executions,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -2333,6 +2337,10 @@ Note that this is not a good signal for KV health. The remote side of the
23332337
RPCs tracked here may experience contention, so an end user can easily
23342338
cause values for this metric to be emitted by leaving a transaction open
23352339
for a long time and contending with it using a second transaction.",Requests,GAUGE,COUNT,AVG,NONE
2340+
APPLICATION,round-trip-default-class-latency,"Distribution of round-trip latencies with other nodes.
2341+
2342+
Similar to round-trip-latency, but only for default class connections.
2343+
",Round-trip time,HISTOGRAM,NANOSECONDS,AVG,NONE
23362344
APPLICATION,round-trip-latency,"Distribution of round-trip latencies with other nodes.
23372345

23382346
This only reflects successful heartbeats and measures gRPC overhead as well as
@@ -2343,6 +2351,18 @@ metrics such as packet loss, retransmits, etc, to conclusively diagnose network
23432351
issues. Heartbeats are not very frequent (~seconds), so they may not capture
23442352
rare or short-lived degradations.
23452353
",Round-trip time,HISTOGRAM,NANOSECONDS,AVG,NONE
2354+
APPLICATION,round-trip-raft-class-latency,"Distribution of round-trip latencies with other nodes.
2355+
2356+
Similar to round-trip-latency, but only for raft class connections.
2357+
",Round-trip time,HISTOGRAM,NANOSECONDS,AVG,NONE
2358+
APPLICATION,round-trip-rangefeed-class-latency,"Distribution of round-trip latencies with other nodes.
2359+
2360+
Similar to round-trip-latency, but only for rangefeed class connections.
2361+
",Round-trip time,HISTOGRAM,NANOSECONDS,AVG,NONE
2362+
APPLICATION,round-trip-system-class-latency,"Distribution of round-trip latencies with other nodes.
2363+
2364+
Similar to round-trip-latency, but only for system class connections.
2365+
",Round-trip time,HISTOGRAM,NANOSECONDS,AVG,NONE
23462366
APPLICATION,rpc.client.bytes.egress,Counter of TCP bytes sent via gRPC on connections we initiated.,Bytes,COUNTER,BYTES,AVG,NON_NEGATIVE_DERIVATIVE
23472367
APPLICATION,rpc.client.bytes.ingress,Counter of TCP bytes received via gRPC on connections we initiated.,Bytes,COUNTER,BYTES,AVG,NON_NEGATIVE_DERIVATIVE
23482368
APPLICATION,rpc.connection.avg_round_trip_latency,"Sum of exponentially weighted moving average of round-trip latencies, as measured through a gRPC RPC.
@@ -2575,6 +2595,7 @@ APPLICATION,sql.savepoint.rollback.started.count.internal,Number of `ROLLBACK TO
25752595
APPLICATION,sql.savepoint.started.count,Number of SQL SAVEPOINT statements started,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25762596
APPLICATION,sql.savepoint.started.count.internal,Number of SQL SAVEPOINT statements started (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25772597
APPLICATION,sql.schema.invalid_objects,Gauge of detected invalid objects within the system.descriptor table (measured by querying crdb_internal.invalid_objects),Objects,GAUGE,COUNT,AVG,NONE
2598+
APPLICATION,sql.schema_changer.object_count,Counter of the number of objects in the cluster,Objects,GAUGE,COUNT,AVG,NONE
25782599
APPLICATION,sql.schema_changer.permanent_errors,Counter of the number of permanent errors experienced by the schema changer,Errors,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25792600
APPLICATION,sql.schema_changer.retry_errors,Counter of the number of retriable errors experienced by the schema changer,Errors,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25802601
APPLICATION,sql.schema_changer.running,Gauge of currently running schema changes,Schema changes,GAUGE,COUNT,AVG,NONE
@@ -2585,8 +2606,12 @@ APPLICATION,sql.select.started.count,Number of SQL SELECT statements started,SQL
25852606
APPLICATION,sql.select.started.count.internal,Number of SQL SELECT statements started (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25862607
APPLICATION,sql.service.latency,Latency of SQL request execution,Latency,HISTOGRAM,NANOSECONDS,AVG,NONE
25872608
APPLICATION,sql.service.latency.internal,Latency of SQL request execution (internal queries),SQL Internal Statements,HISTOGRAM,NANOSECONDS,AVG,NONE
2609+
APPLICATION,sql.statement_timeout.count,Count of statements that failed because they exceeded the statement timeout,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2610+
APPLICATION,sql.statement_timeout.count.internal,Count of statements that failed because they exceeded the statement timeout (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25882611
APPLICATION,sql.statements.active,Number of currently active user SQL statements,Active Statements,GAUGE,COUNT,AVG,NONE
25892612
APPLICATION,sql.statements.active.internal,Number of currently active user SQL statements (internal queries),SQL Internal Statements,GAUGE,COUNT,AVG,NONE
2613+
APPLICATION,sql.statements.auto_retry.count,Number of SQL statement automatic retries,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2614+
APPLICATION,sql.statements.auto_retry.count.internal,Number of SQL statement automatic retries (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25902615
APPLICATION,sql.stats.activity.update.latency,The latency of updates made by the SQL activity updater job. Includes failed update attempts,Nanoseconds,HISTOGRAM,NANOSECONDS,AVG,NONE
25912616
APPLICATION,sql.stats.activity.updates.failed,Number of update attempts made by the SQL activity updater job that failed with errors,failed updates,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
25922617
APPLICATION,sql.stats.activity.updates.successful,Number of successful updates made by the SQL activity updater job,successful updates,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -2606,8 +2631,12 @@ APPLICATION,sql.temp_object_cleaner.active_cleaners,number of cleaner tasks curr
26062631
APPLICATION,sql.temp_object_cleaner.schemas_deletion_error,number of errored schema deletions by the temp object cleaner on this node,Count,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26072632
APPLICATION,sql.temp_object_cleaner.schemas_deletion_success,number of successful schema deletions by the temp object cleaner on this node,Count,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26082633
APPLICATION,sql.temp_object_cleaner.schemas_to_delete,number of schemas to be deleted by the temp object cleaner on this node,Count,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2634+
APPLICATION,sql.transaction_timeout.count,Count of statements that failed because they exceeded the transaction timeout,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2635+
APPLICATION,sql.transaction_timeout.count.internal,Count of statements that failed because they exceeded the transaction timeout (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26092636
APPLICATION,sql.txn.abort.count,Number of SQL transaction abort errors,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26102637
APPLICATION,sql.txn.abort.count.internal,Number of SQL transaction abort errors (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2638+
APPLICATION,sql.txn.auto_retry.count,Number of SQL transaction automatic retries,SQL Transactions,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2639+
APPLICATION,sql.txn.auto_retry.count.internal,Number of SQL transaction automatic retries (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26112640
APPLICATION,sql.txn.begin.count,Number of SQL transaction BEGIN statements successfully executed,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26122641
APPLICATION,sql.txn.begin.count.internal,Number of SQL transaction BEGIN statements successfully executed (internal queries),SQL Internal Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
26132642
APPLICATION,sql.txn.begin.started.count,Number of SQL transaction BEGIN statements started,SQL Statements,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
@@ -2739,6 +2768,38 @@ SERVER,sys.host.net.send.bytes,Bytes sent on all network interfaces since this p
27392768
SERVER,sys.host.net.send.drop,Sending packets that got dropped on all network interfaces since this process started (as reported by the OS),Packets,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
27402769
SERVER,sys.host.net.send.err,Error on sending packets on all network interfaces since this process started (as reported by the OS),Packets,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
27412770
SERVER,sys.host.net.send.packets,Packets sent on all network interfaces since this process started (as reported by the OS),Packets,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2771+
SERVER,sys.host.net.send.tcp.fast_retrans_segs,"Segments retransmitted due to the fast retransmission mechanism in TCP.
2772+
Fast retransmissions occur when the sender learns that intermediate segments have been lost.",Segments,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2773+
SERVER,sys.host.net.send.tcp.loss_probes,"
2774+
Number of TCP tail loss probes sent. Loss probes are an optimization to detect
2775+
loss of the last packet earlier than the retransmission timer, and can indicate
2776+
network issues. Tail loss probes are aggressive, so the base rate is often nonzero
2777+
even in healthy networks.",Probes,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2778+
SERVER,sys.host.net.send.tcp.retrans_segs,"
2779+
The number of TCP segments retransmitted across all network interfaces.
2780+
This can indicate packet loss occurring in the network. However, it can
2781+
also be caused by recipient nodes not consuming packets in a timely manner,
2782+
or the local node overflowing its outgoing buffers, for example due to overload.
2783+
2784+
Retransmissions also occur in the absence of problems, as modern TCP stacks
2785+
err on the side of aggressively retransmitting segments.
2786+
2787+
The linux tool 'ss -i' can show the Linux kernel's smoothed view of round-trip
2788+
latency and variance on a per-connection basis. Additionally, 'netstat -s'
2789+
shows all TCP counters maintained by the kernel.
2790+
",Segments,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2791+
SERVER,sys.host.net.send.tcp.slow_start_retrans,"
2792+
Number of TCP retransmissions in slow start. This can indicate that the network
2793+
is unable to support the initial fast ramp-up in window size, and can be a sign
2794+
of packet loss or congestion.
2795+
",Segments,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
2796+
SERVER,sys.host.net.send.tcp_timeouts,"
2797+
Number of TCP retransmission timeouts. These typically imply that a packet has
2798+
not been acknowledged within at least 200ms. Modern TCP stacks use
2799+
optimizations such as fast retransmissions and loss probes to avoid hitting
2800+
retransmission timeouts. Anecdotally, they still occasionally present themselves
2801+
even in supposedly healthy cloud environments.
2802+
",Timeouts,COUNTER,COUNT,AVG,NON_NEGATIVE_DERIVATIVE
27422803
SERVER,sys.rss,Current process RSS,RSS,GAUGE,BYTES,AVG,NONE
27432804
SERVER,sys.runnable.goroutines.per.cpu,"Average number of goroutines that are waiting to run, normalized by number of cores",goroutines,GAUGE,COUNT,AVG,NONE
27442805
SERVER,sys.totalmem,Total memory (both free and used),Memory,GAUGE,BYTES,AVG,NONE

0 commit comments

Comments
 (0)