2016-02-10 00:12:00 +01:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-16 01:03:42 +02:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2013-10-16 23:59:46 +02:00
|
|
|
//
|
2013-10-05 07:32:05 +02:00
|
|
|
#pragma once
|
2017-04-06 04:02:00 +02:00
|
|
|
#include "monitoring/perf_step_timer.h"
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 08:09:15 +02:00
|
|
|
#include "rocksdb/perf_context.h"
|
|
|
|
#include "util/stop_watch.h"
|
|
|
|
|
2020-02-20 21:07:53 +01:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2017-09-16 02:02:37 +02:00
|
|
|
#if defined(NPERF_CONTEXT) || !defined(ROCKSDB_SUPPORT_THREAD_LOCAL)
|
|
|
|
extern PerfContext perf_context;
|
|
|
|
#else
|
|
|
|
#if defined(OS_SOLARIS)
|
|
|
|
extern __thread PerfContext perf_context_;
|
2018-06-15 21:39:16 +02:00
|
|
|
#define perf_context (*get_perf_context())
|
2017-09-16 02:02:37 +02:00
|
|
|
#else
|
2018-10-17 20:18:00 +02:00
|
|
|
extern thread_local PerfContext perf_context;
|
2017-09-16 02:02:37 +02:00
|
|
|
#endif
|
|
|
|
#endif
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 08:09:15 +02:00
|
|
|
|
2017-06-03 02:12:39 +02:00
|
|
|
#if defined(NPERF_CONTEXT)
|
2014-04-08 19:58:07 +02:00
|
|
|
|
|
|
|
#define PERF_TIMER_STOP(metric)
|
2014-08-23 00:28:58 +02:00
|
|
|
#define PERF_TIMER_START(metric)
|
2019-08-16 23:36:41 +02:00
|
|
|
#define PERF_TIMER_GUARD(metric)
|
|
|
|
#define PERF_TIMER_GUARD_WITH_ENV(metric, env)
|
|
|
|
#define PERF_CPU_TIMER_GUARD(metric, env)
|
2019-09-20 21:00:55 +02:00
|
|
|
#define PERF_CONDITIONAL_TIMER_FOR_MUTEX_GUARD(metric, condition, stats, \
|
|
|
|
ticker_type)
|
2019-08-16 23:36:41 +02:00
|
|
|
#define PERF_TIMER_MEASURE(metric)
|
2014-04-08 19:58:07 +02:00
|
|
|
#define PERF_COUNTER_ADD(metric, value)
|
2019-08-16 23:36:41 +02:00
|
|
|
#define PERF_COUNTER_BY_LEVEL_ADD(metric, value, level)
|
2014-04-08 19:58:07 +02:00
|
|
|
|
2014-04-04 22:11:44 +02:00
|
|
|
#else
|
2014-04-08 19:58:07 +02:00
|
|
|
|
2014-08-23 00:28:58 +02:00
|
|
|
// Stop the timer and update the metric
|
2017-06-03 02:12:39 +02:00
|
|
|
#define PERF_TIMER_STOP(metric) perf_step_timer_##metric.Stop();
|
2014-04-08 19:58:07 +02:00
|
|
|
|
2017-06-03 02:12:39 +02:00
|
|
|
#define PERF_TIMER_START(metric) perf_step_timer_##metric.Start();
|
2014-04-08 19:58:07 +02:00
|
|
|
|
|
|
|
// Declare and set start time of the timer
|
2017-09-16 02:02:37 +02:00
|
|
|
#define PERF_TIMER_GUARD(metric) \
|
|
|
|
PerfStepTimer perf_step_timer_##metric(&(perf_context.metric)); \
|
2017-06-03 02:12:39 +02:00
|
|
|
perf_step_timer_##metric.Start();
|
|
|
|
|
2018-12-20 21:00:40 +01:00
|
|
|
// Declare and set start time of the timer
|
|
|
|
#define PERF_TIMER_GUARD_WITH_ENV(metric, env) \
|
|
|
|
PerfStepTimer perf_step_timer_##metric(&(perf_context.metric), env); \
|
|
|
|
perf_step_timer_##metric.Start();
|
|
|
|
|
|
|
|
// Declare and set start time of the timer
|
|
|
|
#define PERF_CPU_TIMER_GUARD(metric, env) \
|
|
|
|
PerfStepTimer perf_step_timer_##metric( \
|
|
|
|
&(perf_context.metric), env, true, \
|
|
|
|
PerfLevel::kEnableTimeAndCPUTimeExceptForMutex); \
|
|
|
|
perf_step_timer_##metric.Start();
|
|
|
|
|
|
|
|
#define PERF_CONDITIONAL_TIMER_FOR_MUTEX_GUARD(metric, condition, stats, \
|
|
|
|
ticker_type) \
|
|
|
|
PerfStepTimer perf_step_timer_##metric(&(perf_context.metric), nullptr, \
|
|
|
|
false, PerfLevel::kEnableTime, stats, \
|
|
|
|
ticker_type); \
|
|
|
|
if (condition) { \
|
|
|
|
perf_step_timer_##metric.Start(); \
|
2015-10-13 19:41:48 +02:00
|
|
|
}
|
|
|
|
|
2014-04-08 19:58:07 +02:00
|
|
|
// Update metric with time elapsed since last START. start time is reset
|
|
|
|
// to current timestamp.
|
2017-06-03 02:12:39 +02:00
|
|
|
#define PERF_TIMER_MEASURE(metric) perf_step_timer_##metric.Measure();
|
2014-04-08 19:58:07 +02:00
|
|
|
|
|
|
|
// Increase metric value
|
2017-06-27 00:13:35 +02:00
|
|
|
#define PERF_COUNTER_ADD(metric, value) \
|
|
|
|
if (perf_level >= PerfLevel::kEnableCount) { \
|
2017-09-16 02:02:37 +02:00
|
|
|
perf_context.metric += value; \
|
2017-06-27 00:13:35 +02:00
|
|
|
}
|
2014-04-08 19:58:07 +02:00
|
|
|
|
2018-10-17 20:18:00 +02:00
|
|
|
// Increase metric value
|
|
|
|
#define PERF_COUNTER_BY_LEVEL_ADD(metric, value, level) \
|
|
|
|
if (perf_level >= PerfLevel::kEnableCount && \
|
|
|
|
perf_context.per_level_perf_context_enabled && \
|
|
|
|
perf_context.level_to_perf_context) { \
|
|
|
|
if ((*(perf_context.level_to_perf_context)).find(level) != \
|
|
|
|
(*(perf_context.level_to_perf_context)).end()) { \
|
|
|
|
(*(perf_context.level_to_perf_context))[level].metric += value; \
|
|
|
|
} \
|
|
|
|
else { \
|
|
|
|
PerfContextByLevel empty_context; \
|
|
|
|
(*(perf_context.level_to_perf_context))[level] = empty_context; \
|
|
|
|
(*(perf_context.level_to_perf_context))[level].metric += value; \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
|
2014-04-08 19:58:07 +02:00
|
|
|
#endif
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 08:09:15 +02:00
|
|
|
|
2020-02-20 21:07:53 +01:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|