f5b27cbcc7
Type: refactor Change-Id: I1303219f9f2a25d821737665903b0264edd3de32 Signed-off-by: Damjan Marion <damarion@cisco.com>
548 lines
13 KiB
C
548 lines
13 KiB
C
/*
|
|
* perfmon_periodic.c - skeleton plug-in periodic function
|
|
*
|
|
* Copyright (c) <current-year> <your-organization>
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <vlib/vlib.h>
|
|
#include <vppinfra/error.h>
|
|
#include <perfmon/perfmon.h>
|
|
#include <asm/unistd.h>
|
|
#include <sys/ioctl.h>
|
|
|
|
/* "not in glibc" */
|
|
static long
|
|
perf_event_open (struct perf_event_attr *hw_event, pid_t pid, int cpu,
|
|
int group_fd, unsigned long flags)
|
|
{
|
|
int ret;
|
|
|
|
ret = syscall (__NR_perf_event_open, hw_event, pid, cpu, group_fd, flags);
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
read_current_perf_counters (vlib_node_runtime_perf_callback_data_t * data,
|
|
vlib_node_runtime_perf_callback_args_t * args)
|
|
{
|
|
int i;
|
|
perfmon_main_t *pm = &perfmon_main;
|
|
perfmon_thread_t *pt = data->u[0].v;
|
|
u64 c[2] = { 0, 0 };
|
|
u64 *cc;
|
|
|
|
if (PREDICT_FALSE (args->call_type == VLIB_NODE_RUNTIME_PERF_RESET))
|
|
return;
|
|
|
|
if (args->call_type == VLIB_NODE_RUNTIME_PERF_BEFORE)
|
|
cc = pt->c;
|
|
else
|
|
cc = c;
|
|
|
|
for (i = 0; i < pm->n_active; i++)
|
|
{
|
|
if (pt->rdpmc_indices[i] != ~0)
|
|
cc[i] = clib_rdpmc ((int) pt->rdpmc_indices[i]);
|
|
else
|
|
{
|
|
u64 sw_value;
|
|
int read_result;
|
|
if ((read_result = read (pt->pm_fds[i], &sw_value,
|
|
sizeof (sw_value))) != sizeof (sw_value))
|
|
{
|
|
clib_unix_warning
|
|
("counter read returned %d, expected %d",
|
|
read_result, sizeof (sw_value));
|
|
clib_callback_data_enable_disable
|
|
(&args->vm->vlib_node_runtime_perf_callbacks,
|
|
read_current_perf_counters, 0 /* enable */ );
|
|
return;
|
|
}
|
|
cc[i] = sw_value;
|
|
}
|
|
}
|
|
|
|
if (args->call_type == VLIB_NODE_RUNTIME_PERF_AFTER)
|
|
{
|
|
u32 node_index = args->node->node_index;
|
|
vec_validate (pt->counters, node_index);
|
|
pt->counters[node_index].ticks[0] += c[0] - pt->c[0];
|
|
pt->counters[node_index].ticks[1] += c[1] - pt->c[1];
|
|
pt->counters[node_index].vectors += args->packets;
|
|
}
|
|
}
|
|
|
|
static void
|
|
clear_counters (perfmon_main_t * pm)
|
|
{
|
|
int j;
|
|
vlib_main_t *vm = pm->vlib_main;
|
|
vlib_main_t *stat_vm;
|
|
perfmon_thread_t *pt;
|
|
u32 len;
|
|
|
|
|
|
vlib_worker_thread_barrier_sync (vm);
|
|
|
|
for (j = 0; j < vec_len (vlib_mains); j++)
|
|
{
|
|
stat_vm = vlib_mains[j];
|
|
if (stat_vm == 0)
|
|
continue;
|
|
|
|
pt = pm->threads[j];
|
|
len = vec_len (pt->counters);
|
|
if (!len)
|
|
continue;
|
|
|
|
clib_memset (pt->counters, 0, len * sizeof (pt->counters[0]));
|
|
}
|
|
vlib_worker_thread_barrier_release (vm);
|
|
}
|
|
|
|
static void
|
|
enable_current_events (perfmon_main_t * pm)
|
|
{
|
|
struct perf_event_attr pe;
|
|
int fd;
|
|
struct perf_event_mmap_page *p = 0;
|
|
perfmon_event_config_t *c;
|
|
vlib_main_t *vm = vlib_get_main ();
|
|
u32 my_thread_index = vm->thread_index;
|
|
perfmon_thread_t *pt = pm->threads[my_thread_index];
|
|
u32 index;
|
|
int i, limit = 1;
|
|
int cpu;
|
|
vlib_node_runtime_perf_callback_data_t cbdata = { 0 };
|
|
cbdata.fp = read_current_perf_counters;
|
|
cbdata.u[0].v = pt;
|
|
cbdata.u[1].v = vm;
|
|
|
|
if ((pm->current_event + 1) < vec_len (pm->single_events_to_collect))
|
|
limit = 2;
|
|
|
|
for (i = 0; i < limit; i++)
|
|
{
|
|
c = vec_elt_at_index (pm->single_events_to_collect,
|
|
pm->current_event + i);
|
|
|
|
memset (&pe, 0, sizeof (struct perf_event_attr));
|
|
pe.type = c->pe_type;
|
|
pe.size = sizeof (struct perf_event_attr);
|
|
pe.config = c->pe_config;
|
|
pe.disabled = 1;
|
|
pe.pinned = 1;
|
|
/*
|
|
* Note: excluding the kernel makes the
|
|
* (software) context-switch counter read 0...
|
|
*/
|
|
if (pe.type != PERF_TYPE_SOFTWARE)
|
|
{
|
|
/* Exclude kernel and hypervisor */
|
|
pe.exclude_kernel = 1;
|
|
pe.exclude_hv = 1;
|
|
}
|
|
|
|
cpu = vm->cpu_id;
|
|
|
|
fd = perf_event_open (&pe, 0, cpu, -1, 0);
|
|
if (fd == -1)
|
|
{
|
|
clib_unix_warning ("event open: type %d config %d", c->pe_type,
|
|
c->pe_config);
|
|
return;
|
|
}
|
|
|
|
if (pe.type != PERF_TYPE_SOFTWARE)
|
|
{
|
|
p = mmap (0, pm->page_size, PROT_READ, MAP_SHARED, fd, 0);
|
|
if (p == MAP_FAILED)
|
|
{
|
|
clib_unix_warning ("mmap");
|
|
close (fd);
|
|
return;
|
|
}
|
|
CLIB_MEM_UNPOISON (p, pm->page_size);
|
|
}
|
|
else
|
|
p = 0;
|
|
|
|
if (ioctl (fd, PERF_EVENT_IOC_RESET, 0) < 0)
|
|
clib_unix_warning ("reset ioctl");
|
|
|
|
if (ioctl (fd, PERF_EVENT_IOC_ENABLE, 0) < 0)
|
|
clib_unix_warning ("enable ioctl");
|
|
|
|
pt->perf_event_pages[i] = (void *) p;
|
|
pt->pm_fds[i] = fd;
|
|
}
|
|
|
|
/*
|
|
* Hardware events must be all opened and enabled before aquiring
|
|
* pmc indices, otherwise the pmc indices might be out-dated.
|
|
*/
|
|
for (i = 0; i < limit; i++)
|
|
{
|
|
p = (struct perf_event_mmap_page *) pt->perf_event_pages[i];
|
|
|
|
/*
|
|
* Software event counters - and others not capable of being
|
|
* read via the "rdpmc" instruction - will be read
|
|
* by system calls.
|
|
*/
|
|
if (p == 0 || p->cap_user_rdpmc == 0)
|
|
index = ~0;
|
|
else
|
|
index = p->index - 1;
|
|
|
|
pt->rdpmc_indices[i] = index;
|
|
}
|
|
|
|
pm->n_active = i;
|
|
/* Enable the main loop counter snapshot mechanism */
|
|
clib_callback_data_add (&vm->vlib_node_runtime_perf_callbacks, cbdata);
|
|
}
|
|
|
|
static void
|
|
disable_events (perfmon_main_t * pm)
|
|
{
|
|
vlib_main_t *vm = vlib_get_main ();
|
|
u32 my_thread_index = vm->thread_index;
|
|
perfmon_thread_t *pt = pm->threads[my_thread_index];
|
|
int i;
|
|
|
|
/* Stop main loop collection */
|
|
clib_callback_data_remove (&vm->vlib_node_runtime_perf_callbacks,
|
|
read_current_perf_counters);
|
|
|
|
for (i = 0; i < pm->n_active; i++)
|
|
{
|
|
if (pt->pm_fds[i] == 0)
|
|
continue;
|
|
|
|
if (ioctl (pt->pm_fds[i], PERF_EVENT_IOC_DISABLE, 0) < 0)
|
|
clib_unix_warning ("disable ioctl");
|
|
|
|
if (pt->perf_event_pages[i])
|
|
{
|
|
if (munmap (pt->perf_event_pages[i], pm->page_size) < 0)
|
|
clib_unix_warning ("munmap");
|
|
pt->perf_event_pages[i] = 0;
|
|
}
|
|
|
|
(void) close (pt->pm_fds[i]);
|
|
pt->pm_fds[i] = 0;
|
|
}
|
|
}
|
|
|
|
static void
|
|
worker_thread_start_event (vlib_main_t * vm)
|
|
{
|
|
perfmon_main_t *pm = &perfmon_main;
|
|
|
|
clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
|
|
vm->worker_thread_main_loop_callback_tmp,
|
|
vm->worker_thread_main_loop_callback_lock,
|
|
worker_thread_start_event, 0 /* disable */ );
|
|
enable_current_events (pm);
|
|
}
|
|
|
|
static void
|
|
worker_thread_stop_event (vlib_main_t * vm)
|
|
{
|
|
perfmon_main_t *pm = &perfmon_main;
|
|
clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
|
|
vm->worker_thread_main_loop_callback_tmp,
|
|
vm->worker_thread_main_loop_callback_lock,
|
|
worker_thread_stop_event, 0 /* disable */ );
|
|
disable_events (pm);
|
|
}
|
|
|
|
static void
|
|
start_event (perfmon_main_t * pm, f64 now, uword event_data)
|
|
{
|
|
int i;
|
|
int last_set;
|
|
int all = 0;
|
|
pm->current_event = 0;
|
|
|
|
if (vec_len (pm->single_events_to_collect) == 0)
|
|
{
|
|
pm->state = PERFMON_STATE_OFF;
|
|
return;
|
|
}
|
|
|
|
last_set = clib_bitmap_last_set (pm->thread_bitmap);
|
|
all = (last_set == ~0);
|
|
|
|
pm->state = PERFMON_STATE_RUNNING;
|
|
clear_counters (pm);
|
|
|
|
/* Start collection on thread 0? */
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, 0))
|
|
{
|
|
/* Start collection on this thread */
|
|
enable_current_events (pm);
|
|
}
|
|
|
|
/* And also on worker threads */
|
|
for (i = 1; i < vec_len (vlib_mains); i++)
|
|
{
|
|
if (vlib_mains[i] == 0)
|
|
continue;
|
|
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, i))
|
|
clib_callback_enable_disable
|
|
(vlib_mains[i]->worker_thread_main_loop_callbacks,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_tmp,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_lock,
|
|
(void *) worker_thread_start_event, 1 /* enable */ );
|
|
}
|
|
}
|
|
|
|
void
|
|
scrape_and_clear_counters (perfmon_main_t * pm)
|
|
{
|
|
int i, j, k;
|
|
vlib_main_t *vm = pm->vlib_main;
|
|
vlib_main_t *stat_vm;
|
|
vlib_node_main_t *nm;
|
|
perfmon_counters_t *ctr;
|
|
perfmon_counters_t *ctrs;
|
|
perfmon_counters_t **ctr_dups = 0;
|
|
perfmon_thread_t *pt;
|
|
perfmon_capture_t *c;
|
|
perfmon_event_config_t *current_event;
|
|
uword *p;
|
|
u8 *counter_name;
|
|
u32 len;
|
|
|
|
/* snapshoot the nodes, including pm counters */
|
|
vlib_worker_thread_barrier_sync (vm);
|
|
|
|
for (j = 0; j < vec_len (vlib_mains); j++)
|
|
{
|
|
stat_vm = vlib_mains[j];
|
|
if (stat_vm == 0)
|
|
continue;
|
|
|
|
pt = pm->threads[j];
|
|
len = vec_len (pt->counters);
|
|
ctrs = 0;
|
|
if (len)
|
|
{
|
|
vec_validate (ctrs, len - 1);
|
|
clib_memcpy (ctrs, pt->counters, len * sizeof (pt->counters[0]));
|
|
clib_memset (pt->counters, 0, len * sizeof (pt->counters[0]));
|
|
}
|
|
vec_add1 (ctr_dups, ctrs);
|
|
}
|
|
|
|
vlib_worker_thread_barrier_release (vm);
|
|
|
|
for (j = 0; j < vec_len (vlib_mains); j++)
|
|
{
|
|
stat_vm = vlib_mains[j];
|
|
if (stat_vm == 0)
|
|
continue;
|
|
|
|
pt = pm->threads[j];
|
|
ctrs = ctr_dups[j];
|
|
|
|
for (i = 0; i < vec_len (ctrs); i++)
|
|
{
|
|
u8 *capture_name;
|
|
|
|
ctr = &ctrs[i];
|
|
nm = &stat_vm->node_main;
|
|
|
|
if (ctr->ticks[0] == 0 && ctr->ticks[1] == 0)
|
|
continue;
|
|
|
|
for (k = 0; k < 2; k++)
|
|
{
|
|
/*
|
|
* We collect 2 counters at once, except for the
|
|
* last counter when the user asks for an odd number of
|
|
* counters
|
|
*/
|
|
if ((pm->current_event + k)
|
|
>= vec_len (pm->single_events_to_collect))
|
|
break;
|
|
|
|
capture_name = format (0, "t%d-%v%c", j, nm->nodes[i]->name, 0);
|
|
|
|
p = hash_get_mem (pm->capture_by_thread_and_node_name,
|
|
capture_name);
|
|
|
|
if (p == 0)
|
|
{
|
|
pool_get (pm->capture_pool, c);
|
|
memset (c, 0, sizeof (*c));
|
|
c->thread_and_node_name = capture_name;
|
|
hash_set_mem (pm->capture_by_thread_and_node_name,
|
|
capture_name, c - pm->capture_pool);
|
|
}
|
|
else
|
|
{
|
|
c = pool_elt_at_index (pm->capture_pool, p[0]);
|
|
vec_free (capture_name);
|
|
}
|
|
|
|
/* Snapshoot counters, etc. into the capture */
|
|
current_event = pm->single_events_to_collect
|
|
+ pm->current_event + k;
|
|
counter_name = (u8 *) current_event->name;
|
|
|
|
vec_add1 (c->counter_names, counter_name);
|
|
vec_add1 (c->counter_values, ctr->ticks[k]);
|
|
vec_add1 (c->vectors_this_counter, ctr->vectors);
|
|
}
|
|
}
|
|
vec_free (ctrs);
|
|
}
|
|
vec_free (ctr_dups);
|
|
}
|
|
|
|
static void
|
|
handle_timeout (vlib_main_t * vm, perfmon_main_t * pm, f64 now)
|
|
{
|
|
int i;
|
|
int last_set, all;
|
|
|
|
last_set = clib_bitmap_last_set (pm->thread_bitmap);
|
|
all = (last_set == ~0);
|
|
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, 0))
|
|
disable_events (pm);
|
|
|
|
/* And also on worker threads */
|
|
for (i = 1; i < vec_len (vlib_mains); i++)
|
|
{
|
|
if (vlib_mains[i] == 0)
|
|
continue;
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, i))
|
|
clib_callback_enable_disable
|
|
(vlib_mains[i]->worker_thread_main_loop_callbacks,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_tmp,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_lock,
|
|
(void *) worker_thread_stop_event, 1 /* enable */ );
|
|
}
|
|
|
|
/* Make sure workers have stopped collection */
|
|
if (i > 1)
|
|
{
|
|
f64 deadman = vlib_time_now (vm) + 1.0;
|
|
|
|
for (i = 1; i < vec_len (vlib_mains); i++)
|
|
{
|
|
/* Has the worker actually stopped collecting data? */
|
|
while (clib_callback_data_is_set
|
|
(&vm->vlib_node_runtime_perf_callbacks,
|
|
read_current_perf_counters))
|
|
{
|
|
if (vlib_time_now (vm) > deadman)
|
|
{
|
|
clib_warning ("Thread %d deadman timeout!", i);
|
|
break;
|
|
}
|
|
vlib_process_suspend (pm->vlib_main, 1e-3);
|
|
}
|
|
}
|
|
}
|
|
scrape_and_clear_counters (pm);
|
|
pm->current_event += pm->n_active;
|
|
if (pm->current_event >= vec_len (pm->single_events_to_collect))
|
|
{
|
|
pm->current_event = 0;
|
|
pm->state = PERFMON_STATE_OFF;
|
|
return;
|
|
}
|
|
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, 0))
|
|
enable_current_events (pm);
|
|
|
|
/* And also on worker threads */
|
|
for (i = 1; i < vec_len (vlib_mains); i++)
|
|
{
|
|
if (vlib_mains[i] == 0)
|
|
continue;
|
|
if (all || clib_bitmap_get (pm->thread_bitmap, i))
|
|
clib_callback_enable_disable
|
|
(vlib_mains[i]->worker_thread_main_loop_callbacks,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_tmp,
|
|
vlib_mains[i]->worker_thread_main_loop_callback_lock,
|
|
worker_thread_start_event, 0 /* disable */ );
|
|
}
|
|
}
|
|
|
|
static uword
|
|
perfmon_periodic_process (vlib_main_t * vm,
|
|
vlib_node_runtime_t * rt, vlib_frame_t * f)
|
|
{
|
|
perfmon_main_t *pm = &perfmon_main;
|
|
f64 now;
|
|
uword *event_data = 0;
|
|
uword event_type;
|
|
int i;
|
|
|
|
while (1)
|
|
{
|
|
if (pm->state == PERFMON_STATE_RUNNING)
|
|
vlib_process_wait_for_event_or_clock (vm, pm->timeout_interval);
|
|
else
|
|
vlib_process_wait_for_event (vm);
|
|
|
|
now = vlib_time_now (vm);
|
|
|
|
event_type = vlib_process_get_events (vm, (uword **) & event_data);
|
|
|
|
switch (event_type)
|
|
{
|
|
case PERFMON_START:
|
|
for (i = 0; i < vec_len (event_data); i++)
|
|
start_event (pm, now, event_data[i]);
|
|
break;
|
|
|
|
/* Handle timeout */
|
|
case ~0:
|
|
handle_timeout (vm, pm, now);
|
|
break;
|
|
|
|
default:
|
|
clib_warning ("Unexpected event %d", event_type);
|
|
break;
|
|
}
|
|
vec_reset_length (event_data);
|
|
}
|
|
return 0; /* or not */
|
|
}
|
|
|
|
/* *INDENT-OFF* */
|
|
VLIB_REGISTER_NODE (perfmon_periodic_node) =
|
|
{
|
|
.function = perfmon_periodic_process,
|
|
.type = VLIB_NODE_TYPE_PROCESS,
|
|
.name = "perfmon-periodic-process",
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
/*
|
|
* fd.io coding-style-patch-verification: ON
|
|
*
|
|
* Local Variables:
|
|
* eval: (c-set-style "gnu")
|
|
* End:
|
|
*/
|