2015-12-08 15:45:58 -07:00
|
|
|
/*
|
|
|
|
|
* Copyright (c) 2015 Cisco and/or its affiliates.
|
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
|
* You may obtain a copy of the License at:
|
|
|
|
|
*
|
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
*
|
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
|
* limitations under the License.
|
|
|
|
|
*/
|
|
|
|
|
/*
|
|
|
|
|
Copyright (c) 2001, 2002, 2003 Eliot Dresselhaus
|
|
|
|
|
|
|
|
|
|
Permission is hereby granted, free of charge, to any person obtaining
|
|
|
|
|
a copy of this software and associated documentation files (the
|
|
|
|
|
"Software"), to deal in the Software without restriction, including
|
|
|
|
|
without limitation the rights to use, copy, modify, merge, publish,
|
|
|
|
|
distribute, sublicense, and/or sell copies of the Software, and to
|
|
|
|
|
permit persons to whom the Software is furnished to do so, subject to
|
|
|
|
|
the following conditions:
|
|
|
|
|
|
|
|
|
|
The above copyright notice and this permission notice shall be
|
|
|
|
|
included in all copies or substantial portions of the Software.
|
|
|
|
|
|
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
|
|
|
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
|
|
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
|
|
|
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
|
|
|
|
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
|
|
|
|
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
|
|
|
|
WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#ifndef included_time_h
|
|
|
|
|
#define included_time_h
|
|
|
|
|
|
|
|
|
|
#include <vppinfra/clib.h>
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
typedef struct
|
|
|
|
|
{
|
2015-12-08 15:45:58 -07:00
|
|
|
/* Total run time in clock cycles
|
|
|
|
|
since clib_time_init call. */
|
|
|
|
|
u64 total_cpu_time;
|
|
|
|
|
|
|
|
|
|
/* Last recorded time stamp. */
|
|
|
|
|
u64 last_cpu_time;
|
|
|
|
|
|
|
|
|
|
/* CPU clock frequency. */
|
|
|
|
|
f64 clocks_per_second;
|
|
|
|
|
|
|
|
|
|
/* 1 / cpu clock frequency: conversion factor
|
|
|
|
|
from clock cycles into seconds. */
|
|
|
|
|
f64 seconds_per_clock;
|
|
|
|
|
|
|
|
|
|
/* Time stamp of call to clib_time_init call. */
|
|
|
|
|
u64 init_cpu_time;
|
|
|
|
|
|
|
|
|
|
u64 last_verify_cpu_time;
|
|
|
|
|
|
|
|
|
|
/* Same but for reference time (if present). */
|
|
|
|
|
f64 last_verify_reference_time;
|
|
|
|
|
|
|
|
|
|
u32 log2_clocks_per_second, log2_clocks_per_frequency_verify;
|
|
|
|
|
} clib_time_t;
|
|
|
|
|
|
|
|
|
|
/* Return CPU time stamp as 64bit number. */
|
|
|
|
|
#if defined(__x86_64__) || defined(i386)
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u32 a, d;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("rdtsc":"=a" (a), "=d" (d));
|
2015-12-08 15:45:58 -07:00
|
|
|
return (u64) a + ((u64) d << (u64) 32);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#elif defined (__powerpc64__)
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u64 t;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("mftb %0":"=r" (t));
|
2015-12-08 15:45:58 -07:00
|
|
|
return t;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#elif defined (__SPU__)
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
#ifdef _XLC
|
|
|
|
|
return spu_rdch (0x8);
|
|
|
|
|
#else
|
2016-08-15 11:12:27 -04:00
|
|
|
return 0 /* __builtin_si_rdch (0x8) FIXME */ ;
|
2015-12-08 15:45:58 -07:00
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#elif defined (__powerpc__)
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u32 hi1, hi2, lo;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("1:\n"
|
|
|
|
|
"mftbu %[hi1]\n"
|
|
|
|
|
"mftb %[lo]\n"
|
|
|
|
|
"mftbu %[hi2]\n"
|
|
|
|
|
"cmpw %[hi1],%[hi2]\n"
|
|
|
|
|
"bne 1b\n":[hi1] "=r" (hi1),[hi2] "=r" (hi2),[lo] "=r" (lo));
|
2015-12-08 15:45:58 -07:00
|
|
|
return (u64) lo + ((u64) hi2 << (u64) 32);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#elif defined (__arm__)
|
2016-05-13 07:07:28 +00:00
|
|
|
#if defined(__ARM_ARCH_8A__)
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void) /* We may run arm64 in aarch32 mode, to leverage 64bit counter */
|
2016-04-09 12:38:49 +09:00
|
|
|
{
|
|
|
|
|
u64 tsc;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("mrrc p15, 0, %Q0, %R0, c9":"=r" (tsc));
|
2016-04-09 12:38:49 +09:00
|
|
|
return tsc;
|
|
|
|
|
}
|
2016-05-13 07:07:28 +00:00
|
|
|
#elif defined(__ARM_ARCH_7A__)
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2016-05-13 07:07:28 +00:00
|
|
|
{
|
|
|
|
|
u32 tsc;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("mrc p15, 0, %0, c9, c13, 0":"=r" (tsc));
|
|
|
|
|
return (u64) tsc;
|
2016-05-13 07:07:28 +00:00
|
|
|
}
|
2016-04-09 12:38:49 +09:00
|
|
|
#else
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u32 lo;
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("mrc p15, 0, %[lo], c15, c12, 1":[lo] "=r" (lo));
|
2015-12-08 15:45:58 -07:00
|
|
|
return (u64) lo;
|
|
|
|
|
}
|
2016-04-09 12:38:49 +09:00
|
|
|
#endif
|
2015-12-08 15:45:58 -07:00
|
|
|
|
|
|
|
|
#elif defined (__xtensa__)
|
|
|
|
|
|
|
|
|
|
/* Stub for now. */
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
|
|
|
|
{
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2015-12-08 15:45:58 -07:00
|
|
|
|
|
|
|
|
#elif defined (__TMS320C6X__)
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u32 l, h;
|
|
|
|
|
|
|
|
|
|
asm volatile (" dint\n"
|
|
|
|
|
" mvc .s2 TSCL,%0\n"
|
2016-08-15 11:12:27 -04:00
|
|
|
" mvc .s2 TSCH,%1\n" " rint\n":"=b" (l), "=b" (h));
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
return ((u64) h << 32) | l;
|
2015-12-08 15:45:58 -07:00
|
|
|
}
|
|
|
|
|
|
2016-01-22 08:23:09 -05:00
|
|
|
#elif defined (__aarch64__)
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
clib_cpu_time_now (void)
|
2016-01-22 08:23:09 -05:00
|
|
|
{
|
|
|
|
|
u64 tsc;
|
|
|
|
|
|
|
|
|
|
/* Works on Cavium ThunderX. Other platforms: YMMV */
|
2016-08-15 11:12:27 -04:00
|
|
|
asm volatile ("mrs %0, cntvct_el0":"=r" (tsc));
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2016-01-22 08:23:09 -05:00
|
|
|
return tsc;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#else
|
2015-12-08 15:45:58 -07:00
|
|
|
#error "don't know how to read CPU time stamp"
|
|
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
void clib_time_verify_frequency (clib_time_t * c);
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline f64
|
|
|
|
|
clib_time_now_internal (clib_time_t * c, u64 n)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u64 l = c->last_cpu_time;
|
|
|
|
|
u64 t = c->total_cpu_time;
|
|
|
|
|
t += n - l;
|
|
|
|
|
c->total_cpu_time = t;
|
|
|
|
|
c->last_cpu_time = n;
|
2016-08-15 11:12:27 -04:00
|
|
|
if (PREDICT_FALSE
|
|
|
|
|
((c->last_cpu_time -
|
|
|
|
|
c->last_verify_cpu_time) >> c->log2_clocks_per_frequency_verify))
|
2015-12-08 15:45:58 -07:00
|
|
|
clib_time_verify_frequency (c);
|
|
|
|
|
return t * c->seconds_per_clock;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
always_inline f64
|
|
|
|
|
clib_time_now (clib_time_t * c)
|
|
|
|
|
{
|
2016-08-15 11:12:27 -04:00
|
|
|
return clib_time_now_internal (c, clib_cpu_time_now ());
|
2015-12-08 15:45:58 -07:00
|
|
|
}
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline void
|
|
|
|
|
clib_cpu_time_wait (u64 dt)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
u64 t_end = clib_cpu_time_now () + dt;
|
|
|
|
|
while (clib_cpu_time_now () < t_end)
|
|
|
|
|
;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void clib_time_init (clib_time_t * c);
|
|
|
|
|
|
|
|
|
|
#ifdef CLIB_UNIX
|
|
|
|
|
|
|
|
|
|
#include <time.h>
|
|
|
|
|
#include <sys/time.h>
|
|
|
|
|
#include <sys/resource.h>
|
|
|
|
|
#include <unistd.h>
|
|
|
|
|
#include <sys/syscall.h>
|
|
|
|
|
|
|
|
|
|
/* Use 64bit floating point to represent time offset from epoch. */
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline f64
|
|
|
|
|
unix_time_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
/* clock_gettime without indirect syscall uses GLIBC wrappers which
|
|
|
|
|
we don't want. Just the bare metal, please. */
|
|
|
|
|
struct timespec ts;
|
|
|
|
|
syscall (SYS_clock_gettime, CLOCK_REALTIME, &ts);
|
2016-08-15 11:12:27 -04:00
|
|
|
return ts.tv_sec + 1e-9 * ts.tv_nsec;
|
2015-12-08 15:45:58 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* As above but integer number of nano-seconds. */
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
unix_time_now_nsec (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
struct timespec ts;
|
|
|
|
|
syscall (SYS_clock_gettime, CLOCK_REALTIME, &ts);
|
2016-08-15 11:12:27 -04:00
|
|
|
return 1e9 * ts.tv_sec + ts.tv_nsec;
|
2015-12-08 15:45:58 -07:00
|
|
|
}
|
|
|
|
|
|
2017-06-13 21:15:40 +02:00
|
|
|
always_inline void
|
|
|
|
|
unix_time_now_nsec_fraction (u32 * sec, u32 * nsec)
|
|
|
|
|
{
|
|
|
|
|
struct timespec ts;
|
|
|
|
|
syscall (SYS_clock_gettime, CLOCK_REALTIME, &ts);
|
|
|
|
|
*sec = ts.tv_sec;
|
|
|
|
|
*nsec = ts.tv_nsec;
|
|
|
|
|
}
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline f64
|
|
|
|
|
unix_usage_now (void)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
struct rusage u;
|
|
|
|
|
getrusage (RUSAGE_SELF, &u);
|
2016-08-15 11:12:27 -04:00
|
|
|
return u.ru_utime.tv_sec + 1e-6 * u.ru_utime.tv_usec
|
|
|
|
|
+ u.ru_stime.tv_sec + 1e-6 * u.ru_stime.tv_usec;
|
2015-12-08 15:45:58 -07:00
|
|
|
}
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline void
|
|
|
|
|
unix_sleep (f64 dt)
|
2015-12-08 15:45:58 -07:00
|
|
|
{
|
|
|
|
|
struct timespec t;
|
|
|
|
|
t.tv_sec = dt;
|
|
|
|
|
t.tv_nsec = 1e9 * dt;
|
|
|
|
|
nanosleep (&t, 0);
|
|
|
|
|
}
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
#else /* ! CLIB_UNIX */
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline f64
|
|
|
|
|
unix_time_now (void)
|
|
|
|
|
{
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline u64
|
|
|
|
|
unix_time_now_nsec (void)
|
|
|
|
|
{
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2017-06-13 21:15:40 +02:00
|
|
|
always_inline void
|
|
|
|
|
unix_time_now_nsec_fraction (u32 * sec, u32 * nsec)
|
|
|
|
|
{
|
|
|
|
|
}
|
|
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline f64
|
|
|
|
|
unix_usage_now (void)
|
|
|
|
|
{
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2015-12-08 15:45:58 -07:00
|
|
|
|
2016-08-15 11:12:27 -04:00
|
|
|
always_inline void
|
|
|
|
|
unix_sleep (f64 dt)
|
|
|
|
|
{
|
|
|
|
|
}
|
2015-12-08 15:45:58 -07:00
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#endif /* included_time_h */
|
2016-08-15 11:12:27 -04:00
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* fd.io coding-style-patch-verification: ON
|
|
|
|
|
*
|
|
|
|
|
* Local Variables:
|
|
|
|
|
* eval: (c-set-style "gnu")
|
|
|
|
|
* End:
|
|
|
|
|
*/
|