For whatever reason, "typedef struct { ... } *foo_t" gives cgo a
horrible case of indigestion. A minor refactor makes the pain go away.
Type: refactor
Signed-off-by: Dave Barach <dave@barachs.net>
Change-Id: I856b2abf9b16348d2f6145178e683e722914c756
221 lines
4.9 KiB
C
221 lines
4.9 KiB
C
/*
|
|
* Copyright (c) 2017 Cisco and/or its affiliates.
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#ifndef included_clib_lock_h
|
|
#define included_clib_lock_h
|
|
|
|
#include <vppinfra/clib.h>
|
|
#include <vppinfra/atomics.h>
|
|
|
|
#if __x86_64__
|
|
#define CLIB_PAUSE() __builtin_ia32_pause ()
|
|
#elif defined (__aarch64__) || defined (__arm__)
|
|
#define CLIB_PAUSE() __asm__ ("yield")
|
|
#else
|
|
#define CLIB_PAUSE()
|
|
#endif
|
|
|
|
#if CLIB_DEBUG > 1
|
|
#define CLIB_LOCK_DBG(_p) \
|
|
do { \
|
|
(*_p)->frame_address = __builtin_frame_address (0); \
|
|
(*_p)->pid = getpid (); \
|
|
(*_p)->thread_index = os_get_thread_index (); \
|
|
} while (0)
|
|
#define CLIB_LOCK_DBG_CLEAR(_p) \
|
|
do { \
|
|
(*_p)->frame_address = 0; \
|
|
(*_p)->pid = 0; \
|
|
(*_p)->thread_index = 0; \
|
|
} while (0)
|
|
#else
|
|
#define CLIB_LOCK_DBG(_p)
|
|
#define CLIB_LOCK_DBG_CLEAR(_p)
|
|
#endif
|
|
|
|
#define CLIB_SPINLOCK_IS_LOCKED(_p) (*(_p))->lock
|
|
#define CLIB_SPINLOCK_ASSERT_LOCKED(_p) ASSERT(CLIB_SPINLOCK_IS_LOCKED((_p)))
|
|
|
|
struct clib_spinlock_s
|
|
{
|
|
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
|
|
u32 lock;
|
|
#if CLIB_DEBUG > 0
|
|
pid_t pid;
|
|
uword thread_index;
|
|
void *frame_address;
|
|
#endif
|
|
};
|
|
|
|
typedef struct clib_spinlock_s *clib_spinlock_t;
|
|
|
|
static inline void
|
|
clib_spinlock_init (clib_spinlock_t * p)
|
|
{
|
|
*p = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES, CLIB_CACHE_LINE_BYTES);
|
|
clib_memset ((void *) *p, 0, CLIB_CACHE_LINE_BYTES);
|
|
}
|
|
|
|
static inline void
|
|
clib_spinlock_free (clib_spinlock_t * p)
|
|
{
|
|
if (*p)
|
|
{
|
|
clib_mem_free ((void *) *p);
|
|
*p = 0;
|
|
}
|
|
}
|
|
|
|
static_always_inline void
|
|
clib_spinlock_lock (clib_spinlock_t * p)
|
|
{
|
|
u32 free = 0;
|
|
while (!clib_atomic_cmp_and_swap_acq_relax_n (&(*p)->lock, &free, 1, 0))
|
|
{
|
|
/* atomic load limits number of compare_exchange executions */
|
|
while (clib_atomic_load_relax_n (&(*p)->lock))
|
|
CLIB_PAUSE ();
|
|
/* on failure, compare_exchange writes (*p)->lock into free */
|
|
free = 0;
|
|
}
|
|
CLIB_LOCK_DBG (p);
|
|
}
|
|
|
|
static_always_inline int
|
|
clib_spinlock_trylock (clib_spinlock_t * p)
|
|
{
|
|
if (PREDICT_FALSE (CLIB_SPINLOCK_IS_LOCKED (p)))
|
|
return 0;
|
|
clib_spinlock_lock (p);
|
|
return 1;
|
|
}
|
|
|
|
static_always_inline void
|
|
clib_spinlock_lock_if_init (clib_spinlock_t * p)
|
|
{
|
|
if (PREDICT_FALSE (*p != 0))
|
|
clib_spinlock_lock (p);
|
|
}
|
|
|
|
static_always_inline int
|
|
clib_spinlock_trylock_if_init (clib_spinlock_t * p)
|
|
{
|
|
if (PREDICT_FALSE (*p != 0))
|
|
return clib_spinlock_trylock (p);
|
|
return 1;
|
|
}
|
|
|
|
static_always_inline void
|
|
clib_spinlock_unlock (clib_spinlock_t * p)
|
|
{
|
|
CLIB_LOCK_DBG_CLEAR (p);
|
|
/* Make sure all reads/writes are complete before releasing the lock */
|
|
clib_atomic_release (&(*p)->lock);
|
|
}
|
|
|
|
static_always_inline void
|
|
clib_spinlock_unlock_if_init (clib_spinlock_t * p)
|
|
{
|
|
if (PREDICT_FALSE (*p != 0))
|
|
clib_spinlock_unlock (p);
|
|
}
|
|
|
|
/*
|
|
* Readers-Writer Lock
|
|
*/
|
|
|
|
typedef struct clib_rw_lock_
|
|
{
|
|
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
|
|
/* -1 when W lock held, > 0 when R lock held */
|
|
volatile i32 rw_cnt;
|
|
#if CLIB_DEBUG > 0
|
|
pid_t pid;
|
|
uword thread_index;
|
|
void *frame_address;
|
|
#endif
|
|
} *clib_rwlock_t;
|
|
|
|
always_inline void
|
|
clib_rwlock_init (clib_rwlock_t * p)
|
|
{
|
|
*p = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES, CLIB_CACHE_LINE_BYTES);
|
|
clib_memset ((void *) *p, 0, CLIB_CACHE_LINE_BYTES);
|
|
}
|
|
|
|
always_inline void
|
|
clib_rwlock_free (clib_rwlock_t * p)
|
|
{
|
|
if (*p)
|
|
{
|
|
clib_mem_free ((void *) *p);
|
|
*p = 0;
|
|
}
|
|
}
|
|
|
|
always_inline void
|
|
clib_rwlock_reader_lock (clib_rwlock_t * p)
|
|
{
|
|
i32 cnt;
|
|
do
|
|
{
|
|
/* rwlock held by a writer */
|
|
while ((cnt = clib_atomic_load_relax_n (&(*p)->rw_cnt)) < 0)
|
|
CLIB_PAUSE ();
|
|
}
|
|
while (!clib_atomic_cmp_and_swap_acq_relax_n
|
|
(&(*p)->rw_cnt, &cnt, cnt + 1, 1));
|
|
CLIB_LOCK_DBG (p);
|
|
}
|
|
|
|
always_inline void
|
|
clib_rwlock_reader_unlock (clib_rwlock_t * p)
|
|
{
|
|
ASSERT ((*p)->rw_cnt > 0);
|
|
CLIB_LOCK_DBG_CLEAR (p);
|
|
clib_atomic_fetch_sub_rel (&(*p)->rw_cnt, 1);
|
|
}
|
|
|
|
always_inline void
|
|
clib_rwlock_writer_lock (clib_rwlock_t * p)
|
|
{
|
|
i32 cnt = 0;
|
|
do
|
|
{
|
|
/* rwlock held by writer or reader(s) */
|
|
while ((cnt = clib_atomic_load_relax_n (&(*p)->rw_cnt)) != 0)
|
|
CLIB_PAUSE ();
|
|
}
|
|
while (!clib_atomic_cmp_and_swap_acq_relax_n (&(*p)->rw_cnt, &cnt, -1, 1));
|
|
CLIB_LOCK_DBG (p);
|
|
}
|
|
|
|
always_inline void
|
|
clib_rwlock_writer_unlock (clib_rwlock_t * p)
|
|
{
|
|
CLIB_LOCK_DBG_CLEAR (p);
|
|
clib_atomic_release (&(*p)->rw_cnt);
|
|
}
|
|
|
|
#endif
|
|
|
|
/*
|
|
* fd.io coding-style-patch-verification: ON
|
|
*
|
|
* Local Variables:
|
|
* eval: (c-set-style "gnu")
|
|
* End:
|
|
*/
|