2017-01-04 13:19:27 +01:00
|
|
|
/*
|
|
|
|
|
* Copyright (c) 2017 Cisco and/or its affiliates.
|
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
|
* You may obtain a copy of the License at:
|
|
|
|
|
*
|
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
*
|
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
|
* limitations under the License.
|
|
|
|
|
*/
|
|
|
|
|
/*
|
|
|
|
|
* buffer.c: allocate/free network buffers.
|
|
|
|
|
*
|
|
|
|
|
* Copyright (c) 2008 Eliot Dresselhaus
|
|
|
|
|
*
|
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining
|
|
|
|
|
* a copy of this software and associated documentation files (the
|
|
|
|
|
* "Software"), to deal in the Software without restriction, including
|
|
|
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
|
|
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
|
|
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
|
|
|
* the following conditions:
|
|
|
|
|
*
|
|
|
|
|
* The above copyright notice and this permission notice shall be
|
|
|
|
|
* included in all copies or substantial portions of the Software.
|
|
|
|
|
*
|
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
|
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
|
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
|
|
|
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
|
|
|
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
|
|
|
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* @file
|
|
|
|
|
*
|
|
|
|
|
* Allocate/free network buffers.
|
|
|
|
|
*/
|
|
|
|
|
|
2018-03-09 16:05:56 +01:00
|
|
|
#include <unistd.h>
|
|
|
|
|
|
2017-01-04 13:19:27 +01:00
|
|
|
#include <rte_config.h>
|
|
|
|
|
|
|
|
|
|
#include <rte_common.h>
|
|
|
|
|
#include <rte_log.h>
|
|
|
|
|
#include <rte_memory.h>
|
|
|
|
|
#include <rte_memzone.h>
|
|
|
|
|
#include <rte_tailq.h>
|
|
|
|
|
#include <rte_eal.h>
|
|
|
|
|
#include <rte_per_lcore.h>
|
|
|
|
|
#include <rte_launch.h>
|
|
|
|
|
#include <rte_atomic.h>
|
|
|
|
|
#include <rte_cycles.h>
|
|
|
|
|
#include <rte_prefetch.h>
|
|
|
|
|
#include <rte_lcore.h>
|
|
|
|
|
#include <rte_per_lcore.h>
|
|
|
|
|
#include <rte_branch_prediction.h>
|
|
|
|
|
#include <rte_interrupts.h>
|
2018-10-20 23:26:46 +02:00
|
|
|
#include <rte_vfio.h>
|
2017-01-04 13:19:27 +01:00
|
|
|
#include <rte_random.h>
|
|
|
|
|
#include <rte_debug.h>
|
|
|
|
|
#include <rte_ether.h>
|
|
|
|
|
#include <rte_ethdev.h>
|
|
|
|
|
#include <rte_ring.h>
|
|
|
|
|
#include <rte_mempool.h>
|
|
|
|
|
#include <rte_mbuf.h>
|
|
|
|
|
#include <rte_version.h>
|
|
|
|
|
|
|
|
|
|
#include <vlib/vlib.h>
|
2018-03-09 16:05:56 +01:00
|
|
|
#include <vlib/unix/unix.h>
|
2017-01-04 13:19:27 +01:00
|
|
|
#include <vnet/vnet.h>
|
2017-02-28 19:22:22 +01:00
|
|
|
#include <dpdk/device/dpdk.h>
|
|
|
|
|
#include <dpdk/device/dpdk_priv.h>
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
STATIC_ASSERT (VLIB_BUFFER_PRE_DATA_SIZE == RTE_PKTMBUF_HEADROOM,
|
|
|
|
|
"VLIB_BUFFER_PRE_DATA_SIZE must be equal to RTE_PKTMBUF_HEADROOM");
|
|
|
|
|
|
2017-11-22 13:27:03 +01:00
|
|
|
typedef struct
|
|
|
|
|
{
|
2017-11-24 11:15:38 +01:00
|
|
|
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
|
2018-06-02 20:42:07 +02:00
|
|
|
struct rte_mbuf **mbuf_alloc_list;
|
2017-11-22 13:27:03 +01:00
|
|
|
} dpdk_buffer_per_thread_data;
|
|
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
|
{
|
2018-03-09 16:05:56 +01:00
|
|
|
int vfio_container_fd;
|
2017-11-22 13:27:03 +01:00
|
|
|
dpdk_buffer_per_thread_data *ptd;
|
|
|
|
|
} dpdk_buffer_main_t;
|
|
|
|
|
|
|
|
|
|
dpdk_buffer_main_t dpdk_buffer_main;
|
2017-11-09 18:02:54 +01:00
|
|
|
|
2017-01-25 14:18:03 +01:00
|
|
|
static_always_inline void
|
2018-06-03 14:29:42 +02:00
|
|
|
dpdk_rte_pktmbuf_free (vlib_main_t * vm, u32 thread_index, vlib_buffer_t * b,
|
|
|
|
|
int maybe_next)
|
2017-01-25 14:18:03 +01:00
|
|
|
{
|
|
|
|
|
struct rte_mbuf *mb;
|
|
|
|
|
u32 next, flags;
|
|
|
|
|
|
|
|
|
|
next:
|
|
|
|
|
flags = b->flags;
|
|
|
|
|
next = b->next_buffer;
|
|
|
|
|
mb = rte_mbuf_from_vlib_buffer (b);
|
|
|
|
|
|
|
|
|
|
if (PREDICT_FALSE (b->n_add_refs))
|
|
|
|
|
{
|
|
|
|
|
rte_mbuf_refcnt_update (mb, b->n_add_refs);
|
|
|
|
|
b->n_add_refs = 0;
|
|
|
|
|
}
|
|
|
|
|
|
2018-06-03 14:29:42 +02:00
|
|
|
if ((mb = rte_pktmbuf_prefree_seg (mb)))
|
|
|
|
|
rte_mempool_put (mb->pool, mb);
|
2017-01-25 14:18:03 +01:00
|
|
|
|
2018-06-03 14:29:42 +02:00
|
|
|
if (maybe_next && (flags & VLIB_BUFFER_NEXT_PRESENT))
|
2017-01-25 14:18:03 +01:00
|
|
|
{
|
|
|
|
|
b = vlib_get_buffer (vm, next);
|
|
|
|
|
goto next;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2018-05-28 21:26:47 +02:00
|
|
|
#ifndef CLIB_MARCH_VARIANT
|
2017-01-04 13:19:27 +01:00
|
|
|
static void
|
|
|
|
|
del_free_list (vlib_main_t * vm, vlib_buffer_free_list_t * f)
|
|
|
|
|
{
|
|
|
|
|
u32 i;
|
|
|
|
|
vlib_buffer_t *b;
|
2017-11-22 13:27:03 +01:00
|
|
|
u32 thread_index = vlib_get_thread_index ();
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2017-02-05 23:44:42 +01:00
|
|
|
for (i = 0; i < vec_len (f->buffers); i++)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
2017-02-05 23:44:42 +01:00
|
|
|
b = vlib_get_buffer (vm, f->buffers[i]);
|
2018-06-03 14:29:42 +02:00
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b, 1);
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
2017-01-25 14:18:03 +01:00
|
|
|
|
2017-01-04 13:19:27 +01:00
|
|
|
vec_free (f->name);
|
2017-02-05 23:44:42 +01:00
|
|
|
vec_free (f->buffers);
|
2018-03-13 21:32:17 +01:00
|
|
|
/* Poison it. */
|
2018-10-17 10:38:51 -04:00
|
|
|
clib_memset (f, 0xab, sizeof (f[0]));
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Add buffer free list. */
|
|
|
|
|
static void
|
2018-02-01 15:30:13 +01:00
|
|
|
dpdk_buffer_delete_free_list (vlib_main_t * vm,
|
|
|
|
|
vlib_buffer_free_list_index_t free_list_index)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
|
|
|
|
vlib_buffer_free_list_t *f;
|
|
|
|
|
int i;
|
|
|
|
|
|
2017-04-05 19:18:20 +02:00
|
|
|
ASSERT (vlib_get_thread_index () == 0);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
f = vlib_buffer_get_free_list (vm, free_list_index);
|
|
|
|
|
|
|
|
|
|
del_free_list (vm, f);
|
|
|
|
|
|
2018-03-13 21:32:17 +01:00
|
|
|
pool_put (vm->buffer_free_list_pool, f);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
for (i = 1; i < vec_len (vlib_mains); i++)
|
|
|
|
|
{
|
2018-03-13 21:32:17 +01:00
|
|
|
vlib_main_t *wvm = vlib_mains[i];
|
|
|
|
|
f = vlib_buffer_get_free_list (vlib_mains[i], free_list_index);
|
|
|
|
|
del_free_list (wvm, f);
|
|
|
|
|
pool_put (wvm->buffer_free_list_pool, f);
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
}
|
2017-11-10 21:55:45 +01:00
|
|
|
#endif
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
/* Make sure free list has at least given number of free buffers. */
|
2017-11-24 20:15:23 +01:00
|
|
|
uword
|
|
|
|
|
CLIB_MULTIARCH_FN (dpdk_buffer_fill_free_list) (vlib_main_t * vm,
|
|
|
|
|
vlib_buffer_free_list_t * fl,
|
|
|
|
|
uword min_free_buffers)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
|
|
|
|
dpdk_main_t *dm = &dpdk_main;
|
2018-06-02 20:42:07 +02:00
|
|
|
dpdk_buffer_main_t *dbm = &dpdk_buffer_main;
|
|
|
|
|
struct rte_mbuf **mb;
|
|
|
|
|
uword n_left, first;
|
|
|
|
|
word n_alloc;
|
2017-01-04 13:19:27 +01:00
|
|
|
unsigned socket_id = rte_socket_id ();
|
2018-06-02 20:42:07 +02:00
|
|
|
u32 thread_index = vlib_get_thread_index ();
|
|
|
|
|
dpdk_buffer_per_thread_data *d = vec_elt_at_index (dbm->ptd, thread_index);
|
2017-01-04 13:19:27 +01:00
|
|
|
struct rte_mempool *rmp = dm->pktmbuf_pools[socket_id];
|
2017-11-09 18:02:54 +01:00
|
|
|
dpdk_mempool_private_t *privp = rte_mempool_get_priv (rmp);
|
|
|
|
|
vlib_buffer_t bt;
|
2018-06-02 20:42:07 +02:00
|
|
|
u32 *bi;
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
/* Too early? */
|
|
|
|
|
if (PREDICT_FALSE (rmp == 0))
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
/* Already have enough free buffers on free list? */
|
2018-06-02 20:42:07 +02:00
|
|
|
n_alloc = min_free_buffers - vec_len (fl->buffers);
|
|
|
|
|
if (n_alloc <= 0)
|
2017-01-04 13:19:27 +01:00
|
|
|
return min_free_buffers;
|
|
|
|
|
|
|
|
|
|
/* Always allocate round number of buffers. */
|
2018-06-02 20:42:07 +02:00
|
|
|
n_alloc = round_pow2 (n_alloc, CLIB_CACHE_LINE_BYTES / sizeof (u32));
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
/* Always allocate new buffers in reasonably large sized chunks. */
|
2018-06-02 20:42:07 +02:00
|
|
|
n_alloc = clib_max (n_alloc, fl->min_n_buffers_each_alloc);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
vec_validate_aligned (d->mbuf_alloc_list, n_alloc - 1,
|
|
|
|
|
CLIB_CACHE_LINE_BYTES);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
if (rte_mempool_get_bulk (rmp, (void *) d->mbuf_alloc_list, n_alloc) < 0)
|
2017-01-04 13:19:27 +01:00
|
|
|
return 0;
|
|
|
|
|
|
2018-10-17 10:38:51 -04:00
|
|
|
clib_memset (&bt, 0, sizeof (vlib_buffer_t));
|
2017-11-09 18:02:54 +01:00
|
|
|
vlib_buffer_init_for_free_list (&bt, fl);
|
|
|
|
|
bt.buffer_pool_index = privp->buffer_pool_index;
|
2017-10-12 13:09:26 +02:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
_vec_len (d->mbuf_alloc_list) = n_alloc;
|
2017-02-16 20:16:06 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
first = vec_len (fl->buffers);
|
|
|
|
|
vec_resize_aligned (fl->buffers, n_alloc, CLIB_CACHE_LINE_BYTES);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
n_left = n_alloc;
|
|
|
|
|
mb = d->mbuf_alloc_list;
|
|
|
|
|
bi = fl->buffers + first;
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
ASSERT (n_left % 8 == 0);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
while (n_left >= 8)
|
|
|
|
|
{
|
|
|
|
|
if (PREDICT_FALSE (n_left < 24))
|
|
|
|
|
goto no_prefetch;
|
|
|
|
|
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[16]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[17]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[18]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[19]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[20]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[21]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[22]), STORE);
|
|
|
|
|
vlib_prefetch_buffer_header (vlib_buffer_from_rte_mbuf (mb[23]), STORE);
|
|
|
|
|
|
|
|
|
|
no_prefetch:
|
|
|
|
|
vlib_get_buffer_indices_with_offset (vm, (void **) mb, bi, 8,
|
|
|
|
|
sizeof (struct rte_mbuf));
|
|
|
|
|
clib_memcpy64_x4 (vlib_buffer_from_rte_mbuf (mb[0]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[1]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[2]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[3]), &bt);
|
|
|
|
|
clib_memcpy64_x4 (vlib_buffer_from_rte_mbuf (mb[4]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[5]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[6]),
|
|
|
|
|
vlib_buffer_from_rte_mbuf (mb[7]), &bt);
|
|
|
|
|
|
|
|
|
|
n_left -= 8;
|
|
|
|
|
mb += 8;
|
|
|
|
|
bi += 8;
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
if (fl->buffer_init_function)
|
|
|
|
|
fl->buffer_init_function (vm, fl, fl->buffers + first, n_alloc);
|
|
|
|
|
|
|
|
|
|
fl->n_alloc += n_alloc;
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2018-06-02 20:42:07 +02:00
|
|
|
return n_alloc;
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
|
2017-11-09 18:02:54 +01:00
|
|
|
static_always_inline void
|
2018-06-03 14:29:42 +02:00
|
|
|
dpdk_prefetch_buffer (vlib_buffer_t * b)
|
2017-11-09 18:02:54 +01:00
|
|
|
{
|
|
|
|
|
struct rte_mbuf *mb;
|
|
|
|
|
mb = rte_mbuf_from_vlib_buffer (b);
|
2018-06-03 14:29:42 +02:00
|
|
|
CLIB_PREFETCH (mb, 2 * CLIB_CACHE_LINE_BYTES, LOAD);
|
2017-11-09 18:02:54 +01:00
|
|
|
CLIB_PREFETCH (b, CLIB_CACHE_LINE_BYTES, LOAD);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static_always_inline void
|
|
|
|
|
recycle_or_free (vlib_main_t * vm, vlib_buffer_main_t * bm, u32 bi,
|
|
|
|
|
vlib_buffer_t * b)
|
|
|
|
|
{
|
2017-11-22 13:27:03 +01:00
|
|
|
u32 thread_index = vlib_get_thread_index ();
|
2017-11-09 18:02:54 +01:00
|
|
|
|
2018-09-03 15:41:45 +02:00
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b, 1);
|
2017-11-09 18:02:54 +01:00
|
|
|
}
|
|
|
|
|
|
2017-01-04 13:19:27 +01:00
|
|
|
static_always_inline void
|
|
|
|
|
vlib_buffer_free_inline (vlib_main_t * vm,
|
|
|
|
|
u32 * buffers, u32 n_buffers, u32 follow_buffer_next)
|
|
|
|
|
{
|
2018-03-13 21:32:17 +01:00
|
|
|
vlib_buffer_main_t *bm = &buffer_main;
|
2018-06-03 14:29:42 +02:00
|
|
|
vlib_buffer_t *bufp[n_buffers], **b = bufp;
|
2017-11-22 13:27:03 +01:00
|
|
|
u32 thread_index = vlib_get_thread_index ();
|
2017-11-09 18:02:54 +01:00
|
|
|
int i = 0;
|
2018-09-03 15:41:45 +02:00
|
|
|
u32 simple_mask = (VLIB_BUFFER_NON_DEFAULT_FREELIST |
|
2018-06-03 14:29:42 +02:00
|
|
|
VLIB_BUFFER_NEXT_PRESENT);
|
|
|
|
|
u32 n_left, *bi;
|
2017-01-04 13:19:27 +01:00
|
|
|
u32 (*cb) (vlib_main_t * vm, u32 * buffers, u32 n_buffers,
|
|
|
|
|
u32 follow_buffer_next);
|
|
|
|
|
|
|
|
|
|
cb = bm->buffer_free_callback;
|
|
|
|
|
|
|
|
|
|
if (PREDICT_FALSE (cb != 0))
|
|
|
|
|
n_buffers = (*cb) (vm, buffers, n_buffers, follow_buffer_next);
|
|
|
|
|
|
|
|
|
|
if (!n_buffers)
|
|
|
|
|
return;
|
|
|
|
|
|
2018-06-03 14:29:42 +02:00
|
|
|
n_left = n_buffers;
|
|
|
|
|
bi = buffers;
|
|
|
|
|
b = bufp;
|
|
|
|
|
vlib_get_buffers (vm, bi, b, n_buffers);
|
|
|
|
|
|
|
|
|
|
while (n_left >= 4)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
2018-06-03 14:29:42 +02:00
|
|
|
u32 or_flags;
|
|
|
|
|
vlib_buffer_t **p;
|
|
|
|
|
|
|
|
|
|
if (n_left < 16)
|
|
|
|
|
goto no_prefetch;
|
|
|
|
|
|
|
|
|
|
p = b + 12;
|
|
|
|
|
dpdk_prefetch_buffer (p[0]);
|
|
|
|
|
dpdk_prefetch_buffer (p[1]);
|
|
|
|
|
dpdk_prefetch_buffer (p[2]);
|
|
|
|
|
dpdk_prefetch_buffer (p[3]);
|
|
|
|
|
no_prefetch:
|
|
|
|
|
|
|
|
|
|
for (i = 0; i < 4; i++)
|
|
|
|
|
VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b[i]);
|
|
|
|
|
|
|
|
|
|
or_flags = b[0]->flags | b[1]->flags | b[2]->flags | b[3]->flags;
|
|
|
|
|
|
|
|
|
|
if (or_flags & simple_mask)
|
|
|
|
|
{
|
|
|
|
|
recycle_or_free (vm, bm, bi[0], b[0]);
|
|
|
|
|
recycle_or_free (vm, bm, bi[1], b[1]);
|
|
|
|
|
recycle_or_free (vm, bm, bi[2], b[2]);
|
|
|
|
|
recycle_or_free (vm, bm, bi[3], b[3]);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b[0], 0);
|
|
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b[1], 0);
|
|
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b[2], 0);
|
|
|
|
|
dpdk_rte_pktmbuf_free (vm, thread_index, b[3], 0);
|
|
|
|
|
}
|
|
|
|
|
bi += 4;
|
|
|
|
|
b += 4;
|
|
|
|
|
n_left -= 4;
|
2017-11-09 18:02:54 +01:00
|
|
|
}
|
2018-06-03 14:29:42 +02:00
|
|
|
while (n_left)
|
2017-11-09 18:02:54 +01:00
|
|
|
{
|
2018-06-03 14:29:42 +02:00
|
|
|
VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b[0]);
|
|
|
|
|
recycle_or_free (vm, bm, bi[0], b[0]);
|
|
|
|
|
bi += 1;
|
|
|
|
|
b += 1;
|
|
|
|
|
n_left -= 1;
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2017-11-10 21:55:45 +01:00
|
|
|
void
|
|
|
|
|
CLIB_MULTIARCH_FN (dpdk_buffer_free) (vlib_main_t * vm, u32 * buffers,
|
|
|
|
|
u32 n_buffers)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
|
|
|
|
vlib_buffer_free_inline (vm, buffers, n_buffers, /* follow_buffer_next */
|
|
|
|
|
1);
|
|
|
|
|
}
|
|
|
|
|
|
2017-11-10 21:55:45 +01:00
|
|
|
void
|
|
|
|
|
CLIB_MULTIARCH_FN (dpdk_buffer_free_no_next) (vlib_main_t * vm, u32 * buffers,
|
|
|
|
|
u32 n_buffers)
|
2017-01-04 13:19:27 +01:00
|
|
|
{
|
|
|
|
|
vlib_buffer_free_inline (vm, buffers, n_buffers, /* follow_buffer_next */
|
|
|
|
|
0);
|
|
|
|
|
}
|
|
|
|
|
|
2018-05-28 21:26:47 +02:00
|
|
|
#ifndef CLIB_MARCH_VARIANT
|
2017-11-15 15:43:35 +00:00
|
|
|
clib_error_t *
|
|
|
|
|
dpdk_pool_create (vlib_main_t * vm, u8 * pool_name, u32 elt_size,
|
|
|
|
|
u32 num_elts, u32 pool_priv_size, u16 cache_size, u8 numa,
|
2018-09-30 18:26:20 +02:00
|
|
|
struct rte_mempool **_mp, u32 * map_index)
|
2017-11-15 15:43:35 +00:00
|
|
|
{
|
|
|
|
|
struct rte_mempool *mp;
|
2018-10-20 23:26:46 +02:00
|
|
|
enum rte_iova_mode iova_mode;
|
2018-05-28 14:45:34 +05:30
|
|
|
dpdk_mempool_private_t priv;
|
2018-09-30 18:26:20 +02:00
|
|
|
vlib_physmem_map_t *pm;
|
2017-11-15 15:43:35 +00:00
|
|
|
clib_error_t *error = 0;
|
2018-08-12 00:26:45 +02:00
|
|
|
size_t min_chunk_size, align;
|
2018-10-20 23:26:46 +02:00
|
|
|
int map_dma = 1;
|
2018-08-12 00:26:45 +02:00
|
|
|
u32 size;
|
2017-11-15 15:43:35 +00:00
|
|
|
i32 ret;
|
2018-06-06 14:23:42 +02:00
|
|
|
uword i;
|
2017-11-15 15:43:35 +00:00
|
|
|
|
2018-08-12 00:26:45 +02:00
|
|
|
mp = rte_mempool_create_empty ((char *) pool_name, num_elts, elt_size,
|
|
|
|
|
512, pool_priv_size, numa, 0);
|
|
|
|
|
if (!mp)
|
|
|
|
|
return clib_error_return (0, "failed to create %s", pool_name);
|
|
|
|
|
|
|
|
|
|
rte_mempool_set_ops_byname (mp, RTE_MBUF_DEFAULT_MEMPOOL_OPS, NULL);
|
|
|
|
|
|
|
|
|
|
size = rte_mempool_op_calc_mem_size_default (mp, num_elts, 21,
|
|
|
|
|
&min_chunk_size, &align);
|
2017-11-15 15:43:35 +00:00
|
|
|
|
2018-09-30 18:26:20 +02:00
|
|
|
if ((error = vlib_physmem_shared_map_create (vm, (char *) pool_name, size,
|
2018-10-24 17:08:26 +02:00
|
|
|
0, numa, map_index)))
|
2018-08-12 00:26:45 +02:00
|
|
|
{
|
|
|
|
|
rte_mempool_free (mp);
|
|
|
|
|
return error;
|
|
|
|
|
}
|
2018-09-30 18:26:20 +02:00
|
|
|
pm = vlib_physmem_get_map (vm, *map_index);
|
2017-11-15 15:43:35 +00:00
|
|
|
|
2018-05-28 14:45:34 +05:30
|
|
|
/* Call the mempool priv initializer */
|
|
|
|
|
priv.mbp_priv.mbuf_data_room_size = VLIB_BUFFER_PRE_DATA_SIZE +
|
|
|
|
|
VLIB_BUFFER_DATA_SIZE;
|
|
|
|
|
priv.mbp_priv.mbuf_priv_size = VLIB_BUFFER_HDR_SIZE;
|
|
|
|
|
rte_pktmbuf_pool_init (mp, &priv);
|
|
|
|
|
|
2018-10-20 23:26:46 +02:00
|
|
|
if (rte_eth_dev_count_avail () == 0)
|
|
|
|
|
map_dma = 0;
|
|
|
|
|
|
|
|
|
|
iova_mode = rte_eal_iova_mode ();
|
2018-09-30 18:26:20 +02:00
|
|
|
for (i = 0; i < pm->n_pages; i++)
|
2017-11-15 15:43:35 +00:00
|
|
|
{
|
2018-09-30 18:26:20 +02:00
|
|
|
size_t page_sz = 1ULL << pm->log2_page_size;
|
|
|
|
|
char *va = ((char *) pm->base) + i * page_sz;
|
2018-10-20 23:26:46 +02:00
|
|
|
uword pa = iova_mode == RTE_IOVA_VA ?
|
2018-09-30 18:26:20 +02:00
|
|
|
pointer_to_uword (va) : pm->page_table[i];
|
2018-10-20 23:26:46 +02:00
|
|
|
ret = rte_mempool_populate_iova (mp, va, pa, page_sz, 0, 0);
|
2018-06-06 14:23:42 +02:00
|
|
|
if (ret < 0)
|
|
|
|
|
{
|
|
|
|
|
rte_mempool_free (mp);
|
|
|
|
|
return clib_error_return (0, "failed to populate %s", pool_name);
|
|
|
|
|
}
|
2018-10-20 23:26:46 +02:00
|
|
|
/* -1 likely means there is no PCI devices assigned to vfio
|
|
|
|
|
container or noiommu mode is used so we stop trying */
|
|
|
|
|
if (map_dma && rte_vfio_dma_map (pointer_to_uword (va), pa, page_sz))
|
|
|
|
|
map_dma = 0;
|
2017-11-15 15:43:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
_mp[0] = mp;
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
2017-01-04 13:19:27 +01:00
|
|
|
clib_error_t *
|
2017-07-07 14:57:07 -04:00
|
|
|
dpdk_buffer_pool_create (vlib_main_t * vm, unsigned num_mbufs,
|
2017-01-04 13:19:27 +01:00
|
|
|
unsigned socket_id)
|
|
|
|
|
{
|
|
|
|
|
dpdk_main_t *dm = &dpdk_main;
|
|
|
|
|
struct rte_mempool *rmp;
|
2017-11-15 15:43:35 +00:00
|
|
|
clib_error_t *error = 0;
|
2017-07-20 18:10:35 +02:00
|
|
|
u8 *pool_name;
|
2017-11-15 15:43:35 +00:00
|
|
|
u32 elt_size, i;
|
2018-09-30 18:26:20 +02:00
|
|
|
u32 map_index;
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
vec_validate_aligned (dm->pktmbuf_pools, socket_id, CLIB_CACHE_LINE_BYTES);
|
|
|
|
|
|
|
|
|
|
/* pool already exists, nothing to do */
|
|
|
|
|
if (dm->pktmbuf_pools[socket_id])
|
|
|
|
|
return 0;
|
|
|
|
|
|
2017-07-20 18:10:35 +02:00
|
|
|
pool_name = format (0, "dpdk_mbuf_pool_socket%u%c", socket_id, 0);
|
|
|
|
|
|
|
|
|
|
elt_size = sizeof (struct rte_mbuf) +
|
|
|
|
|
VLIB_BUFFER_HDR_SIZE /* priv size */ +
|
|
|
|
|
VLIB_BUFFER_PRE_DATA_SIZE + VLIB_BUFFER_DATA_SIZE; /*data room size */
|
|
|
|
|
|
2018-09-30 18:26:20 +02:00
|
|
|
error = dpdk_pool_create (vm, pool_name, elt_size, num_mbufs,
|
|
|
|
|
sizeof (dpdk_mempool_private_t), 512, socket_id,
|
|
|
|
|
&rmp, &map_index);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
vec_free (pool_name);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
if (!error)
|
2017-07-20 18:10:35 +02:00
|
|
|
{
|
2017-11-15 15:43:35 +00:00
|
|
|
/* call the object initializers */
|
|
|
|
|
rte_mempool_obj_iter (rmp, rte_pktmbuf_init, 0);
|
2017-07-20 18:10:35 +02:00
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
dpdk_mempool_private_t *privp = rte_mempool_get_priv (rmp);
|
2018-09-30 18:26:20 +02:00
|
|
|
privp->buffer_pool_index =
|
|
|
|
|
vlib_buffer_register_physmem_map (vm, map_index);
|
2017-07-20 18:10:35 +02:00
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
dm->pktmbuf_pools[socket_id] = rmp;
|
2017-07-20 18:10:35 +02:00
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
return 0;
|
2017-01-04 13:19:27 +01:00
|
|
|
}
|
|
|
|
|
|
2017-11-15 15:43:35 +00:00
|
|
|
clib_error_report (error);
|
2017-01-04 13:19:27 +01:00
|
|
|
|
|
|
|
|
/* no usable pool for this socket, try to use pool from another one */
|
|
|
|
|
for (i = 0; i < vec_len (dm->pktmbuf_pools); i++)
|
|
|
|
|
{
|
|
|
|
|
if (dm->pktmbuf_pools[i])
|
|
|
|
|
{
|
2017-07-20 18:10:35 +02:00
|
|
|
clib_warning ("WARNING: Failed to allocate mempool for CPU socket "
|
|
|
|
|
"%u. Threads running on socket %u will use socket %u "
|
|
|
|
|
"mempool.", socket_id, socket_id, i);
|
2017-01-04 13:19:27 +01:00
|
|
|
dm->pktmbuf_pools[socket_id] = dm->pktmbuf_pools[i];
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return clib_error_return (0, "failed to allocate mempool on socket %u",
|
|
|
|
|
socket_id);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#if CLIB_DEBUG > 0
|
|
|
|
|
|
|
|
|
|
u32 *vlib_buffer_state_validation_lock;
|
|
|
|
|
uword *vlib_buffer_state_validation_hash;
|
|
|
|
|
void *vlib_buffer_state_heap;
|
|
|
|
|
|
|
|
|
|
static clib_error_t *
|
|
|
|
|
buffer_state_validation_init (vlib_main_t * vm)
|
|
|
|
|
{
|
|
|
|
|
void *oldheap;
|
|
|
|
|
|
2018-07-04 10:56:23 -04:00
|
|
|
vlib_buffer_state_heap =
|
|
|
|
|
mheap_alloc_with_lock (0, 10 << 20, 0 /* locked */ );
|
2017-01-04 13:19:27 +01:00
|
|
|
oldheap = clib_mem_set_heap (vlib_buffer_state_heap);
|
|
|
|
|
|
|
|
|
|
vlib_buffer_state_validation_hash = hash_create (0, sizeof (uword));
|
|
|
|
|
vec_validate_aligned (vlib_buffer_state_validation_lock, 0,
|
|
|
|
|
CLIB_CACHE_LINE_BYTES);
|
|
|
|
|
clib_mem_set_heap (oldheap);
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
VLIB_INIT_FUNCTION (buffer_state_validation_init);
|
|
|
|
|
#endif
|
|
|
|
|
|
2017-09-11 20:54:15 -04:00
|
|
|
#if CLI_DEBUG
|
|
|
|
|
struct dpdk_validate_buf_result
|
|
|
|
|
{
|
|
|
|
|
u32 invalid;
|
|
|
|
|
u32 uninitialized;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
#define DPDK_TRAJECTORY_POISON 31
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
dpdk_buffer_validate_trajectory (struct rte_mempool *mp, void *opaque,
|
|
|
|
|
void *obj, unsigned obj_idx)
|
|
|
|
|
{
|
|
|
|
|
vlib_buffer_t *b;
|
|
|
|
|
struct dpdk_validate_buf_result *counter = opaque;
|
|
|
|
|
b = vlib_buffer_from_rte_mbuf ((struct rte_mbuf *) obj);
|
|
|
|
|
if (b->pre_data[0] != 0)
|
|
|
|
|
{
|
|
|
|
|
if (b->pre_data[0] == DPDK_TRAJECTORY_POISON)
|
|
|
|
|
counter->uninitialized++;
|
|
|
|
|
else
|
|
|
|
|
counter->invalid++;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
int
|
|
|
|
|
dpdk_buffer_validate_trajectory_all (u32 * uninitialized)
|
|
|
|
|
{
|
|
|
|
|
dpdk_main_t *dm = &dpdk_main;
|
|
|
|
|
struct dpdk_validate_buf_result counter = { 0 };
|
|
|
|
|
int i;
|
|
|
|
|
|
|
|
|
|
for (i = 0; i < vec_len (dm->pktmbuf_pools); i++)
|
|
|
|
|
rte_mempool_obj_iter (dm->pktmbuf_pools[i],
|
|
|
|
|
dpdk_buffer_validate_trajectory, &counter);
|
|
|
|
|
if (uninitialized)
|
|
|
|
|
*uninitialized = counter.uninitialized;
|
|
|
|
|
return counter.invalid;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
dpdk_buffer_poison_trajectory (struct rte_mempool *mp, void *opaque,
|
|
|
|
|
void *obj, unsigned obj_idx)
|
|
|
|
|
{
|
|
|
|
|
vlib_buffer_t *b;
|
|
|
|
|
b = vlib_buffer_from_rte_mbuf ((struct rte_mbuf *) obj);
|
|
|
|
|
b->pre_data[0] = DPDK_TRAJECTORY_POISON;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
dpdk_buffer_poison_trajectory_all (void)
|
|
|
|
|
{
|
|
|
|
|
dpdk_main_t *dm = &dpdk_main;
|
|
|
|
|
int i;
|
|
|
|
|
|
|
|
|
|
for (i = 0; i < vec_len (dm->pktmbuf_pools); i++)
|
|
|
|
|
rte_mempool_obj_iter (dm->pktmbuf_pools[i], dpdk_buffer_poison_trajectory,
|
|
|
|
|
0);
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
2017-11-22 13:27:03 +01:00
|
|
|
static clib_error_t *
|
|
|
|
|
dpdk_buffer_init (vlib_main_t * vm)
|
|
|
|
|
{
|
|
|
|
|
dpdk_buffer_main_t *dbm = &dpdk_buffer_main;
|
|
|
|
|
vlib_thread_main_t *tm = vlib_get_thread_main ();
|
2018-03-09 16:05:56 +01:00
|
|
|
|
2017-11-22 13:27:03 +01:00
|
|
|
vec_validate_aligned (dbm->ptd, tm->n_vlib_mains - 1,
|
|
|
|
|
CLIB_CACHE_LINE_BYTES);
|
2018-03-09 16:05:56 +01:00
|
|
|
|
|
|
|
|
dbm->vfio_container_fd = -1;
|
|
|
|
|
|
2017-11-22 13:27:03 +01:00
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
VLIB_INIT_FUNCTION (dpdk_buffer_init);
|
|
|
|
|
|
2017-07-10 15:06:17 +02:00
|
|
|
/* *INDENT-OFF* */
|
|
|
|
|
VLIB_BUFFER_REGISTER_CALLBACKS (dpdk, static) = {
|
2017-11-24 20:15:23 +01:00
|
|
|
.vlib_buffer_fill_free_list_cb = &dpdk_buffer_fill_free_list,
|
2017-01-04 13:19:27 +01:00
|
|
|
.vlib_buffer_free_cb = &dpdk_buffer_free,
|
|
|
|
|
.vlib_buffer_free_no_next_cb = &dpdk_buffer_free_no_next,
|
|
|
|
|
.vlib_buffer_delete_free_list_cb = &dpdk_buffer_delete_free_list,
|
|
|
|
|
};
|
2017-07-10 15:06:17 +02:00
|
|
|
/* *INDENT-ON* */
|
2017-01-04 13:19:27 +01:00
|
|
|
|
2017-11-10 21:55:45 +01:00
|
|
|
#if __x86_64__
|
2017-11-24 20:15:23 +01:00
|
|
|
vlib_buffer_fill_free_list_cb_t __clib_weak dpdk_buffer_fill_free_list_avx512;
|
|
|
|
|
vlib_buffer_fill_free_list_cb_t __clib_weak dpdk_buffer_fill_free_list_avx2;
|
2017-11-14 15:21:19 +01:00
|
|
|
vlib_buffer_free_cb_t __clib_weak dpdk_buffer_free_avx512;
|
|
|
|
|
vlib_buffer_free_cb_t __clib_weak dpdk_buffer_free_avx2;
|
|
|
|
|
vlib_buffer_free_no_next_cb_t __clib_weak dpdk_buffer_free_no_next_avx512;
|
|
|
|
|
vlib_buffer_free_no_next_cb_t __clib_weak dpdk_buffer_free_no_next_avx2;
|
2017-11-10 21:55:45 +01:00
|
|
|
|
|
|
|
|
static void __clib_constructor
|
|
|
|
|
dpdk_input_multiarch_select (void)
|
|
|
|
|
{
|
|
|
|
|
vlib_buffer_callbacks_t *cb = &__dpdk_buffer_callbacks;
|
2017-11-24 20:15:23 +01:00
|
|
|
if (dpdk_buffer_fill_free_list_avx512 && clib_cpu_supports_avx512f ())
|
2017-11-10 21:55:45 +01:00
|
|
|
{
|
2017-11-24 20:15:23 +01:00
|
|
|
cb->vlib_buffer_fill_free_list_cb = dpdk_buffer_fill_free_list_avx512;
|
2017-11-14 15:21:19 +01:00
|
|
|
cb->vlib_buffer_free_cb = dpdk_buffer_free_avx512;
|
|
|
|
|
cb->vlib_buffer_free_no_next_cb = dpdk_buffer_free_no_next_avx512;
|
2017-11-10 21:55:45 +01:00
|
|
|
}
|
2017-11-24 20:15:23 +01:00
|
|
|
else if (dpdk_buffer_fill_free_list_avx2 && clib_cpu_supports_avx2 ())
|
2017-11-10 21:55:45 +01:00
|
|
|
{
|
2017-11-24 20:15:23 +01:00
|
|
|
cb->vlib_buffer_fill_free_list_cb = dpdk_buffer_fill_free_list_avx2;
|
2017-11-14 15:21:19 +01:00
|
|
|
cb->vlib_buffer_free_cb = dpdk_buffer_free_avx2;
|
|
|
|
|
cb->vlib_buffer_free_no_next_cb = dpdk_buffer_free_no_next_avx2;
|
2017-11-10 21:55:45 +01:00
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
#endif
|
|
|
|
|
|
2017-01-04 13:19:27 +01:00
|
|
|
/** @endcond */
|
|
|
|
|
/*
|
|
|
|
|
* fd.io coding-style-patch-verification: ON
|
|
|
|
|
*
|
|
|
|
|
* Local Variables:
|
|
|
|
|
* eval: (c-set-style "gnu")
|
|
|
|
|
* End:
|
|
|
|
|
*/
|