vlib: don't inline vlib_buffer_enqueue_to_*

Function call in this case is not bad thing, as those functions
are called once per frame.

Type: improvement
Change-Id: I7bd61cf746b905e2c8e3085ebb2ff001c0a52da8
Signed-off-by: Damjan Marion <damarion@cisco.com>
diff --git a/src/vlib/buffer_node.h b/src/vlib/buffer_node.h
index 17eb54e4..d2bd32f 100644
--- a/src/vlib/buffer_node.h
+++ b/src/vlib/buffer_node.h
@@ -339,120 +339,9 @@
 vlib_buffer_enqueue_to_next (vlib_main_t * vm, vlib_node_runtime_t * node,
 			     u32 * buffers, u16 * nexts, uword count)
 {
-  u32 *to_next, n_left_to_next, max;
-  u16 next_index;
-
-  next_index = nexts[0];
-  vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-  max = clib_min (n_left_to_next, count);
-
-  while (count)
-    {
-      u32 n_enqueued;
-      if ((nexts[0] != next_index) || n_left_to_next == 0)
-	{
-	  vlib_put_next_frame (vm, node, next_index, n_left_to_next);
-	  next_index = nexts[0];
-	  vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-	  max = clib_min (n_left_to_next, count);
-	}
-#if defined(CLIB_HAVE_VEC512)
-      u16x32 next32 = CLIB_MEM_OVERFLOW_LOAD (u16x32_load_unaligned, nexts);
-      next32 = (next32 == u16x32_splat (next32[0]));
-      u64 bitmap = u16x32_msb_mask (next32);
-      n_enqueued = count_trailing_zeros (~bitmap);
-#elif defined(CLIB_HAVE_VEC256)
-      u16x16 next16 = CLIB_MEM_OVERFLOW_LOAD (u16x16_load_unaligned, nexts);
-      next16 = (next16 == u16x16_splat (next16[0]));
-      u64 bitmap = u8x32_msb_mask ((u8x32) next16);
-      n_enqueued = count_trailing_zeros (~bitmap) / 2;
-#elif defined(CLIB_HAVE_VEC128) && defined(CLIB_HAVE_VEC128_MSB_MASK)
-      u16x8 next8 = CLIB_MEM_OVERFLOW_LOAD (u16x8_load_unaligned, nexts);
-      next8 = (next8 == u16x8_splat (next8[0]));
-      u64 bitmap = u8x16_msb_mask ((u8x16) next8);
-      n_enqueued = count_trailing_zeros (~bitmap) / 2;
-#else
-      u16 x = 0;
-      if (count + 3 < max)
-	{
-	  x |= next_index ^ nexts[1];
-	  x |= next_index ^ nexts[2];
-	  x |= next_index ^ nexts[3];
-	  n_enqueued = (x == 0) ? 4 : 1;
-	}
-      else
-	n_enqueued = 1;
-#endif
-
-      if (PREDICT_FALSE (n_enqueued > max))
-	n_enqueued = max;
-
-#ifdef CLIB_HAVE_VEC512
-      if (n_enqueued >= 32)
-	{
-	  vlib_buffer_copy_indices (to_next, buffers, 32);
-	  nexts += 32;
-	  to_next += 32;
-	  buffers += 32;
-	  n_left_to_next -= 32;
-	  count -= 32;
-	  max -= 32;
-	  continue;
-	}
-#endif
-
-#ifdef CLIB_HAVE_VEC256
-      if (n_enqueued >= 16)
-	{
-	  vlib_buffer_copy_indices (to_next, buffers, 16);
-	  nexts += 16;
-	  to_next += 16;
-	  buffers += 16;
-	  n_left_to_next -= 16;
-	  count -= 16;
-	  max -= 16;
-	  continue;
-	}
-#endif
-
-#ifdef CLIB_HAVE_VEC128
-      if (n_enqueued >= 8)
-	{
-	  vlib_buffer_copy_indices (to_next, buffers, 8);
-	  nexts += 8;
-	  to_next += 8;
-	  buffers += 8;
-	  n_left_to_next -= 8;
-	  count -= 8;
-	  max -= 8;
-	  continue;
-	}
-#endif
-
-      if (n_enqueued >= 4)
-	{
-	  vlib_buffer_copy_indices (to_next, buffers, 4);
-	  nexts += 4;
-	  to_next += 4;
-	  buffers += 4;
-	  n_left_to_next -= 4;
-	  count -= 4;
-	  max -= 4;
-	  continue;
-	}
-
-      /* copy */
-      to_next[0] = buffers[0];
-
-      /* next */
-      nexts += 1;
-      to_next += 1;
-      buffers += 1;
-      n_left_to_next -= 1;
-      count -= 1;
-      max -= 1;
-    }
-  vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+  vlib_buffer_enqueue_to_next_fn_t *fn;
+  fn = vlib_buffer_func_main.buffer_enqueue_to_next_fn;
+  (fn) (vm, node, buffers, nexts, count);
 }
 
 static_always_inline void
@@ -460,34 +349,9 @@
 				    vlib_node_runtime_t * node, u32 * buffers,
 				    u16 next_index, u32 count)
 {
-  u32 *to_next, n_left_to_next, n_enq;
-
-  vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
-  if (PREDICT_TRUE (n_left_to_next >= count))
-    {
-      vlib_buffer_copy_indices (to_next, buffers, count);
-      n_left_to_next -= count;
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
-      return;
-    }
-
-  n_enq = n_left_to_next;
-next:
-  vlib_buffer_copy_indices (to_next, buffers, n_enq);
-  n_left_to_next -= n_enq;
-
-  if (PREDICT_FALSE (count > n_enq))
-    {
-      count -= n_enq;
-      buffers += n_enq;
-
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
-      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-      n_enq = clib_min (n_left_to_next, count);
-      goto next;
-    }
-  vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+  vlib_buffer_enqueue_to_single_next_fn_t *fn;
+  fn = vlib_buffer_func_main.buffer_enqueue_to_single_next_fn;
+  (fn) (vm, node, buffers, next_index, count);
 }
 
 static_always_inline u32
@@ -495,100 +359,10 @@
 			       u32 * buffer_indices, u16 * thread_indices,
 			       u32 n_packets, int drop_on_congestion)
 {
-  vlib_thread_main_t *tm = vlib_get_thread_main ();
-  vlib_frame_queue_main_t *fqm;
-  vlib_frame_queue_per_thread_data_t *ptd;
-  u32 n_left = n_packets;
-  u32 drop_list[VLIB_FRAME_SIZE], *dbi = drop_list, n_drop = 0;
-  vlib_frame_queue_elt_t *hf = 0;
-  u32 n_left_to_next_thread = 0, *to_next_thread = 0;
-  u32 next_thread_index, current_thread_index = ~0;
-  int i;
-
-  fqm = vec_elt_at_index (tm->frame_queue_mains, frame_queue_index);
-  ptd = vec_elt_at_index (fqm->per_thread_data, vm->thread_index);
-
-  while (n_left)
-    {
-      next_thread_index = thread_indices[0];
-
-      if (next_thread_index != current_thread_index)
-	{
-	  if (drop_on_congestion &&
-	      is_vlib_frame_queue_congested
-	      (frame_queue_index, next_thread_index, fqm->queue_hi_thresh,
-	       ptd->congested_handoff_queue_by_thread_index))
-	    {
-	      dbi[0] = buffer_indices[0];
-	      dbi++;
-	      n_drop++;
-	      goto next;
-	    }
-
-	  if (hf)
-	    hf->n_vectors = VLIB_FRAME_SIZE - n_left_to_next_thread;
-
-	  hf = vlib_get_worker_handoff_queue_elt (frame_queue_index,
-						  next_thread_index,
-						  ptd->handoff_queue_elt_by_thread_index);
-
-	  n_left_to_next_thread = VLIB_FRAME_SIZE - hf->n_vectors;
-	  to_next_thread = &hf->buffer_index[hf->n_vectors];
-	  current_thread_index = next_thread_index;
-	}
-
-      to_next_thread[0] = buffer_indices[0];
-      to_next_thread++;
-      n_left_to_next_thread--;
-
-      if (n_left_to_next_thread == 0)
-	{
-	  hf->n_vectors = VLIB_FRAME_SIZE;
-	  vlib_put_frame_queue_elt (hf);
-	  vlib_get_main_by_index (current_thread_index)->check_frame_queues =
-	    1;
-	  current_thread_index = ~0;
-	  ptd->handoff_queue_elt_by_thread_index[next_thread_index] = 0;
-	  hf = 0;
-	}
-
-      /* next */
-    next:
-      thread_indices += 1;
-      buffer_indices += 1;
-      n_left -= 1;
-    }
-
-  if (hf)
-    hf->n_vectors = VLIB_FRAME_SIZE - n_left_to_next_thread;
-
-  /* Ship frames to the thread nodes */
-  for (i = 0; i < vec_len (ptd->handoff_queue_elt_by_thread_index); i++)
-    {
-      if (ptd->handoff_queue_elt_by_thread_index[i])
-	{
-	  hf = ptd->handoff_queue_elt_by_thread_index[i];
-	  /*
-	   * It works better to let the handoff node
-	   * rate-adapt, always ship the handoff queue element.
-	   */
-	  if (1 || hf->n_vectors == hf->last_n_vectors)
-	    {
-	      vlib_put_frame_queue_elt (hf);
-	      vlib_get_main_by_index (i)->check_frame_queues = 1;
-	      ptd->handoff_queue_elt_by_thread_index[i] = 0;
-	    }
-	  else
-	    hf->last_n_vectors = hf->n_vectors;
-	}
-      ptd->congested_handoff_queue_by_thread_index[i] =
-	(vlib_frame_queue_t *) (~0);
-    }
-
-  if (drop_on_congestion && n_drop)
-    vlib_buffer_free (vm, drop_list, n_drop);
-
-  return n_packets - n_drop;
+  vlib_buffer_enqueue_to_thread_fn_t *fn;
+  fn = vlib_buffer_func_main.buffer_enqueue_to_thread_fn;
+  return (fn) (vm, frame_queue_index, buffer_indices, thread_indices,
+	       n_packets, drop_on_congestion);
 }
 
 #endif /* included_vlib_buffer_node_h */