vxlan:vxlan.c conform coding style

Change-Id: I9937912cd760698e39044e8ae022a90b58c8db30
Signed-off-by: Eyal Bari <ebari@cisco.com>
diff --git a/src/vnet/vxlan/vxlan.c b/src/vnet/vxlan/vxlan.c
index 7516a05..4f966fc 100644
--- a/src/vnet/vxlan/vxlan.c
+++ b/src/vnet/vxlan/vxlan.c
@@ -41,7 +41,8 @@
 
 vxlan_main_t vxlan_main;
 
-static u8 * format_decap_next (u8 * s, va_list * args)
+static u8 *
+format_decap_next (u8 * s, va_list * args)
 {
   u32 next_index = va_arg (*args, u32);
 
@@ -52,16 +53,17 @@
   return s;
 }
 
-u8 * format_vxlan_tunnel (u8 * s, va_list * args)
+u8 *
+format_vxlan_tunnel (u8 * s, va_list * args)
 {
-  vxlan_tunnel_t * t = va_arg (*args, vxlan_tunnel_t *);
+  vxlan_tunnel_t *t = va_arg (*args, vxlan_tunnel_t *);
 
   s = format (s,
 	      "[%d] instance %d src %U dst %U vni %d fib-idx %d sw-if-idx %d ",
 	      t->dev_instance, t->user_instance,
-              format_ip46_address, &t->src, IP46_TYPE_ANY,
-              format_ip46_address, &t->dst, IP46_TYPE_ANY,
-              t->vni, t->encap_fib_index, t->sw_if_index);
+	      format_ip46_address, &t->src, IP46_TYPE_ANY,
+	      format_ip46_address, &t->dst, IP46_TYPE_ANY,
+	      t->vni, t->encap_fib_index, t->sw_if_index);
 
   s = format (s, "encap-dpo-idx %d ", t->next_dpo.dpoi_index);
 
@@ -74,19 +76,20 @@
   return s;
 }
 
-static u8 * format_vxlan_name (u8 * s, va_list * args)
+static u8 *
+format_vxlan_name (u8 * s, va_list * args)
 {
   u32 dev_instance = va_arg (*args, u32);
-  vxlan_main_t * vxm = &vxlan_main;
+  vxlan_main_t *vxm = &vxlan_main;
   vxlan_tunnel_t *t;
 
   if (dev_instance == ~0)
-    return format(s, "<cached-unused>");
+    return format (s, "<cached-unused>");
 
-  if (dev_instance >= vec_len(vxm->tunnels))
-    return format(s, "<improperly-referenced>");
+  if (dev_instance >= vec_len (vxm->tunnels))
+    return format (s, "<improperly-referenced>");
 
-  t = pool_elt_at_index(vxm->tunnels, dev_instance);
+  t = pool_elt_at_index (vxm->tunnels, dev_instance);
 
   return format (s, "vxlan_tunnel%d", t->user_instance);
 }
@@ -101,33 +104,38 @@
   return /* no error */ 0;
 }
 
-VNET_DEVICE_CLASS (vxlan_device_class,static) = {
+/* *INDENT-OFF* */
+VNET_DEVICE_CLASS (vxlan_device_class, static) = {
   .name = "VXLAN",
   .format_device_name = format_vxlan_name,
   .format_tx_trace = format_vxlan_encap_trace,
   .admin_up_down_function = vxlan_interface_admin_up_down,
 };
+/* *INDENT-ON* */
 
-static u8 * format_vxlan_header_with_length (u8 * s, va_list * args)
+static u8 *
+format_vxlan_header_with_length (u8 * s, va_list * args)
 {
   u32 dev_instance = va_arg (*args, u32);
   s = format (s, "unimplemented dev %u", dev_instance);
   return s;
 }
 
+/* *INDENT-OFF* */
 VNET_HW_INTERFACE_CLASS (vxlan_hw_class) = {
   .name = "VXLAN",
   .format_header = format_vxlan_header_with_length,
   .build_rewrite = default_build_rewrite,
 };
+/* *INDENT-ON* */
 
 static void
-vxlan_tunnel_restack_dpo(vxlan_tunnel_t * t)
+vxlan_tunnel_restack_dpo (vxlan_tunnel_t * t)
 {
-  u8 is_ip4 = ip46_address_is_ip4(&t->dst);
+  u8 is_ip4 = ip46_address_is_ip4 (&t->dst);
   dpo_id_t dpo = DPO_INVALID;
   fib_forward_chain_type_t forw_type = is_ip4 ?
-      FIB_FORW_CHAIN_TYPE_UNICAST_IP4 : FIB_FORW_CHAIN_TYPE_UNICAST_IP6;
+    FIB_FORW_CHAIN_TYPE_UNICAST_IP4 : FIB_FORW_CHAIN_TYPE_UNICAST_IP6;
 
   fib_entry_contribute_forwarding (t->fib_entry_index, forw_type, &dpo);
 
@@ -138,23 +146,23 @@
     {
       load_balance_t *lb = load_balance_get (dpo.dpoi_index);
       if (lb->lb_n_buckets > 1)
-        break;
+	break;
 
       dpo_copy (&dpo, load_balance_get_bucket_i (lb, 0));
     }
 
   u32 encap_index = is_ip4 ?
-      vxlan4_encap_node.index : vxlan6_encap_node.index;
+    vxlan4_encap_node.index : vxlan6_encap_node.index;
   dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
-  dpo_reset(&dpo);
+  dpo_reset (&dpo);
 }
 
 static vxlan_tunnel_t *
-vxlan_tunnel_from_fib_node (fib_node_t *node)
+vxlan_tunnel_from_fib_node (fib_node_t * node)
 {
-    ASSERT(FIB_NODE_TYPE_VXLAN_TUNNEL == node->fn_type);
-    return ((vxlan_tunnel_t*) (((char*)node) -
-			       STRUCT_OFFSET_OF(vxlan_tunnel_t, node)));
+  ASSERT (FIB_NODE_TYPE_VXLAN_TUNNEL == node->fn_type);
+  return ((vxlan_tunnel_t *) (((char *) node) -
+			      STRUCT_OFFSET_OF (vxlan_tunnel_t, node)));
 }
 
 /**
@@ -162,38 +170,37 @@
  * Here we will restack the new dpo of VXLAN DIP to encap node.
  */
 static fib_node_back_walk_rc_t
-vxlan_tunnel_back_walk (fib_node_t *node,
-			fib_node_back_walk_ctx_t *ctx)
+vxlan_tunnel_back_walk (fib_node_t * node, fib_node_back_walk_ctx_t * ctx)
 {
-    vxlan_tunnel_restack_dpo(vxlan_tunnel_from_fib_node(node));
-    return (FIB_NODE_BACK_WALK_CONTINUE);
+  vxlan_tunnel_restack_dpo (vxlan_tunnel_from_fib_node (node));
+  return (FIB_NODE_BACK_WALK_CONTINUE);
 }
 
 /**
  * Function definition to get a FIB node from its index
  */
-static fib_node_t*
+static fib_node_t *
 vxlan_tunnel_fib_node_get (fib_node_index_t index)
 {
-    vxlan_tunnel_t * t;
-    vxlan_main_t * vxm = &vxlan_main;
+  vxlan_tunnel_t *t;
+  vxlan_main_t *vxm = &vxlan_main;
 
-    t = pool_elt_at_index(vxm->tunnels, index);
+  t = pool_elt_at_index (vxm->tunnels, index);
 
-    return (&t->node);
+  return (&t->node);
 }
 
 /**
  * Function definition to inform the FIB node that its last lock has gone.
  */
 static void
-vxlan_tunnel_last_lock_gone (fib_node_t *node)
+vxlan_tunnel_last_lock_gone (fib_node_t * node)
 {
-    /*
-     * The VXLAN tunnel is a root of the graph. As such
-     * it never has children and thus is never locked.
-     */
-    ASSERT(0);
+  /*
+   * The VXLAN tunnel is a root of the graph. As such
+   * it never has children and thus is never locked.
+   */
+  ASSERT (0);
 }
 
 /*
@@ -201,9 +208,9 @@
  * for participation in the FIB object graph.
  */
 const static fib_node_vft_t vxlan_vft = {
-    .fnv_get = vxlan_tunnel_fib_node_get,
-    .fnv_last_lock = vxlan_tunnel_last_lock_gone,
-    .fnv_back_walk = vxlan_tunnel_back_walk,
+  .fnv_get = vxlan_tunnel_fib_node_get,
+  .fnv_last_lock = vxlan_tunnel_last_lock_gone,
+  .fnv_back_walk = vxlan_tunnel_back_walk,
 };
 
 
@@ -218,25 +225,26 @@
 static void
 vxlan_rewrite (vxlan_tunnel_t * t, bool is_ip6)
 {
-  union {
+  union
+  {
     ip4_vxlan_header_t h4;
     ip6_vxlan_header_t h6;
   } h;
   int len = is_ip6 ? sizeof h.h6 : sizeof h.h4;
 
-  udp_header_t * udp;
-  vxlan_header_t * vxlan;
+  udp_header_t *udp;
+  vxlan_header_t *vxlan;
   /* Fixed portion of the (outer) ip header */
 
-  memset (&h, 0, sizeof(h));
-  if (!is_ip6) 
+  memset (&h, 0, sizeof (h));
+  if (!is_ip6)
     {
-      ip4_header_t * ip = &h.h4.ip4;
+      ip4_header_t *ip = &h.h4.ip4;
       udp = &h.h4.udp, vxlan = &h.h4.vxlan;
       ip->ip_version_and_header_length = 0x45;
       ip->ttl = 254;
       ip->protocol = IP_PROTOCOL_UDP;
-    
+
       ip->src_address = t->src.ip4;
       ip->dst_address = t->dst.ip4;
 
@@ -245,12 +253,13 @@
     }
   else
     {
-      ip6_header_t * ip = &h.h6.ip6;
+      ip6_header_t *ip = &h.h6.ip6;
       udp = &h.h6.udp, vxlan = &h.h6.vxlan;
-      ip->ip_version_traffic_class_and_flow_label = clib_host_to_net_u32(6 << 28);
+      ip->ip_version_traffic_class_and_flow_label =
+	clib_host_to_net_u32 (6 << 28);
       ip->hop_limit = 255;
       ip->protocol = IP_PROTOCOL_UDP;
-    
+
       ip->src_address = t->src.ip6;
       ip->dst_address = t->dst.ip6;
     }
@@ -260,98 +269,102 @@
   udp->dst_port = clib_host_to_net_u16 (UDP_DST_PORT_vxlan);
 
   /* VXLAN header */
-  vnet_set_vni_and_flags(vxlan, t->vni);
+  vnet_set_vni_and_flags (vxlan, t->vni);
   vnet_rewrite_set_data (*t, &h, len);
 }
 
 static bool
-vxlan_decap_next_is_valid (vxlan_main_t * vxm, u32 is_ip6, u32 decap_next_index)
+vxlan_decap_next_is_valid (vxlan_main_t * vxm, u32 is_ip6,
+			   u32 decap_next_index)
 {
-  vlib_main_t * vm = vxm->vlib_main;
-  u32 input_idx = (!is_ip6) ? vxlan4_input_node.index : vxlan6_input_node.index;
+  vlib_main_t *vm = vxm->vlib_main;
+  u32 input_idx = (!is_ip6) ?
+    vxlan4_input_node.index : vxlan6_input_node.index;
   vlib_node_runtime_t *r = vlib_node_get_runtime (vm, input_idx);
 
   return decap_next_index < r->n_next_nodes;
 }
 
 static uword
-vtep_addr_ref(ip46_address_t *ip)
+vtep_addr_ref (ip46_address_t * ip)
 {
-	uword *vtep = ip46_address_is_ip4(ip) ?
-	                 hash_get (vxlan_main.vtep4, ip->ip4.as_u32) :
-	                 hash_get_mem (vxlan_main.vtep6, &ip->ip6);
-	if (vtep) 
-		return ++(*vtep);
-	ip46_address_is_ip4(ip) ?
-	                 hash_set (vxlan_main.vtep4, ip->ip4.as_u32, 1) :
-	                 hash_set_mem_alloc (&vxlan_main.vtep6, &ip->ip6, 1);
-	return 1;
+  uword *vtep = ip46_address_is_ip4 (ip) ?
+    hash_get (vxlan_main.vtep4, ip->ip4.as_u32) :
+    hash_get_mem (vxlan_main.vtep6, &ip->ip6);
+  if (vtep)
+    return ++(*vtep);
+  ip46_address_is_ip4 (ip) ?
+    hash_set (vxlan_main.vtep4, ip->ip4.as_u32, 1) :
+    hash_set_mem_alloc (&vxlan_main.vtep6, &ip->ip6, 1);
+  return 1;
 }
 
 static uword
-vtep_addr_unref(ip46_address_t *ip)
+vtep_addr_unref (ip46_address_t * ip)
 {
-	uword *vtep = ip46_address_is_ip4(ip) ?
-	                 hash_get (vxlan_main.vtep4, ip->ip4.as_u32) :
-	                 hash_get_mem (vxlan_main.vtep6, &ip->ip6);
-        ASSERT(vtep);
-	if (--(*vtep) != 0)
-		return *vtep;
-	ip46_address_is_ip4(ip) ?
-		hash_unset (vxlan_main.vtep4, ip->ip4.as_u32) :
-		hash_unset_mem_free (&vxlan_main.vtep6, &ip->ip6);
-	return 0;
+  uword *vtep = ip46_address_is_ip4 (ip) ?
+    hash_get (vxlan_main.vtep4, ip->ip4.as_u32) :
+    hash_get_mem (vxlan_main.vtep6, &ip->ip6);
+  ASSERT (vtep);
+  if (--(*vtep) != 0)
+    return *vtep;
+  ip46_address_is_ip4 (ip) ?
+    hash_unset (vxlan_main.vtep4, ip->ip4.as_u32) :
+    hash_unset_mem_free (&vxlan_main.vtep6, &ip->ip6);
+  return 0;
 }
 
-typedef CLIB_PACKED(union {
-  struct {
+/* *INDENT-OFF* */
+typedef CLIB_PACKED(union
+{
+  struct
+  {
     fib_node_index_t mfib_entry_index;
     adj_index_t mcast_adj_index;
   };
   u64 as_u64;
 }) mcast_shared_t;
+/* *INDENT-ON* */
 
 static inline mcast_shared_t
-mcast_shared_get(ip46_address_t * ip)
+mcast_shared_get (ip46_address_t * ip)
 {
-        ASSERT(ip46_address_is_multicast(ip));
-	uword * p = hash_get_mem (vxlan_main.mcast_shared, ip);
-        ASSERT(p);
-	return (mcast_shared_t) { .as_u64 = *p };
+  ASSERT (ip46_address_is_multicast (ip));
+  uword *p = hash_get_mem (vxlan_main.mcast_shared, ip);
+  ASSERT (p);
+  mcast_shared_t ret = {.as_u64 = *p };
+  return ret;
 }
 
 static inline void
-mcast_shared_add(ip46_address_t *dst,
-                 fib_node_index_t mfei,
-                 adj_index_t ai)
+mcast_shared_add (ip46_address_t * dst, fib_node_index_t mfei, adj_index_t ai)
 {
-    mcast_shared_t new_ep = {
-        .mcast_adj_index = ai,
-        .mfib_entry_index = mfei,
-    };
+  mcast_shared_t new_ep = {
+    .mcast_adj_index = ai,
+    .mfib_entry_index = mfei,
+  };
 
-    hash_set_mem_alloc (&vxlan_main.mcast_shared, dst, new_ep.as_u64);
+  hash_set_mem_alloc (&vxlan_main.mcast_shared, dst, new_ep.as_u64);
 }
 
 static inline void
-mcast_shared_remove(ip46_address_t *dst)
+mcast_shared_remove (ip46_address_t * dst)
 {
-    mcast_shared_t ep = mcast_shared_get(dst);
+  mcast_shared_t ep = mcast_shared_get (dst);
 
-    adj_unlock(ep.mcast_adj_index);
-    mfib_table_entry_delete_index(ep.mfib_entry_index,
-                                  MFIB_SOURCE_VXLAN);
+  adj_unlock (ep.mcast_adj_index);
+  mfib_table_entry_delete_index (ep.mfib_entry_index, MFIB_SOURCE_VXLAN);
 
-    hash_unset_mem_free (&vxlan_main.mcast_shared, dst);
+  hash_unset_mem_free (&vxlan_main.mcast_shared, dst);
 }
 
-int vnet_vxlan_add_del_tunnel 
-(vnet_vxlan_add_del_tunnel_args_t *a, u32 * sw_if_indexp)
+int vnet_vxlan_add_del_tunnel
+  (vnet_vxlan_add_del_tunnel_args_t * a, u32 * sw_if_indexp)
 {
-  vxlan_main_t * vxm = &vxlan_main;
+  vxlan_main_t *vxm = &vxlan_main;
   vxlan_tunnel_t *t = 0;
-  vnet_main_t * vnm = vxm->vnet_main;
-  uword * p;
+  vnet_main_t *vnm = vxm->vnet_main;
+  uword *p;
   u32 sw_if_index = ~0;
   vxlan4_tunnel_key_t key4;
   vxlan6_tunnel_key_t key6;
@@ -359,11 +372,11 @@
 
   if (!is_ip6)
     {
-      key4.src = a->dst.ip4.as_u32; /* decap src in key is encap dst in config */
+      key4.src = a->dst.ip4.as_u32;	/* decap src in key is encap dst in config */
       key4.vni = clib_host_to_net_u32 (a->vni << 8);
       p = hash_get (vxm->vxlan4_tunnel_by_key, key4.as_u64);
-    } 
-  else 
+    }
+  else
     {
       key6.src = a->dst.ip6;
       key6.vni = clib_host_to_net_u32 (a->vni << 8);
@@ -372,22 +385,22 @@
 
   if (a->is_add)
     {
-      l2input_main_t * l2im = &l2input_main;
+      l2input_main_t *l2im = &l2input_main;
       u32 dev_instance;		/* real dev instance tunnel index */
       u32 user_instance;	/* request and actual instance number */
 
       /* adding a tunnel: tunnel must not already exist */
       if (p)
-        return VNET_API_ERROR_TUNNEL_EXIST;
+	return VNET_API_ERROR_TUNNEL_EXIST;
 
       /*if not set explicitly, default to l2 */
-      if(a->decap_next_index == ~0)
+      if (a->decap_next_index == ~0)
 	a->decap_next_index = VXLAN_INPUT_NEXT_L2_INPUT;
-      if (!vxlan_decap_next_is_valid(vxm, is_ip6, a->decap_next_index))
-	  return VNET_API_ERROR_INVALID_DECAP_NEXT;
+      if (!vxlan_decap_next_is_valid (vxm, is_ip6, a->decap_next_index))
+	return VNET_API_ERROR_INVALID_DECAP_NEXT;
 
       pool_get_aligned (vxm->tunnels, t, CLIB_CACHE_LINE_BYTES);
-      memset (t, 0, sizeof(*t));
+      memset (t, 0, sizeof (*t));
       dev_instance = t - vxm->tunnels;
 
       /* copy from arg structure */
@@ -400,7 +413,7 @@
        * Reconcile the real dev_instance and a possible requested instance.
        */
       user_instance = a->instance;
-      if (user_instance == ~0) 
+      if (user_instance == ~0)
 	user_instance = dev_instance;
       if (hash_get (vxm->instance_used, user_instance))
 	{
@@ -409,28 +422,29 @@
 	}
       hash_set (vxm->instance_used, user_instance, 1);
 
-      t->dev_instance = dev_instance;		/* actual */
-      t->user_instance = user_instance;		/* name */
+      t->dev_instance = dev_instance;	/* actual */
+      t->user_instance = user_instance;	/* name */
 
       /* copy the key */
       if (is_ip6)
-        hash_set_mem_alloc (&vxm->vxlan6_tunnel_by_key, &key6, dev_instance);
+	hash_set_mem_alloc (&vxm->vxlan6_tunnel_by_key, &key6, dev_instance);
       else
-        hash_set (vxm->vxlan4_tunnel_by_key, key4.as_u64, dev_instance);
+	hash_set (vxm->vxlan4_tunnel_by_key, key4.as_u64, dev_instance);
 
       t->hw_if_index = vnet_register_interface
-        (vnm, vxlan_device_class.index, dev_instance,
-         vxlan_hw_class.index, dev_instance);
-      vnet_hw_interface_t * hi = vnet_get_hw_interface (vnm, t->hw_if_index);
+	(vnm, vxlan_device_class.index, dev_instance,
+	 vxlan_hw_class.index, dev_instance);
+      vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, t->hw_if_index);
 
       /* Set vxlan tunnel output node */
       u32 encap_index = !is_ip6 ?
-	  vxlan4_encap_node.index : vxlan6_encap_node.index;
+	vxlan4_encap_node.index : vxlan6_encap_node.index;
       vnet_set_interface_output_node (vnm, t->hw_if_index, encap_index);
 
       t->sw_if_index = sw_if_index = hi->sw_if_index;
 
-      vec_validate_init_empty (vxm->tunnel_index_by_sw_if_index, sw_if_index, ~0);
+      vec_validate_init_empty (vxm->tunnel_index_by_sw_if_index, sw_if_index,
+			       ~0);
       vxm->tunnel_index_by_sw_if_index[sw_if_index] = dev_instance;
 
       /* setup l2 input config with l2 feature and bd 0 to drop packet */
@@ -438,172 +452,173 @@
       l2im->configs[sw_if_index].feature_bitmap = L2INPUT_FEAT_DROP;
       l2im->configs[sw_if_index].bd_index = 0;
 
-      vnet_sw_interface_t * si = vnet_get_sw_interface (vnm, sw_if_index);
+      vnet_sw_interface_t *si = vnet_get_sw_interface (vnm, sw_if_index);
       si->flags &= ~VNET_SW_INTERFACE_FLAG_HIDDEN;
       vnet_sw_interface_set_flags (vnm, sw_if_index,
-                                   VNET_SW_INTERFACE_FLAG_ADMIN_UP);
+				   VNET_SW_INTERFACE_FLAG_ADMIN_UP);
 
-      fib_node_init(&t->node, FIB_NODE_TYPE_VXLAN_TUNNEL);
+      fib_node_init (&t->node, FIB_NODE_TYPE_VXLAN_TUNNEL);
       fib_prefix_t tun_dst_pfx;
       vnet_flood_class_t flood_class = VNET_FLOOD_CLASS_TUNNEL_NORMAL;
 
-      fib_prefix_from_ip46_addr(&t->dst, &tun_dst_pfx);
-      if (!ip46_address_is_multicast(&t->dst))
-        {
-          /* Unicast tunnel -
-           * source the FIB entry for the tunnel's destination
-           * and become a child thereof. The tunnel will then get poked
-           * when the forwarding for the entry updates, and the tunnel can
-           * re-stack accordingly
-           */
-          vtep_addr_ref(&t->src);
-          t->fib_entry_index = fib_table_entry_special_add
-            (t->encap_fib_index, &tun_dst_pfx, FIB_SOURCE_RR,
+      fib_prefix_from_ip46_addr (&t->dst, &tun_dst_pfx);
+      if (!ip46_address_is_multicast (&t->dst))
+	{
+	  /* Unicast tunnel -
+	   * source the FIB entry for the tunnel's destination
+	   * and become a child thereof. The tunnel will then get poked
+	   * when the forwarding for the entry updates, and the tunnel can
+	   * re-stack accordingly
+	   */
+	  vtep_addr_ref (&t->src);
+	  t->fib_entry_index = fib_table_entry_special_add
+	    (t->encap_fib_index, &tun_dst_pfx, FIB_SOURCE_RR,
 	     FIB_ENTRY_FLAG_NONE);
-          t->sibling_index = fib_entry_child_add
-            (t->fib_entry_index, FIB_NODE_TYPE_VXLAN_TUNNEL, dev_instance);
-          vxlan_tunnel_restack_dpo(t);
-	} 
+	  t->sibling_index = fib_entry_child_add
+	    (t->fib_entry_index, FIB_NODE_TYPE_VXLAN_TUNNEL, dev_instance);
+	  vxlan_tunnel_restack_dpo (t);
+	}
       else
-        {
- 	  /* Multicast tunnel -
+	{
+	  /* Multicast tunnel -
 	   * as the same mcast group can be used for mutiple mcast tunnels
 	   * with different VNIs, create the output fib adjecency only if
 	   * it does not already exist
 	   */
-          fib_protocol_t fp = fib_ip_proto(is_ip6);
+	  fib_protocol_t fp = fib_ip_proto (is_ip6);
 
-	  if (vtep_addr_ref(&t->dst) == 1)
-          {
-              fib_node_index_t mfei;
-              adj_index_t ai;
-              fib_route_path_t path = {
-                  .frp_proto = fib_proto_to_dpo(fp),
-                  .frp_addr = zero_addr,
-                  .frp_sw_if_index = 0xffffffff,
-                  .frp_fib_index = ~0,
-                  .frp_weight = 0,
-                  .frp_flags = FIB_ROUTE_PATH_LOCAL,
-              };
-              const mfib_prefix_t mpfx = {
-                  .fp_proto = fp,
-                  .fp_len = (is_ip6 ? 128 : 32),
-                  .fp_grp_addr = tun_dst_pfx.fp_addr,
-              };
+	  if (vtep_addr_ref (&t->dst) == 1)
+	    {
+	      fib_node_index_t mfei;
+	      adj_index_t ai;
+	      fib_route_path_t path = {
+		.frp_proto = fib_proto_to_dpo (fp),
+		.frp_addr = zero_addr,
+		.frp_sw_if_index = 0xffffffff,
+		.frp_fib_index = ~0,
+		.frp_weight = 0,
+		.frp_flags = FIB_ROUTE_PATH_LOCAL,
+	      };
+	      const mfib_prefix_t mpfx = {
+		.fp_proto = fp,
+		.fp_len = (is_ip6 ? 128 : 32),
+		.fp_grp_addr = tun_dst_pfx.fp_addr,
+	      };
 
-              /*
-               * Setup the (*,G) to receive traffic on the mcast group
-               *  - the forwarding interface is for-us
-               *  - the accepting interface is that from the API
-               */
-              mfib_table_entry_path_update(t->encap_fib_index,
-                                           &mpfx,
-                                           MFIB_SOURCE_VXLAN,
-                                           &path,
-                                           MFIB_ITF_FLAG_FORWARD);
+	      /*
+	       * Setup the (*,G) to receive traffic on the mcast group
+	       *  - the forwarding interface is for-us
+	       *  - the accepting interface is that from the API
+	       */
+	      mfib_table_entry_path_update (t->encap_fib_index,
+					    &mpfx,
+					    MFIB_SOURCE_VXLAN,
+					    &path, MFIB_ITF_FLAG_FORWARD);
 
-              path.frp_sw_if_index = a->mcast_sw_if_index;
-              path.frp_flags = FIB_ROUTE_PATH_FLAG_NONE;
-              mfei = mfib_table_entry_path_update(t->encap_fib_index,
-                                                  &mpfx,
-                                                  MFIB_SOURCE_VXLAN,
-                                                  &path,
-                                                  MFIB_ITF_FLAG_ACCEPT);
+	      path.frp_sw_if_index = a->mcast_sw_if_index;
+	      path.frp_flags = FIB_ROUTE_PATH_FLAG_NONE;
+	      mfei = mfib_table_entry_path_update (t->encap_fib_index,
+						   &mpfx,
+						   MFIB_SOURCE_VXLAN,
+						   &path,
+						   MFIB_ITF_FLAG_ACCEPT);
 
-              /*
-               * Create the mcast adjacency to send traffic to the group
-               */
-              ai = adj_mcast_add_or_lock(fp,
-                                         fib_proto_to_link(fp),
-                                         a->mcast_sw_if_index);
+	      /*
+	       * Create the mcast adjacency to send traffic to the group
+	       */
+	      ai = adj_mcast_add_or_lock (fp,
+					  fib_proto_to_link (fp),
+					  a->mcast_sw_if_index);
 
-              /*
-               * create a new end-point
-               */
-              mcast_shared_add(&t->dst, mfei, ai);
-          }
+	      /*
+	       * create a new end-point
+	       */
+	      mcast_shared_add (&t->dst, mfei, ai);
+	    }
 
-          dpo_id_t dpo = DPO_INVALID;
-          mcast_shared_t ep = mcast_shared_get(&t->dst);
+	  dpo_id_t dpo = DPO_INVALID;
+	  mcast_shared_t ep = mcast_shared_get (&t->dst);
 
-          /* Stack shared mcast dst mac addr rewrite on encap */
-          dpo_set (&dpo, DPO_ADJACENCY_MCAST,
-                   fib_proto_to_dpo(fp),
-                   ep.mcast_adj_index);
+	  /* Stack shared mcast dst mac addr rewrite on encap */
+	  dpo_set (&dpo, DPO_ADJACENCY_MCAST,
+		   fib_proto_to_dpo (fp), ep.mcast_adj_index);
 
-          dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
-          dpo_reset (&dpo);
+	  dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
+	  dpo_reset (&dpo);
 	  flood_class = VNET_FLOOD_CLASS_TUNNEL_MASTER;
 	}
 
-      vnet_get_sw_interface (vnet_get_main(), sw_if_index)->flood_class = flood_class;
+      vnet_get_sw_interface (vnet_get_main (), sw_if_index)->flood_class =
+	flood_class;
     }
   else
     {
       /* deleting a tunnel: tunnel must exist */
       if (!p)
-        return VNET_API_ERROR_NO_SUCH_ENTRY;
+	return VNET_API_ERROR_NO_SUCH_ENTRY;
 
       u32 instance = p[0];
       t = pool_elt_at_index (vxm->tunnels, instance);
 
       sw_if_index = t->sw_if_index;
-      vnet_sw_interface_set_flags (vnm, sw_if_index, 0 /* down */);
+      vnet_sw_interface_set_flags (vnm, sw_if_index, 0 /* down */ );
 
       vxm->tunnel_index_by_sw_if_index[sw_if_index] = ~0;
 
       if (!is_ip6)
-        hash_unset (vxm->vxlan4_tunnel_by_key, key4.as_u64);
+	hash_unset (vxm->vxlan4_tunnel_by_key, key4.as_u64);
       else
 	hash_unset_mem_free (&vxm->vxlan6_tunnel_by_key, &key6);
 
-      if (!ip46_address_is_multicast(&t->dst))
-        {
-	  vtep_addr_unref(&t->src);
-	  fib_entry_child_remove(t->fib_entry_index, t->sibling_index);
-	  fib_table_entry_delete_index(t->fib_entry_index, FIB_SOURCE_RR);
-        }
-      else if (vtep_addr_unref(&t->dst) == 0)
-        {
-	  mcast_shared_remove(&t->dst);
-        }
+      if (!ip46_address_is_multicast (&t->dst))
+	{
+	  vtep_addr_unref (&t->src);
+	  fib_entry_child_remove (t->fib_entry_index, t->sibling_index);
+	  fib_table_entry_delete_index (t->fib_entry_index, FIB_SOURCE_RR);
+	}
+      else if (vtep_addr_unref (&t->dst) == 0)
+	{
+	  mcast_shared_remove (&t->dst);
+	}
 
       vnet_delete_hw_interface (vnm, t->hw_if_index);
       hash_unset (vxm->instance_used, t->user_instance);
 
-      fib_node_deinit(&t->node);
+      fib_node_deinit (&t->node);
       pool_put (vxm->tunnels, t);
     }
 
   if (sw_if_indexp)
-      *sw_if_indexp = sw_if_index;
+    *sw_if_indexp = sw_if_index;
 
   return 0;
 }
 
-static uword get_decap_next_for_node(u32 node_index, u32 ipv4_set)
+static uword
+get_decap_next_for_node (u32 node_index, u32 ipv4_set)
 {
-  vxlan_main_t * vxm = &vxlan_main;
-  vlib_main_t * vm = vxm->vlib_main;
+  vxlan_main_t *vxm = &vxlan_main;
+  vlib_main_t *vm = vxm->vlib_main;
   uword input_node = (ipv4_set) ? vxlan4_input_node.index :
     vxlan6_input_node.index;
 
   return vlib_node_add_next (vm, input_node, node_index);
 }
 
-static uword unformat_decap_next (unformat_input_t * input, va_list * args)
+static uword
+unformat_decap_next (unformat_input_t * input, va_list * args)
 {
-  u32 * result = va_arg (*args, u32 *);
+  u32 *result = va_arg (*args, u32 *);
   u32 ipv4_set = va_arg (*args, int);
-  vxlan_main_t * vxm = &vxlan_main;
-  vlib_main_t * vm = vxm->vlib_main;
+  vxlan_main_t *vxm = &vxlan_main;
+  vlib_main_t *vm = vxm->vlib_main;
   u32 node_index;
   u32 tmp;
 
   if (unformat (input, "l2"))
     *result = VXLAN_INPUT_NEXT_L2_INPUT;
   else if (unformat (input, "node %U", unformat_vlib_node, vm, &node_index))
-    *result = get_decap_next_for_node(node_index, ipv4_set);
+    *result = get_decap_next_for_node (node_index, ipv4_set);
   else if (unformat (input, "%d", &tmp))
     *result = tmp;
   else
@@ -613,11 +628,12 @@
 
 static clib_error_t *
 vxlan_add_del_tunnel_command_fn (vlib_main_t * vm,
-                                   unformat_input_t * input,
-                                   vlib_cli_command_t * cmd)
+				 unformat_input_t * input,
+				 vlib_cli_command_t * cmd)
 {
-  unformat_input_t _line_input, * line_input = &_line_input;
-  ip46_address_t src = ip46_address_initializer, dst = ip46_address_initializer;
+  unformat_input_t _line_input, *line_input = &_line_input;
+  ip46_address_t src = ip46_address_initializer, dst =
+    ip46_address_initializer;
   u8 is_add = 1;
   u8 src_set = 0;
   u8 dst_set = 0;
@@ -633,57 +649,61 @@
   clib_error_t *error = NULL;
 
   /* Get a line of input. */
-  if (! unformat_user (input, unformat_line_input, line_input))
+  if (!unformat_user (input, unformat_line_input, line_input))
     return 0;
 
-  while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
-    if (unformat (line_input, "del"))
-      {
-        is_add = 0;
-      }
-    else if (unformat (line_input, "instance %d", &instance))
-      ;
-    else if (unformat (line_input, "src %U",
-                       unformat_ip46_address, &src, IP46_TYPE_ANY))
-      {
-        src_set = 1;
-        ip46_address_is_ip4(&src) ? (ipv4_set = 1) : (ipv6_set = 1);
-      }
-    else if (unformat (line_input, "dst %U",
-                       unformat_ip46_address, &dst, IP46_TYPE_ANY))
-      {
-        dst_set = 1;
-        ip46_address_is_ip4(&dst) ? (ipv4_set = 1) : (ipv6_set = 1);
-      }
-    else if (unformat (line_input, "group %U %U",
-                       unformat_ip46_address, &dst, IP46_TYPE_ANY,
-		       unformat_vnet_sw_interface,
-		       vnet_get_main(), &mcast_sw_if_index))
-      {
-        grp_set = dst_set = 1;
-        ip46_address_is_ip4(&dst) ? (ipv4_set = 1) : (ipv6_set = 1);
-      }
-    else if (unformat (line_input, "encap-vrf-id %d", &table_id))
-      {
-        encap_fib_index = fib_table_find (fib_ip_proto (ipv6_set), table_id);
-        if (encap_fib_index == ~0)
-          {
-            error = clib_error_return (0, "nonexistent encap-vrf-id %d", table_id);
-            break;
-          }
-      }
-    else if (unformat (line_input, "decap-next %U", unformat_decap_next, 
-                       &decap_next_index, ipv4_set))
-      ;
-    else if (unformat (line_input, "vni %d", &vni))
-      ;
-    else 
-      {
-        error = clib_error_return (0, "parse error: '%U'",
-                                   format_unformat_error, line_input);
-        break;
-      }
-  }
+  while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
+    {
+      if (unformat (line_input, "del"))
+	{
+	  is_add = 0;
+	}
+      else if (unformat (line_input, "instance %d", &instance))
+	;
+      else if (unformat (line_input, "src %U",
+			 unformat_ip46_address, &src, IP46_TYPE_ANY))
+	{
+	  src_set = 1;
+	  ip46_address_is_ip4 (&src) ? (ipv4_set = 1) : (ipv6_set = 1);
+	}
+      else if (unformat (line_input, "dst %U",
+			 unformat_ip46_address, &dst, IP46_TYPE_ANY))
+	{
+	  dst_set = 1;
+	  ip46_address_is_ip4 (&dst) ? (ipv4_set = 1) : (ipv6_set = 1);
+	}
+      else if (unformat (line_input, "group %U %U",
+			 unformat_ip46_address, &dst, IP46_TYPE_ANY,
+			 unformat_vnet_sw_interface,
+			 vnet_get_main (), &mcast_sw_if_index))
+	{
+	  grp_set = dst_set = 1;
+	  ip46_address_is_ip4 (&dst) ? (ipv4_set = 1) : (ipv6_set = 1);
+	}
+      else if (unformat (line_input, "encap-vrf-id %d", &table_id))
+	{
+	  encap_fib_index =
+	    fib_table_find (fib_ip_proto (ipv6_set), table_id);
+	  if (encap_fib_index == ~0)
+	    {
+	      error =
+		clib_error_return (0, "nonexistent encap-vrf-id %d",
+				   table_id);
+	      break;
+	    }
+	}
+      else if (unformat (line_input, "decap-next %U", unformat_decap_next,
+			 &decap_next_index, ipv4_set))
+	;
+      else if (unformat (line_input, "vni %d", &vni))
+	;
+      else
+	{
+	  error = clib_error_return (0, "parse error: '%U'",
+				     format_unformat_error, line_input);
+	  break;
+	}
+    }
 
   unformat_free (line_input);
 
@@ -696,10 +716,10 @@
   if (dst_set == 0)
     return clib_error_return (0, "tunnel dst address not specified");
 
-  if (grp_set && !ip46_address_is_multicast(&dst))
+  if (grp_set && !ip46_address_is_multicast (&dst))
     return clib_error_return (0, "tunnel group address not multicast");
 
-  if (grp_set == 0 && ip46_address_is_multicast(&dst))
+  if (grp_set == 0 && ip46_address_is_multicast (&dst))
     return clib_error_return (0, "dst address must be unicast");
 
   if (grp_set && mcast_sw_if_index == ~0)
@@ -708,7 +728,7 @@
   if (ipv4_set && ipv6_set)
     return clib_error_return (0, "both IPv4 and IPv6 addresses specified");
 
-  if (ip46_address_cmp(&src, &dst) == 0)
+  if (ip46_address_cmp (&src, &dst) == 0)
     return clib_error_return (0, "src and dst addresses are identical");
 
   if (decap_next_index == ~0)
@@ -717,27 +737,27 @@
   if (vni == 0)
     return clib_error_return (0, "vni not specified");
 
-  if (vni >> 24)  
+  if (vni >> 24)
     return clib_error_return (0, "vni %d out of range", vni);
 
   vnet_vxlan_add_del_tunnel_args_t a = {
-  .is_add = is_add,
-  .is_ip6 = ipv6_set,
-  .instance = instance,
+    .is_add = is_add,
+    .is_ip6 = ipv6_set,
+    .instance = instance,
 #define _(x) .x = x,
-  foreach_copy_field
+    foreach_copy_field
 #undef _
   };
 
   u32 tunnel_sw_if_index;
   int rv = vnet_vxlan_add_del_tunnel (&a, &tunnel_sw_if_index);
 
-  switch(rv)
+  switch (rv)
     {
     case 0:
       if (is_add)
-        vlib_cli_output(vm, "%U\n", format_vnet_sw_if_index_name, 
-			vnet_get_main(), tunnel_sw_if_index);
+	vlib_cli_output (vm, "%U\n", format_vnet_sw_if_index_name,
+			 vnet_get_main (), tunnel_sw_if_index);
       break;
 
     case VNET_API_ERROR_TUNNEL_EXIST:
@@ -751,7 +771,7 @@
 
     default:
       return clib_error_return
-        (0, "vnet_vxlan_add_del_tunnel returned %d", rv);
+	(0, "vnet_vxlan_add_del_tunnel returned %d", rv);
     }
 
   return error;
@@ -782,7 +802,7 @@
 /* *INDENT-OFF* */
 VLIB_CLI_COMMAND (create_vxlan_tunnel_command, static) = {
   .path = "create vxlan tunnel",
-  .short_help = 
+  .short_help =
   "create vxlan tunnel src <local-vtep-addr>"
   " {dst <remote-vtep-addr>|group <mcast-vtep-addr> <intf-name>} vni <nn>"
   " [instance <id>]"
@@ -793,20 +813,22 @@
 
 static clib_error_t *
 show_vxlan_tunnel_command_fn (vlib_main_t * vm,
-                                unformat_input_t * input,
-                                vlib_cli_command_t * cmd)
+			      unformat_input_t * input,
+			      vlib_cli_command_t * cmd)
 {
-  vxlan_main_t * vxm = &vxlan_main;
-  vxlan_tunnel_t * t;
-  
+  vxlan_main_t *vxm = &vxlan_main;
+  vxlan_tunnel_t *t;
+
   if (pool_elts (vxm->tunnels) == 0)
     vlib_cli_output (vm, "No vxlan tunnels configured...");
 
+/* *INDENT-OFF* */
   pool_foreach (t, vxm->tunnels,
   ({
     vlib_cli_output (vm, "%U", format_vxlan_tunnel, t);
   }));
-  
+/* *INDENT-ON* */
+
   return 0;
 }
 
@@ -828,9 +850,8 @@
 /* *INDENT-ON* */
 
 
-void vnet_int_vxlan_bypass_mode (u32 sw_if_index,
-				 u8 is_ip6,
-				 u8 is_enable)
+void
+vnet_int_vxlan_bypass_mode (u32 sw_if_index, u8 is_ip6, u8 is_enable)
 {
   if (is_ip6)
     vnet_feature_enable_disable ("ip6-unicast", "ip6-vxlan-bypass",
@@ -843,28 +864,28 @@
 
 static clib_error_t *
 set_ip_vxlan_bypass (u32 is_ip6,
-		     unformat_input_t * input,
-		     vlib_cli_command_t * cmd)
+		     unformat_input_t * input, vlib_cli_command_t * cmd)
 {
-  unformat_input_t _line_input, * line_input = &_line_input;
-  vnet_main_t * vnm = vnet_get_main();
-  clib_error_t * error = 0;
+  unformat_input_t _line_input, *line_input = &_line_input;
+  vnet_main_t *vnm = vnet_get_main ();
+  clib_error_t *error = 0;
   u32 sw_if_index, is_enable;
 
   sw_if_index = ~0;
   is_enable = 1;
 
-  if (! unformat_user (input, unformat_line_input, line_input))
+  if (!unformat_user (input, unformat_line_input, line_input))
     return 0;
 
   while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
     {
-      if (unformat_user (line_input, unformat_vnet_sw_interface, vnm, &sw_if_index))
-	  ;
+      if (unformat_user
+	  (line_input, unformat_vnet_sw_interface, vnm, &sw_if_index))
+	;
       else if (unformat (line_input, "del"))
-        is_enable = 0;
+	is_enable = 0;
       else
-        {
+	{
 	  error = unformat_parse_error (line_input);
 	  goto done;
 	}
@@ -879,7 +900,7 @@
 
   vnet_int_vxlan_bypass_mode (sw_if_index, is_ip6, is_enable);
 
- done:
+done:
   unformat_free (line_input);
 
   return error;
@@ -887,17 +908,16 @@
 
 static clib_error_t *
 set_ip4_vxlan_bypass (vlib_main_t * vm,
-		      unformat_input_t * input,
-		      vlib_cli_command_t * cmd)
+		      unformat_input_t * input, vlib_cli_command_t * cmd)
 {
   return set_ip_vxlan_bypass (0, input, cmd);
 }
 
 /*?
- * This command adds the 'ip4-vxlan-bypass' graph node for a given interface. 
+ * This command adds the 'ip4-vxlan-bypass' graph node for a given interface.
  * By adding the IPv4 vxlan-bypass graph node to an interface, the node checks
- *  for and validate input vxlan packet and bypass ip4-lookup, ip4-local, 
- * ip4-udp-lookup nodes to speedup vxlan packet forwarding. This node will 
+ *  for and validate input vxlan packet and bypass ip4-lookup, ip4-local,
+ * ip4-udp-lookup nodes to speedup vxlan packet forwarding. This node will
  * cause extra overhead to for non-vxlan packets which is kept at a minimum.
  *
  * @cliexpar
@@ -907,7 +927,7 @@
  *            Name                      Next                    Previous
  * ip4-vxlan-bypass                error-drop [0]
  *                                vxlan4-input [1]
- *                                 ip4-lookup [2]      
+ *                                 ip4-lookup [2]
  * @cliexend
  *
  * Example of how to enable ip4-vxlan-bypass on an interface:
@@ -915,10 +935,10 @@
  *
  * Example of graph node after ip4-vxlan-bypass is enabled:
  * @cliexstart{show vlib graph ip4-vxlan-bypass}
- *            Name                      Next                    Previous         
- * ip4-vxlan-bypass                error-drop [0]               ip4-input        
- *                                vxlan4-input [1]        ip4-input-no-checksum  
- *                                 ip4-lookup [2]      
+ *            Name                      Next                    Previous
+ * ip4-vxlan-bypass                error-drop [0]               ip4-input
+ *                                vxlan4-input [1]        ip4-input-no-checksum
+ *                                 ip4-lookup [2]
  * @cliexend
  *
  * Example of how to display the feature enabed on an interface:
@@ -945,17 +965,16 @@
 
 static clib_error_t *
 set_ip6_vxlan_bypass (vlib_main_t * vm,
-		      unformat_input_t * input,
-		      vlib_cli_command_t * cmd)
+		      unformat_input_t * input, vlib_cli_command_t * cmd)
 {
   return set_ip_vxlan_bypass (1, input, cmd);
 }
 
 /*?
- * This command adds the 'ip6-vxlan-bypass' graph node for a given interface. 
+ * This command adds the 'ip6-vxlan-bypass' graph node for a given interface.
  * By adding the IPv6 vxlan-bypass graph node to an interface, the node checks
- *  for and validate input vxlan packet and bypass ip6-lookup, ip6-local, 
- * ip6-udp-lookup nodes to speedup vxlan packet forwarding. This node will 
+ *  for and validate input vxlan packet and bypass ip6-lookup, ip6-local,
+ * ip6-udp-lookup nodes to speedup vxlan packet forwarding. This node will
  * cause extra overhead to for non-vxlan packets which is kept at a minimum.
  *
  * @cliexpar
@@ -965,7 +984,7 @@
  *            Name                      Next                    Previous
  * ip6-vxlan-bypass                error-drop [0]
  *                                vxlan6-input [1]
- *                                 ip6-lookup [2]      
+ *                                 ip6-lookup [2]
  * @cliexend
  *
  * Example of how to enable ip6-vxlan-bypass on an interface:
@@ -973,10 +992,10 @@
  *
  * Example of graph node after ip6-vxlan-bypass is enabled:
  * @cliexstart{show vlib graph ip6-vxlan-bypass}
- *            Name                      Next                    Previous         
- * ip6-vxlan-bypass                error-drop [0]               ip6-input        
- *                                vxlan6-input [1]        ip4-input-no-checksum  
- *                                 ip6-lookup [2]      
+ *            Name                      Next                    Previous
+ * ip6-vxlan-bypass                error-drop [0]               ip6-input
+ *                                vxlan6-input [1]        ip4-input-no-checksum
+ *                                 ip6-lookup [2]
  * @cliexend
  *
  * Example of how to display the feature enabed on an interface:
@@ -1001,37 +1020,38 @@
 };
 /* *INDENT-ON* */
 
-clib_error_t *vxlan_init (vlib_main_t *vm)
+clib_error_t *
+vxlan_init (vlib_main_t * vm)
 {
-  vxlan_main_t * vxm = &vxlan_main;
-  
-  vxm->vnet_main = vnet_get_main();
+  vxlan_main_t *vxm = &vxlan_main;
+
+  vxm->vnet_main = vnet_get_main ();
   vxm->vlib_main = vm;
 
   /* initialize the ip6 hash */
-  vxm->vxlan6_tunnel_by_key = hash_create_mem(0,
-        sizeof(vxlan6_tunnel_key_t),
-        sizeof(uword));
-  vxm->vtep6 = hash_create_mem(0,
-        sizeof(ip6_address_t),
-	sizeof(uword));
-  vxm->mcast_shared = hash_create_mem(0,
-        sizeof(ip46_address_t),
-	sizeof(mcast_shared_t));
+  vxm->vxlan6_tunnel_by_key = hash_create_mem (0,
+					       sizeof (vxlan6_tunnel_key_t),
+					       sizeof (uword));
+  vxm->vtep6 = hash_create_mem (0, sizeof (ip6_address_t), sizeof (uword));
+  vxm->mcast_shared = hash_create_mem (0,
+				       sizeof (ip46_address_t),
+				       sizeof (mcast_shared_t));
 
-  udp_register_dst_port (vm, UDP_DST_PORT_vxlan, 
-                         vxlan4_input_node.index, /* is_ip4 */ 1);
+  udp_register_dst_port (vm, UDP_DST_PORT_vxlan,
+			 vxlan4_input_node.index, /* is_ip4 */ 1);
   udp_register_dst_port (vm, UDP_DST_PORT_vxlan6,
-                         vxlan6_input_node.index, /* is_ip4 */ 0);
+			 vxlan6_input_node.index, /* is_ip4 */ 0);
 
-  fib_node_register_type(FIB_NODE_TYPE_VXLAN_TUNNEL, &vxlan_vft);
+  fib_node_register_type (FIB_NODE_TYPE_VXLAN_TUNNEL, &vxlan_vft);
 
   return 0;
 }
 
-VLIB_INIT_FUNCTION(vxlan_init);
+VLIB_INIT_FUNCTION (vxlan_init);
 
 /*
+ * fd.io coding-style-patch-verification: ON
+ *
  * Local Variables:
  * eval: (c-set-style "gnu")
  * End: