udp/session: refactor to support dgram mode

- adds session layer support for datagram based protocols
- updates udp to work in pure connectionless and datagram mode. The
  existing connected mode is now 'accessible' for apps as a dummy UDPC,
  as in, connected udp, protocol.
- updates udp_echo, echo client, echo server code to work in datagram
  mode.

Change-Id: I2960c0d2d246cb166005f545794ec31fe0d546dd
Signed-off-by: Florin Coras <fcoras@cisco.com>
diff --git a/src/svm/svm_fifo.c b/src/svm/svm_fifo.c
index b2c8e5b..3552192 100644
--- a/src/svm/svm_fifo.c
+++ b/src/svm/svm_fifo.c
@@ -169,6 +169,9 @@
   svm_fifo_t *f = va_arg (*args, svm_fifo_t *);
   int verbose = va_arg (*args, int);
 
+  if (!s)
+    return s;
+
   s = format (s, "cursize %u nitems %u has_event %d\n",
 	      f->cursize, f->nitems, f->has_event);
   s = format (s, " head %d tail %d\n", f->head, f->tail);
@@ -459,7 +462,7 @@
   f->ooos_newest = OOO_SEGMENT_INVALID_INDEX;
 
   if (PREDICT_FALSE (cursize == f->nitems))
-    return -2;			/* fifo stuffed */
+    return SVM_FIFO_FULL;
 
   nitems = f->nitems;
 
@@ -615,6 +618,20 @@
 						copy_from_here);
 }
 
+void
+svm_fifo_overwrite_head (svm_fifo_t * f, u8 * data, u32 len)
+{
+  u32 first_chunk;
+  ASSERT (len <= f->nitems);
+  if (len < f->nitems - f->head)
+    clib_memcpy (&f->data[f->head], data, len);
+  else
+    {
+      first_chunk = len - (f->nitems - f->head);
+      clib_memcpy (&f->data[f->head], data, first_chunk);
+      clib_memcpy (f->data, data + first_chunk, len - first_chunk);
+    }
+}
 
 static int
 svm_fifo_dequeue_internal (svm_fifo_t * f, u32 max_bytes, u8 * copy_here)
diff --git a/src/svm/svm_fifo.h b/src/svm/svm_fifo.h
index d06d77a..39cdcc0 100644
--- a/src/svm/svm_fifo.h
+++ b/src/svm/svm_fifo.h
@@ -80,6 +80,11 @@
     CLIB_CACHE_LINE_ALIGN_MARK (data);
 } svm_fifo_t;
 
+typedef enum
+{
+  SVM_FIFO_FULL = -2,
+} svm_fifo_err_t;
+
 #if SVM_FIFO_TRACE
 #define svm_fifo_trace_add(_f, _s, _l, _t)		\
 {							\
@@ -150,6 +155,7 @@
 u32 svm_fifo_number_ooo_segments (svm_fifo_t * f);
 ooo_segment_t *svm_fifo_first_ooo_segment (svm_fifo_t * f);
 void svm_fifo_init_pointers (svm_fifo_t * f, u32 pointer);
+void svm_fifo_overwrite_head (svm_fifo_t * f, u8 * data, u32 len);
 
 format_function_t format_svm_fifo;
 
diff --git a/src/tests/vnet/session/udp_echo.c b/src/tests/vnet/session/udp_echo.c
index d1363fc..fe5461d 100644
--- a/src/tests/vnet/session/udp_echo.c
+++ b/src/tests/vnet/session/udp_echo.c
@@ -58,12 +58,6 @@
 
 typedef struct
 {
-  svm_fifo_t *server_rx_fifo;
-  svm_fifo_t *server_tx_fifo;
-} session_t;
-
-typedef struct
-{
   /* vpe input queue */
   svm_queue_t *vl_input_queue;
 
@@ -71,10 +65,13 @@
   u32 my_client_index;
 
   /* The URI we're playing with */
-  u8 *uri;
+  u8 *listen_uri;
+
+  /* URI for connect */
+  u8 *connect_uri;
 
   /* Session pool */
-  session_t *sessions;
+  app_session_t *sessions;
 
   /* Hash table for disconnect processing */
   uword *session_index_by_vpp_handles;
@@ -85,10 +82,9 @@
   /* intermediate rx buffer */
   u8 *rx_buf;
 
-  /* URI for connect */
-  u8 *connect_uri;
-
-  int i_am_master;
+  u32 fifo_size;
+  int i_am_server;
+  u8 is_connected;
 
   /* Our event queue */
   svm_queue_t *our_event_queue;
@@ -128,7 +124,8 @@
   u8 *connect_test_data;
 
   uword *segments_table;
-} uri_udp_test_main_t;
+  u8 do_echo;
+} udp_echo_main_t;
 
 #if CLIB_DEBUG > 0
 #define NITER 10000
@@ -136,12 +133,12 @@
 #define NITER 4000000
 #endif
 
-uri_udp_test_main_t uri_udp_test_main;
+udp_echo_main_t udp_echo_main;
 
 static void
 stop_signal (int signum)
 {
-  uri_udp_test_main_t *um = &uri_udp_test_main;
+  udp_echo_main_t *um = &udp_echo_main;
 
   um->time_to_stop = 1;
 }
@@ -149,7 +146,7 @@
 static void
 stats_signal (int signum)
 {
-  uri_udp_test_main_t *um = &uri_udp_test_main;
+  udp_echo_main_t *um = &udp_echo_main;
 
   um->time_to_print_stats = 1;
 }
@@ -164,11 +161,146 @@
   return 0;
 }
 
+uword
+unformat_ip4_address (unformat_input_t * input, va_list * args)
+{
+  u8 *result = va_arg (*args, u8 *);
+  unsigned a[4];
+
+  if (!unformat (input, "%d.%d.%d.%d", &a[0], &a[1], &a[2], &a[3]))
+    return 0;
+
+  if (a[0] >= 256 || a[1] >= 256 || a[2] >= 256 || a[3] >= 256)
+    return 0;
+
+  result[0] = a[0];
+  result[1] = a[1];
+  result[2] = a[2];
+  result[3] = a[3];
+
+  return 1;
+}
+
+uword
+unformat_ip6_address (unformat_input_t * input, va_list * args)
+{
+  ip6_address_t *result = va_arg (*args, ip6_address_t *);
+  u16 hex_quads[8];
+  uword hex_quad, n_hex_quads, hex_digit, n_hex_digits;
+  uword c, n_colon, double_colon_index;
+
+  n_hex_quads = hex_quad = n_hex_digits = n_colon = 0;
+  double_colon_index = ARRAY_LEN (hex_quads);
+  while ((c = unformat_get_input (input)) != UNFORMAT_END_OF_INPUT)
+    {
+      hex_digit = 16;
+      if (c >= '0' && c <= '9')
+	hex_digit = c - '0';
+      else if (c >= 'a' && c <= 'f')
+	hex_digit = c + 10 - 'a';
+      else if (c >= 'A' && c <= 'F')
+	hex_digit = c + 10 - 'A';
+      else if (c == ':' && n_colon < 2)
+	n_colon++;
+      else
+	{
+	  unformat_put_input (input);
+	  break;
+	}
+
+      /* Too many hex quads. */
+      if (n_hex_quads >= ARRAY_LEN (hex_quads))
+	return 0;
+
+      if (hex_digit < 16)
+	{
+	  hex_quad = (hex_quad << 4) | hex_digit;
+
+	  /* Hex quad must fit in 16 bits. */
+	  if (n_hex_digits >= 4)
+	    return 0;
+
+	  n_colon = 0;
+	  n_hex_digits++;
+	}
+
+      /* Save position of :: */
+      if (n_colon == 2)
+	{
+	  /* More than one :: ? */
+	  if (double_colon_index < ARRAY_LEN (hex_quads))
+	    return 0;
+	  double_colon_index = n_hex_quads;
+	}
+
+      if (n_colon > 0 && n_hex_digits > 0)
+	{
+	  hex_quads[n_hex_quads++] = hex_quad;
+	  hex_quad = 0;
+	  n_hex_digits = 0;
+	}
+    }
+
+  if (n_hex_digits > 0)
+    hex_quads[n_hex_quads++] = hex_quad;
+
+  {
+    word i;
+
+    /* Expand :: to appropriate number of zero hex quads. */
+    if (double_colon_index < ARRAY_LEN (hex_quads))
+      {
+	word n_zero = ARRAY_LEN (hex_quads) - n_hex_quads;
+
+	for (i = n_hex_quads - 1; i >= (signed) double_colon_index; i--)
+	  hex_quads[n_zero + i] = hex_quads[i];
+
+	for (i = 0; i < n_zero; i++)
+	  hex_quads[double_colon_index + i] = 0;
+
+	n_hex_quads = ARRAY_LEN (hex_quads);
+      }
+
+    /* Too few hex quads given. */
+    if (n_hex_quads < ARRAY_LEN (hex_quads))
+      return 0;
+
+    for (i = 0; i < ARRAY_LEN (hex_quads); i++)
+      result->as_u16[i] = clib_host_to_net_u16 (hex_quads[i]);
+
+    return 1;
+  }
+}
+
+uword
+unformat_uri (unformat_input_t * input, va_list * args)
+{
+  session_endpoint_extended_t *sep = va_arg (*args,
+					     session_endpoint_extended_t *);
+  u32 port;
+  char *tmp;
+
+  if (unformat (input, "%s://%U/%d", &tmp, unformat_ip4_address, &sep->ip.ip4,
+		&port))
+    {
+      sep->port = clib_host_to_net_u16 (port);
+      sep->is_ip4 = 1;
+      return 1;
+    }
+  else if (unformat (input, "%s://%U/%d", &tmp, unformat_ip6_address,
+		     &sep->ip.ip6, &port))
+    {
+      sep->port = clib_host_to_net_u16 (port);
+      sep->is_ip4 = 0;
+      return 1;
+    }
+  return 0;
+}
+
 void
-application_send_attach (uri_udp_test_main_t * utm)
+application_send_attach (udp_echo_main_t * utm)
 {
   vl_api_application_attach_t *bmp;
-  u32 fifo_size = 1 << 20;
   bmp = vl_msg_api_alloc (sizeof (*bmp));
   memset (bmp, 0, sizeof (*bmp));
 
@@ -179,8 +311,8 @@
   bmp->options[APP_OPTIONS_FLAGS] |= APP_OPTIONS_FLAGS_USE_GLOBAL_SCOPE;
   bmp->options[APP_OPTIONS_FLAGS] |= APP_OPTIONS_FLAGS_USE_LOCAL_SCOPE;
   bmp->options[APP_OPTIONS_PREALLOC_FIFO_PAIRS] = 2;
-  bmp->options[APP_OPTIONS_RX_FIFO_SIZE] = fifo_size;
-  bmp->options[APP_OPTIONS_TX_FIFO_SIZE] = fifo_size;
+  bmp->options[APP_OPTIONS_RX_FIFO_SIZE] = utm->fifo_size;
+  bmp->options[APP_OPTIONS_TX_FIFO_SIZE] = utm->fifo_size;
   bmp->options[APP_OPTIONS_ADD_SEGMENT_SIZE] = 128 << 20;
   bmp->options[APP_OPTIONS_SEGMENT_SIZE] = 256 << 20;
   bmp->options[APP_OPTIONS_EVT_QUEUE_SIZE] = 16768;
@@ -188,7 +320,7 @@
 }
 
 void
-application_detach (uri_udp_test_main_t * utm)
+application_detach (udp_echo_main_t * utm)
 {
   vl_api_application_detach_t *bmp;
   bmp = vl_msg_api_alloc (sizeof (*bmp));
@@ -204,7 +336,7 @@
 vl_api_application_attach_reply_t_handler (vl_api_application_attach_reply_t *
 					   mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   svm_fifo_segment_create_args_t _a = { 0 }, *a = &_a;
   int rv;
 
@@ -250,7 +382,7 @@
 u8 *
 format_api_error (u8 * s, va_list * args)
 {
-  uri_udp_test_main_t *utm = va_arg (*args, uri_udp_test_main_t *);
+  udp_echo_main_t *utm = va_arg (*args, udp_echo_main_t *);
   i32 error = va_arg (*args, u32);
   uword *p;
 
@@ -264,7 +396,7 @@
 }
 
 int
-wait_for_state_change (uri_udp_test_main_t * utm, connection_state_t state)
+wait_for_state_change (udp_echo_main_t * utm, connection_state_t state)
 {
 #if CLIB_DEBUG > 0
 #define TIMEOUT 600.0
@@ -287,11 +419,11 @@
 static void *
 cut_through_thread_fn (void *arg)
 {
-  session_t *s;
+  app_session_t *s;
   svm_fifo_t *rx_fifo;
   svm_fifo_t *tx_fifo;
   u8 *my_copy_buffer = 0;
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   i32 actual_transfer;
   int rv;
   u32 buffer_offset;
@@ -301,8 +433,8 @@
 
   s = pool_elt_at_index (utm->sessions, utm->cut_through_session_index);
 
-  rx_fifo = s->server_rx_fifo;
-  tx_fifo = s->server_tx_fifo;
+  rx_fifo = s->rx_fifo;
+  tx_fifo = s->tx_fifo;
 
   vec_validate (my_copy_buffer, 64 * 1024 - 1);
 
@@ -340,7 +472,7 @@
 }
 
 static void
-udp_client_connect (uri_udp_test_main_t * utm)
+udp_client_connect (udp_echo_main_t * utm)
 {
   vl_api_connect_uri_t *cmp;
   cmp = vl_msg_api_alloc (sizeof (*cmp));
@@ -354,7 +486,7 @@
 }
 
 static void
-client_send_cut_through (uri_udp_test_main_t * utm, session_t * session)
+client_send_cut_through (udp_echo_main_t * utm, app_session_t * session)
 {
   int i;
   u8 *test_data = 0;
@@ -372,8 +504,8 @@
   for (i = 0; i < vec_len (test_data); i++)
     test_data[i] = i & 0xff;
 
-  rx_fifo = session->server_rx_fifo;
-  tx_fifo = session->server_tx_fifo;
+  rx_fifo = session->rx_fifo;
+  tx_fifo = session->tx_fifo;
 
   before = clib_time_now (&utm->clib_time);
 
@@ -456,15 +588,12 @@
 }
 
 static void
-send_test_chunk (uri_udp_test_main_t * utm, svm_fifo_t * tx_fifo, int mypid,
-		 u32 bytes)
+send_test_chunk (udp_echo_main_t * utm, app_session_t * s, u32 bytes)
 {
   u8 *test_data = utm->connect_test_data;
-  u64 bytes_sent = 0;
   int test_buf_offset = 0;
+  u64 bytes_sent = 0;
   u32 bytes_to_snd;
-  u32 queue_max_chunk = 128 << 10, actual_write;
-  session_fifo_event_t evt;
   int rv;
 
   bytes_to_snd = (bytes == 0) ? vec_len (test_data) : bytes;
@@ -473,62 +602,27 @@
 
   while (bytes_to_snd > 0 && !utm->time_to_stop)
     {
-      actual_write = (bytes_to_snd > queue_max_chunk) ?
-	queue_max_chunk : bytes_to_snd;
-      rv = svm_fifo_enqueue_nowait (tx_fifo, actual_write,
-				    test_data + test_buf_offset);
-
+      rv = app_send (s, test_data + test_buf_offset, bytes_to_snd, 0);
       if (rv > 0)
 	{
 	  bytes_to_snd -= rv;
 	  test_buf_offset += rv;
 	  bytes_sent += rv;
-
-	  if (svm_fifo_set_event (tx_fifo))
-	    {
-	      /* Fabricate TX event, send to vpp */
-	      evt.fifo = tx_fifo;
-	      evt.event_type = FIFO_EVENT_APP_TX;
-
-	      svm_queue_add (utm->vpp_event_queue,
-			     (u8 *) & evt, 0 /* do wait for mutex */ );
-	    }
 	}
     }
 }
 
 static void
-recv_test_chunk (uri_udp_test_main_t * utm, session_t * session)
+recv_test_chunk (udp_echo_main_t * utm, app_session_t * s)
 {
-  svm_fifo_t *rx_fifo;
-  int buffer_offset, bytes_to_read = 0, rv;
-
-  rx_fifo = session->server_rx_fifo;
-  bytes_to_read = svm_fifo_max_dequeue (rx_fifo);
-  bytes_to_read =
-    vec_len (utm->rx_buf) > bytes_to_read ?
-    bytes_to_read : vec_len (utm->rx_buf);
-
-  buffer_offset = 0;
-  while (bytes_to_read > 0)
-    {
-      rv = svm_fifo_dequeue_nowait (rx_fifo, bytes_to_read,
-				    utm->rx_buf + buffer_offset);
-      if (rv > 0)
-	{
-	  bytes_to_read -= rv;
-	  buffer_offset += rv;
-	}
-    }
+  app_recv (s, utm->rx_buf, vec_len (utm->rx_buf));
 }
 
 void
-client_send_data (uri_udp_test_main_t * utm)
+client_send_data (udp_echo_main_t * utm)
 {
   u8 *test_data;
-  int mypid = getpid ();
-  session_t *session;
-  svm_fifo_t *tx_fifo;
+  app_session_t *session;
   u32 n_iterations;
   int i;
 
@@ -538,8 +632,6 @@
 
   test_data = utm->connect_test_data;
   session = pool_elt_at_index (utm->sessions, utm->connected_session);
-  tx_fifo = session->server_tx_fifo;
-
   ASSERT (vec_len (test_data) > 0);
 
   vec_validate (utm->rx_buf, vec_len (test_data) - 1);
@@ -547,7 +639,7 @@
 
   for (i = 0; i < n_iterations; i++)
     {
-      send_test_chunk (utm, tx_fifo, mypid, 0);
+      send_test_chunk (utm, session, 0);
       recv_test_chunk (utm, session);
       if (utm->time_to_stop)
 	break;
@@ -562,9 +654,9 @@
 }
 
 static void
-client_test (uri_udp_test_main_t * utm)
+client_test (udp_echo_main_t * utm)
 {
-  session_t *session;
+  app_session_t *session;
 
   application_send_attach (utm);
   udp_client_connect (utm);
@@ -593,7 +685,10 @@
 static void
 vl_api_bind_uri_reply_t_handler (vl_api_bind_uri_reply_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
+  svm_fifo_t *rx_fifo, *tx_fifo;
+  app_session_t *session;
+  u32 session_index;
 
   if (mp->retval)
     {
@@ -602,13 +697,30 @@
       return;
     }
 
-  utm->state = STATE_BOUND;
+  rx_fifo = uword_to_pointer (mp->rx_fifo, svm_fifo_t *);
+  tx_fifo = uword_to_pointer (mp->tx_fifo, svm_fifo_t *);
+
+  pool_get (utm->sessions, session);
+  memset (session, 0, sizeof (*session));
+  session_index = session - utm->sessions;
+
+  rx_fifo->client_session_index = session_index;
+  tx_fifo->client_session_index = session_index;
+  session->rx_fifo = rx_fifo;
+  session->tx_fifo = tx_fifo;
+  clib_memcpy (&session->transport.lcl_ip, mp->lcl_ip,
+	       sizeof (ip46_address_t));
+  session->transport.is_ip4 = mp->lcl_is_ip4;
+  session->transport.lcl_port = mp->lcl_port;
+  session->vpp_evt_q = uword_to_pointer (mp->vpp_evt_q, svm_queue_t *);
+
+  utm->state = utm->is_connected ? STATE_BOUND : STATE_READY;
 }
 
 static void
 vl_api_map_another_segment_t_handler (vl_api_map_another_segment_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   svm_fifo_segment_create_args_t _a, *a = &_a;
   svm_fifo_segment_private_t *seg;
   u8 *seg_name;
@@ -636,7 +748,7 @@
 static void
 vl_api_unmap_segment_t_handler (vl_api_unmap_segment_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   svm_fifo_segment_private_t *seg;
   u64 *seg_indexp;
   u8 *seg_name;
@@ -663,13 +775,13 @@
 vl_api_connect_uri_t_handler (vl_api_connect_uri_t * mp)
 {
   u32 segment_index;
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   svm_fifo_segment_main_t *sm = &svm_fifo_segment_main;
   svm_fifo_segment_create_args_t _a, *a = &_a;
   svm_fifo_segment_private_t *seg;
   svm_queue_t *client_q;
   vl_api_connect_session_reply_t *rmp;
-  session_t *session = 0;
+  app_session_t *session = 0;
   int rv = 0;
 
   /* Create the segment */
@@ -692,16 +804,16 @@
 
   pool_get (utm->sessions, session);
 
-  session->server_rx_fifo = svm_fifo_segment_alloc_fifo
+  session->rx_fifo = svm_fifo_segment_alloc_fifo
     (utm->seg, 128 * 1024, FIFO_SEGMENT_RX_FREELIST);
-  ASSERT (session->server_rx_fifo);
+  ASSERT (session->rx_fifo);
 
-  session->server_tx_fifo = svm_fifo_segment_alloc_fifo
+  session->tx_fifo = svm_fifo_segment_alloc_fifo
     (utm->seg, 128 * 1024, FIFO_SEGMENT_TX_FREELIST);
-  ASSERT (session->server_tx_fifo);
+  ASSERT (session->tx_fifo);
 
-  session->server_rx_fifo->master_session_index = session - utm->sessions;
-  session->server_tx_fifo->master_session_index = session - utm->sessions;
+  session->rx_fifo->master_session_index = session - utm->sessions;
+  session->tx_fifo->master_session_index = session - utm->sessions;
   utm->cut_through_session_index = session - utm->sessions;
 
   rv = pthread_create (&utm->cut_through_thread_handle,
@@ -722,8 +834,8 @@
   rmp->segment_name_length = vec_len (a->segment_name);
   if (session)
     {
-      rmp->server_rx_fifo = pointer_to_uword (session->server_rx_fifo);
-      rmp->server_tx_fifo = pointer_to_uword (session->server_tx_fifo);
+      rmp->server_rx_fifo = pointer_to_uword (session->rx_fifo);
+      rmp->server_tx_fifo = pointer_to_uword (session->tx_fifo);
     }
 
   memcpy (rmp->segment_name, a->segment_name, vec_len (a->segment_name));
@@ -737,7 +849,7 @@
 static void
 vl_api_unbind_uri_reply_t_handler (vl_api_unbind_uri_reply_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
 
   if (mp->retval != 0)
     clib_warning ("returned %d", ntohl (mp->retval));
@@ -748,10 +860,10 @@
 static void
 vl_api_accept_session_t_handler (vl_api_accept_session_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   vl_api_accept_session_reply_t *rmp;
   svm_fifo_t *rx_fifo, *tx_fifo;
-  session_t *session;
+  app_session_t *session;
   static f64 start_time;
   u32 session_index;
   int rv = 0;
@@ -759,9 +871,8 @@
   if (start_time == 0.0)
     start_time = clib_time_now (&utm->clib_time);
 
-  utm->vpp_event_queue =
-    uword_to_pointer (mp->vpp_event_queue_address, svm_queue_t *);
-
+  utm->vpp_event_queue = uword_to_pointer (mp->vpp_event_queue_address,
+					   svm_queue_t *);
   rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *);
   tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *);
 
@@ -778,8 +889,8 @@
       rx_fifo->master_session_index = session_index;
       tx_fifo->master_session_index = session_index;
       utm->cut_through_session_index = session_index;
-      session->server_rx_fifo = rx_fifo;
-      session->server_tx_fifo = tx_fifo;
+      session->rx_fifo = rx_fifo;
+      session->tx_fifo = tx_fifo;
 
       rv = pthread_create (&utm->cut_through_thread_handle,
 			   NULL /*attr */ , cut_through_thread_fn, 0);
@@ -788,13 +899,18 @@
 	  clib_warning ("pthread_create returned %d", rv);
 	  rv = VNET_API_ERROR_SYSCALL_ERROR_1;
 	}
+      utm->do_echo = 1;
     }
   else
     {
       rx_fifo->client_session_index = session_index;
       tx_fifo->client_session_index = session_index;
-      session->server_rx_fifo = rx_fifo;
-      session->server_tx_fifo = tx_fifo;
+      session->rx_fifo = rx_fifo;
+      session->tx_fifo = tx_fifo;
+      clib_memcpy (&session->transport.rmt_ip, mp->ip,
+		   sizeof (ip46_address_t));
+      session->transport.is_ip4 = mp->is_ip4;
+      session->transport.rmt_port = mp->port;
     }
 
   hash_set (utm->session_index_by_vpp_handles, mp->handle, session_index);
@@ -821,8 +937,8 @@
 static void
 vl_api_disconnect_session_t_handler (vl_api_disconnect_session_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
-  session_t *session;
+  udp_echo_main_t *utm = &udp_echo_main;
+  app_session_t *session;
   vl_api_disconnect_session_reply_t *rmp;
   uword *p;
   int rv = 0;
@@ -853,10 +969,12 @@
 static void
 vl_api_connect_session_reply_t_handler (vl_api_connect_session_reply_t * mp)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
-  session_t *session;
+  udp_echo_main_t *utm = &udp_echo_main;
+  unformat_input_t _input, *input = &_input;
+  session_endpoint_extended_t _sep, *sep = &_sep;
+  app_session_t *session;
 
-  ASSERT (utm->i_am_master == 0);
+  ASSERT (utm->i_am_server == 0);
 
   if (mp->retval)
     {
@@ -864,14 +982,13 @@
       return;
     }
 
-  pool_get (utm->sessions, session);
-  session->server_rx_fifo = uword_to_pointer (mp->server_rx_fifo,
-					      svm_fifo_t *);
-  ASSERT (session->server_rx_fifo);
-  session->server_tx_fifo = uword_to_pointer (mp->server_tx_fifo,
-					      svm_fifo_t *);
-  ASSERT (session->server_tx_fifo);
+  ASSERT (mp->server_rx_fifo && mp->server_tx_fifo);
 
+  pool_get (utm->sessions, session);
+  session->rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *);
+  session->tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *);
+  session->vpp_evt_q = uword_to_pointer (mp->vpp_event_queue_address,
+					 svm_queue_t *);
   /* Cut-through case */
   if (mp->client_event_queue_address)
     {
@@ -881,12 +998,32 @@
 					       svm_queue_t *);
       utm->our_event_queue = uword_to_pointer (mp->client_event_queue_address,
 					       svm_queue_t *);
+      utm->do_echo = 1;
     }
   else
     {
       utm->connected_session = session - utm->sessions;
       utm->vpp_event_queue = uword_to_pointer (mp->vpp_event_queue_address,
 					       svm_queue_t *);
+
+      clib_memcpy (&session->transport.lcl_ip, mp->lcl_ip,
+		   sizeof (ip46_address_t));
+      session->transport.is_ip4 = mp->is_ip4;
+      session->transport.lcl_port = mp->lcl_port;
+
+      unformat_init_vector (input, utm->connect_uri);
+      if (!unformat (input, "%U", unformat_uri, sep))
+	{
+	  clib_warning ("can't figure out remote ip and port");
+	  utm->state = STATE_FAILED;
+	  unformat_free (input);
+	  return;
+	}
+      unformat_free (input);
+      clib_memcpy (&session->transport.rmt_ip, &sep->ip,
+		   sizeof (ip46_address_t));
+      session->transport.rmt_port = sep->port;
+      session->is_dgram = !utm->is_connected;
     }
   utm->state = STATE_READY;
 }
@@ -904,7 +1041,7 @@
 _(APPLICATION_DETACH_REPLY, application_detach_reply)	\
 
 void
-tcp_echo_api_hookup (uri_udp_test_main_t * utm)
+tcp_echo_api_hookup (udp_echo_main_t * utm)
 {
 #define _(N,n)                                                  \
     vl_msg_api_set_handlers(VL_API_##N, #n,                     \
@@ -921,7 +1058,7 @@
 int
 connect_to_vpp (char *name)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
+  udp_echo_main_t *utm = &udp_echo_main;
   api_main_t *am = &api_main;
 
   if (vl_client_connect_to_vlib ("/vpe-api", name, 32) < 0)
@@ -940,7 +1077,7 @@
 }
 
 static void
-init_error_string_table (uri_udp_test_main_t * utm)
+init_error_string_table (udp_echo_main_t * utm)
 {
   utm->error_string_by_error_number = hash_create (0, sizeof (uword));
 
@@ -952,43 +1089,26 @@
 }
 
 void
-server_handle_fifo_event_rx (uri_udp_test_main_t * utm,
-			     session_fifo_event_t * e)
+server_handle_fifo_event_rx (udp_echo_main_t * utm, session_fifo_event_t * e)
 {
-  svm_fifo_t *rx_fifo, *tx_fifo;
-  int nbytes;
-  session_fifo_event_t evt;
-  svm_queue_t *q;
+  app_session_t *s;
   int rv;
 
-  rx_fifo = e->fifo;
-  tx_fifo = utm->sessions[rx_fifo->client_session_index].server_tx_fifo;
-  svm_fifo_unset_event (rx_fifo);
+  s = pool_elt_at_index (utm->sessions, e->fifo->client_session_index);
+  app_recv (s, utm->rx_buf, vec_len (utm->rx_buf));
 
-  do
+  if (utm->do_echo)
     {
-      nbytes = svm_fifo_dequeue_nowait (rx_fifo, vec_len (utm->rx_buf),
-					utm->rx_buf);
-    }
-  while (nbytes <= 0);
-  do
-    {
-      rv = svm_fifo_enqueue_nowait (tx_fifo, nbytes, utm->rx_buf);
-    }
-  while (rv == -2);
-
-  if (svm_fifo_set_event (tx_fifo))
-    {
-      /* Fabricate TX event, send to vpp */
-      evt.fifo = tx_fifo;
-      evt.event_type = FIFO_EVENT_APP_TX;
-      q = utm->vpp_event_queue;
-      svm_queue_add (q, (u8 *) & evt, 0 /* do wait for mutex */ );
+      do
+	{
+	  rv = app_send_stream (s, utm->rx_buf, vec_len (utm->rx_buf), 0);
+	}
+      while (rv == SVM_FIFO_FULL);
     }
 }
 
 void
-server_handle_event_queue (uri_udp_test_main_t * utm)
+server_handle_event_queue (udp_echo_main_t * utm)
 {
   session_fifo_event_t _e, *e = &_e;
 
@@ -1022,7 +1142,7 @@
 }
 
 static void
-server_unbind (uri_udp_test_main_t * utm)
+server_unbind (udp_echo_main_t * utm)
 {
   vl_api_unbind_uri_t *ump;
 
@@ -1031,12 +1151,12 @@
 
   ump->_vl_msg_id = ntohs (VL_API_UNBIND_URI);
   ump->client_index = utm->my_client_index;
-  memcpy (ump->uri, utm->uri, vec_len (utm->uri));
+  memcpy (ump->uri, utm->listen_uri, vec_len (utm->listen_uri));
   vl_msg_api_send_shmem (utm->vl_input_queue, (u8 *) & ump);
 }
 
 static void
-server_bind (uri_udp_test_main_t * utm)
+server_bind (udp_echo_main_t * utm)
 {
   vl_api_bind_uri_t *bmp;
 
@@ -1046,22 +1166,22 @@
   bmp->_vl_msg_id = ntohs (VL_API_BIND_URI);
   bmp->client_index = utm->my_client_index;
   bmp->context = ntohl (0xfeedface);
-  memcpy (bmp->uri, utm->uri, vec_len (utm->uri));
+  memcpy (bmp->uri, utm->listen_uri, vec_len (utm->listen_uri));
   vl_msg_api_send_shmem (utm->vl_input_queue, (u8 *) & bmp);
 }
 
 void
-udp_server_test (uri_udp_test_main_t * utm)
+udp_server_test (udp_echo_main_t * utm)
 {
-
+  u8 wait_for_state = utm->is_connected ? STATE_BOUND : STATE_READY;
   application_send_attach (utm);
 
   /* Bind to uri */
   server_bind (utm);
 
-  if (wait_for_state_change (utm, STATE_BOUND))
+  if (wait_for_state_change (utm, wait_for_state))
     {
-      clib_warning ("timeout waiting for STATE_BOUND");
+      clib_warning ("timeout waiting for state change");
       return;
     }
 
@@ -1084,11 +1204,11 @@
 int
 main (int argc, char **argv)
 {
-  uri_udp_test_main_t *utm = &uri_udp_test_main;
-  u8 *bind_name = (u8 *) "udp://0.0.0.0/1234";
+  udp_echo_main_t *utm = &udp_echo_main;
+  u8 *uri = (u8 *) "udp://0.0.0.0/1234";
   unformat_input_t _argv, *a = &_argv;
-  int i_am_master = 1;
-  session_t *session;
+  int i_am_server = 1;
+  app_session_t *session;
   u8 *chroot_prefix;
   char *app_name;
   mheap_t *h;
@@ -1116,39 +1236,50 @@
   svm_fifo_segment_main_init (0x200000000ULL, 20);
   unformat_init_command_line (a, argv);
 
+  utm->fifo_size = 128 << 10;
+
   while (unformat_check_input (a) != UNFORMAT_END_OF_INPUT)
     {
       if (unformat (a, "chroot prefix %s", &chroot_prefix))
 	{
 	  vl_set_memory_root_path ((char *) chroot_prefix);
 	}
-      else if (unformat (a, "uri %s", &bind_name))
+      else if (unformat (a, "uri %s", &uri))
 	;
       else if (unformat (a, "segment-size %dM", &tmp))
 	utm->configured_segment_size = tmp << 20;
       else if (unformat (a, "segment-size %dG", &tmp))
 	utm->configured_segment_size = tmp << 30;
-      else if (unformat (a, "master"))
-	i_am_master = 1;
-      else if (unformat (a, "slave"))
-	i_am_master = 0;
+      else if (unformat (a, "server"))
+	i_am_server = 1;
+      else if (unformat (a, "client"))
+	i_am_server = 0;
       else
 	{
-	  fformat (stderr, "%s: usage [master|slave]\n");
+	  fformat (stderr, "%s: usage [server|client]\n");
 	  exit (1);
 	}
     }
 
   utm->cut_through_session_index = ~0;
-  utm->uri = format (0, "%s%c", bind_name, 0);
-  utm->i_am_master = i_am_master;
+  utm->i_am_server = i_am_server;
   utm->segment_main = &svm_fifo_segment_main;
-  utm->connect_uri = format (0, "udp://6.0.1.2/1234%c", 0);
 
   setup_signal_handlers ();
   tcp_echo_api_hookup (utm);
 
-  app_name = i_am_master ? "udp_echo_master" : "udp_echo_slave";
+  if (i_am_server)
+    {
+      utm->listen_uri = format (0, "%s%c", uri, 0);
+      utm->is_connected = (utm->listen_uri[4] == 'c');
+      app_name = "udp_echo_server";
+    }
+  else
+    {
+      app_name = "udp_echo_client";
+      utm->connect_uri = format (0, "%s%c", uri, 0);
+      utm->is_connected = (utm->connect_uri[4] == 'c');
+    }
   if (connect_to_vpp (app_name) < 0)
     {
       svm_region_exit ();
@@ -1156,7 +1287,7 @@
       exit (1);
     }
 
-  if (i_am_master == 0)
+  if (i_am_server == 0)
     {
       client_test (utm);
       exit (0);
diff --git a/src/vnet/session-apps/echo_client.c b/src/vnet/session-apps/echo_client.c
index 2cfb471..11528ef 100644
--- a/src/vnet/session-apps/echo_client.c
+++ b/src/vnet/session-apps/echo_client.c
@@ -46,23 +46,21 @@
 send_data_chunk (echo_client_main_t * ecm, session_t * s)
 {
   u8 *test_data = ecm->connect_test_data;
-  int test_buf_offset;
+  int test_buf_len, test_buf_offset, rv;
   u32 bytes_this_chunk;
-  session_fifo_event_t evt;
-  svm_fifo_t *txf;
-  int rv;
 
   ASSERT (vec_len (test_data) > 0);
+  test_buf_len = vec_len (test_data);
+  test_buf_offset = s->bytes_sent % test_buf_len;
+  bytes_this_chunk = clib_min (test_buf_len - test_buf_offset,
+			       s->bytes_to_send);
 
-  test_buf_offset = s->bytes_sent % vec_len (test_data);
-  bytes_this_chunk = vec_len (test_data) - test_buf_offset;
-
-  bytes_this_chunk = bytes_this_chunk < s->bytes_to_send
-    ? bytes_this_chunk : s->bytes_to_send;
-
-  txf = s->server_tx_fifo;
-  rv = svm_fifo_enqueue_nowait (txf, bytes_this_chunk,
-				test_data + test_buf_offset);
+  if (!ecm->is_dgram)
+    rv = app_send_stream (&s->data, test_data + test_buf_offset,
+			  bytes_this_chunk, 0);
+  else
+    rv = app_send_dgram (&s->data, test_data + test_buf_offset,
+			 bytes_this_chunk, 0);
 
   /* If we managed to enqueue data... */
   if (rv > 0)
@@ -89,35 +87,24 @@
 	  ed->data[1] = s->bytes_sent;
 	  ed->data[2] = s->bytes_to_send;
 	}
-
-      /* Poke the session layer */
-      if (svm_fifo_set_event (txf))
-	{
-	  /* Fabricate TX event, send to vpp */
-	  evt.fifo = txf;
-	  evt.event_type = FIFO_EVENT_APP_TX;
-
-	  if (svm_queue_add
-	      (ecm->vpp_event_queue[txf->master_thread_index], (u8 *) & evt,
-	       0 /* do wait for mutex */ ))
-	    clib_warning ("could not enqueue event");
-	}
     }
 }
 
 static void
 receive_data_chunk (echo_client_main_t * ecm, session_t * s)
 {
-  svm_fifo_t *rx_fifo = s->server_rx_fifo;
-  u32 my_thread_index = vlib_get_thread_index ();
+  svm_fifo_t *rx_fifo = s->data.rx_fifo;
+  u32 thread_index = vlib_get_thread_index ();
   int n_read, i;
 
   if (ecm->test_bytes)
     {
-      n_read = svm_fifo_dequeue_nowait (rx_fifo,
-					vec_len (ecm->rx_buf
-						 [my_thread_index]),
-					ecm->rx_buf[my_thread_index]);
+      if (!ecm->is_dgram)
+	n_read = app_recv_stream (&s->data, ecm->rx_buf[thread_index],
+				  vec_len (ecm->rx_buf[thread_index]));
+      else
+	n_read = app_recv_dgram (&s->data, ecm->rx_buf[thread_index],
+				 vec_len (ecm->rx_buf[thread_index]));
     }
   else
     {
@@ -148,17 +135,18 @@
 	{
 	  for (i = 0; i < n_read; i++)
 	    {
-	      if (ecm->rx_buf[my_thread_index][i]
+	      if (ecm->rx_buf[thread_index][i]
 		  != ((s->bytes_received + i) & 0xff))
 		{
 		  clib_warning ("read %d error at byte %lld, 0x%x not 0x%x",
 				n_read, s->bytes_received + i,
-				ecm->rx_buf[my_thread_index][i],
+				ecm->rx_buf[thread_index][i],
 				((s->bytes_received + i) & 0xff));
 		  ecm->test_failed = 1;
 		}
 	    }
 	}
+      ASSERT (n_read <= s->bytes_to_receive);
       s->bytes_to_receive -= n_read;
       s->bytes_received += n_read;
     }
@@ -230,20 +218,15 @@
 	}
       if (sp->bytes_to_receive > 0)
 	{
-	  receive_data_chunk (ecm, sp);
 	  delete_session = 0;
 	}
       if (PREDICT_FALSE (delete_session == 1))
 	{
-	  u32 index, thread_index;
 	  stream_session_t *s;
 
 	  __sync_fetch_and_add (&ecm->tx_total, sp->bytes_sent);
 	  __sync_fetch_and_add (&ecm->rx_total, sp->bytes_received);
-
-	  session_parse_handle (sp->vpp_session_handle,
-				&index, &thread_index);
-	  s = session_get_if_valid (index, thread_index);
+	  s = session_get_from_handle_if_valid (sp->vpp_session_handle);
 
 	  if (s)
 	    {
@@ -312,8 +295,8 @@
 
   num_threads = 1 /* main thread */  + vtm->n_threads;
 
-  /* Init test data. Bigecmuffer */
-  vec_validate (ecm->connect_test_data, 1024 * 1024 - 1);
+  /* Init test data. Big buffer */
+  vec_validate (ecm->connect_test_data, 4 * 1024 * 1024 - 1);
   for (i = 0; i < vec_len (ecm->connect_test_data); i++)
     ecm->connect_test_data[i] = i & 0xff;
 
@@ -363,12 +346,22 @@
   session_index = session - ecm->sessions;
   session->bytes_to_send = ecm->bytes_to_send;
   session->bytes_to_receive = ecm->no_return ? 0ULL : ecm->bytes_to_send;
-  session->server_rx_fifo = s->server_rx_fifo;
-  session->server_rx_fifo->client_session_index = session_index;
-  session->server_tx_fifo = s->server_tx_fifo;
-  session->server_tx_fifo->client_session_index = session_index;
+  session->data.rx_fifo = s->server_rx_fifo;
+  session->data.rx_fifo->client_session_index = session_index;
+  session->data.tx_fifo = s->server_tx_fifo;
+  session->data.tx_fifo->client_session_index = session_index;
+  session->data.vpp_evt_q = ecm->vpp_event_queue[thread_index];
   session->vpp_session_handle = session_handle (s);
 
+  if (ecm->is_dgram)
+    {
+      transport_connection_t *tc;
+      tc = session_get_transport (s);
+      clib_memcpy (&session->data.transport, tc,
+		   sizeof (session->data.transport));
+      session->data.is_dgram = 1;
+    }
+
   vec_add1 (ecm->connection_index_by_thread[thread_index], session_index);
   __sync_fetch_and_add (&ecm->ready_connections, 1);
   if (ecm->ready_connections == ecm->expected_connections)
@@ -410,6 +403,28 @@
 static int
 echo_clients_rx_callback (stream_session_t * s)
 {
+  echo_client_main_t *ecm = &echo_client_main;
+  session_t *sp;
+
+  sp = pool_elt_at_index (ecm->sessions,
+			  s->server_rx_fifo->client_session_index);
+  receive_data_chunk (ecm, sp);
+
+  if (svm_fifo_max_dequeue (s->server_rx_fifo))
+    {
+      session_fifo_event_t evt;
+      svm_queue_t *q;
+      if (svm_fifo_set_event (s->server_rx_fifo))
+	{
+	  evt.fifo = s->server_rx_fifo;
+	  evt.event_type = FIFO_EVENT_BUILTIN_RX;
+	  q = session_manager_get_vpp_event_queue (s->thread_index);
+	  if (PREDICT_FALSE (q->cursize == q->maxsize))
+	    clib_warning ("out of event queue space");
+	  else if (svm_queue_add (q, (u8 *) & evt, 0))
+	    clib_warning ("failed to enqueue self-tap");
+	}
+    }
   return 0;
 }
 
@@ -544,7 +559,7 @@
 }
 
 #define ec_cli_output(_fmt, _args...) 			\
-  if (!ecm->no_output)  					\
+  if (!ecm->no_output)  				\
     vlib_cli_output(vm, _fmt, ##_args)
 
 static clib_error_t *
@@ -663,6 +678,9 @@
       ecm->connect_uri = format (0, "%s%c", default_uri, 0);
     }
 
+  if (ecm->connect_uri[0] == 'u' && ecm->connect_uri[3] != 'c')
+    ecm->is_dgram = 1;
+
 #if ECHO_CLIENT_PTHREAD
   echo_clients_start_tx_pthread ();
 #endif
diff --git a/src/vnet/session-apps/echo_client.h b/src/vnet/session-apps/echo_client.h
index 5712da5..344e438 100644
--- a/src/vnet/session-apps/echo_client.h
+++ b/src/vnet/session-apps/echo_client.h
@@ -30,15 +30,13 @@
 
 typedef struct
 {
+  app_session_t data;
   u64 bytes_to_send;
   u64 bytes_sent;
   u64 bytes_to_receive;
   u64 bytes_received;
-
-  svm_fifo_t *server_rx_fifo;
-  svm_fifo_t *server_tx_fifo;
-
   u64 vpp_session_handle;
+  u8 thread_index;
 } session_t;
 
 typedef struct
@@ -46,7 +44,7 @@
   /*
    * Application setup parameters
    */
-  svm_queue_t *vl_input_queue;	/**< vpe input queue */
+  svm_queue_t *vl_input_queue;		/**< vpe input queue */
   svm_queue_t **vpp_event_queue;
 
   u32 cli_node_index;			/**< cli process node index */
@@ -65,6 +63,7 @@
   u32 private_segment_count;		/**< Number of private fifo segs */
   u32 private_segment_size;		/**< size of private fifo segs */
   u32 tls_engine;			/**< TLS engine mbedtls/openssl */
+  u8 is_dgram;
   /*
    * Test state variables
    */
diff --git a/src/vnet/session-apps/echo_server.c b/src/vnet/session-apps/echo_server.c
index 0bde2e4..85e6c29 100644
--- a/src/vnet/session-apps/echo_server.c
+++ b/src/vnet/session-apps/echo_server.c
@@ -34,13 +34,14 @@
    * Config params
    */
   u8 no_echo;			/**< Don't echo traffic */
-  u32 fifo_size;			/**< Fifo size */
+  u32 fifo_size;		/**< Fifo size */
   u32 rcv_buffer_size;		/**< Rcv buffer size */
   u32 prealloc_fifos;		/**< Preallocate fifos */
   u32 private_segment_count;	/**< Number of private segments  */
   u32 private_segment_size;	/**< Size of private segments  */
   char *server_uri;		/**< Server URI */
   u32 tls_engine;		/**< TLS engine: mbedtls/openssl */
+  u8 is_dgram;			/**< set if transport is dgram */
   /*
    * Test state
    */
@@ -126,25 +127,13 @@
 }
 
 /*
- * If no-echo, just read the data and be done with it
+ * If no-echo, just drop the data and be done with it.
  */
 int
 echo_server_builtin_server_rx_callback_no_echo (stream_session_t * s)
 {
-  echo_server_main_t *esm = &echo_server_main;
-  u32 my_thread_id = vlib_get_thread_index ();
-  int actual_transfer;
-  svm_fifo_t *rx_fifo;
-
-  rx_fifo = s->server_rx_fifo;
-
-  do
-    {
-      actual_transfer =
-	svm_fifo_dequeue_nowait (rx_fifo, esm->rcv_buffer_size,
-				 esm->rx_buf[my_thread_id]);
-    }
-  while (actual_transfer > 0);
+  svm_fifo_t *rx_fifo = s->server_rx_fifo;
+  svm_fifo_dequeue_drop (rx_fifo, svm_fifo_max_dequeue (rx_fifo));
   return 0;
 }
 
@@ -157,6 +146,8 @@
   echo_server_main_t *esm = &echo_server_main;
   session_fifo_event_t evt;
   u32 thread_index = vlib_get_thread_index ();
+  app_session_transport_t at;
+  svm_queue_t *q;
 
   ASSERT (s->thread_index == thread_index);
 
@@ -166,14 +157,29 @@
   ASSERT (rx_fifo->master_thread_index == thread_index);
   ASSERT (tx_fifo->master_thread_index == thread_index);
 
-  max_dequeue = svm_fifo_max_dequeue (s->server_rx_fifo);
-  max_enqueue = svm_fifo_max_enqueue (s->server_tx_fifo);
+  max_enqueue = svm_fifo_max_enqueue (tx_fifo);
+  if (!esm->is_dgram)
+    {
+      max_dequeue = svm_fifo_max_dequeue (rx_fifo);
+    }
+  else
+    {
+      session_dgram_pre_hdr_t ph;
+      svm_fifo_peek (rx_fifo, 0, sizeof (ph), (u8 *) & ph);
+      max_dequeue = ph.data_length - ph.data_offset;
+      if (!esm->vpp_queue[s->thread_index])
+	{
+	  q = session_manager_get_vpp_event_queue (s->thread_index);
+	  esm->vpp_queue[s->thread_index] = q;
+	}
+      max_enqueue -= sizeof (session_dgram_hdr_t);
+    }
 
   if (PREDICT_FALSE (max_dequeue == 0))
     return 0;
 
   /* Number of bytes we're going to copy */
-  max_transfer = (max_dequeue < max_enqueue) ? max_dequeue : max_enqueue;
+  max_transfer = clib_min (max_dequeue, max_enqueue);
 
   /* No space in tx fifo */
   if (PREDICT_FALSE (max_transfer == 0))
@@ -184,16 +190,16 @@
       /* Program self-tap to retry */
       if (svm_fifo_set_event (rx_fifo))
 	{
-	  svm_queue_t *q;
 	  evt.fifo = rx_fifo;
 	  evt.event_type = FIFO_EVENT_BUILTIN_RX;
 
-	  q = esm->vpp_queue[thread_index];
+	  q = esm->vpp_queue[s->thread_index];
 	  if (PREDICT_FALSE (q->cursize == q->maxsize))
 	    clib_warning ("out of event queue space");
 	  else if (svm_queue_add (q, (u8 *) & evt, 0))
 	    clib_warning ("failed to enqueue self-tap");
 
+	  vec_validate (esm->rx_retries[s->thread_index], s->session_index);
 	  if (esm->rx_retries[thread_index][s->session_index] == 500000)
 	    {
 	      clib_warning ("session stuck: %U", format_stream_session, s, 2);
@@ -205,36 +211,47 @@
       return 0;
     }
 
-  _vec_len (esm->rx_buf[thread_index]) = max_transfer;
-
-  actual_transfer = svm_fifo_dequeue_nowait (rx_fifo, max_transfer,
-					     esm->rx_buf[thread_index]);
+  vec_validate (esm->rx_buf[thread_index], max_transfer);
+  if (!esm->is_dgram)
+    {
+      actual_transfer = app_recv_stream_raw (rx_fifo,
+					     esm->rx_buf[thread_index],
+					     max_transfer,
+					     0 /* don't clear event */ );
+    }
+  else
+    {
+      actual_transfer = app_recv_dgram_raw (rx_fifo,
+					    esm->rx_buf[thread_index],
+					    max_transfer, &at,
+					    0 /* don't clear event */ );
+    }
   ASSERT (actual_transfer == max_transfer);
-
-//  test_bytes (esm, actual_transfer);
+  /* test_bytes (esm, actual_transfer); */
 
   /*
    * Echo back
    */
 
-  n_written = svm_fifo_enqueue_nowait (tx_fifo, actual_transfer,
-				       esm->rx_buf[thread_index]);
-
-  if (n_written != max_transfer)
-    clib_warning ("short trout!");
-
-  if (svm_fifo_set_event (tx_fifo))
+  if (!esm->is_dgram)
     {
-      /* Fabricate TX event, send to vpp */
-      evt.fifo = tx_fifo;
-      evt.event_type = FIFO_EVENT_APP_TX;
-
-      if (svm_queue_add (esm->vpp_queue[s->thread_index],
-			 (u8 *) & evt, 0 /* do wait for mutex */ ))
-	clib_warning ("failed to enqueue tx evt");
+      n_written = app_send_stream_raw (tx_fifo,
+				       esm->vpp_queue[thread_index],
+				       esm->rx_buf[thread_index],
+				       actual_transfer, 0);
+    }
+  else
+    {
+      n_written = app_send_dgram_raw (tx_fifo, &at,
+				      esm->vpp_queue[s->thread_index],
+				      esm->rx_buf[thread_index],
+				      actual_transfer, 0);
     }
 
-  if (PREDICT_FALSE (n_written < max_dequeue))
+  if (n_written != max_transfer)
+    clib_warning ("short trout! written %u read %u", n_written, max_transfer);
+
+  if (PREDICT_FALSE (svm_fifo_max_dequeue (rx_fifo)))
     goto rx_event;
 
   return 0;
@@ -411,6 +428,7 @@
   esm->private_segment_count = 0;
   esm->private_segment_size = 0;
   esm->tls_engine = TLS_ENGINE_OPENSSL;
+  esm->is_dgram = 0;
   vec_free (esm->server_uri);
 
   while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
@@ -479,6 +497,8 @@
       clib_warning ("No uri provided! Using default: %s", default_uri);
       esm->server_uri = (char *) format (0, "%s%c", default_uri, 0);
     }
+  if (esm->server_uri[0] == 'u' && esm->server_uri[3] != 'c')
+    esm->is_dgram = 1;
 
   rv = echo_server_create (vm, appns_id, appns_flags, appns_secret);
   vec_free (appns_id);
diff --git a/src/vnet/session/application.c b/src/vnet/session/application.c
index 1c46e78..68bbd59 100644
--- a/src/vnet/session/application.c
+++ b/src/vnet/session/application.c
@@ -445,13 +445,9 @@
   s = listen_session_new (0, sst);
   s->app_index = srv->index;
 
-  if (stream_session_listen (s, sep))
-    goto err;
-
   /* Allocate segment manager. All sessions derived out of a listen session
    * have fifos allocated by the same segment manager. */
-  sm = application_alloc_segment_manager (srv);
-  if (sm == 0)
+  if (!(sm = application_alloc_segment_manager (srv)))
     goto err;
 
   /* Add to app's listener table. Useful to find all child listeners
@@ -459,6 +455,13 @@
   handle = listen_session_get_handle (s);
   hash_set (srv->listeners_table, handle, segment_manager_index (sm));
 
+  if (stream_session_listen (s, sep))
+    {
+      segment_manager_del (sm);
+      hash_unset (srv->listeners_table, handle);
+      goto err;
+    }
+
   *res = handle;
   return 0;
 
diff --git a/src/vnet/session/application_interface.h b/src/vnet/session/application_interface.h
index 5dc237f..5fd2185 100644
--- a/src/vnet/session/application_interface.h
+++ b/src/vnet/session/application_interface.h
@@ -173,6 +173,163 @@
 extern const char test_srv_key_rsa[];
 extern const u32 test_srv_key_rsa_len;
 
+typedef struct app_session_transport_
+{
+  ip46_address_t rmt_ip;	/**< remote ip */
+  ip46_address_t lcl_ip;	/**< local ip */
+  u16 rmt_port;			/**< remote port */
+  u16 lcl_port;			/**< local port */
+  u8 is_ip4;			/**< set if uses ip4 networking */
+} app_session_transport_t;
+
+typedef struct app_session_
+{
+  svm_fifo_t *rx_fifo;			/**< rx fifo */
+  svm_fifo_t *tx_fifo;			/**< tx fifo */
+  session_type_t session_type;		/**< session type */
+  volatile u8 session_state;		/**< session state */
+  u32 session_index;			/**< index in owning pool */
+  app_session_transport_t transport;	/**< transport info */
+  svm_queue_t *vpp_evt_q;		/**< vpp event queue for session */
+  u8 is_dgram;				/**< set if it works in dgram mode */
+} app_session_t;
+
+always_inline int
+app_send_dgram_raw (svm_fifo_t * f, app_session_transport_t * at,
+		    svm_queue_t * vpp_evt_q, u8 * data, u32 len, u8 noblock)
+{
+  u32 max_enqueue, actual_write;
+  session_dgram_hdr_t hdr;
+  session_fifo_event_t evt;
+  int rv;
+
+  max_enqueue = svm_fifo_max_enqueue (f);
+  if (svm_fifo_max_enqueue (f) <= sizeof (session_dgram_hdr_t))
+    return 0;
+
+  max_enqueue -= sizeof (session_dgram_hdr_t);
+  actual_write = clib_min (len, max_enqueue);
+  hdr.data_length = actual_write;
+  hdr.data_offset = 0;
+  clib_memcpy (&hdr.rmt_ip, &at->rmt_ip, sizeof (ip46_address_t));
+  hdr.is_ip4 = at->is_ip4;
+  hdr.rmt_port = at->rmt_port;
+  clib_memcpy (&hdr.lcl_ip, &at->lcl_ip, sizeof (ip46_address_t));
+  hdr.lcl_port = at->lcl_port;
+  rv = svm_fifo_enqueue_nowait (f, sizeof (hdr), (u8 *) & hdr);
+  if (rv <= 0)
+    return 0;
+
+  ASSERT (rv == sizeof (hdr));
+
+  if ((rv = svm_fifo_enqueue_nowait (f, actual_write, data)) > 0)
+    {
+      if (svm_fifo_set_event (f))
+	{
+	  evt.fifo = f;
+	  evt.event_type = FIFO_EVENT_APP_TX;
+	  svm_queue_add (vpp_evt_q, (u8 *) & evt, noblock);
+	}
+    }
+  return rv;
+}
+
+always_inline int
+app_send_dgram (app_session_t * s, u8 * data, u32 len, u8 noblock)
+{
+  return app_send_dgram_raw (s->tx_fifo, &s->transport, s->vpp_evt_q, data,
+			     len, noblock);
+}
+
+always_inline int
+app_send_stream_raw (svm_fifo_t * f, svm_queue_t * vpp_evt_q, u8 * data,
+		     u32 len, u8 noblock)
+{
+  session_fifo_event_t evt;
+  int rv;
+
+  if ((rv = svm_fifo_enqueue_nowait (f, len, data)) > 0)
+    {
+      if (svm_fifo_set_event (f))
+	{
+	  evt.fifo = f;
+	  evt.event_type = FIFO_EVENT_APP_TX;
+	  svm_queue_add (vpp_evt_q, (u8 *) & evt, noblock);
+	}
+    }
+  return rv;
+}
+
+always_inline int
+app_send_stream (app_session_t * s, u8 * data, u32 len, u8 noblock)
+{
+  return app_send_stream_raw (s->tx_fifo, s->vpp_evt_q, data, len, noblock);
+}
+
+always_inline int
+app_send (app_session_t * s, u8 * data, u32 len, u8 noblock)
+{
+  if (s->is_dgram)
+    return app_send_dgram (s, data, len, noblock);
+  return app_send_stream (s, data, len, noblock);
+}
+
+always_inline int
+app_recv_dgram_raw (svm_fifo_t * f, u8 * buf, u32 len,
+		    app_session_transport_t * at, u8 clear_evt)
+{
+  session_dgram_pre_hdr_t ph;
+  u32 max_deq;
+  int rv;
+
+  if (clear_evt)
+    svm_fifo_unset_event (f);
+  max_deq = svm_fifo_max_dequeue (f);
+  if (max_deq < sizeof (session_dgram_hdr_t))
+    return 0;
+
+  svm_fifo_peek (f, 0, sizeof (ph), (u8 *) & ph);
+  ASSERT (ph.data_length >= ph.data_offset);
+  if (!ph.data_offset)
+    svm_fifo_peek (f, sizeof (ph), sizeof (*at), (u8 *) at);
+  len = clib_min (len, ph.data_length - ph.data_offset);
+  rv = svm_fifo_peek (f, ph.data_offset + SESSION_CONN_HDR_LEN, len, buf);
+  ph.data_offset += rv;
+  if (ph.data_offset == ph.data_length)
+    svm_fifo_dequeue_drop (f, ph.data_length + SESSION_CONN_HDR_LEN);
+  else
+    svm_fifo_overwrite_head (f, (u8 *) & ph, sizeof (ph));
+  return rv;
+}
+
+always_inline int
+app_recv_dgram (app_session_t * s, u8 * buf, u32 len)
+{
+  return app_recv_dgram_raw (s->rx_fifo, buf, len, &s->transport, 1);
+}
+
+always_inline int
+app_recv_stream_raw (svm_fifo_t * f, u8 * buf, u32 len, u8 clear_evt)
+{
+  if (clear_evt)
+    svm_fifo_unset_event (f);
+  return svm_fifo_dequeue_nowait (f, len, buf);
+}
+
+always_inline int
+app_recv_stream (app_session_t * s, u8 * buf, u32 len)
+{
+  return app_recv_stream_raw (s->rx_fifo, buf, len, 1);
+}
+
+always_inline int
+app_recv (app_session_t * s, u8 * data, u32 len)
+{
+  if (s->is_dgram)
+    return app_recv_dgram (s, data, len);
+  return app_recv_stream (s, data, len);
+}
+
 #endif /* __included_uri_h__ */
 
 /*
diff --git a/src/vnet/session/session.api b/src/vnet/session/session.api
index bf88e82..98748d8 100644
--- a/src/vnet/session/session.api
+++ b/src/vnet/session/session.api
@@ -13,7 +13,7 @@
  * limitations under the License.
  */
 
-option version = "1.0.2";
+option version = "1.0.3";
 
 /** \brief client->vpp, attach application to session layer
     @param client_index - opaque cookie to identify the sender
@@ -119,13 +119,25 @@
                  "tcp://::/0/80" [ipv6] etc.
     @param options - socket options, fifo sizes, etc.
 */
-autoreply define bind_uri {
+define bind_uri {
   u32 client_index;
   u32 context;
   u32 accept_cookie;
   u8 uri[128];
 };
 
+define bind_uri_reply {
+  u32 context;
+  u64 handle;
+  i32 retval;
+  u64 rx_fifo;
+  u64 tx_fifo;
+  u8 lcl_is_ip4;
+  u8 lcl_ip[16];
+  u16 lcl_port;
+  u64 vpp_evt_q;
+};
+
 /** \brief Unbind a given URI
     @param client_index - opaque cookie to identify the sender
     @param context - sender context, to match reply w/ request
@@ -314,8 +326,12 @@
     @param context - sender context, to match reply w/ request
     @param handle - bind handle
     @param retval - return code for the request
-    @param event_queue_address - vpp event queue address or 0 if this 
-                                 connection shouldn't send events
+    @param lcl_is_ip4 - local ip address type
+    @param lcl_ip6 - local ip address
+    @param lcl_port - local port
+    @param rx_fifo - rx fifo address if allocated (connectionless)
+    @param tx_fifo - tx fifo address if allocated (connectionless) 
+    @param vpp_evt_q - vpp event queue address (connectionless)
     @param segment_name_length - length of segment name 
     @param segment_name - name of segment client needs to attach to
 */
@@ -323,10 +339,12 @@
   u32 context;
   u64 handle;
   i32 retval;
-  u64 server_event_queue_address;
   u8 lcl_is_ip4;
   u8 lcl_ip[16];
   u16 lcl_port;
+  u64 rx_fifo;
+  u64 tx_fifo;
+  u64 vpp_evt_q;
   u32 segment_size;
   u8 segment_name_length;
   u8 segment_name[128];
diff --git a/src/vnet/session/session.c b/src/vnet/session/session.c
index d258b82..dfc967b 100644
--- a/src/vnet/session/session.c
+++ b/src/vnet/session/session.c
@@ -355,14 +355,19 @@
   return enqueued;
 }
 
+
 int
-session_enqueue_dgram_connection (stream_session_t * s, vlib_buffer_t * b,
-				  u8 proto, u8 queue_event)
+session_enqueue_dgram_connection (stream_session_t * s,
+				  session_dgram_hdr_t * hdr,
+				  vlib_buffer_t * b, u8 proto, u8 queue_event)
 {
   int enqueued = 0, rv, in_order_off;
 
-  if (svm_fifo_max_enqueue (s->server_rx_fifo) < b->current_length)
-    return -1;
+  ASSERT (svm_fifo_max_enqueue (s->server_rx_fifo)
+	  >= b->current_length + sizeof (*hdr));
+
+  svm_fifo_enqueue_nowait (s->server_rx_fifo, sizeof (session_dgram_hdr_t),
+			   (u8 *) hdr);
   enqueued = svm_fifo_enqueue_nowait (s->server_rx_fifo, b->current_length,
 				      vlib_buffer_get_current (b));
   if (PREDICT_FALSE ((b->flags & VLIB_BUFFER_NEXT_PRESENT) && enqueued >= 0))
@@ -530,6 +535,16 @@
   return errors;
 }
 
+int
+session_manager_flush_all_enqueue_events (u8 transport_proto)
+{
+  vlib_thread_main_t *vtm = vlib_get_thread_main ();
+  int i, errors = 0;
+  for (i = 0; i < 1 + vtm->n_threads; i++)
+    errors += session_manager_flush_enqueue_events (transport_proto, i);
+  return errors;
+}
+
 /**
  * Init fifo tail and head pointers
  *
@@ -825,7 +840,7 @@
   if (session_alloc_and_init (sm, tc, 1, &s))
     return -1;
   s->app_index = app->index;
-  s->session_state = SESSION_STATE_CONNECTING_READY;
+  s->session_state = SESSION_STATE_OPENED;
 
   /* Tell the app about the new event fifo for this session */
   app->cb_fns.session_connected_callback (app->index, opaque, s, 0);
@@ -841,10 +856,6 @@
   u64 handle;
   int rv;
 
-  /* TODO until udp is fixed */
-  if (rmt->transport_proto == TRANSPORT_PROTO_UDP)
-    return session_open_cl (app_index, rmt, opaque);
-
   tep = session_endpoint_to_transport (rmt);
   rv = tp_vfts[rmt->transport_proto].open (tep);
   if (rv < 0)
@@ -912,14 +923,6 @@
   return session_open_srv_fns[tst] (app_index, rmt, opaque);
 }
 
-/**
- * Ask transport to listen on local transport endpoint.
- *
- * @param s Session for which listen will be called. Note that unlike
- * 	    established sessions, listen sessions are not associated to a
- * 	    thread.
- * @param tep Local endpoint to be listened on.
- */
 int
 session_listen_vc (stream_session_t * s, session_endpoint_t * sep)
 {
@@ -948,6 +951,40 @@
 }
 
 int
+session_listen_cl (stream_session_t * s, session_endpoint_t * sep)
+{
+  transport_connection_t *tc;
+  application_t *server;
+  segment_manager_t *sm;
+  u32 tci;
+
+  /* Transport bind/listen  */
+  tci = tp_vfts[sep->transport_proto].bind (s->session_index,
+					    session_endpoint_to_transport
+					    (sep));
+
+  if (tci == (u32) ~ 0)
+    return -1;
+
+  /* Attach transport to session */
+  s->connection_index = tci;
+  tc = tp_vfts[sep->transport_proto].get_listener (tci);
+
+  /* Weird but handle it ... */
+  if (tc == 0)
+    return -1;
+
+  server = application_get (s->app_index);
+  sm = application_get_listen_segment_manager (server, s);
+  if (session_alloc_fifos (sm, s))
+    return -1;
+
+  /* Add to the main lookup table */
+  session_lookup_add_connection (tc, s->session_index);
+  return 0;
+}
+
+int
 session_listen_app (stream_session_t * s, session_endpoint_t * sep)
 {
   session_endpoint_extended_t esep;
@@ -965,11 +1002,19 @@
 static session_listen_service_fn
 session_listen_srv_fns[TRANSPORT_N_SERVICES] = {
   session_listen_vc,
-  session_listen_vc,
+  session_listen_cl,
   session_listen_app,
 };
 /* *INDENT-ON* */
 
+/**
+ * Ask transport to listen on local transport endpoint.
+ *
+ * @param s Session for which listen will be called. Note that unlike
+ * 	    established sessions, listen sessions are not associated to a
+ * 	    thread.
+ * @param tep Local endpoint to be listened on.
+ */
 int
 stream_session_listen (stream_session_t * s, session_endpoint_t * sep)
 {
@@ -1125,7 +1170,8 @@
 static session_fifo_rx_fn *session_tx_fns[TRANSPORT_TX_N_FNS] = {
     session_tx_fifo_peek_and_snd,
     session_tx_fifo_dequeue_and_snd,
-    session_tx_fifo_dequeue_internal
+    session_tx_fifo_dequeue_internal,
+    session_tx_fifo_dequeue_and_snd
 };
 /* *INDENT-ON* */
 
@@ -1228,11 +1274,12 @@
   vec_validate (smm->peekers_rw_locks, num_threads - 1);
 
   for (i = 0; i < TRANSPORT_N_PROTO; i++)
-    for (j = 0; j < num_threads; j++)
-      {
-	vec_validate (smm->session_to_enqueue[i], num_threads - 1);
-	vec_validate (smm->current_enqueue_epoch[i], num_threads - 1);
-      }
+    {
+      vec_validate (smm->current_enqueue_epoch[i], num_threads - 1);
+      vec_validate (smm->session_to_enqueue[i], num_threads - 1);
+      for (j = 0; j < num_threads; j++)
+	smm->current_enqueue_epoch[i][j] = 1;
+    }
 
   for (i = 0; i < num_threads; i++)
     {
diff --git a/src/vnet/session/session.h b/src/vnet/session/session.h
index 0b53f61..9d534ae 100644
--- a/src/vnet/session/session.h
+++ b/src/vnet/session/session.h
@@ -204,6 +204,31 @@
 
 };
 
+typedef struct session_dgram_pre_hdr_
+{
+  u32 data_length;
+  u32 data_offset;
+} session_dgram_pre_hdr_t;
+
+/* *INDENT-OFF* */
+typedef CLIB_PACKED (struct session_dgram_header_
+{
+  u32 data_length;
+  u32 data_offset;
+  ip46_address_t rmt_ip;
+  ip46_address_t lcl_ip;
+  u16 rmt_port;
+  u16 lcl_port;
+  u8 is_ip4;
+}) session_dgram_hdr_t;
+/* *INDENT-ON* */
+
+#define SESSION_CONN_ID_LEN 37
+#define SESSION_CONN_HDR_LEN 45
+
+STATIC_ASSERT (sizeof (session_dgram_hdr_t) == (SESSION_CONN_ID_LEN + 8),
+	       "session conn id wrong length");
+
 extern session_manager_main_t session_manager_main;
 extern vlib_node_registration_t session_queue_node;
 
@@ -342,6 +367,14 @@
   return (session_get_transport_proto (s) != TRANSPORT_PROTO_NONE);
 }
 
+always_inline transport_service_type_t
+session_transport_service_type (stream_session_t * s)
+{
+  transport_proto_t tp;
+  tp = session_get_transport_proto (s);
+  return transport_protocol_service_type (tp);
+}
+
 /**
  * Acquires a lock that blocks a session pool from expanding.
  *
@@ -442,8 +475,10 @@
 session_enqueue_stream_connection (transport_connection_t * tc,
 				   vlib_buffer_t * b, u32 offset,
 				   u8 queue_event, u8 is_in_order);
-int session_enqueue_dgram_connection (stream_session_t * s, vlib_buffer_t * b,
-				      u8 proto, u8 queue_event);
+int session_enqueue_dgram_connection (stream_session_t * s,
+				      session_dgram_hdr_t * hdr,
+				      vlib_buffer_t * b, u8 proto,
+				      u8 queue_event);
 int stream_session_peek_bytes (transport_connection_t * tc, u8 * buffer,
 			       u32 offset, u32 max_bytes);
 u32 stream_session_dequeue_drop (transport_connection_t * tc, u32 max_bytes);
@@ -490,6 +525,7 @@
 }
 
 int session_manager_flush_enqueue_events (u8 proto, u32 thread_index);
+int session_manager_flush_all_enqueue_events (u8 transport_proto);
 
 always_inline u64
 listen_session_get_handle (stream_session_t * s)
diff --git a/src/vnet/session/session_api.c b/src/vnet/session/session_api.c
index 2a74a19..67c42fa 100755
--- a/src/vnet/session/session_api.c
+++ b/src/vnet/session/session_api.c
@@ -480,9 +480,12 @@
 static void
 vl_api_bind_uri_t_handler (vl_api_bind_uri_t * mp)
 {
-  vl_api_bind_uri_reply_t *rmp;
+  transport_connection_t *tc = 0;
   vnet_bind_args_t _a, *a = &_a;
-  application_t *app;
+  vl_api_bind_uri_reply_t *rmp;
+  stream_session_t *s;
+  application_t *app = 0;
+  svm_queue_t *vpp_evt_q;
   int rv;
 
   if (session_manager_is_enabled () == 0)
@@ -505,7 +508,30 @@
     }
 
 done:
-  REPLY_MACRO (VL_API_BIND_URI_REPLY);
+
+  /* *INDENT-OFF* */
+  REPLY_MACRO2 (VL_API_BIND_URI_REPLY, ({
+    if (!rv)
+      {
+        rmp->handle = a->handle;
+	rmp->lcl_is_ip4 = tc->is_ip4;
+  	rmp->lcl_port = tc->lcl_port;
+        if (app && application_has_global_scope (app))
+            {
+              s = listen_session_get_from_handle (a->handle);
+              tc = listen_session_get_transport (s);
+              clib_memcpy (rmp->lcl_ip, &tc->lcl_ip, sizeof(tc->lcl_ip));
+              if (session_transport_service_type (s) == TRANSPORT_SERVICE_CL)
+                {
+                  rmp->rx_fifo = pointer_to_uword (s->server_rx_fifo);
+                  rmp->tx_fifo = pointer_to_uword (s->server_tx_fifo);
+                  vpp_evt_q = session_manager_get_vpp_event_queue (0);
+                  rmp->vpp_evt_q = pointer_to_uword (vpp_evt_q);
+                }
+            }
+      }
+  }));
+  /* *INDENT-ON* */
 }
 
 static void
@@ -733,6 +759,7 @@
   stream_session_t *s;
   transport_connection_t *tc = 0;
   ip46_address_t *ip46;
+  svm_queue_t *vpp_evt_q;
 
   if (session_manager_is_enabled () == 0)
     {
@@ -775,8 +802,14 @@
 	  {
 	    s = listen_session_get_from_handle (a->handle);
 	    tc = listen_session_get_transport (s);
-            rmp->lcl_is_ip4 = tc->is_ip4;
             clib_memcpy (rmp->lcl_ip, &tc->lcl_ip, sizeof (tc->lcl_ip));
+            if (session_transport_service_type (s) == TRANSPORT_SERVICE_CL)
+              {
+        	rmp->rx_fifo = pointer_to_uword (s->server_rx_fifo);
+        	rmp->tx_fifo = pointer_to_uword (s->server_tx_fifo);
+        	vpp_evt_q = session_manager_get_vpp_event_queue (0);
+        	rmp->vpp_evt_q = pointer_to_uword (vpp_evt_q);
+              }
 	  }
       }
   }));
diff --git a/src/vnet/session/session_cli.c b/src/vnet/session/session_cli.c
index 5283355..201f6f1 100755
--- a/src/vnet/session/session_cli.c
+++ b/src/vnet/session/session_cli.c
@@ -23,6 +23,9 @@
   session_fifo_event_t _e, *e = &_e;
   u8 found;
 
+  if (!ss->server_rx_fifo || !ss->server_tx_fifo)
+    return s;
+
   s = format (s, " Rx fifo: %U", format_svm_fifo, ss->server_rx_fifo, 1);
   if (verbose > 2 && ss->server_rx_fifo->has_event)
     {
@@ -76,6 +79,8 @@
     {
       s = format (s, "%-40U%v", tp_vft->format_listener, ss->connection_index,
 		  str);
+      if (verbose > 1)
+	s = format (s, "\n%U", format_stream_session_fifos, ss, verbose);
     }
   else if (ss->session_state == SESSION_STATE_CONNECTING)
     {
diff --git a/src/vnet/session/session_node.c b/src/vnet/session/session_node.c
index b8f429e..1471696 100644
--- a/src/vnet/session/session_node.c
+++ b/src/vnet/session/session_node.c
@@ -70,7 +70,7 @@
 			    vlib_buffer_t * b0, u32 bi0, u8 n_bufs_per_seg,
 			    u32 left_from_seg, u32 * left_to_snd0,
 			    u16 * n_bufs, u32 * tx_offset, u16 deq_per_buf,
-			    u8 peek_data)
+			    u8 peek_data, transport_tx_fn_type_t tx_type)
 {
   vlib_buffer_t *chain_b0, *prev_b0;
   u32 chain_bi0, to_deq;
@@ -102,7 +102,23 @@
 	}
       else
 	{
-	  n_bytes_read = svm_fifo_dequeue_nowait (fifo, len_to_deq0, data0);
+	  if (tx_type == TRANSPORT_TX_DGRAM)
+	    {
+	      session_dgram_hdr_t *hdr;
+	      u16 deq_now;
+	      hdr = (session_dgram_hdr_t *) svm_fifo_head (fifo);
+	      deq_now = clib_min (hdr->data_length - hdr->data_offset,
+				  len_to_deq0);
+	      n_bytes_read = svm_fifo_peek (fifo, hdr->data_offset, deq_now,
+					    data0);
+	      ASSERT (n_bytes_read > 0);
+
+	      hdr->data_offset += n_bytes_read;
+	      if (hdr->data_offset == hdr->data_length)
+		svm_fifo_dequeue_drop (fifo, hdr->data_length);
+	    }
+	  else
+	    n_bytes_read = svm_fifo_dequeue_nowait (fifo, len_to_deq0, data0);
 	}
       ASSERT (n_bytes_read == len_to_deq0);
       chain_b0->current_length = n_bytes_read;
@@ -145,12 +161,35 @@
   int i, n_bytes_read;
   u32 n_bytes_per_buf, deq_per_buf, deq_per_first_buf;
   u32 bufs_alloc, bufs_now;
+  session_dgram_hdr_t hdr;
 
   next_index = next0 = smm->session_type_to_next[s0->session_type];
-
   tp = session_get_transport_proto (s0);
   transport_vft = transport_protocol_get_vft (tp);
-  tc0 = transport_vft->get_connection (s0->connection_index, thread_index);
+  if (peek_data)
+    {
+      if (PREDICT_FALSE (s0->session_state < SESSION_STATE_READY))
+	{
+	  /* Can retransmit for closed sessions but can't send new data if
+	   * session is not ready or closed */
+	  vec_add1 (smm->pending_event_vector[thread_index], *e0);
+	  return 0;
+	}
+      tc0 =
+	transport_vft->get_connection (s0->connection_index, thread_index);
+    }
+  else
+    {
+      if (s0->session_state == SESSION_STATE_LISTENING)
+	tc0 = transport_vft->get_listener (s0->connection_index);
+      else
+	{
+	  if (PREDICT_FALSE (s0->session_state == SESSION_STATE_CLOSED))
+	    return 0;
+	  tc0 = transport_vft->get_connection (s0->connection_index,
+					       thread_index);
+	}
+    }
 
   /* Make sure we have space to send and there's something to dequeue */
   snd_mss0 = transport_vft->send_mss (tc0);
@@ -168,20 +207,26 @@
 
   /* Check how much we can pull. */
   max_dequeue0 = svm_fifo_max_dequeue (s0->server_tx_fifo);
-
   if (peek_data)
     {
       /* Offset in rx fifo from where to peek data */
       tx_offset = transport_vft->tx_fifo_offset (tc0);
       if (PREDICT_FALSE (tx_offset >= max_dequeue0))
-	max_dequeue0 = 0;
-      else
-	max_dequeue0 -= tx_offset;
+	return 0;
+      max_dequeue0 -= tx_offset;
     }
-
-  /* Nothing to read return */
-  if (max_dequeue0 == 0)
-    return 0;
+  else
+    {
+      if (transport_vft->tx_type == TRANSPORT_TX_DGRAM)
+	{
+	  if (max_dequeue0 < sizeof (hdr))
+	    return 0;
+	  svm_fifo_peek (s0->server_tx_fifo, 0, sizeof (hdr), (u8 *) & hdr);
+	  ASSERT (hdr.data_length > hdr.data_offset);
+	  max_dequeue0 = hdr.data_length - hdr.data_offset;
+	}
+    }
+  ASSERT (max_dequeue0 > 0);
 
   /* Ensure we're not writing more than transport window allows */
   if (max_dequeue0 < snd_space0)
@@ -286,14 +331,42 @@
 	    }
 	  else
 	    {
-	      n_bytes_read = svm_fifo_dequeue_nowait (s0->server_tx_fifo,
-						      len_to_deq0, data0);
-	      if (n_bytes_read <= 0)
-		goto dequeue_fail;
+	      if (transport_vft->tx_type == TRANSPORT_TX_DGRAM)
+		{
+		  svm_fifo_t *f = s0->server_tx_fifo;
+		  u16 deq_now;
+		  u32 offset;
+
+		  ASSERT (hdr.data_length > hdr.data_offset);
+		  deq_now = clib_min (hdr.data_length - hdr.data_offset,
+				      len_to_deq0);
+		  offset = hdr.data_offset + SESSION_CONN_HDR_LEN;
+		  n_bytes_read = svm_fifo_peek (f, offset, deq_now, data0);
+		  if (PREDICT_FALSE (n_bytes_read <= 0))
+		    goto dequeue_fail;
+
+		  if (s0->session_state == SESSION_STATE_LISTENING)
+		    {
+		      ip_copy (&tc0->rmt_ip, &hdr.rmt_ip, tc0->is_ip4);
+		      tc0->rmt_port = hdr.rmt_port;
+		    }
+		  hdr.data_offset += n_bytes_read;
+		  if (hdr.data_offset == hdr.data_length)
+		    {
+		      offset = hdr.data_length + SESSION_CONN_HDR_LEN;
+		      svm_fifo_dequeue_drop (f, offset);
+		    }
+		}
+	      else
+		{
+		  n_bytes_read = svm_fifo_dequeue_nowait (s0->server_tx_fifo,
+							  len_to_deq0, data0);
+		  if (n_bytes_read <= 0)
+		    goto dequeue_fail;
+		}
 	    }
 
 	  b0->current_length = n_bytes_read;
-
 	  left_to_snd0 -= n_bytes_read;
 	  *n_tx_packets = *n_tx_packets + 1;
 
@@ -307,7 +380,8 @@
 					  s0->server_tx_fifo, b0, bi0,
 					  n_bufs_per_seg, left_for_seg,
 					  &left_to_snd0, &n_bufs, &tx_offset,
-					  deq_per_buf, peek_data);
+					  deq_per_buf, peek_data,
+					  transport_vft->tx_type);
 	    }
 
 	  /* Ask transport to push header after current_length and
@@ -345,12 +419,18 @@
 
   /* If we couldn't dequeue all bytes mark as partially read */
   if (max_len_to_snd0 < max_dequeue0)
+    if (svm_fifo_set_event (s0->server_tx_fifo))
+      vec_add1 (smm->pending_event_vector[thread_index], *e0);
+
+  if (!peek_data && transport_vft->tx_type == TRANSPORT_TX_DGRAM)
     {
-      /* If we don't already have new event */
-      if (svm_fifo_set_event (s0->server_tx_fifo))
-	{
-	  vec_add1 (smm->pending_event_vector[thread_index], *e0);
-	}
+      /* Fix dgram pre header */
+      if (max_len_to_snd0 < max_dequeue0)
+	svm_fifo_overwrite_head (s0->server_tx_fifo, (u8 *) & hdr,
+				 sizeof (session_dgram_pre_hdr_t));
+      /* More data needs to be read */
+      else if (svm_fifo_max_dequeue (s0->server_tx_fifo) > 0)
+	vec_add1 (smm->pending_event_vector[thread_index], *e0);
     }
   return 0;
 
@@ -360,7 +440,6 @@
    * read, return buff to free list and return
    */
   clib_warning ("dequeue fail");
-
   if (svm_fifo_set_event (s0->server_tx_fifo))
     {
       vec_add1 (smm->pending_event_vector[thread_index], *e0);
@@ -638,13 +717,6 @@
 	      clib_warning ("It's dead, Jim!");
 	      continue;
 	    }
-	  /* Can retransmit for closed sessions but can't do anything if
-	   * session is not ready or closed */
-	  if (PREDICT_FALSE (s0->session_state < SESSION_STATE_READY))
-	    {
-	      vec_add1 (smm->pending_event_vector[my_thread_index], *e0);
-	      continue;
-	    }
 	  /* Spray packets in per session type frames, since they go to
 	   * different nodes */
 	  rv = (smm->session_tx_fns[s0->session_type]) (vm, node, smm, e0, s0,
diff --git a/src/vnet/session/stream_session.h b/src/vnet/session/stream_session.h
index b7a5eee..9e0e4d9 100644
--- a/src/vnet/session/stream_session.h
+++ b/src/vnet/session/stream_session.h
@@ -31,30 +31,19 @@
   SESSION_STATE_CONNECTING,
   SESSION_STATE_ACCEPTING,
   SESSION_STATE_READY,
-  SESSION_STATE_CONNECTING_READY,
+  SESSION_STATE_OPENED,
   SESSION_STATE_CLOSED,
   SESSION_STATE_N_STATES,
 } stream_session_state_t;
 
-/* TODO convert to macro once cleanup completed */
-typedef struct app_session_
+typedef struct generic_session_
 {
-  /** fifo pointers. Once allocated, these do not move */
-  svm_fifo_t *server_rx_fifo;
-  svm_fifo_t *server_tx_fifo;
-
-  /** Type */
-  session_type_t session_type;
-
-  /** State */
-  volatile u8 session_state;
-
-  /** Session index in owning pool */
-  u32 session_index;
-
-  /** Application index */
-  u32 app_index;
-} app_session_t;
+  svm_fifo_t *rx_fifo;		/**< rx fifo */
+  svm_fifo_t *tx_fifo;		/**< tx fifo */
+  session_type_t session_type;	/**< session type */
+  volatile u8 session_state;	/**< session state */
+  u32 session_index;		/**< index in owning pool */
+} generic_session_t;
 
 typedef struct _stream_session_t
 {
diff --git a/src/vnet/session/transport.c b/src/vnet/session/transport.c
index 797bdad..20b9129 100644
--- a/src/vnet/session/transport.c
+++ b/src/vnet/session/transport.c
@@ -57,6 +57,9 @@
     case TRANSPORT_PROTO_SCTP:
       s = format (s, "SCTP");
       break;
+    case TRANSPORT_PROTO_UDPC:
+      s = format (s, "UDPC");
+      break;
     }
   return s;
 }
@@ -76,6 +79,9 @@
     case TRANSPORT_PROTO_SCTP:
       s = format (s, "S");
       break;
+    case TRANSPORT_PROTO_UDPC:
+      s = format (s, "U");
+      break;
     }
   return s;
 }
@@ -100,6 +106,10 @@
     *proto = TRANSPORT_PROTO_TLS;
   else if (unformat (input, "TLS"))
     *proto = TRANSPORT_PROTO_TLS;
+  else if (unformat (input, "udpc"))
+    *proto = TRANSPORT_PROTO_UDPC;
+  else if (unformat (input, "UDPC"))
+    *proto = TRANSPORT_PROTO_UDPC;
   else
     return 0;
   return 1;
@@ -185,6 +195,12 @@
   return &tp_vfts[transport_proto];
 }
 
+transport_service_type_t
+transport_protocol_service_type (transport_proto_t tp)
+{
+  return tp_vfts[tp].service_type;
+}
+
 #define PORT_MASK ((1 << 16)- 1)
 
 void
diff --git a/src/vnet/session/transport.h b/src/vnet/session/transport.h
index ed9eb02..8340fd8 100644
--- a/src/vnet/session/transport.h
+++ b/src/vnet/session/transport.h
@@ -35,10 +35,10 @@
     {
       ip46_address_t rmt_ip;	/**< Remote IP */
       ip46_address_t lcl_ip;	/**< Local IP */
-      u16 lcl_port;		/**< Local port */
       u16 rmt_port;		/**< Remote port */
-      u8 proto;			/**< Protocol id */
+      u16 lcl_port;		/**< Local port */
       u8 is_ip4;		/**< Flag if IP4 connection */
+      u8 proto;			/**< Protocol id */
       u32 fib_index;		/**< Network namespace */
     };
     /*
@@ -88,6 +88,7 @@
   TRANSPORT_PROTO_SCTP,
   TRANSPORT_PROTO_NONE,
   TRANSPORT_PROTO_TLS,
+  TRANSPORT_PROTO_UDPC,
   TRANSPORT_N_PROTO
 } transport_proto_t;
 
@@ -99,7 +100,7 @@
   _(u32, sw_if_index) 	/**< interface endpoint is associated with  */	\
   _(ip46_address_t, ip) /**< ip address */				\
   _(u32, fib_index)	/**< fib table endpoint is associated with */	\
-  _(u8, is_ip4)	/**< 1 if ip4 */					\
+  _(u8, is_ip4)		/**< set if ip4 */				\
   _(u16, port)		/**< port in net order */			\
 
 typedef struct _transport_endpoint
@@ -125,18 +126,6 @@
   return tep->is_ip4 ? FIB_PROTOCOL_IP4 : FIB_PROTOCOL_IP6;
 }
 
-always_inline u8
-transport_is_stream (u8 proto)
-{
-  return ((proto == TRANSPORT_PROTO_TCP) || (proto == TRANSPORT_PROTO_SCTP));
-}
-
-always_inline u8
-transport_is_dgram (u8 proto)
-{
-  return (proto == TRANSPORT_PROTO_UDP);
-}
-
 int transport_alloc_local_port (u8 proto, ip46_address_t * ip);
 int transport_alloc_local_endpoint (u8 proto, transport_endpoint_t * rmt,
 				    ip46_address_t * lcl_addr,
diff --git a/src/vnet/session/transport_interface.h b/src/vnet/session/transport_interface.h
index 04a5ff2..f21e483 100644
--- a/src/vnet/session/transport_interface.h
+++ b/src/vnet/session/transport_interface.h
@@ -23,7 +23,8 @@
 {
   TRANSPORT_TX_PEEK,		/**< reliable transport protos */
   TRANSPORT_TX_DEQUEUE,		/**< unreliable transport protos */
-  TRANSPORT_TX_INTERNAL,		/**< apps acting as transports */
+  TRANSPORT_TX_INTERNAL,	/**< apps acting as transports */
+  TRANSPORT_TX_DGRAM,		/**< datagram mode */
   TRANSPORT_TX_N_FNS
 } transport_tx_fn_type_t;
 
@@ -31,7 +32,7 @@
 {
   TRANSPORT_SERVICE_VC,		/**< virtual circuit service */
   TRANSPORT_SERVICE_CL,		/**< connectionless service */
-  TRANSPORT_SERVICE_APP,		/**< app transport service */
+  TRANSPORT_SERVICE_APP,	/**< app transport service */
   TRANSPORT_N_SERVICES
 } transport_service_type_t;
 
@@ -96,6 +97,7 @@
 				  const transport_proto_vft_t * vft,
 				  fib_protocol_t fib_proto, u32 output_node);
 transport_proto_vft_t *transport_protocol_get_vft (transport_proto_t tp);
+transport_service_type_t transport_protocol_service_type (transport_proto_t);
 void transport_update_time (f64 time_now, u8 thread_index);
 void transport_enable_disable (vlib_main_t * vm, u8 is_en);
 
diff --git a/src/vnet/udp/udp.c b/src/vnet/udp/udp.c
index 3b8b707..947cc1e 100644
--- a/src/vnet/udp/udp.c
+++ b/src/vnet/udp/udp.c
@@ -49,6 +49,7 @@
   uc->c_c_index = uc - um->connections[thread_index];
   uc->c_thread_index = thread_index;
   uc->c_proto = TRANSPORT_PROTO_UDP;
+  clib_spinlock_init (&uc->rx_lock);
   return uc;
 }
 
@@ -92,6 +93,7 @@
   listener->c_proto = TRANSPORT_PROTO_UDP;
   listener->c_s_index = session_index;
   listener->c_fib_index = lcl->fib_index;
+  clib_spinlock_init (&listener->rx_lock);
 
   node_index = lcl->is_ip4 ? udp4_input_node.index : udp6_input_node.index;
   udp_register_dst_port (vm, clib_net_to_host_u16 (lcl->port), node_index,
@@ -140,7 +142,7 @@
       vnet_buffer (b)->l3_hdr_offset = (u8 *) ih - b->data;
     }
   vnet_buffer (b)->sw_if_index[VLIB_RX] = 0;
-  vnet_buffer (b)->sw_if_index[VLIB_TX] = ~0;
+  vnet_buffer (b)->sw_if_index[VLIB_TX] = uc->c_fib_index;
   b->flags |= VNET_BUFFER_F_LOCALLY_ORIGINATED;
 
   return 0;
@@ -210,7 +212,7 @@
   if (verbose)
     {
       if (verbose == 1)
-	s = format (s, "%-15s", "-");
+	s = format (s, "%-15s\n", "-");
       else
 	s = format (s, "\n");
     }
@@ -322,8 +324,52 @@
   .format_connection = format_udp_session,
   .format_half_open = format_udp_half_open_session,
   .format_listener = format_udp_listener_session,
+  .tx_type = TRANSPORT_TX_DGRAM,
+  .service_type = TRANSPORT_SERVICE_CL,
+};
+/* *INDENT-ON* */
+
+
+int
+udpc_connection_open (transport_endpoint_t * rmt)
+{
+  udp_connection_t *uc;
+  u32 uc_index;
+  uc_index = udp_open_connection (rmt);
+  uc = udp_connection_get (uc_index, vlib_get_thread_index ());
+  uc->is_connected = 1;
+  return uc_index;
+}
+
+u32
+udpc_connection_listen (u32 session_index, transport_endpoint_t * lcl)
+{
+  udp_connection_t *listener;
+  u32 li;
+  li = udp_session_bind (session_index, lcl);
+  listener = udp_listener_get (li);
+  listener->is_connected = 1;
+  return li;
+}
+
+/* *INDENT-OFF* */
+const static transport_proto_vft_t udpc_proto = {
+  .bind = udpc_connection_listen,
+  .open = udpc_connection_open,
+  .unbind = udp_session_unbind,
+  .push_header = udp_push_header,
+  .get_connection = udp_session_get,
+  .get_listener = udp_session_get_listener,
+  .get_half_open = udp_half_open_session_get_transport,
+  .close = udp_session_close,
+  .cleanup = udp_session_cleanup,
+  .send_mss = udp_send_mss,
+  .send_space = udp_send_space,
+  .format_connection = format_udp_session,
+  .format_half_open = format_udp_half_open_session,
+  .format_listener = format_udp_listener_session,
   .tx_type = TRANSPORT_TX_DEQUEUE,
-  .service_type = TRANSPORT_SERVICE_VC,
+  .service_type = TRANSPORT_SERVICE_CL,
 };
 /* *INDENT-ON* */
 
@@ -361,6 +407,10 @@
 			       FIB_PROTOCOL_IP4, ip4_lookup_node.index);
   transport_register_protocol (TRANSPORT_PROTO_UDP, &udp_proto,
 			       FIB_PROTOCOL_IP6, ip6_lookup_node.index);
+  transport_register_protocol (TRANSPORT_PROTO_UDPC, &udpc_proto,
+			       FIB_PROTOCOL_IP4, ip4_lookup_node.index);
+  transport_register_protocol (TRANSPORT_PROTO_UDPC, &udpc_proto,
+			       FIB_PROTOCOL_IP6, ip6_lookup_node.index);
 
   /*
    * Initialize data structures
diff --git a/src/vnet/udp/udp.h b/src/vnet/udp/udp.h
index af6c6b8..8e3ab9e 100644
--- a/src/vnet/udp/udp.h
+++ b/src/vnet/udp/udp.h
@@ -36,9 +36,9 @@
 
 typedef struct
 {
-  transport_connection_t connection;	      /** must be first */
-  /** ersatz MTU to limit fifo pushes to test data size */
-  u32 mtu;
+  transport_connection_t connection;	/**< must be first */
+  clib_spinlock_t rx_lock;		/**< rx fifo lock */
+  u8 is_connected;			/**< connected mode */
 } udp_connection_t;
 
 #define foreach_udp4_dst_port			\
@@ -207,7 +207,7 @@
 }
 
 always_inline udp_connection_t *
-udp_conenction_clone_safe (u32 connection_index, u32 thread_index)
+udp_connection_clone_safe (u32 connection_index, u32 thread_index)
 {
   udp_connection_t *old_c, *new_c;
   u32 current_thread_index = vlib_get_thread_index ();
diff --git a/src/vnet/udp/udp_input.c b/src/vnet/udp/udp_input.c
index 8170cfb..ad469f6 100644
--- a/src/vnet/udp/udp_input.c
+++ b/src/vnet/udp/udp_input.c
@@ -13,20 +13,19 @@
  * limitations under the License.
  */
 
+#include <vlibmemory/api.h>
 #include <vlib/vlib.h>
-#include <vnet/vnet.h>
-#include <vnet/pg/pg.h>
-#include <vnet/ip/ip.h>
 
-#include <vnet/udp/udp.h>
 #include <vppinfra/hash.h>
 #include <vppinfra/error.h>
 #include <vppinfra/elog.h>
 
+#include <vnet/vnet.h>
+#include <vnet/pg/pg.h>
+#include <vnet/ip/ip.h>
+#include <vnet/udp/udp.h>
 #include <vnet/udp/udp_packet.h>
-
-#include <vlibmemory/api.h>
-#include "../session/application_interface.h"
+#include <vnet/session/session.h>
 
 static char *udp_error_strings[] = {
 #define udp_error(n,s) s,
@@ -106,9 +105,11 @@
 	  ip6_header_t *ip60;
 	  u8 *data0;
 	  stream_session_t *s0;
-	  transport_connection_t *tc0 = 0;
-	  udp_connection_t *child0, *new_uc0;
-	  int written0;
+	  udp_connection_t *uc0, *child0, *new_uc0;
+	  transport_connection_t *tc0;
+	  int wrote0;
+	  void *rmt_addr, *lcl_addr;
+	  session_dgram_hdr_t hdr0;
 
 	  /* speculatively enqueue b0 to the current next frame */
 	  bi0 = from[0];
@@ -127,11 +128,14 @@
 
 	  if (is_ip4)
 	    {
-	      /* $$$$ fixme: udp_local doesn't do ip options correctly anyhow */
+	      /* TODO: must fix once udp_local does ip options correctly */
 	      ip40 = (ip4_header_t *) (((u8 *) udp0) - sizeof (*ip40));
 	      s0 = session_lookup_safe4 (fib_index0, &ip40->dst_address,
 					 &ip40->src_address, udp0->dst_port,
 					 udp0->src_port, TRANSPORT_PROTO_UDP);
+	      lcl_addr = &ip40->dst_address;
+	      rmt_addr = &ip40->src_address;
+
 	    }
 	  else
 	    {
@@ -139,67 +143,83 @@
 	      s0 = session_lookup_safe6 (fib_index0, &ip60->dst_address,
 					 &ip60->src_address, udp0->dst_port,
 					 udp0->src_port, TRANSPORT_PROTO_UDP);
+	      lcl_addr = &ip60->dst_address;
+	      rmt_addr = &ip60->src_address;
 	    }
 
-	  if (PREDICT_FALSE (s0 == 0))
+	  if (PREDICT_FALSE (!s0))
 	    {
 	      error0 = UDP_ERROR_NO_LISTENER;
 	      goto trace0;
 	    }
 
-	  if (PREDICT_TRUE (s0->session_state == SESSION_STATE_READY))
+	  if (s0->session_state == SESSION_STATE_OPENED)
+	    {
+	      /* TODO optimization: move cl session to right thread
+	       * However, since such a move would affect the session handle,
+	       * which we pass 'raw' to the app, we'd also have notify the
+	       * app of the change or change the way we pass handles to apps.
+	       */
+	      tc0 = session_get_transport (s0);
+	      uc0 = udp_get_connection_from_transport (tc0);
+	      if (uc0->is_connected)
+		{
+		  /*
+		   * Clone the transport. It will be cleaned up with the
+		   * session once we notify the session layer.
+		   */
+		  new_uc0 = udp_connection_clone_safe (s0->connection_index,
+						       s0->thread_index);
+		  ASSERT (s0->session_index == new_uc0->c_s_index);
+
+		  /*
+		   * Drop the 'lock' on pool resize
+		   */
+		  session_pool_remove_peeker (s0->thread_index);
+		  session_dgram_connect_notify (&new_uc0->connection,
+						s0->thread_index, &s0);
+		  tc0 = &new_uc0->connection;
+		}
+	    }
+	  else if (s0->session_state == SESSION_STATE_READY)
 	    {
 	      tc0 = session_get_transport (s0);
-	    }
-	  else if (s0->session_state == SESSION_STATE_CONNECTING_READY)
-	    {
-	      /*
-	       * Clone the transport. It will be cleaned up with the
-	       * session once we notify the session layer.
-	       */
-	      new_uc0 = udp_conenction_clone_safe (s0->connection_index,
-						   s0->thread_index);
-	      ASSERT (s0->session_index == new_uc0->c_s_index);
-
-	      /*
-	       * Drop the 'lock' on pool resize
-	       */
-	      session_pool_remove_peeker (s0->thread_index);
-	      session_dgram_connect_notify (&new_uc0->connection,
-					    s0->thread_index, &s0);
-	      tc0 = &new_uc0->connection;
+	      uc0 = udp_get_connection_from_transport (tc0);
 	    }
 	  else if (s0->session_state == SESSION_STATE_LISTENING)
 	    {
 	      tc0 = listen_session_get_transport (s0);
-
-	      child0 = udp_connection_alloc (my_thread_index);
-	      if (is_ip4)
+	      uc0 = udp_get_connection_from_transport (tc0);
+	      if (uc0->is_connected)
 		{
-		  ip_set (&child0->c_lcl_ip, &ip40->dst_address, 1);
-		  ip_set (&child0->c_rmt_ip, &ip40->src_address, 1);
-		}
-	      else
-		{
-		  ip_set (&child0->c_lcl_ip, &ip60->dst_address, 0);
-		  ip_set (&child0->c_rmt_ip, &ip60->src_address, 0);
-		}
-	      child0->c_lcl_port = udp0->dst_port;
-	      child0->c_rmt_port = udp0->src_port;
-	      child0->c_is_ip4 = is_ip4;
-	      child0->mtu = 1460;	/* $$$$ policy */
+		  child0 = udp_connection_alloc (my_thread_index);
+		  if (is_ip4)
+		    {
+		      ip_set (&child0->c_lcl_ip, &ip40->dst_address, 1);
+		      ip_set (&child0->c_rmt_ip, &ip40->src_address, 1);
+		    }
+		  else
+		    {
+		      ip_set (&child0->c_lcl_ip, &ip60->dst_address, 0);
+		      ip_set (&child0->c_rmt_ip, &ip60->src_address, 0);
+		    }
+		  child0->c_lcl_port = udp0->dst_port;
+		  child0->c_rmt_port = udp0->src_port;
+		  child0->c_is_ip4 = is_ip4;
 
-	      if (stream_session_accept
-		  (&child0->connection, tc0->s_index, 1))
-		{
-		  error0 = UDP_ERROR_CREATE_SESSION;
-		  goto trace0;
+		  if (stream_session_accept (&child0->connection,
+					     tc0->s_index, 1))
+		    {
+		      error0 = UDP_ERROR_CREATE_SESSION;
+		      goto trace0;
+		    }
+		  s0 =
+		    session_get (child0->c_s_index, child0->c_thread_index);
+		  s0->session_state = SESSION_STATE_READY;
+		  tc0 = &child0->connection;
+		  uc0 = udp_get_connection_from_transport (tc0);
+		  error0 = UDP_ERROR_LISTENER;
 		}
-	      s0 = session_get (child0->c_s_index, child0->c_thread_index);
-	      s0->session_state = SESSION_STATE_READY;
-	      tc0 = &child0->connection;
-
-	      error0 = UDP_ERROR_LISTENER;
 	    }
 	  else
 	    {
@@ -207,15 +227,48 @@
 	      goto trace0;
 	    }
 
-	  written0 = session_enqueue_dgram_connection (s0, b0, tc0->proto,
-						       1 /* queue evt */ );
-	  if (PREDICT_FALSE (written0 < 0))
+	  if (!uc0->is_connected)
 	    {
-	      error0 = UDP_ERROR_FIFO_FULL;
-	      goto trace0;
+	      if (svm_fifo_max_enqueue (s0->server_rx_fifo)
+		  < b0->current_length + sizeof (session_dgram_hdr_t))
+		{
+		  error0 = UDP_ERROR_FIFO_FULL;
+		  goto trace0;
+		}
+	      hdr0.data_length = b0->current_length;
+	      hdr0.data_offset = 0;
+	      ip_set (&hdr0.lcl_ip, lcl_addr, is_ip4);
+	      ip_set (&hdr0.rmt_ip, rmt_addr, is_ip4);
+	      hdr0.lcl_port = udp0->dst_port;
+	      hdr0.rmt_port = udp0->src_port;
+	      hdr0.is_ip4 = is_ip4;
+
+	      clib_spinlock_lock (&uc0->rx_lock);
+	      wrote0 = session_enqueue_dgram_connection (s0, &hdr0, b0,
+							 TRANSPORT_PROTO_UDP,
+							 1 /* queue evt */ );
+	      clib_spinlock_unlock (&uc0->rx_lock);
+	      ASSERT (wrote0 > 0);
+
+	      if (s0->session_state != SESSION_STATE_LISTENING)
+		session_pool_remove_peeker (s0->thread_index);
+	    }
+	  else
+	    {
+	      if (svm_fifo_max_enqueue (s0->server_rx_fifo)
+		  < b0->current_length)
+		{
+		  error0 = UDP_ERROR_FIFO_FULL;
+		  goto trace0;
+		}
+	      wrote0 = session_enqueue_stream_connection (tc0, b0, 0,
+							  1 /* queue evt */ ,
+							  1 /* in order */ );
+	      ASSERT (wrote0 > 0);
 	    }
 
 	trace0:
+
 	  b0->error = node->errors[error0];
 
 	  if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
@@ -224,7 +277,7 @@
 	      udp_input_trace_t *t = vlib_add_trace (vm, node, b0,
 						     sizeof (*t));
 
-	      t->connection = tc0 ? tc0->c_index : ~0;
+	      t->connection = s0 ? s0->connection_index : ~0;
 	      t->disposition = error0;
 	      t->thread_index = my_thread_index;
 	    }
@@ -237,14 +290,11 @@
       vlib_put_next_frame (vm, node, next_index, n_left_to_next);
     }
 
-  errors = session_manager_flush_enqueue_events (TRANSPORT_PROTO_UDP,
-						 my_thread_index);
+  errors = session_manager_flush_all_enqueue_events (TRANSPORT_PROTO_UDP);
   udp_input_inc_counter (vm, is_ip4, UDP_ERROR_EVENT_FIFO_FULL, errors);
   return frame->n_vectors;
 }
 
-vlib_node_registration_t udp4_input_node;
-vlib_node_registration_t udp6_input_node;
 
 static uword
 udp4_input (vlib_main_t * vm, vlib_node_runtime_t * node,