* Copyright (c) 2010, Lei Xia <lxia@northwestern.edu>
* Copyright (c) 2009, Yuan Tang <ytang@northwestern.edu>
* Copyright (c) 2009, The V3VEE Project <http://www.v3vee.org>
- * All rights reserved.
+ * All rights reserved
*
* Author: Lei Xia <lxia@northwestern.edu>
* Yuan Tang <ytang@northwestern.edu>
#include <vnet/vnet_host.h>
#include <vnet/vnet_vmm.h>
+#include <palacios/vmm_queue.h>
+
#ifndef V3_CONFIG_DEBUG_VNET
-#undef Vnet_Debug
-#define Vnet_Debug(fmt, args...)
+#undef PrintDebug
+#define PrintDebug(fmt, args...)
#endif
-int vnet_debug = 0;
+#define VNET_NOPROGRESS_LIMIT 1000
+#define VNET_YIELD_USEC 1000
+
+int net_debug = 0;
struct eth_hdr {
uint8_t dst_mac[ETH_ALEN];
uint8_t mac_addr[ETH_ALEN];
struct v3_vm_info * vm;
struct v3_vnet_dev_ops dev_ops;
+
+ int poll;
+
+#define VNET_MAX_QUOTE 64
+ int quote;
+
void * private_data;
struct list_head node;
struct vnet_dev * dst_dev;
struct vnet_dev * src_dev;
+ uint32_t idx;
+
struct list_head node;
struct list_head match_node; // used for route matching
};
uint32_t size_alloc;
};
-#define VNET_QUEUE_SIZE 1024
-struct vnet_queue {
- struct queue_entry buf[VNET_QUEUE_SIZE];
- int head, tail;
- int count;
- vnet_lock_t lock;
-};
static struct {
struct list_head routes;
struct list_head devs;
-
- int num_routes;
- int num_devs;
+
+ uint8_t status;
+
+ uint32_t num_routes;
+ uint32_t route_idx;
+ uint32_t num_devs;
+ uint32_t dev_idx;
struct vnet_brg_dev * bridge;
vnet_lock_t lock;
struct vnet_stat stats;
- struct vnet_thread * pkt_flush_thread;
+ /* device queue that are waiting to be polled */
+ struct v3_queue * poll_devs;
- struct vnet_queue pkt_q;
+ struct vnet_thread * pkt_flush_thread;
struct hashtable * route_cache;
} vnet_state;
#ifdef V3_CONFIG_DEBUG_VNET
-static inline void mac_to_string(uint8_t * mac, char * buf) {
+static inline void mac2str(uint8_t * mac, char * buf) {
snprintf(buf, 100, "%2x:%2x:%2x:%2x:%2x:%2x",
mac[0], mac[1], mac[2],
mac[3], mac[4], mac[5]);
static void print_route(struct v3_vnet_route * route){
char str[50];
- mac_to_string(route->src_mac, str);
- Vnet_Debug("Src Mac (%s), src_qual (%d)\n",
+ mac2str(route->src_mac, str);
+ PrintDebug("Src Mac (%s), src_qual (%d)\n",
str, route->src_mac_qual);
- mac_to_string(route->dst_mac, str);
- Vnet_Debug("Dst Mac (%s), dst_qual (%d)\n",
+ mac2str(route->dst_mac, str);
+ PrintDebug("Dst Mac (%s), dst_qual (%d)\n",
str, route->dst_mac_qual);
- Vnet_Debug("Src dev id (%d), src type (%d)",
+ PrintDebug("Src dev id (%d), src type (%d)",
route->src_id,
route->src_type);
- Vnet_Debug("Dst dev id (%d), dst type (%d)\n",
+ PrintDebug("Dst dev id (%d), dst type (%d)\n",
route->dst_id,
route->dst_type);
}
static void dump_routes(){
struct vnet_route_info *route;
- int i = 0;
- Vnet_Debug("\n========Dump routes starts ============\n");
+ PrintDebug("\n========Dump routes starts ============\n");
list_for_each_entry(route, &(vnet_state.routes), node) {
- Vnet_Debug("\nroute %d:\n", i++);
+ PrintDebug("\nroute %d:\n", route->idx);
print_route(&(route->route_def));
if (route->route_def.dst_type == LINK_INTERFACE) {
- Vnet_Debug("dst_dev (%p), dst_dev_id (%d), dst_dev_ops(%p), dst_dev_data (%p)\n",
+ PrintDebug("dst_dev (%p), dst_dev_id (%d), dst_dev_ops(%p), dst_dev_data (%p)\n",
route->dst_dev,
route->dst_dev->dev_id,
(void *)&(route->dst_dev->dev_ops),
}
}
- Vnet_Debug("\n========Dump routes end ============\n");
+ PrintDebug("\n========Dump routes end ============\n");
}
#endif
struct vnet_dev * dev = NULL;
list_for_each_entry(dev, &(vnet_state.devs), node) {
- int dev_id = dev->dev_id;
-
- if (dev_id == idx)
+ if (dev->dev_id == idx) {
return dev;
+ }
}
return NULL;
int v3_vnet_add_route(struct v3_vnet_route route) {
struct vnet_route_info * new_route = NULL;
- unsigned long flags;
+ vnet_intr_flags_t flags;
new_route = (struct vnet_route_info *)Vnet_Malloc(sizeof(struct vnet_route_info));
+
+ if (!new_route) {
+ PrintError("Cannot allocate new route\n");
+ return -1;
+ }
+
memset(new_route, 0, sizeof(struct vnet_route_info));
#ifdef V3_CONFIG_DEBUG_VNET
- Vnet_Debug("VNET/P Core: add_route_entry:\n");
+ PrintDebug("VNET/P Core: add_route_entry:\n");
print_route(&route);
#endif
flags = vnet_lock_irqsave(vnet_state.lock);
list_add(&(new_route->node), &(vnet_state.routes));
- clear_hash_cache();
-
+ new_route->idx = ++ vnet_state.route_idx;
+ vnet_state.num_routes ++;
+
vnet_unlock_irqrestore(vnet_state.lock, flags);
-
+
+ clear_hash_cache();
#ifdef V3_CONFIG_DEBUG_VNET
dump_routes();
#endif
- return 0;
+ return new_route->idx;
}
-/* delete all route entries with specfied src or dst device id */
-static void inline del_routes_by_dev(int dev_id){
+void v3_vnet_del_route(uint32_t route_idx){
struct vnet_route_info * route = NULL;
- unsigned long flags;
+ vnet_intr_flags_t flags;
flags = vnet_lock_irqsave(vnet_state.lock);
list_for_each_entry(route, &(vnet_state.routes), node) {
+ Vnet_Print(0, "v3_vnet_del_route, route idx: %d\n", route->idx);
+ if(route->idx == route_idx){
+ list_del(&(route->node));
+ Vnet_Free(route);
+ break;
+ }
+ }
+
+ vnet_unlock_irqrestore(vnet_state.lock, flags);
+ clear_hash_cache();
+
+#ifdef V3_CONFIG_DEBUG_VNET
+ dump_routes();
+#endif
+}
+
+
+/* delete all route entries with specfied src or dst device id */
+static void inline del_routes_by_dev(int dev_id){
+ struct vnet_route_info * route, *tmp_route;
+ vnet_intr_flags_t flags;
+
+ flags = vnet_lock_irqsave(vnet_state.lock);
+
+ list_for_each_entry_safe(route, tmp_route, &(vnet_state.routes), node) {
if((route->route_def.dst_type == LINK_INTERFACE &&
route->route_def.dst_id == dev_id) ||
(route->route_def.src_type == LINK_INTERFACE &&
Vnet_Free(route);
}
}
-
+
vnet_unlock_irqrestore(vnet_state.lock, flags);
}
-/* At the end allocate a route_list
- * This list will be inserted into the cache so we don't need to free it
- */
-static struct route_list * match_route(const struct v3_vnet_pkt * pkt) {
+
+// Match classes, must be in order
+#define NUM_MATCH_CLASSES 4
+#define NUM_MATCH_CLASSES_BOUND 3
+#define NONE 0
+#define NOT 1
+#define ANY 2
+#define DIRECT 3
+
+
+static inline uint8_t match_mac(uint8_t test_mac[ETH_ALEN],
+ uint8_t route_mac[ETH_ALEN],
+ uint8_t route_qual)
+{
+ switch (route_qual) {
+ case MAC_NOSET:
+ return NONE;
+ break;
+ case MAC_NONE:
+ return NONE;
+ break;
+ case MAC_ANY:
+ return ANY;
+ break;
+ case MAC_NOT:
+ if (memcmp(test_mac,route_mac,ETH_ALEN)) {
+ return NOT;
+ } else {
+ return NONE;
+ }
+ break;
+ case MAC_ADDR:
+ if (memcmp(test_mac,route_mac,ETH_ALEN)) {
+ return NONE;
+ } else {
+ return DIRECT;
+ }
+ break;
+ default:
+ PrintError("Unknown qualifier %u\n",route_qual);
+ return NONE;
+ break;
+ }
+
+}
+
+#define QUAL_TO_STR(q) ( \
+(q)==MAC_NOSET ? "MAC_NOSET" : \
+(q)==MAC_NONE ? "MAC_NONE" : \
+(q)==MAC_ANY ? "MAC_ANY" : \
+(q)==MAC_NOT ? "MAC_NOT" : \
+(q)==MAC_ADDR ? "MAC_ADDR" : \
+"***UNDEFINED****" \
+ ) \
+
+#define MATCH_CLASS_TO_STR(c) ( \
+(c)==NONE ? "NONE" : \
+(c)==NOT ? "NOT" : \
+(c)==ANY ? "ANY" : \
+(c)==DIRECT ? "DIRECT" : \
+"***UNDEFINED****" \
+ ) \
+
+
+
+/*
+
+Original priority behavior...
+
+priority src srcqual dst dstqual
+3 ANY ANY
+4 X NONE
+5 ANY X NOT
+5 X NOT ANY
+6 X ~NOT ANY
+6 ANY X ~NOT
+7 X ~NOT X NOT
+7 X NOT X ~NOT
+8 X ~NOT X ~NOT
+8 X ~NOT X ~NOT
+
+*/
+
+/*
+ Current priority order is given in the following table
+*/
+
+// [src][dst] => priority
+static int priority_map[NUM_MATCH_CLASSES][NUM_MATCH_CLASSES] =
+{
+ [NONE] = { [ 0 ... NUM_MATCH_CLASSES_BOUND ] = -1}, // ignore if it's not a source match
+ [NOT][NONE] = -1, // ignore it if there is no destination match
+ [NOT][NOT] = 3,
+ [NOT][ANY] = 5,
+ [NOT][DIRECT] = 7,
+ [ANY][NONE] = -1, // ignore if there is no destination match
+ [ANY][NOT] = 5,
+ [ANY][ANY] = 6,
+ [ANY][DIRECT] = 6,
+ [DIRECT][NONE] = -1, // ignore if there is no destination match
+ [DIRECT][NOT] = 7,
+ [DIRECT][ANY] = 8,
+ [DIRECT][DIRECT] = 8,
+};
+
+
+
+
+static inline int match_priority(uint8_t src_mac[ETH_ALEN],
+ uint8_t dst_mac[ETH_ALEN],
+ uint8_t route_src_mac[ETH_ALEN],
+ uint8_t route_src_qual,
+ uint8_t route_dst_mac[ETH_ALEN],
+ uint8_t route_dst_qual)
+
+{
+
+ return priority_map[match_mac(src_mac,route_src_mac,route_src_qual)][match_mac(dst_mac,route_dst_mac,route_dst_qual)];
+}
+
+
+/*
+ Route matching will return the list of the highest priority routes that
+ match. It's a list because it's possible to have multiple high priority routes
+ */
+static struct route_list * match_route(const struct v3_vnet_pkt * pkt)
+{
+ int i;
struct vnet_route_info * route = NULL;
struct route_list * matches = NULL;
int num_matches = 0;
- int max_rank = 0;
+ int max_priority = -1;
struct list_head match_list;
struct eth_hdr * hdr = (struct eth_hdr *)(pkt->data);
- // uint8_t src_type = pkt->src_type;
- // uint32_t src_link = pkt->src_id;
+ //
+ //
+ // NOTE: USING THE MATCH_NODE in the route list to record a match list
+ // IS A DISASTER WAITING TO HAPPEN
+ //
+
#ifdef V3_CONFIG_DEBUG_VNET
{
- char dst_str[100];
- char src_str[100];
-
- mac_to_string(hdr->src_mac, src_str);
- mac_to_string(hdr->dst_mac, dst_str);
- Vnet_Debug("VNET/P Core: match_route. pkt: SRC(%s), DEST(%s)\n", src_str, dst_str);
+ char dst_str[32], src_str[32];
+ mac2str(hdr->src_mac, src_str);
+ mac2str(hdr->dst_mac, dst_str);
+ PrintDebug("VNET/P Core: match_route. pkt: SRC(%s), DEST(%s)\n", src_str, dst_str);
}
#endif
-
- INIT_LIST_HEAD(&match_list);
-#define UPDATE_MATCHES(rank) do { \
- if (max_rank < (rank)) { \
- max_rank = (rank); \
- INIT_LIST_HEAD(&match_list); \
- \
- list_add(&(route->match_node), &match_list); \
- num_matches = 1; \
- } else if (max_rank == (rank)) { \
- list_add(&(route->match_node), &match_list); \
- num_matches++; \
- } \
- } while (0)
+ INIT_LIST_HEAD(&match_list);
+
-
list_for_each_entry(route, &(vnet_state.routes), node) {
+
struct v3_vnet_route * route_def = &(route->route_def);
+
+ int priority;
+
+ priority = match_priority(hdr->src_mac,
+ hdr->dst_mac,
+ route_def->src_mac,
+ route_def->src_mac_qual,
+ route_def->dst_mac,
+ route_def->dst_mac_qual);
-/*
- // CHECK SOURCE TYPE HERE
- if ( (route_def->src_type != LINK_ANY) &&
- ( (route_def->src_type != src_type) ||
- ( (route_def->src_id != src_link) &&
- (route_def->src_id != -1)))) {
- continue;
+
+
+#ifdef V3_CONFIG_DEBUG_VNET
+ {
+ char dst_str[32];
+ char src_str[32];
+
+ mac2str(route_def->src_mac, src_str);
+ mac2str(route_def->dst_mac, dst_str);
+
+ PrintDebug("Tested match against SRC(%s) SRC_QUAL(%s), DEST(%s) DST_QUAL(%s): "
+ "SRC_MATCH=%s DEST_MATCH=%s PRIORITY=%d\n",
+ src_str, QUAL_TO_STR(route_def->src_mac_qual),
+ dst_str, QUAL_TO_STR(route_def->dst_mac_qual),
+ MATCH_CLASS_TO_STR(match_mac(hdr->src_mac,route_def->src_mac,route_def->src_mac_qual)),
+ MATCH_CLASS_TO_STR(match_mac(hdr->dst_mac,route_def->dst_mac,route_def->dst_mac_qual)),
+ priority);
}
-*/
+#endif
- if ((route_def->dst_mac_qual == MAC_ANY) &&
- (route_def->src_mac_qual == MAC_ANY)) {
- UPDATE_MATCHES(3);
+ if (priority<0) {
+ PrintDebug("No match to this rule\n");
+ continue;
}
-
- if (memcmp(route_def->src_mac, hdr->src_mac, 6) == 0) {
- if (route_def->src_mac_qual != MAC_NOT) {
- if (route_def->dst_mac_qual == MAC_ANY) {
- UPDATE_MATCHES(6);
- } else if (route_def->dst_mac_qual != MAC_NOT &&
- memcmp(route_def->dst_mac, hdr->dst_mac, 6) == 0) {
- UPDATE_MATCHES(8);
- }
+
+ if (priority > max_priority) {
+ PrintDebug("New highest priority match, reseting list\n");
+ max_priority = priority;
+
+ struct vnet_route_info *my_route, *tmp_route;
+
+ list_for_each_entry_safe(my_route, tmp_route, &match_list,match_node) {
+ list_del(&(my_route->match_node));
}
- }
+
+ list_add(&(route->match_node), &match_list);
+ num_matches = 1;
- if (memcmp(route_def->dst_mac, hdr->dst_mac, 6) == 0) {
- if (route_def->dst_mac_qual != MAC_NOT) {
- if (route_def->src_mac_qual == MAC_ANY) {
- UPDATE_MATCHES(6);
- } else if ((route_def->src_mac_qual != MAC_NOT) &&
- (memcmp(route_def->src_mac, hdr->src_mac, 6) == 0)) {
- UPDATE_MATCHES(8);
- }
- }
- }
+ } else if (priority == max_priority) {
+ PrintDebug("Equal priority match, adding to list\n");
- if ((route_def->dst_mac_qual == MAC_NOT) &&
- (memcmp(route_def->dst_mac, hdr->dst_mac, 6) != 0)) {
- if (route_def->src_mac_qual == MAC_ANY) {
- UPDATE_MATCHES(5);
- } else if ((route_def->src_mac_qual != MAC_NOT) &&
- (memcmp(route_def->src_mac, hdr->src_mac, 6) == 0)) {
- UPDATE_MATCHES(7);
- }
- }
-
- if ((route_def->src_mac_qual == MAC_NOT) &&
- (memcmp(route_def->src_mac, hdr->src_mac, 6) != 0)) {
- if (route_def->dst_mac_qual == MAC_ANY) {
- UPDATE_MATCHES(5);
- } else if ((route_def->dst_mac_qual != MAC_NOT) &&
- (memcmp(route_def->dst_mac, hdr->dst_mac, 6) == 0)) {
- UPDATE_MATCHES(7);
- }
- }
+ list_add(&(route->match_node), &match_list);
+ num_matches++;
+ }
- // Default route
- if ( (memcmp(route_def->src_mac, hdr->src_mac, 6) == 0) &&
- (route_def->dst_mac_qual == MAC_NONE)) {
- UPDATE_MATCHES(4);
- }
}
- Vnet_Debug("VNET/P Core: match_route: Matches=%d\n", num_matches);
+ PrintDebug("VNET/P Core: match_route: Matches=%d\n", num_matches);
- if (num_matches == 0) {
+ if (num_matches <= 0) {
return NULL;
}
-
+
matches = (struct route_list *)Vnet_Malloc(sizeof(struct route_list) +
(sizeof(struct vnet_route_info *) * num_matches));
+
+ if (!matches) {
+ PrintError("VNET/P Core: Unable to allocate matches\n");
+ return NULL;
+ }
+
matches->num_routes = num_matches;
- {
- int i = 0;
- list_for_each_entry(route, &match_list, match_node) {
+ i=0;
+ list_for_each_entry(route, &match_list, match_node) {
+ if (i==num_matches) {
+ // the list should never have more than num_matches on it...
+ PrintError("Weird list behavior\n");
+ break;
+ } else {
matches->routes[i++] = route;
}
+
}
return matches;
}
+int v3_vnet_query_header(uint8_t src_mac[ETH_ALEN],
+ uint8_t dest_mac[ETH_ALEN],
+ int recv, // 0 = send, 1=recv
+ struct v3_vnet_header *header)
+{
+ struct route_list *routes;
+ struct vnet_route_info *r;
+ struct v3_vnet_pkt p;
+ void *flags;
+
+ p.size=14;
+ p.data=p.header;
+ memcpy(p.header,dest_mac,6);
+ memcpy(p.header+6,src_mac,6);
+ memset(p.header+12,0,2);
+
+ p.src_type = LINK_EDGE;
+ p.src_id = 0;
+
+ memcpy(header->src_mac,src_mac,6);
+ memcpy(header->dst_mac,dest_mac,6);
+
+
+ flags = vnet_lock_irqsave(vnet_state.lock);
+
+ look_into_cache(&p,&routes);
+
+ if (!routes) {
+ routes = match_route(&p);
+ if (!routes) {
+ vnet_unlock_irqrestore(vnet_state.lock,flags);
+ PrintError("Cannot match route\n");
+ header->header_type=VNET_HEADER_NOMATCH;
+ header->header_len=0;
+ return -1;
+ } else {
+ add_route_to_cache(&p,routes);
+ }
+ }
+
+ vnet_unlock_irqrestore(vnet_state.lock,flags);
+
+ if (routes->num_routes<1) {
+ PrintError("Less than one route\n");
+ header->header_type=VNET_HEADER_NOMATCH;
+ header->header_len=0;
+ return -1;
+ }
+
+ if (routes->num_routes>1) {
+ PrintError("More than one route, building header for the first one only\n");
+ }
+
+ r=routes->routes[0];
+
+ switch (r->route_def.dst_type) {
+ case LINK_EDGE: {
+ // switch based on the link type
+ // for mac-in-udp, we would want to generate a mac, ip, and udp header
+ // direct transmission
+
+ // for now we will say we have no encapsulation
+ //
+ header->header_type=VNET_HEADER_NONE;
+ header->header_len=0;
+ header->src_mac_qual=r->route_def.src_mac_qual;
+ header->dst_mac_qual=r->route_def.dst_mac_qual;
+
+ }
+
+ return 0;
+ break;
+
+
+ case LINK_INTERFACE:
+ // direct transmission
+ // let's guess that it goes to the same interface...
+ header->header_type=VNET_HEADER_NONE;
+ header->header_len=0;
+ header->src_mac_qual=r->route_def.src_mac_qual;
+ header->dst_mac_qual=r->route_def.dst_mac_qual;
+
+ return 0;
+ break;
-int vnet_tx_one_pkt(struct v3_vnet_pkt * pkt, void * private_data) {
+ default:
+ PrintError("Unknown destination type\n");
+ return -1;
+ break;
+
+ }
+
+}
+
+
+
+
+int v3_vnet_send_pkt(struct v3_vnet_pkt * pkt, void * private_data) {
struct route_list * matched_routes = NULL;
- unsigned long flags;
+ vnet_intr_flags_t flags;
int i;
int cpu = V3_Get_CPU();
+
Vnet_Print(2, "VNET/P Core: cpu %d: pkt (size %d, src_id:%d, src_type: %d, dst_id: %d, dst_type: %d)\n",
- cpu, pkt->size, pkt->src_id,
- pkt->src_type, pkt->dst_id, pkt->dst_type);
- if(vnet_debug >= 4){
- v3_hexdump(pkt->data, pkt->size, NULL, 0);
+ cpu, pkt->size, pkt->src_id,
+ pkt->src_type, pkt->dst_id, pkt->dst_type);
+
+ if(net_debug >= 4){
+ v3_hexdump(pkt->data, pkt->size, NULL, 0);
}
flags = vnet_lock_irqsave(vnet_state.lock);
vnet_state.stats.rx_pkts++;
look_into_cache(pkt, &matched_routes);
+
if (matched_routes == NULL) {
- Vnet_Debug("VNET/P Core: send pkt Looking into routing table\n");
+ PrintDebug("VNET/P Core: sending pkt - matching route\n");
matched_routes = match_route(pkt);
if (matched_routes) {
add_route_to_cache(pkt, matched_routes);
} else {
- Vnet_Debug("VNET/P Core: Could not find route for packet... discards packet\n");
+ PrintDebug("VNET/P Core: Could not find route for packet... discarding packet\n");
vnet_unlock_irqrestore(vnet_state.lock, flags);
return 0; /* do we return -1 here?*/
}
vnet_unlock_irqrestore(vnet_state.lock, flags);
- Vnet_Debug("VNET/P Core: send pkt route matches %d\n", matched_routes->num_routes);
+ PrintDebug("VNET/P Core: send pkt route matches %d\n", matched_routes->num_routes);
for (i = 0; i < matched_routes->num_routes; i++) {
struct vnet_route_info * route = matched_routes->routes[i];
}
-static int vnet_pkt_enqueue(struct v3_vnet_pkt * pkt){
- unsigned long flags;
- struct queue_entry * entry;
- struct vnet_queue * q = &(vnet_state.pkt_q);
- uint16_t num_pages;
-
- flags = vnet_lock_irqsave(q->lock);
-
- if (q->count >= VNET_QUEUE_SIZE){
- Vnet_Print(1, "VNET Queue overflow!\n");
- vnet_unlock_irqrestore(q->lock, flags);
- return -1;
- }
-
- q->count ++;
- entry = &(q->buf[q->tail++]);
- q->tail %= VNET_QUEUE_SIZE;
-
- vnet_unlock_irqrestore(q->lock, flags);
-
- /* this is ugly, but should happen very unlikely */
- while(entry->use);
-
- if(entry->size_alloc < pkt->size){
- if(entry->data != NULL){
- Vnet_FreePages(Vnet_PAddr(entry->data), (entry->size_alloc / PAGE_SIZE));
- entry->data = NULL;
- }
-
- num_pages = 1 + (pkt->size / PAGE_SIZE);
- entry->data = Vnet_VAddr(Vnet_AllocPages(num_pages));
- if(entry->data == NULL){
- return -1;
- }
- entry->size_alloc = PAGE_SIZE * num_pages;
- }
-
- entry->pkt.data = entry->data;
- memcpy(&(entry->pkt), pkt, sizeof(struct v3_vnet_pkt));
- memcpy(entry->data, pkt->data, pkt->size);
-
- entry->use = 1;
-
- return 0;
-}
-
-
-int v3_vnet_send_pkt(struct v3_vnet_pkt * pkt, void * private_data, int synchronize) {
- if(synchronize){
- vnet_tx_one_pkt(pkt, NULL);
- }else {
- vnet_pkt_enqueue(pkt);
- Vnet_Print(2, "VNET/P Core: Put pkt into Queue: pkt size %d\n", pkt->size);
- }
-
- return 0;
-}
-
int v3_vnet_add_dev(struct v3_vm_info * vm, uint8_t * mac,
- struct v3_vnet_dev_ops *ops,
+ struct v3_vnet_dev_ops * ops, int quote, int poll_state,
void * priv_data){
struct vnet_dev * new_dev = NULL;
- unsigned long flags;
+ vnet_intr_flags_t flags;
new_dev = (struct vnet_dev *)Vnet_Malloc(sizeof(struct vnet_dev));
if (new_dev == NULL) {
- Vnet_Print(0, "Malloc fails\n");
+ Vnet_Print(0, "VNET/P Core: Unable to allocate a new device\n");
return -1;
}
memcpy(new_dev->mac_addr, mac, 6);
new_dev->dev_ops.input = ops->input;
+ new_dev->dev_ops.poll = ops->poll;
new_dev->private_data = priv_data;
new_dev->vm = vm;
new_dev->dev_id = 0;
+ new_dev->quote = quote<VNET_MAX_QUOTE ? quote : VNET_MAX_QUOTE;
+ new_dev->poll = poll_state;
flags = vnet_lock_irqsave(vnet_state.lock);
if (dev_by_mac(mac) == NULL) {
list_add(&(new_dev->node), &(vnet_state.devs));
- new_dev->dev_id = ++vnet_state.num_devs;
+ new_dev->dev_id = ++ vnet_state.dev_idx;
+ vnet_state.num_devs ++;
+
+ if(new_dev->poll) {
+ v3_enqueue(vnet_state.poll_devs, (addr_t)new_dev);
+ }
+ } else {
+ PrintError("VNET/P: Device with the same MAC has already been added\n");
}
vnet_unlock_irqrestore(vnet_state.lock, flags);
return -1;
}
- Vnet_Debug("VNET/P Core: Add Device: dev_id %d\n", new_dev->dev_id);
+ PrintDebug("VNET/P Core: Add Device: dev_id %d\n", new_dev->dev_id);
return new_dev->dev_id;
}
int v3_vnet_del_dev(int dev_id){
struct vnet_dev * dev = NULL;
- unsigned long flags;
+ vnet_intr_flags_t flags;
flags = vnet_lock_irqsave(vnet_state.lock);
dev = dev_by_id(dev_id);
if (dev != NULL){
list_del(&(dev->node));
- del_routes_by_dev(dev_id);
+ //del_routes_by_dev(dev_id);
+ vnet_state.num_devs --;
}
vnet_unlock_irqrestore(vnet_state.lock, flags);
Vnet_Free(dev);
- Vnet_Debug("VNET/P Core: Remove Device: dev_id %d\n", dev_id);
+ PrintDebug("VNET/P Core: Removed Device: dev_id %d\n", dev_id);
return 0;
}
int v3_vnet_stat(struct vnet_stat * stats){
-
stats->rx_bytes = vnet_state.stats.rx_bytes;
stats->rx_pkts = vnet_state.stats.rx_pkts;
stats->tx_bytes = vnet_state.stats.tx_bytes;
return 0;
}
-static void free_devices(){
- struct vnet_dev * dev = NULL;
+static void deinit_devices_list(){
+ struct vnet_dev * dev, * tmp;
- list_for_each_entry(dev, &(vnet_state.devs), node) {
+ list_for_each_entry_safe(dev, tmp, &(vnet_state.devs), node) {
list_del(&(dev->node));
Vnet_Free(dev);
}
}
-static void free_routes(){
- struct vnet_route_info * route = NULL;
+static void deinit_routes_list(){
+ struct vnet_route_info * route, * tmp;
- list_for_each_entry(route, &(vnet_state.routes), node) {
+ list_for_each_entry_safe(route, tmp, &(vnet_state.routes), node) {
list_del(&(route->node));
list_del(&(route->match_node));
Vnet_Free(route);
struct v3_vnet_bridge_ops * ops,
uint8_t type,
void * priv_data) {
- unsigned long flags;
+ vnet_intr_flags_t flags;
int bridge_free = 0;
struct vnet_brg_dev * tmp_bridge = NULL;
tmp_bridge = (struct vnet_brg_dev *)Vnet_Malloc(sizeof(struct vnet_brg_dev));
if (tmp_bridge == NULL) {
- PrintError("Malloc Fails\n");
+ PrintError("VNET/P Core: Unable to allocate new bridge\n");
vnet_state.bridge = NULL;
return -1;
}
return 0;
}
-static int vnet_tx_flush(void *args){
- unsigned long flags;
- struct queue_entry * entry;
- struct vnet_queue * q = &(vnet_state.pkt_q);
- Vnet_Print(0, "VNET/P Handing Pkt Thread Starting ....\n");
+void v3_vnet_del_bridge(uint8_t type) {
+ vnet_intr_flags_t flags;
+ struct vnet_brg_dev * tmp_bridge = NULL;
+
+ flags = vnet_lock_irqsave(vnet_state.lock);
+
+ if (vnet_state.bridge != NULL && vnet_state.bridge->type == type) {
+ tmp_bridge = vnet_state.bridge;
+ vnet_state.bridge = NULL;
+ }
+
+ vnet_unlock_irqrestore(vnet_state.lock, flags);
+
+ if (tmp_bridge) {
+ Vnet_Free(tmp_bridge);
+ }
+}
+
+
+/* can be instanieoued to multiple threads
+ * that runs on multiple cores
+ * or it could be running on a dedicated side core
+ */
+static int vnet_tx_flush(void * args){
+ struct vnet_dev * dev = NULL;
+ int more;
+ int rc;
+ uint64_t noprogress_count;
- /* we need thread sleep/wakeup in Palacios */
- while(!vnet_thread_should_stop()){
- flags = vnet_lock_irqsave(q->lock);
+ Vnet_Print(0, "VNET/P Polling Thread Starting ....\n");
- if (q->count <= 0){
- vnet_unlock_irqrestore(q->lock, flags);
- Vnet_Yield();
- }else {
- q->count --;
- entry = &(q->buf[q->head++]);
- q->head %= VNET_QUEUE_SIZE;
+ // since there are multiple instances of this thread, and only
+ // one queue of pollable devices, our model here will be to synchronize
+ // on that queue, removing devices as we go, and keeping them
+ // then putting them back on the queue when we are done
+ // in this way, multiple instances of this function will never
+ // be polling the same device at the same time
- vnet_unlock_irqrestore(q->lock, flags);
+ struct v3_queue * tq = v3_create_queue();
- /* this is ugly, but should happen very unlikely */
- while(!entry->use);
- vnet_tx_one_pkt(&(entry->pkt), NULL);
+ if (!tq) {
+ PrintError("VNET/P polling thread cannot allocate queue\n");
+ return -1;
+ }
- /* asynchronizely release allocated memory for buffer entry here */
- entry->use = 0;
+ noprogress_count=0;
- Vnet_Print(2, "vnet_tx_flush: pkt (size %d)\n", entry->pkt.size);
+ while (!vnet_thread_should_stop()) {
+
+ more=0; // will indicate if any device has more work for us to do
+
+ while ((dev = (struct vnet_dev *)v3_dequeue(vnet_state.poll_devs))) {
+ // we are handling this device
+ v3_enqueue(tq,(addr_t)dev);
+
+ if (dev->poll && dev->dev_ops.poll) {
+ // The device's poll function MUST NOT BLOCK
+ rc = dev->dev_ops.poll(dev->vm, dev->quote, dev->private_data);
+
+ if (rc<0) {
+ Vnet_Print(0, "VNET/P: poll from device %p error (ignoring) !\n", dev);
+ } else {
+ more |= rc;
+ }
+ }
+ }
+
+ while ((dev = (struct vnet_dev *)v3_dequeue(tq))) {
+ // now someone else can handle it
+ v3_enqueue(vnet_state.poll_devs, (addr_t)dev);
+ }
+
+
+ if (more) {
+ noprogress_count=0;
+ } else {
+ if ( ! ((noprogress_count+1) < noprogress_count)) {
+ noprogress_count++;
+ }
+ }
+
+ // adaptively yield
+ if (noprogress_count < VNET_NOPROGRESS_LIMIT) {
+ V3_Yield();
+ } else {
+ V3_Yield_Timed(VNET_YIELD_USEC);
}
+
}
+ Vnet_Free(tq);
+
+ Vnet_Print(0, "VNET/P Polling Thread Done.\n");
+
return 0;
}
vnet_state.num_routes = 0;
if (vnet_lock_init(&(vnet_state.lock)) == -1){
- PrintError("VNET/P Core: Fails to initiate lock\n");
+ PrintError("VNET/P: Fails to initiate lock\n");
}
vnet_state.route_cache = vnet_create_htable(0, &hash_fn, &hash_eq);
if (vnet_state.route_cache == NULL) {
- PrintError("VNET/P Core: Fails to initiate route cache\n");
+ PrintError("VNET/P: Fails to initiate route cache\n");
return -1;
}
- vnet_lock_init(&(vnet_state.pkt_q.lock));
+ vnet_state.poll_devs = v3_create_queue();
- vnet_state.pkt_flush_thread = vnet_start_thread(vnet_tx_flush, NULL, "VNET_Pkts");
+ vnet_state.pkt_flush_thread = vnet_start_thread(vnet_tx_flush, NULL, "vnetd-1");
- Vnet_Debug("VNET/P Core is initiated\n");
+ PrintDebug("VNET/P is initiated\n");
return 0;
}
-void v3_deinit_vnet(){
+void v3_deinit_vnet() {
+
+ v3_deinit_queue(vnet_state.poll_devs);
+ Vnet_Free(vnet_state.poll_devs);
+
+ PrintDebug("Stopping flush thread\n");
+ // This will pause until the flush thread is gone
+ vnet_thread_stop(vnet_state.pkt_flush_thread);
+ // At this point there should be no lock-holder
+
+ Vnet_Free(vnet_state.poll_devs);
- vnet_lock_deinit(&(vnet_state.lock));
- free_devices();
- free_routes();
+ PrintDebug("Deiniting Device List\n");
+ // close any devices we have open
+ deinit_devices_list();
+
+ PrintDebug("Deiniting Route List\n");
+ // remove any routes we have
+ deinit_routes_list();
+ PrintDebug("Freeing hash table\n");
+ // remove the hash table
vnet_free_htable(vnet_state.route_cache, 1, 1);
- Vnet_Free(vnet_state.bridge);
+
+
+ PrintDebug("Removing Bridge\n");
+ // remove bridge if it was added
+ if (vnet_state.bridge) {
+ Vnet_Free(vnet_state.bridge);
+ }
+
+ PrintDebug("Deleting lock\n");
+ // eliminate the lock
+ vnet_lock_deinit(&(vnet_state.lock));
+
}