uint8_t mac_addr[6];
struct v3_vm_info * vm;
- int (*input)(struct v3_vm_info * vm, struct v3_vnet_pkt * pkt, void * private_data);
+ int (*input)(struct v3_vm_info * vm, struct v3_vnet_pkt * pkt, void * private_data);
void * private_data;
int dev_id;
} __attribute__((packed));
+#define BRIDGE_BUF_SIZE 1024
+struct bridge_pkts_buf {
+ int start, end;
+ int num;
+ v3_lock_t lock;
+ struct v3_vnet_pkt pkts[BRIDGE_BUF_SIZE];
+ uint8_t datas[ETHERNET_PACKET_LEN*BRIDGE_BUF_SIZE];
+};
+
+struct vnet_brg_dev {
+ struct v3_vm_info * vm;
+
+ int (*input)(struct v3_vm_info * vm, struct v3_vnet_pkt pkt[], uint16_t pkt_num, void * private_data);
+ void (*xcall_input)(void *data);
+
+ struct bridge_pkts_buf recv_buf; //packets from Vnet to vnet_bridge device
+
+ struct bridge_pkts_buf send_buf; //packets from vnet_bridge device to Vnet
+
+ int disabled;
+
+ uint16_t max_delayed_pkts;
+ long max_latency; //in cycles
+ void * private_data;
+} __attribute__((packed));
+
+
};
-
-
struct route_list {
uint8_t hash_buf[VNET_HASH_SIZE];
static struct {
struct list_head routes;
struct list_head devs;
-
+
int num_routes;
int num_devs;
+ struct vnet_brg_dev *bridge;
+
v3_lock_t lock;
struct gen_queue * inpkt_q;
#ifdef CONFIG_DEBUG_VNET
static inline void mac_to_string(char mac[6], char * buf) {
- snprintf(buf, 50, "%x:%x:%x:%x:%x:%x",
+ snprintf(buf, 100, "%d:%d:%d:%d:%d:%d",
mac[0], mac[1], mac[2],
mac[3], mac[4], mac[5]);
}
PrintDebug("Src dev id (%d), src type (%d)",
route->route_def.src_id,
route->route_def.src_type);
- PrintDebug("Dst dev id (%d), dst type (%d)",
+ PrintDebug("Dst dev id (%d), dst type (%d)\n",
route->route_def.dst_id,
route->route_def.dst_type);
- PrintDebug("dst_dev (%p), dst_dev_id (%d), dst_dev_input (%p), dst_dev_data (%p)\n",
+ if (route->route_def.dst_type == LINK_INTERFACE) {
+ PrintDebug("dst_dev (%p), dst_dev_id (%d), dst_dev_input (%p), dst_dev_data (%p)\n",
route->dst_dev,
route->dst_dev->dev_id,
route->dst_dev->input,
route->dst_dev->private_data);
-
+ }
}
static void dump_routes(){
* This means we can generate the hash from an offset into the pkt struct
*/
static inline uint_t hash_fn(addr_t hdr_ptr) {
- uint8_t * hdr_buf = (uint8_t *)&(hdr_ptr);
-
+ uint8_t * hdr_buf = (uint8_t *)hdr_ptr;
+
return v3_hash_buffer(hdr_buf, VNET_HASH_SIZE);
}
-static inline int hash_eq(addr_t key1, addr_t key2) {
+static inline int hash_eq(addr_t key1, addr_t key2) {
return (memcmp((uint8_t *)key1, (uint8_t *)key2, VNET_HASH_SIZE) == 0);
}
-
-static int add_route_to_cache(struct v3_vnet_pkt * pkt, struct route_list * routes) {
+static int add_route_to_cache(const struct v3_vnet_pkt * pkt, struct route_list * routes) {
memcpy(routes->hash_buf, pkt->hash_buf, VNET_HASH_SIZE);
if (v3_htable_insert(vnet_state.route_cache, (addr_t)routes->hash_buf, (addr_t)routes) == 0) {
return 0;
}
-static int look_into_cache(struct v3_vnet_pkt * pkt, struct route_list ** routes) {
+static int look_into_cache(const struct v3_vnet_pkt * pkt, struct route_list ** routes) {
- *routes = (struct route_list *)v3_htable_search(vnet_state.route_cache, (addr_t)pkt);
+ *routes = (struct route_list *)v3_htable_search(vnet_state.route_cache, (addr_t)(pkt->hash_buf));
return 0;
}
}
flags = v3_lock_irqsave(vnet_state.lock);
+
list_add(&(new_route->node), &(vnet_state.routes));
+ clear_hash_cache();
+
v3_unlock_irqrestore(vnet_state.lock, flags);
- clear_hash_cache();
#ifdef CONFIG_DEBUG_VNET
dump_routes();
// At the end allocate a route_list
// This list will be inserted into the cache so we don't need to free it
-static struct route_list * match_route(struct v3_vnet_pkt * pkt) {
+static struct route_list * match_route(const struct v3_vnet_pkt * pkt) {
struct vnet_route_info * route = NULL;
struct route_list * matches = NULL;
int num_matches = 0;
#ifdef CONFIG_DEBUG_VNET
{
- char dst_str[50];
- char src_str[50];
+ char dst_str[100];
+ char src_str[100];
mac_to_string(hdr->src_mac, src_str);
mac_to_string(hdr->dst_mac, dst_str);
return NULL;
}
- matches = V3_Malloc(sizeof(struct route_list) +
- (sizeof(struct vnet_route_info *) * num_matches));
+ matches = (struct route_list *)V3_Malloc(sizeof(struct route_list) +
+ (sizeof(struct vnet_route_info *) * num_matches));
matches->num_routes = num_matches;
return matches;
}
-static int handle_one_pkt(struct v3_vnet_pkt * pkt, void *private_data) {
+
+static int flush_bridge_pkts(struct vnet_brg_dev *bridge){
+ unsigned long flags;
+ int num, start, send;
+ struct v3_vnet_bridge_input_args args;
+ int cpu_id = bridge->vm->cores[0].cpu_id;
+ int current_core = V3_Get_CPU();
+
+ if (bridge == NULL) {
+ PrintDebug("VNET: No bridge to sent data to links\n");
+ return -1;
+ }
+
+ flags = v3_lock_irqsave(bridge->recv_buf.lock);
+
+ num = bridge->recv_buf.num;
+ start = bridge->recv_buf.start;
+
+ bridge->recv_buf.num -= num;
+ bridge->recv_buf.start += num;
+ bridge->recv_buf.start %= BRIDGE_BUF_SIZE;
+
+ v3_unlock_irqrestore(bridge->recv_buf.lock, flags);
+
+
+ if(bridge->disabled){
+ PrintDebug("VNET: In flush bridge pkts: Bridge is disabled\n");
+ return -1;
+ }
+
+ if(num <= 2 && num > 0){
+ PrintDebug("VNET: In flush bridge pkts: %d\n", num);
+ }
+
+ if(num > 0) {
+ PrintDebug("VNET: In flush bridge pkts to bridge, cur_cpu %d, brige_core: %d\n", current_core, cpu_id);
+ if (current_core == cpu_id){
+ if ((start + num) < BRIDGE_BUF_SIZE){
+ bridge->input(bridge->vm, &(bridge->recv_buf.pkts[start]), num, bridge->private_data);
+ }else {
+ bridge->input(bridge->vm, &(bridge->recv_buf.pkts[start]), (BRIDGE_BUF_SIZE - start), bridge->private_data);
+ send = num - (BRIDGE_BUF_SIZE - start);
+ bridge->input(bridge->vm, &(bridge->recv_buf.pkts[0]), send, bridge->private_data);
+ }
+ }else {
+ args.vm = bridge->vm;
+ args.private_data = bridge->private_data;
+
+ if ((start + num) < BRIDGE_BUF_SIZE){
+ args.pkt_num = num;
+ args.vnet_pkts = &(bridge->recv_buf.pkts[start]);
+ V3_Call_On_CPU(cpu_id, bridge->xcall_input, (void *)&args);
+ }else {
+ args.pkt_num = BRIDGE_BUF_SIZE - start;
+ args.vnet_pkts = &(bridge->recv_buf.pkts[start]);
+ V3_Call_On_CPU(cpu_id, bridge->xcall_input, (void *)&args);
+
+ send = num - (BRIDGE_BUF_SIZE - start);
+ args.pkt_num = send;
+ args.vnet_pkts = &(bridge->recv_buf.pkts[0]);
+ V3_Call_On_CPU(cpu_id, bridge->xcall_input, (void *)&args);
+ }
+ }
+
+ PrintDebug("VNET: flush bridge pkts %d\n", num);
+ }
+
+ return 0;
+}
+
+
+static int send_to_bridge(struct v3_vnet_pkt * pkt){
+ struct vnet_brg_dev *bridge = vnet_state.bridge;
+ int cpu_id = bridge->vm->cores[0].cpu_id;
+ struct v3_vnet_bridge_input_args args;
+
+ if (bridge == NULL) {
+ PrintDebug("VNET: No bridge to sent data to links\n");
+ return -1;
+ }
+
+ if(bridge->max_delayed_pkts <= 1){
+ if(bridge->disabled){
+ PrintDebug("VNET: Bridge diabled\n");
+ return -1;
+ }
+
+ args.pkt_num = 1;
+ args.vm = bridge->vm;
+ args.vnet_pkts = pkt;
+ args.private_data = bridge->private_data;
+
+ V3_Call_On_CPU(cpu_id, bridge->xcall_input, (void *)&args);
+ PrintDebug("VNET: sent one packet to the bridge\n");
+ return 0;
+ }
+
+ unsigned long flags;
+ int end, num=0;
+ struct v3_vnet_pkt *buf;
+
+ PrintDebug("VNET: send_to_bridge\n");
+
+ flags = v3_lock_irqsave(bridge->recv_buf.lock);
+
+ if(bridge->disabled && bridge->recv_buf.num >= BRIDGE_BUF_SIZE){
+ PrintDebug("Bridge diabled and bridge receive buffer full\n");
+ v3_unlock_irqrestore(bridge->recv_buf.lock, flags);//maybe should move this after copy
+ num = bridge->recv_buf.num;
+ goto exit;
+ }
+
+ end = bridge->recv_buf.end;
+ buf = &(bridge->recv_buf.pkts[end]);
+
+ bridge->recv_buf.num ++;
+ bridge->recv_buf.end ++;
+ bridge->recv_buf.end %= BRIDGE_BUF_SIZE;
+
+ num = bridge->recv_buf.num;
+
+ v3_unlock_irqrestore(bridge->recv_buf.lock, flags);//maybe should move this after copy
+
+
+ buf->size = pkt->size;
+ buf->dst_id = pkt->dst_id;
+ buf->src_id = pkt->src_id;
+ buf->src_type = pkt->src_type;
+ buf->dst_type = pkt->dst_type;
+ memcpy(buf->header, pkt->header, ETHERNET_HEADER_LEN);
+ memcpy(buf->data, pkt->data, pkt->size);
+
+exit:
+
+ if (num >= bridge->max_delayed_pkts){
+ flush_bridge_pkts(bridge);
+ }
+
+ return 0;
+}
+
+int v3_vnet_send_pkt(struct v3_vnet_pkt * pkt, void * private_data) {
struct route_list * matched_routes = NULL;
unsigned long flags;
int i;
-
#ifdef CONFIG_DEBUG_VNET
{
- struct eth_hdr * hdr = (struct eth_hdr *)(pkt->data);
- char dest_str[30];
- char src_str[30];
+ struct eth_hdr * hdr = (struct eth_hdr *)(pkt->header);
+ char dest_str[100];
+ char src_str[100];
mac_to_string(hdr->src_mac, src_str);
mac_to_string(hdr->dst_mac, dest_str);
- PrintDebug("Vnet: HandleDataOverLink. SRC(%s), DEST(%s)\n", src_str, dest_str);
+ int cpu = V3_Get_CPU();
+ PrintDebug("Vnet: on cpu %d, HandleDataOverLink. SRC(%s), DEST(%s), pkt size: %d\n", cpu, src_str, dest_str, pkt->size);
}
#endif
-#ifdef CONFIG_VNET_PROFILE
- struct guest_info *core = (struct guest_info *)private_data;
- uint64_t start, end;
- rdtscll(start);
-#endif
-
flags = v3_lock_irqsave(vnet_state.lock);
look_into_cache(pkt, &matched_routes);
if (matched_routes == NULL) {
+ PrintDebug("Vnet: send pkt Looking into routing table\n");
+
matched_routes = match_route(pkt);
-
+
if (matched_routes) {
add_route_to_cache(pkt, matched_routes);
} else {
- PrintDebug("Could not find route for packet...\n");
+ PrintDebug("Could not find route for packet... discards packet\n");
v3_unlock_irqrestore(vnet_state.lock, flags);
return -1;
}
v3_unlock_irqrestore(vnet_state.lock, flags);
-#ifdef CONFIG_VNET_PROFILE
- {
- rdtscll(end);
- core->vnet_times.time_route_lookup = end - start;
- }
-#endif
+ PrintDebug("Vnet: send pkt route matches %d\n", matched_routes->num_routes);
- PrintDebug("Vnet: HandleOnePacket: %d\n", matched_routes->num_routes);
for (i = 0; i < matched_routes->num_routes; i++) {
struct vnet_route_info * route = matched_routes->routes[i];
if (route->route_def.dst_type == LINK_EDGE) {
+ pkt->dst_type = LINK_EDGE;
+ pkt->dst_id = route->route_def.dst_id;
+ if (send_to_bridge(pkt) == -1) {
+ PrintDebug("VNET: Packet not sent properly to bridge\n");
+ continue;
+ }
+
} else if (route->route_def.dst_type == LINK_INTERFACE) {
if (route->dst_dev->input(route->dst_dev->vm, pkt, route->dst_dev->private_data) == -1) {
PrintDebug("VNET: Packet not sent properly\n");
continue;
}
- PrintDebug("Vnet: HandleOnePacket: Forward packet according to Route\n");
- }
-
-#ifdef CONFIG_VNET_PROFILE
- {
- rdtscll(start);
- core->vnet_times.time_copy_to_guest = start - end;
+ PrintDebug("Vnet: v3_vnet_send_pkt: Forward packet according to Route %d\n", i);
}
-#endif
return 0;
}
-int v3_vnet_send_pkt(struct v3_vnet_pkt * pkt, void *private_data) {
- PrintDebug("In Vnet Send: pkt size: %d\n", pkt->size);
+void v3_vnet_send_pkt_xcall(void * data){
+ struct v3_vnet_pkt * pkt = (struct v3_vnet_pkt *)data;
+ v3_vnet_send_pkt(pkt, NULL);
+}
+
+
+void v3_vnet_bridge_polling()
+{
+ unsigned long flags;
+ int num, start;
+ struct v3_vnet_pkt *buf;
+ struct vnet_brg_dev *bridge = vnet_state.bridge;
+
+ PrintDebug("In vnet bridge pollling: cpu %d\n", V3_Get_CPU());
+
+ if(bridge == NULL){
+ PrintDebug("VNET: Bridge is not set\n");
+ return;
+ }
+
+ flags = v3_lock_irqsave(bridge->send_buf.lock);
- if (handle_one_pkt(pkt, private_data) != -1) {
- PrintDebug("VNET: send one packet! pt length %d\n", pkt->size);
- } else {
- PrintDebug("VNET: Fail to forward one packet, discard it!\n");
+ num = bridge->send_buf.num;
+ start = bridge->send_buf.start;
+
+ PrintDebug("VNET: bridge polling pkts %d\n", num);
+
+ while(num > 0) {
+ buf = &(bridge->send_buf.pkts[bridge->send_buf.start]);
+
+ v3_vnet_send_pkt(buf, NULL);
+
+ bridge->send_buf.num --;
+ bridge->send_buf.start ++;
+ bridge->send_buf.start %= BRIDGE_BUF_SIZE;
+ num --;
+ }
+
+ v3_unlock_irqrestore(bridge->send_buf.lock, flags);
+
+ return;
+}
+
+
+int v3_vnet_bridge_rx(uchar_t *buf, uint16_t size, uint16_t src_link){
+ struct vnet_brg_dev *bridge = vnet_state.bridge;
+ unsigned long flags;
+ int end;
+ struct v3_vnet_pkt *pkt;
+
+ if (bridge == NULL) {
+ PrintDebug("VNET: No bridge is set\n");
+ return -1;
+ }
+
+ flags = v3_lock_irqsave(bridge->send_buf.lock);
+
+ end = bridge->send_buf.end;
+ pkt = &(bridge->send_buf.pkts[end]);
+
+ if(bridge->send_buf.num > BRIDGE_BUF_SIZE){
+ PrintDebug("VNET: bridge rx: buffer full\n");
+ goto exit;
}
+ bridge->send_buf.num ++;
+ bridge->send_buf.end ++;
+ bridge->send_buf.end %= BRIDGE_BUF_SIZE;
+
+ pkt->size = size;
+ pkt->src_id = src_link;
+ pkt->src_type = LINK_EDGE;
+ memcpy(pkt->header, buf, ETHERNET_HEADER_LEN);
+ memcpy(pkt->data, buf, size);
+
+exit:
+
+ v3_unlock_irqrestore(bridge->send_buf.lock, flags);
+
return 0;
}
+
-int v3_vnet_add_dev(struct v3_vm_info *vm,uint8_t mac[6],
+int v3_vnet_add_dev(struct v3_vm_info *vm, uint8_t mac[6],
int (*netif_input)(struct v3_vm_info * vm, struct v3_vnet_pkt * pkt, void * private_data),
void * priv_data){
struct vnet_dev * new_dev = NULL;
unsigned long flags;
- int dev_id;
-
- flags = v3_lock_irqsave(vnet_state.lock);
-
- new_dev = find_dev_by_mac(mac);
- if (new_dev) {
- PrintDebug("VNET: register device: Already has device with the same mac\n");
- dev_id = -1;
- goto exit;
- }
-
new_dev = (struct vnet_dev *)V3_Malloc(sizeof(struct vnet_dev));
if (new_dev == NULL) {
PrintError("VNET: Malloc fails\n");
- dev_id = -1;
- goto exit;
+ return -1;
}
memcpy(new_dev->mac_addr, mac, 6);
new_dev->input = netif_input;
new_dev->private_data = priv_data;
new_dev->vm = vm;
-
- list_add(&(new_dev->node), &(vnet_state.devs));
- vnet_state.num_devs ++;
- new_dev->dev_id = vnet_state.num_devs;
- dev_id = new_dev->dev_id;
+ new_dev->dev_id = 0;
+
+ flags = v3_lock_irqsave(vnet_state.lock);
+
+ if (!find_dev_by_mac(mac)) {
+ list_add(&(new_dev->node), &(vnet_state.devs));
+ new_dev->dev_id = ++vnet_state.num_devs;
+ }
+
+ v3_unlock_irqrestore(vnet_state.lock, flags);
+
+ // if the device was found previosly the id should still be 0
+ if (new_dev->dev_id == 0) {
+ PrintError("Device Alrady exists\n");
+ return -1;
+ }
PrintDebug("Vnet: Add Device: dev_id %d, input : %p, private_data %p\n",
new_dev->dev_id, new_dev->input, new_dev->private_data);
-exit:
+ return new_dev->dev_id;
+}
+
+
+void v3_vnet_heartbeat(struct guest_info *core){
+ static long last_time, cur_time;
+
+ if(vnet_state.bridge == NULL)
+ return;
+
+ if(vnet_state.bridge->max_delayed_pkts <= 1)
+ return;
+
+ if(V3_Get_CPU() != vnet_state.bridge->vm->cores[0].cpu_id){
+ rdtscll(cur_time);
+ }
+
+ if ((cur_time - last_time) >= vnet_state.bridge->max_latency) {
+ last_time = cur_time;
+ flush_bridge_pkts(vnet_state.bridge);
+ }
+}
+
+int v3_vnet_add_bridge(struct v3_vm_info * vm,
+ int (*input)(struct v3_vm_info * vm, struct v3_vnet_pkt pkt[], uint16_t pkt_num, void * private_data),
+ void (*xcall_input)(void *data),
+ uint16_t max_delayed_pkts,
+ long max_latency,
+ void * priv_data) {
+ unsigned long flags;
+ int bridge_free = 0;
+ struct vnet_brg_dev * tmp_bridge = NULL;
+ int i;
+
+
+ flags = v3_lock_irqsave(vnet_state.lock);
+
+ if (vnet_state.bridge == NULL) {
+ bridge_free = 1;
+ vnet_state.bridge = (void *)1;
+ }
+
+ v3_unlock_irqrestore(vnet_state.lock, flags);
+
+ if (bridge_free == 0) {
+ PrintError("Bridge already set\n");
+ return -1;
+ }
+
+ tmp_bridge = (struct vnet_brg_dev *)V3_Malloc(sizeof(struct vnet_brg_dev));
+
+ if (tmp_bridge == NULL) {
+ PrintError("Malloc Fails\n");
+ return -1;
+ }
+
+ tmp_bridge->vm = vm;
+ tmp_bridge->input = input;
+ tmp_bridge->xcall_input = xcall_input;
+ tmp_bridge->private_data = priv_data;
+ tmp_bridge->disabled = 0;
+
+ //initial receving buffer
+ tmp_bridge->recv_buf.start = 0;
+ tmp_bridge->recv_buf.end = 0;
+ tmp_bridge->recv_buf.num = 0;
+ if(v3_lock_init(&(tmp_bridge->recv_buf.lock)) == -1){
+ PrintError("VNET: add bridge, error to initiate recv buf lock\n");
+ }
+ tmp_bridge->max_delayed_pkts = (max_delayed_pkts<BRIDGE_BUF_SIZE)?max_delayed_pkts : BRIDGE_BUF_SIZE;
+ tmp_bridge->max_latency = max_latency;
+ for(i = 0; i<BRIDGE_BUF_SIZE; i++){
+ tmp_bridge->recv_buf.pkts[i].data = &(tmp_bridge->recv_buf.datas[i*ETHERNET_PACKET_LEN]);
+ }
+
+ //initial sending buffer
+ tmp_bridge->send_buf.start = 0;
+ tmp_bridge->send_buf.end = 0;
+ tmp_bridge->send_buf.num = 0;
+ if(v3_lock_init(&(tmp_bridge->send_buf.lock)) == -1){
+ PrintError("VNET: add bridge, error to initiate send buf lock\n");
+ }
+ for(i = 0; i<BRIDGE_BUF_SIZE; i++){
+ tmp_bridge->send_buf.pkts[i].data = &(tmp_bridge->send_buf.datas[i*ETHERNET_PACKET_LEN]);
+ }
+ // make this atomic to avoid possible race conditions
+ flags = v3_lock_irqsave(vnet_state.lock);
+ vnet_state.bridge = tmp_bridge;
v3_unlock_irqrestore(vnet_state.lock, flags);
-
- return dev_id;
+
+ return 0;
+}
+
+
+int v3_vnet_disable_bridge() {
+ unsigned long flags;
+
+ flags = v3_lock_irqsave(vnet_state.lock);
+
+ if (vnet_state.bridge != NULL) {
+ vnet_state.bridge->disabled = 1;
+ }
+
+ v3_unlock_irqrestore(vnet_state.lock, flags);
+
+ return 0;
}
+int v3_vnet_enable_bridge() {
+ unsigned long flags;
+
+ flags = v3_lock_irqsave(vnet_state.lock);
+
+ if (vnet_state.bridge != NULL) {
+ vnet_state.bridge->disabled = 0;
+ }
+
+ v3_unlock_irqrestore(vnet_state.lock, flags);
+
+ return 0;
+}
+
+
+
int V3_init_vnet() {
INIT_LIST_HEAD(&(vnet_state.routes));