X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Fsrc%2Fdevices%2Flnx_virtio_nic.c;h=ad0b9e9d369e4b66f3a439977ae6235635bc3553;hb=d775bbfa668ce9968bacc0e4257cf86e5ab88e90;hp=58e9e58ce14e2c2fcea4555530a9e7f5d66b9a08;hpb=262c92c9f620f87500f350b4744d2f31b6de7da5;p=palacios.git diff --git a/palacios/src/devices/lnx_virtio_nic.c b/palacios/src/devices/lnx_virtio_nic.c index 58e9e58..ad0b9e9 100644 --- a/palacios/src/devices/lnx_virtio_nic.c +++ b/palacios/src/devices/lnx_virtio_nic.c @@ -1,3 +1,4 @@ + /* * This file is part of the Palacios Virtual Machine Monitor developed * by the V3VEE Project with funding from the United States National @@ -7,13 +8,13 @@ * and the University of New Mexico. You can find out more at * http://www.v3vee.org * - * Copyright (c) 2008, Lei Xia - * Copyright (c) 2008, Cui Zheng - * Copyright (c) 2008, The V3VEE Project + * Copyright (c) 2010, Lei Xia + * Copyright (c) 2010, Cui Zheng + * Copyright (c) 2010, The V3VEE Project * All rights reserved. * * Author: Lei Xia - * Cui Zheng + * Cui Zheng * * * This is free software. You are permitted to use, @@ -25,86 +26,138 @@ #include #include #include -#include +#include #include - +#include #include +#include +#include -#ifndef CONFIG_DEBUG_VIRTIO_NET +#ifndef V3_CONFIG_DEBUG_VIRTIO_NET #undef PrintDebug #define PrintDebug(fmt, args...) #endif -#define VIRTIO_NIC_PROFILE +#ifndef V3_CONFIG_VNET +static int net_debug = 0; +#endif + +#define TX_QUEUE_SIZE 4096 +#define RX_QUEUE_SIZE 4096 +#define CTRL_QUEUE_SIZE 64 + +/* The feature bitmap for virtio nic + * from Linux */ +#define VIRTIO_NET_F_CSUM 0 /* Host handles pkts w/ partial csum */ +#define VIRTIO_NET_F_GUEST_CSUM 1 /* Guest handles pkts w/ partial csum */ +#define VIRTIO_NET_F_MAC 5 /* Host has given MAC address. */ +#define VIRTIO_NET_F_GSO 6 /* Host handles pkts w/ any GSO type */ +#define VIRTIO_NET_F_GUEST_TSO4 7 /* Guest can handle TSOv4 in. */ +#define VIRTIO_NET_F_GUEST_TSO6 8 /* Guest can handle TSOv6 in. */ +#define VIRTIO_NET_F_GUEST_ECN 9 /* Guest can handle TSO[6] w/ ECN in. */ +#define VIRTIO_NET_F_GUEST_UFO 10 /* Guest can handle UFO in. */ +#define VIRTIO_NET_F_HOST_TSO4 11 /* Host can handle TSOv4 in. */ +#define VIRTIO_NET_F_HOST_TSO6 12 /* Host can handle TSOv6 in. */ +#define VIRTIO_NET_F_HOST_ECN 13 /* Host can handle TSO[6] w/ ECN in. */ +#define VIRTIO_NET_F_HOST_UFO 14 /* Host can handle UFO in. */ +#define VIRTIO_NET_F_MRG_RXBUF 15 /* Host can merge receive buffers. */ +#define VIRTIO_NET_F_STATUS 16 /* virtio_net_config.status available */ + +/* Port to get virtio config */ +#define VIRTIO_NET_CONFIG 20 -#define VIRTIO_NET_S_LINK_UP 1 /* Link is up */ #define VIRTIO_NET_MAX_BUFSIZE (sizeof(struct virtio_net_hdr) + (64 << 10)) -struct virtio_net_hdr { - uint8_t flags; +/* for gso_type in virtio_net_hdr */ +#define VIRTIO_NET_HDR_GSO_NONE 0 +#define VIRTIO_NET_HDR_GSO_TCPV4 1 /* GSO frame, IPv4 TCP (TSO) */ +#define VIRTIO_NET_HDR_GSO_UDP 3 /* GSO frame, IPv4 UDP (UFO) */ +#define VIRTIO_NET_HDR_GSO_TCPV6 4 /* GSO frame, IPv6 TCP */ +#define VIRTIO_NET_HDR_GSO_ECN 0x80 /* TCP has ECN set */ + + +/* for flags in virtio_net_hdr */ +#define VIRTIO_NET_HDR_F_NEEDS_CSUM 1 /* Use csum_start, csum_offset */ + -#define VIRTIO_NET_HDR_GSO_NONE 0 /* Not a GSO frame */ - uint8_t gso_type; - uint16_t hdr_len; /* Ethernet + IP + tcp/udp hdrs */ - uint16_t gso_size; /* Bytes to append to hdr_len per frame */ - uint16_t csum_start; /* Position to start checksumming from */ - uint16_t csum_offset; /* Offset after that to place checksum */ +/* First element of the scatter-gather list, used with GSO or CSUM features */ +struct virtio_net_hdr +{ + uint8_t flags; + uint8_t gso_type; + uint16_t hdr_len; /* Ethernet + IP + tcp/udp hdrs */ + uint16_t gso_size; /* Bytes to append to hdr_len per frame */ + uint16_t csum_start; /* Position to start checksumming from */ + uint16_t csum_offset; /* Offset after that to place checksum */ }__attribute__((packed)); - -#define QUEUE_SIZE 1024 -#define CTRL_QUEUE_SIZE 64 -#define ETH_ALEN 6 + +/* The header to use when the MRG_RXBUF + * feature has been negotiated. */ +struct virtio_net_hdr_mrg_rxbuf { + struct virtio_net_hdr hdr; + uint16_t num_buffers; /* Number of merged rx buffers */ +}; struct virtio_net_config { - uint8_t mac[ETH_ALEN]; //VIRTIO_NET_F_MAC + uint8_t mac[ETH_ALEN]; /* VIRTIO_NET_F_MAC */ uint16_t status; } __attribute__((packed)); struct virtio_dev_state { + struct vm_device * pci_bus; struct list_head dev_list; - struct guest_info * vm; + struct v3_vm_info *vm; + + enum {GUEST_DRIVEN=0, VMM_DRIVEN, ADAPTIVE} model; + uint64_t lower_thresh_pps, upper_thresh_pps, period_us; + + uint8_t mac[ETH_ALEN]; }; struct virtio_net_state { + struct virtio_net_config net_cfg; struct virtio_config virtio_cfg; + struct v3_vm_info * vm; struct vm_device * dev; struct pci_device * pci_dev; int io_range_size; + + uint16_t status; - struct virtio_queue rx_vq; //index 0, rvq in Linux virtio driver, handle packet to guest - struct virtio_queue tx_vq; //index 1, svq in Linux virtio driver, handle packet from guest - struct virtio_queue ctrl_vq; //index 2, ctrol info from guest + struct virtio_queue rx_vq; /* idx 0*/ + struct virtio_queue tx_vq; /* idx 1*/ + struct virtio_queue ctrl_vq; /* idx 2*/ + + uint8_t mergeable_rx_bufs; - ulong_t pkt_sent, pkt_recv, pkt_drop; - struct gen_queue * inpkt_q; + struct v3_timer * timer; + struct nic_statistics stats; struct v3_dev_net_ops * net_ops; + v3_lock_t rx_lock, tx_lock; + + uint8_t tx_notify, rx_notify; + uint32_t tx_pkts, rx_pkts; + uint64_t past_us; void * backend_data; struct virtio_dev_state * virtio_dev; struct list_head dev_link; }; -#define ETHERNET_PACKET_LEN 1514 -struct eth_pkt { - uint32_t size; //size of data - char data[ETHERNET_PACKET_LEN]; -}__attribute__((packed)); - -static int virtio_free(struct vm_device * dev) -{ - - return -1; -} static int virtio_init_state(struct virtio_net_state * virtio) { + virtio->rx_vq.queue_size = RX_QUEUE_SIZE; + virtio->tx_vq.queue_size = TX_QUEUE_SIZE; + virtio->ctrl_vq.queue_size = CTRL_QUEUE_SIZE; + virtio->rx_vq.ring_desc_addr = 0; virtio->rx_vq.ring_avail_addr = 0; virtio->rx_vq.ring_used_addr = 0; @@ -123,179 +176,87 @@ static int virtio_init_state(struct virtio_net_state * virtio) virtio->ctrl_vq.pfn = 0; virtio->ctrl_vq.cur_avail_idx = 0; - virtio->virtio_cfg.host_features = 0; - //virtio->virtio_cfg.status = VIRTIO_NET_S_LINK_UP; virtio->virtio_cfg.pci_isr = 0; - virtio->pkt_sent = virtio->pkt_recv = virtio->pkt_drop = 0; + virtio->mergeable_rx_bufs = 1; + + virtio->virtio_cfg.host_features = 0 | (1 << VIRTIO_NET_F_MAC); + if(virtio->mergeable_rx_bufs) { + virtio->virtio_cfg.host_features |= (1 << VIRTIO_NET_F_MRG_RXBUF); + } - virtio->inpkt_q = v3_create_queue(); - v3_init_queue(virtio->inpkt_q); + if ((v3_lock_init(&(virtio->rx_lock)) == -1) || + (v3_lock_init(&(virtio->tx_lock)) == -1)){ + PrintError(VM_NONE, VCORE_NONE, "Virtio NIC: Failure to init locks for net_state\n"); + } return 0; } -static int pkt_tx(struct virtio_net_state * virtio, struct vring_desc * buf_desc) +static int virtio_deinit_state(struct guest_info *core, struct virtio_net_state *ns) { - uint8_t * buf = NULL; - uint32_t len = buf_desc->length; - - PrintDebug("Virtio NIC: Handling Virtio Write, net_state: %p\n", virtio); - - if (guest_pa_to_host_va(virtio->virtio_dev->vm, buf_desc->addr_gpa, (addr_t *)&(buf)) == -1) { - PrintError("Could not translate buffer address\n"); - return -1; + if (ns->timer) { + v3_remove_timer(core,ns->timer); } - if (virtio->net_ops->send(buf, len, (void *)virtio, NULL) == -1) { - return -1; - } - - return 0; -} + v3_lock_deinit(&(ns->rx_lock)); + v3_lock_deinit(&(ns->tx_lock)); -static int build_receive_header(struct virtio_net_hdr * hdr, const void * buf, int raw) { - hdr->flags = 0; - - if (!raw) { - memcpy(hdr, buf, sizeof(struct virtio_net_hdr)); - } else { - memset(hdr, 0, sizeof(struct virtio_net_hdr)); - } return 0; } -static int copy_data_to_desc(struct virtio_net_state * virtio_state, struct vring_desc * desc, uchar_t * buf, uint_t buf_len) +static int tx_one_pkt(struct guest_info * core, + struct virtio_net_state * virtio, + struct vring_desc * buf_desc) { - uint32_t len; - uint8_t * desc_buf = NULL; + uint8_t * buf = NULL; + uint32_t len = buf_desc->length; - if (guest_pa_to_host_va(virtio_state->virtio_dev->vm, desc->addr_gpa, (addr_t *)&(desc_buf)) == -1) { - PrintError("Could not translate buffer address\n"); + if (v3_gpa_to_hva(core, buf_desc->addr_gpa, (addr_t *)&(buf)) == -1) { + PrintDebug(core->vm_info, core, "Could not translate buffer address\n"); return -1; } - len = (desc->length < buf_len)?desc->length:buf_len; - memcpy(desc_buf, buf, len); - - return len; -} - -//send data to guest -int send_pkt_to_guest(struct virtio_net_state * virtio, uchar_t * buf, uint_t size, int raw, void * private_data) -{ - struct virtio_queue * q = &(virtio->rx_vq); - struct virtio_net_hdr hdr; - uint32_t hdr_len = sizeof(struct virtio_net_hdr); - uint32_t data_len = size; - uint32_t offset = 0; - - PrintDebug("VIRTIO NIC: sending packet to virtio nic %p, size:%d", virtio, size); - - virtio->pkt_recv ++; - if (!raw) { - data_len -= hdr_len; + + V3_Net_Print(2, "Virtio-NIC: virtio_tx: size: %d\n", len); + if(net_debug >= 4){ + v3_hexdump(buf, len, NULL, 0); } - build_receive_header(&hdr, buf, 1); - - if (q->ring_avail_addr == 0) { - PrintError("Queue is not set\n"); + if(virtio->net_ops->send(buf, len, virtio->backend_data) < 0){ + virtio->stats.tx_dropped ++; return -1; } - - if (q->last_avail_idx > q->avail->index) - q->idx_overflow = true; - q->last_avail_idx = q->avail->index; - - if (q->cur_avail_idx < q->avail->index || (q->idx_overflow && q->cur_avail_idx < q->avail->index+65536)){ - addr_t hdr_addr = 0; - uint16_t hdr_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; - uint16_t buf_idx = 0; - struct vring_desc * hdr_desc = NULL; - - hdr_desc = &(q->desc[hdr_idx]); - if (guest_pa_to_host_va(virtio->virtio_dev->vm, hdr_desc->addr_gpa, &(hdr_addr)) == -1) { - PrintError("Could not translate receive buffer address\n"); - return -1; - } - - memcpy((void *)hdr_addr, &hdr, sizeof(struct virtio_net_hdr)); - if (offset >= data_len) { - hdr_desc->flags &= ~VIRTIO_NEXT_FLAG; - } - - for (buf_idx = hdr_desc->next; offset < data_len; buf_idx = q->desc[hdr_idx].next) { - struct vring_desc * buf_desc = &(q->desc[buf_idx]); - uint32_t len = 0; - - len = copy_data_to_desc(virtio, buf_desc, buf + offset, data_len - offset); - offset += len; - if (offset < data_len) { - buf_desc->flags = VIRTIO_NEXT_FLAG; - } - buf_desc->length = len; - } - - q->used->ring[q->used->index % q->queue_size].id = q->avail->ring[q->cur_avail_idx % q->queue_size]; - q->used->ring[q->used->index % q->queue_size].length = data_len + hdr_len; // This should be the total length of data sent to guest (header+pkt_data) - q->used->index++; - - int last_idx = q->cur_avail_idx; - q->cur_avail_idx++; - if (q->cur_avail_idx < last_idx) - q->idx_overflow = false; - } else { - virtio->pkt_drop++; - -#ifdef VIRTIO_NIC_PROFILE - PrintError("Virtio NIC: %p, one pkt dropped receieved: %ld, dropped: %ld, sent: %ld curidx: %d, avaiIdx: %d\n", - virtio, virtio->pkt_recv, virtio->pkt_drop, virtio->pkt_sent, q->cur_avail_idx, q->avail->index); -#endif - } - - if (!(q->avail->flags & VIRTIO_NO_IRQ_FLAG)) { - PrintDebug("Raising IRQ %d\n", virtio->pci_dev->config_header.intr_line); - v3_pci_raise_irq(virtio->virtio_dev->pci_bus, 0, virtio->pci_dev); - virtio->virtio_cfg.pci_isr = 0x1; - } - -#ifdef VIRTIO_NIC_PROFILE - if ((virtio->pkt_recv % 10000) == 0){ - PrintError("Virtio NIC: %p, receieved: %ld, dropped: %ld, sent: %ld\n", - virtio, virtio->pkt_recv, virtio->pkt_drop, virtio->pkt_sent); - } -#endif - - return offset; + virtio->stats.tx_pkts ++; + virtio->stats.tx_bytes += len; + + return 0; } -static int virtio_sendto_buf(struct virtio_net_state * net_state, uchar_t * buf, uint_t size) { - struct eth_pkt *pkt; - pkt = (struct eth_pkt *)V3_Malloc(sizeof(struct eth_pkt)); - if(pkt == NULL){ - PrintError("Virtio NIC: Memory allocate fails\n"); - return -1; +/*copy data into ring buffer */ +static inline int copy_data_to_desc(struct guest_info * core, + struct virtio_net_state * virtio_state, + struct vring_desc * desc, + uchar_t * buf, + uint_t buf_len, + uint_t dst_offset){ + uint32_t len; + uint8_t * desc_buf = NULL; + + if (v3_gpa_to_hva(core, desc->addr_gpa, (addr_t *)&(desc_buf)) == -1) { + PrintDebug(core->vm_info, core, "Could not translate buffer address\n"); + return -1; } - - pkt->size = size; - memcpy(pkt->data, buf, size); - v3_enqueue(net_state->inpkt_q, (addr_t)pkt); - - PrintDebug("Virtio NIC: __virtio_sendto_buf: transmitting packet: (size:%d)\n", (int)pkt->size); + len = (desc->length < (buf_len+dst_offset))?(desc->length - dst_offset):buf_len; + memcpy(desc_buf + dst_offset, buf, len); - return pkt->size; + return len; } -int virtio_dev_send(uchar_t * buf, uint32_t size, void *private_data) { - struct virtio_net_state *virtio_state = (struct virtio_net_state *)private_data; - - return virtio_sendto_buf(virtio_state, buf, size); -} -static int get_desc_count(struct virtio_queue * q, int index) { +static inline int get_desc_count(struct virtio_queue * q, int index) { struct vring_desc * tmp_desc = &(q->desc[index]); int cnt = 1; @@ -307,73 +268,100 @@ static int get_desc_count(struct virtio_queue * q, int index) { return cnt; } -static int handle_ctrl(struct virtio_net_state * dev) { - return 0; +static inline void enable_cb(struct virtio_queue *queue){ + if(queue->used){ + queue->used->flags &= ~ VRING_NO_NOTIFY_FLAG; + } +} + +static inline void disable_cb(struct virtio_queue *queue) { + if(queue->used){ + queue->used->flags |= VRING_NO_NOTIFY_FLAG; + } } -static int handle_pkt_tx(struct virtio_net_state * virtio_state) +static int handle_pkt_tx(struct guest_info * core, + struct virtio_net_state * virtio_state, + int quote) { - struct virtio_queue * q = &(virtio_state->tx_vq); - struct virtio_net_hdr * hdr = NULL; + struct virtio_queue * q; + int txed = 0, left = 0; + unsigned long flags; - if (q->avail->index < q->last_avail_idx) - q->idx_overflow = true; - q->last_avail_idx = q->avail->index; + q = &(virtio_state->tx_vq); + if (!q->ring_avail_addr) { + return -1; + } - while (q->cur_avail_idx < q->avail->index || - (q->idx_overflow && q->cur_avail_idx < (q->avail->index + 65536))) { + while (1) { struct vring_desc * hdr_desc = NULL; addr_t hdr_addr = 0; - uint16_t desc_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; - int desc_cnt = get_desc_count(q, desc_idx); - uint32_t req_len = 0; - int i = 0; + uint16_t desc_idx, tmp_idx; + int desc_cnt; + + flags = v3_lock_irqsave(virtio_state->tx_lock); - hdr_desc = &(q->desc[desc_idx]); - if (guest_pa_to_host_va(virtio_state->virtio_dev->vm, hdr_desc->addr_gpa, &(hdr_addr)) == -1) { - PrintError("Could not translate block header address\n"); - return -1; + if(q->cur_avail_idx == q->avail->index || + (quote > 0 && txed >= quote)) { + left = (q->cur_avail_idx != q->avail->index); + v3_unlock_irqrestore(virtio_state->tx_lock, flags); + break; } - - hdr = (struct virtio_net_hdr*)hdr_addr; - desc_idx = hdr_desc->next; - for (i = 0; i < desc_cnt - 1; i++) { - struct vring_desc * buf_desc = &(q->desc[desc_idx]); - if (pkt_tx(virtio_state, buf_desc) == -1) { - PrintError("Error handling nic operation\n"); - return -1; - } + desc_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; + tmp_idx = q->cur_avail_idx ++; + + v3_unlock_irqrestore(virtio_state->tx_lock, flags); - req_len += buf_desc->length; - desc_idx = buf_desc->next; + desc_cnt = get_desc_count(q, desc_idx); + if(desc_cnt != 2){ + PrintError(core->vm_info, core, "VNIC: merged rx buffer not supported, desc_cnt %d\n", desc_cnt); } - virtio_state->pkt_sent ++; - q->used->ring[q->used->index % q->queue_size].id = q->avail->ring[q->cur_avail_idx % q->queue_size]; - q->used->ring[q->used->index % q->queue_size].length = req_len; // What do we set this to???? - q->used->index++; + hdr_desc = &(q->desc[desc_idx]); + if (v3_gpa_to_hva(core, hdr_desc->addr_gpa, &(hdr_addr)) != -1) { + struct virtio_net_hdr_mrg_rxbuf * hdr; + struct vring_desc * buf_desc; + + hdr = (struct virtio_net_hdr_mrg_rxbuf *)hdr_addr; + desc_idx = hdr_desc->next; - int last_idx = q->cur_avail_idx; - q->cur_avail_idx ++; - if (q->cur_avail_idx < last_idx) - q->idx_overflow = false; - } + /* here we assumed that one ethernet pkt is not splitted into multiple buffer */ + buf_desc = &(q->desc[desc_idx]); + if (tx_one_pkt(core, virtio_state, buf_desc) == -1) { + PrintError(core->vm_info, core, "Virtio NIC: Fails to send packet\n"); + } + } else { + PrintError(core->vm_info, core, "Could not translate block header address\n"); + } + + flags = v3_lock_irqsave(virtio_state->tx_lock); + + q->used->ring[q->used->index % q->queue_size].id = + q->avail->ring[tmp_idx % q->queue_size]; + + q->used->index ++; + + v3_unlock_irqrestore(virtio_state->tx_lock, flags); - if (!(q->avail->flags & VIRTIO_NO_IRQ_FLAG)) { - v3_pci_raise_irq(virtio_state->virtio_dev->pci_bus, 0, virtio_state->pci_dev); + txed ++; + } + + if (txed && !(q->avail->flags & VIRTIO_NO_IRQ_FLAG)) { + v3_pci_raise_irq(virtio_state->virtio_dev->pci_bus, + virtio_state->pci_dev, 0); virtio_state->virtio_cfg.pci_isr = 0x1; + virtio_state->stats.rx_interrupts ++; } -#ifdef VIRTIO_NIC_PROFILE - if(virtio_state->pkt_sent % 10000 == 0) - PrintError("Virtio NIC: %p, pkt_sent: %ld\n", virtio_state, virtio_state->pkt_sent); -#endif - - return 0; + return left; } -static int virtio_setup_queue(struct virtio_net_state * virtio_state, struct virtio_queue * queue, addr_t pfn, addr_t page_addr) { + +static int virtio_setup_queue(struct guest_info *core, + struct virtio_net_state * virtio_state, + struct virtio_queue * queue, + addr_t pfn, addr_t page_addr) { queue->pfn = pfn; queue->ring_desc_addr = page_addr; @@ -381,56 +369,58 @@ static int virtio_setup_queue(struct virtio_net_state * virtio_state, struct vir queue->ring_used_addr = ((queue->ring_avail_addr) + (sizeof(struct vring_avail)) + (queue->queue_size * sizeof(uint16_t))); - + // round up to next page boundary. queue->ring_used_addr = (queue->ring_used_addr + 0xfff) & ~0xfff; - if (guest_pa_to_host_va(virtio_state->virtio_dev->vm, queue->ring_desc_addr, (addr_t *)&(queue->desc)) == -1) { - PrintError("Could not translate ring descriptor address\n"); + if (v3_gpa_to_hva(core, queue->ring_desc_addr, (addr_t *)&(queue->desc)) == -1) { + PrintError(core->vm_info, core, "Could not translate ring descriptor address\n"); return -1; } - if (guest_pa_to_host_va(virtio_state->virtio_dev->vm, queue->ring_avail_addr, (addr_t *)&(queue->avail)) == -1) { - PrintError("Could not translate ring available address\n"); + if (v3_gpa_to_hva(core, queue->ring_avail_addr, (addr_t *)&(queue->avail)) == -1) { + PrintError(core->vm_info, core, "Could not translate ring available address\n"); return -1; } - if (guest_pa_to_host_va(virtio_state->virtio_dev->vm, queue->ring_used_addr, (addr_t *)&(queue->used)) == -1) { - PrintError("Could not translate ring used address\n"); + if (v3_gpa_to_hva(core, queue->ring_used_addr, (addr_t *)&(queue->used)) == -1) { + PrintError(core->vm_info, core, "Could not translate ring used address\n"); return -1; } - PrintDebug("RingDesc_addr=%p, Avail_addr=%p, Used_addr=%p\n", + PrintDebug(core->vm_info, core, "RingDesc_addr=%p, Avail_addr=%p, Used_addr=%p\n", (void *)(queue->ring_desc_addr), (void *)(queue->ring_avail_addr), (void *)(queue->ring_used_addr)); - PrintDebug("RingDesc=%p, Avail=%p, Used=%p\n", + PrintDebug(core->vm_info, core, "RingDesc=%p, Avail=%p, Used=%p\n", queue->desc, queue->avail, queue->used); return 0; } -static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, uint_t length, void * private_data) +static int virtio_io_write(struct guest_info *core, + uint16_t port, void * src, + uint_t length, void * private_data) { struct virtio_net_state * virtio = (struct virtio_net_state *)private_data; int port_idx = port % virtio->io_range_size; - PrintDebug("VIRTIO NIC %p Write for port %d (index=%d) len=%d, value=%x\n", private_data, - port, port_idx, length, *(uint32_t *)src); + PrintDebug(core->vm_info, core, "VIRTIO NIC %p Write for port %d (index=%d) len=%d, value=%x\n", + private_data, port, port_idx, + length, *(uint32_t *)src); switch (port_idx) { case GUEST_FEATURES_PORT: if (length != 4) { - PrintError("Illegal write length for guest features\n"); + PrintError(core->vm_info, core, "Illegal write length for guest features\n"); return -1; } virtio->virtio_cfg.guest_features = *(uint32_t *)src; - PrintDebug("Setting Guest Features to %x\n", virtio->virtio_cfg.guest_features); break; case VRING_PG_NUM_PORT: if (length != 4) { - PrintError("Illegal write length for page frame number\n"); + PrintError(core->vm_info, core, "Illegal write length for page frame number\n"); return -1; } addr_t pfn = *(uint32_t *)src; @@ -438,13 +428,23 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u uint16_t queue_idx = virtio->virtio_cfg.vring_queue_selector; switch (queue_idx) { case 0: - virtio_setup_queue(virtio, &virtio->rx_vq, pfn, page_addr); + virtio_setup_queue(core, virtio, + &virtio->rx_vq, + pfn, page_addr); break; case 1: - virtio_setup_queue(virtio, &virtio->tx_vq, pfn, page_addr); + virtio_setup_queue(core, virtio, + &virtio->tx_vq, + pfn, page_addr); + if(virtio->tx_notify == 0){ + disable_cb(&virtio->tx_vq); + } + virtio->status = 1; break; case 2: - virtio_setup_queue(virtio, &virtio->ctrl_vq, pfn, page_addr); + virtio_setup_queue(core, virtio, + &virtio->ctrl_vq, + pfn, page_addr); break; default: break; @@ -454,7 +454,7 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u case VRING_Q_SEL_PORT: virtio->virtio_cfg.vring_queue_selector = *(uint16_t *)src; if (virtio->virtio_cfg.vring_queue_selector > 2) { - PrintError("Virtio NIC device only uses 3 queue, selected %d\n", + PrintError(core->vm_info, core, "Virtio NIC: wrong queue idx: %d\n", virtio->virtio_cfg.vring_queue_selector); return -1; } @@ -464,20 +464,18 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u { uint16_t queue_idx = *(uint16_t *)src; if (queue_idx == 0){ - PrintDebug("receive queue notification 0, packet get by Guest\n"); + /* receive queue refill */ + virtio->stats.tx_interrupts ++; } else if (queue_idx == 1){ - if (handle_pkt_tx(virtio) == -1) { - PrintError("Could not handle NIC Notification\n"); + if (handle_pkt_tx(core, virtio, 0) < 0) { + PrintError(core->vm_info, core, "Virtio NIC: Error to handle packet TX\n"); return -1; } + virtio->stats.tx_interrupts ++; } else if (queue_idx == 2){ - if (handle_ctrl(virtio) == -1) { - PrintError("Could not handle NIC Notification\n"); - return -1; - } + /* ctrl */ } else { - PrintError("Virtio NIC device only uses 3 queue, selected %d\n", - queue_idx); + PrintError(core->vm_info, core, "Virtio NIC: Wrong queue index %d\n", queue_idx); } break; } @@ -485,7 +483,6 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u case VIRTIO_STATUS_PORT: virtio->virtio_cfg.status = *(uint8_t *)src; if (virtio->virtio_cfg.status == 0) { - PrintDebug("Resetting device\n"); virtio_init_state(virtio); } break; @@ -493,6 +490,7 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u case VIRTIO_ISR_PORT: virtio->virtio_cfg.pci_isr = *(uint8_t *)src; break; + default: return -1; break; @@ -501,27 +499,29 @@ static int virtio_io_write(struct guest_info *core, uint16_t port, void * src, u return length; } -static int virtio_io_read(struct guest_info *core, uint16_t port, void * dst, uint_t length, void * private_data) +static int virtio_io_read(struct guest_info *core, + uint16_t port, void * dst, + uint_t length, void * private_data) { struct virtio_net_state * virtio = (struct virtio_net_state *)private_data; int port_idx = port % virtio->io_range_size; uint16_t queue_idx = virtio->virtio_cfg.vring_queue_selector; - PrintDebug("Virtio NIC %p: Read for port %d (index =%d), length=%d", private_data, - port, port_idx, length); + PrintDebug(core->vm_info, core, "Virtio NIC %p: Read for port 0x%x (index =%d), length=%d\n", + private_data, port, port_idx, length); switch (port_idx) { case HOST_FEATURES_PORT: if (length != 4) { - PrintError("Illegal read length for host features\n"); - return -1; + PrintError(core->vm_info, core, "Virtio NIC: Illegal read length for host features\n"); + //return -1; } *(uint32_t *)dst = virtio->virtio_cfg.host_features; break; case VRING_PG_NUM_PORT: if (length != 4) { - PrintError("Illegal read length for page frame number\n"); + PrintError(core->vm_info, core, "Virtio NIC: Illegal read length for page frame number\n"); return -1; } switch (queue_idx) { @@ -541,7 +541,7 @@ static int virtio_io_read(struct guest_info *core, uint16_t port, void * dst, ui case VRING_SIZE_PORT: if (length != 2) { - PrintError("Illegal read length for vring size\n"); + PrintError(core->vm_info, core, "Virtio NIC: Illegal read length for vring size\n"); return -1; } switch (queue_idx) { @@ -557,12 +557,11 @@ static int virtio_io_read(struct guest_info *core, uint16_t port, void * dst, ui default: break; } - PrintDebug("queue index: %d, value=0x%x\n", (int)queue_idx, *(uint16_t *)dst); break; case VIRTIO_STATUS_PORT: if (length != 1) { - PrintError("Illegal read length for status\n"); + PrintError(core->vm_info, core, "Virtio NIC: Illegal read length for status\n"); return -1; } *(uint8_t *)dst = virtio->virtio_cfg.status; @@ -571,11 +570,17 @@ static int virtio_io_read(struct guest_info *core, uint16_t port, void * dst, ui case VIRTIO_ISR_PORT: *(uint8_t *)dst = virtio->virtio_cfg.pci_isr; virtio->virtio_cfg.pci_isr = 0; - v3_pci_lower_irq(virtio->virtio_dev->pci_bus, 0, virtio->pci_dev); + v3_pci_lower_irq(virtio->virtio_dev->pci_bus, + virtio->pci_dev, 0); + break; + + case VIRTIO_NET_CONFIG ... VIRTIO_NET_CONFIG + ETH_ALEN - 1: + *(uint8_t *)dst = virtio->net_cfg.mac[port_idx-VIRTIO_NET_CONFIG]; break; default: - PrintError("Virtio NIC: Read of Unhandled Virtio Read\n"); + PrintError(core->vm_info, core, "Virtio NIC: Read of Unhandled Virtio Read:%d\n", + port_idx); return -1; } @@ -583,14 +588,200 @@ static int virtio_io_read(struct guest_info *core, uint16_t port, void * dst, ui } +/* receiving raw ethernet pkt from backend */ +static int virtio_rx(uint8_t * buf, uint32_t size, void * private_data) { + struct virtio_net_state * virtio = (struct virtio_net_state *)private_data; + struct virtio_queue * q = &(virtio->rx_vq); + struct virtio_net_hdr_mrg_rxbuf hdr; + unsigned long flags; + uint8_t kick_guest = 0; + + V3_Net_Print(2, "Virtio NIC: virtio_rx: size: %d\n", size); + + if (!q->ring_avail_addr) { + V3_Net_Print(2, "Virtio NIC: RX Queue not set\n"); + virtio->stats.rx_dropped ++; + + return -1; + } + + memset(&hdr, 0, sizeof(struct virtio_net_hdr_mrg_rxbuf)); + + flags = v3_lock_irqsave(virtio->rx_lock); + + if (q->cur_avail_idx != q->avail->index){ + uint16_t buf_idx; + struct vring_desc * buf_desc; + uint32_t hdr_len, len; + uint32_t offset = 0; + + hdr_len = (virtio->mergeable_rx_bufs)? + sizeof(struct virtio_net_hdr_mrg_rxbuf): + sizeof(struct virtio_net_hdr); + + if(virtio->mergeable_rx_bufs){/* merged buffer */ + struct vring_desc * hdr_desc; + uint16_t old_idx = q->cur_avail_idx; + + buf_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; + hdr_desc = &(q->desc[buf_idx]); + + len = copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, hdr_desc, buf, size, hdr_len); + if(len < 0){ + goto err_exit; + } + offset += len; + + q->used->ring[q->used->index % q->queue_size].id = q->avail->ring[q->cur_avail_idx % q->queue_size]; + q->used->ring[q->used->index % q->queue_size].length = hdr_len + offset; + q->cur_avail_idx ++; + hdr.num_buffers ++; + + while(offset < size) { + buf_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; + buf_desc = &(q->desc[buf_idx]); + + len = copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, buf_desc, buf+offset, size-offset, 0); + if (len < 0){ + V3_Net_Print(2, "Virtio NIC: merged buffer, %d buffer size %d\n", + hdr.num_buffers, len); + q->cur_avail_idx = old_idx; + goto err_exit; + } + offset += len; + buf_desc->flags &= ~VIRTIO_NEXT_FLAG; + + q->used->ring[(q->used->index + hdr.num_buffers) % q->queue_size].id = q->avail->ring[q->cur_avail_idx % q->queue_size]; + q->used->ring[(q->used->index + hdr.num_buffers) % q->queue_size].length = len; + q->cur_avail_idx ++; + + hdr.num_buffers ++; + } + + copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, hdr_desc, (uchar_t *)&hdr, hdr_len, 0); + q->used->index += hdr.num_buffers; + }else{ + buf_idx = q->avail->ring[q->cur_avail_idx % q->queue_size]; + buf_desc = &(q->desc[buf_idx]); + + /* copy header */ + len = copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, buf_desc, (uchar_t *)&(hdr.hdr), hdr_len, 0); + if(len < hdr_len){ + V3_Net_Print(2, "Virtio NIC: rx copy header error %d, hdr_len %d\n", + len, hdr_len); + goto err_exit; + } + + len = copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, buf_desc, buf, size, hdr_len); + if(len < 0){ + V3_Net_Print(2, "Virtio NIC: rx copy data error %d\n", len); + goto err_exit; + } + offset += len; + + /* copy rest of data */ + while(offset < size && + (buf_desc->flags & VIRTIO_NEXT_FLAG)){ + buf_desc = &(q->desc[buf_desc->next]); + len = copy_data_to_desc(&(virtio->virtio_dev->vm->cores[0]), + virtio, buf_desc, buf+offset, size-offset, 0); + if (len < 0) { + break; + } + offset += len; + } + buf_desc->flags &= ~VIRTIO_NEXT_FLAG; + + if(offset < size){ + V3_Net_Print(2, "Virtio NIC: rx not enough ring buffer, buffer size %d\n", + len); + goto err_exit; + } + + q->used->ring[q->used->index % q->queue_size].id = q->avail->ring[q->cur_avail_idx % q->queue_size]; + q->used->ring[q->used->index % q->queue_size].length = size + hdr_len; /* This should be the total length of data sent to guest (header+pkt_data) */ + q->used->index ++; + q->cur_avail_idx ++; + } + + virtio->stats.rx_pkts ++; + virtio->stats.rx_bytes += size; + } else { + V3_Net_Print(2, "Virtio NIC: Guest RX queue is full\n"); + virtio->stats.rx_dropped ++; + + /* kick guest to refill RX queue */ + kick_guest = 1; + } + + v3_unlock_irqrestore(virtio->rx_lock, flags); + + if (!(q->avail->flags & VIRTIO_NO_IRQ_FLAG) || kick_guest) { + V3_Net_Print(2, "Virtio NIC: RX Raising IRQ %d\n", + virtio->pci_dev->config_header.intr_line); + + virtio->virtio_cfg.pci_isr = 0x1; + v3_pci_raise_irq(virtio->virtio_dev->pci_bus, virtio->pci_dev, 0); + virtio->stats.rx_interrupts ++; + } + + /* notify guest if it is in guest mode */ + if((kick_guest || virtio->rx_notify == 1) && + V3_Get_CPU() != virtio->virtio_dev->vm->cores[0].pcpu_id){ + v3_interrupt_cpu(virtio->virtio_dev->vm, + virtio->virtio_dev->vm->cores[0].pcpu_id, + 0); + } + + return 0; + +err_exit: + virtio->stats.rx_dropped ++; + v3_unlock_irqrestore(virtio->rx_lock, flags); + + return -1; +} + +static int virtio_free(struct virtio_dev_state * virtio) { + struct virtio_net_state * backend = NULL; + struct virtio_net_state * tmp = NULL; + + + list_for_each_entry_safe(backend, tmp, &(virtio->dev_list), dev_link) { + virtio_deinit_state(&(virtio->vm->cores[0]),backend); + list_del(&(backend->dev_link)); + V3_Free(backend); + } + + V3_Free(virtio); + + return 0; +} + + static struct v3_device_ops dev_ops = { - .free = virtio_free, - .reset = NULL, - .start = NULL, - .stop = NULL, + .free = (int (*)(void *))virtio_free, }; -static int register_dev(struct virtio_dev_state * virtio, struct virtio_net_state * net_state) + +static int virtio_poll(int quote, void * data){ + struct virtio_net_state * virtio = (struct virtio_net_state *)data; + + if (virtio->status) { + + return handle_pkt_tx(&(virtio->vm->cores[0]), virtio, quote); + } + + return 0; +} + +static int register_dev(struct virtio_dev_state * virtio, + struct virtio_net_state * net_state) { struct pci_device * pci_dev = NULL; struct v3_pci_bar bars[6]; @@ -598,15 +789,16 @@ static int register_dev(struct virtio_dev_state * virtio, struct virtio_net_stat int tmp_ports = num_ports; int i; - // This gets the number of ports, rounded up to a power of 2 - net_state->io_range_size = 1; // must be a power of 2 + /* This gets the number of ports, rounded up to a power of 2 */ + net_state->io_range_size = 1; while (tmp_ports > 0) { tmp_ports >>= 1; net_state->io_range_size <<= 1; } - // this is to account for any low order bits being set in num_ports - // if there are none, then num_ports was already a power of 2 so we shift right to reset it + /* this is to account for any low order bits being set in num_ports + * if there are none, then num_ports was already a power of 2 so we shift right to reset it + */ if ((num_ports & ((net_state->io_range_size >> 1) - 1)) == 0) { net_state->io_range_size >>= 1; } @@ -615,7 +807,8 @@ static int register_dev(struct virtio_dev_state * virtio, struct virtio_net_stat bars[i].type = PCI_BAR_NONE; } - PrintDebug("Virtio-NIC io_range_size = %d\n", net_state->io_range_size); + PrintDebug(VM_NONE, VCORE_NONE, "Virtio NIC: io_range_size = %d\n", + net_state->io_range_size); bars[0].type = PCI_BAR_IO; bars[0].default_base_port = -1; @@ -627,14 +820,14 @@ static int register_dev(struct virtio_dev_state * virtio, struct virtio_net_stat pci_dev = v3_pci_register_device(virtio->pci_bus, PCI_STD_DEVICE, 0, PCI_AUTO_DEV_NUM, 0, "LNX_VIRTIO_NIC", bars, - NULL, NULL, NULL, net_state); + NULL, NULL, NULL, NULL, net_state); if (!pci_dev) { - PrintError("Virtio NIC: Could not register PCI Device\n"); + PrintError(VM_NONE, VCORE_NONE, "Virtio NIC: Could not register PCI Device\n"); return -1; } - PrintDebug("Virtio NIC: registered to PCI bus\n"); + PrintDebug(VM_NONE, VCORE_NONE, "Virtio NIC: registered to PCI bus\n"); pci_dev->config_header.vendor_id = VIRTIO_VENDOR_ID; pci_dev->config_header.subsystem_vendor_id = VIRTIO_SUBVENDOR_ID; @@ -647,18 +840,99 @@ static int register_dev(struct virtio_dev_state * virtio, struct virtio_net_stat pci_dev->config_header.intr_pin = 1; pci_dev->config_header.max_latency = 1; // ?? (qemu does it...) - net_state->pci_dev = pci_dev; - net_state->virtio_cfg.host_features = 0; //no features support now - net_state->rx_vq.queue_size = QUEUE_SIZE; - net_state->tx_vq.queue_size = QUEUE_SIZE; - net_state->ctrl_vq.queue_size = CTRL_QUEUE_SIZE; + net_state->pci_dev = pci_dev; net_state->virtio_dev = virtio; - + + memcpy(net_state->net_cfg.mac, virtio->mac, 6); + virtio_init_state(net_state); + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: Registered Intr Line %d\n", pci_dev->config_header.intr_line); + + /* Add backend to list of devices */ + list_add(&(net_state->dev_link), &(virtio->dev_list)); + return 0; } +#define RATE_UPPER_THRESHOLD_DEFAULT 10000 /* 10000 pkts per second, around 100Mbits */ +#define RATE_LOWER_THRESHOLD_DEFAULT 1000 /* 1000 pkts per second, around 10Mbits */ +#define PROFILE_PERIOD_DEFAULT 10000 /* us */ + +static void virtio_nic_timer(struct guest_info * core, + uint64_t cpu_cycles, uint64_t cpu_freq, + void * priv_data) { + struct virtio_net_state * net_state = (struct virtio_net_state *)priv_data; + uint64_t period_us; + static int profile_ms = 0; + uint64_t target_period_us = net_state->virtio_dev->period_us; + uint64_t upper_thresh_pps = net_state->virtio_dev->upper_thresh_pps; + uint64_t lower_thresh_pps = net_state->virtio_dev->lower_thresh_pps; + + + if(!net_state->status){ /* VNIC is not in working status */ + return; + } + + period_us = (1000*cpu_cycles)/cpu_freq; + net_state->past_us += period_us; + + if (net_state->past_us > target_period_us) { + + uint64_t tx_count, rx_count; + uint64_t lb_tx_count, lb_rx_count; + uint64_t ub_tx_count, ub_rx_count; + + lb_tx_count = lb_rx_count = (lower_thresh_pps * 1000000) / net_state->past_us; // packets expected in this interval + ub_tx_count = ub_rx_count = (upper_thresh_pps * 1000000) / net_state->past_us; + + tx_count = net_state->stats.tx_pkts - net_state->tx_pkts; + rx_count = net_state->stats.rx_pkts - net_state->rx_pkts; + + net_state->tx_pkts = net_state->stats.tx_pkts; + net_state->rx_pkts = net_state->stats.rx_pkts; + + if(tx_count > ub_tx_count && net_state->tx_notify == 1) { + PrintDebug(core->vm_info, core, "Virtio NIC: Switch TX to VMM driven mode\n"); + disable_cb(&(net_state->tx_vq)); + net_state->tx_notify = 0; + } + + if(tx_count < lb_tx_count && net_state->tx_notify == 0) { + PrintDebug(core->vm_info, core, "Virtio NIC: Switch TX to Guest driven mode\n"); + enable_cb(&(net_state->tx_vq)); + net_state->tx_notify = 1; + } + + if(rx_count > ub_rx_count && net_state->rx_notify == 1) { + PrintDebug(core->vm_info, core, "Virtio NIC: Switch RX to VMM None notify mode\n"); + net_state->rx_notify = 0; + } + + if(rx_count < lb_rx_count && net_state->rx_notify == 0) { + V3_Print(core->vm_info, core, "Virtio NIC: Switch RX to VMM notify mode\n"); + net_state->rx_notify = 1; + } + + net_state->past_us = 0; + } + + profile_ms += period_us/1000; + if(profile_ms > 20000){ + PrintDebug(core->vm_info, core, "Virtio NIC: TX: Pkt: %lld, Bytes: %lld\n\t\tRX Pkt: %lld. Bytes: %lld\n\t\tDropped: tx %lld, rx %lld\nInterrupts: tx %d, rx %d\nTotal Exit: %lld\n", + net_state->stats.tx_pkts, net_state->stats.tx_bytes, + net_state->stats.rx_pkts, net_state->stats.rx_bytes, + net_state->stats.tx_dropped, net_state->stats.rx_dropped, + net_state->stats.tx_interrupts, net_state->stats.rx_interrupts, + net_state->vm->cores[0].num_exits); + profile_ms = 0; + } +} + +static struct v3_timer_ops timer_ops = { + .update_timer = virtio_nic_timer, +}; + static int connect_fn(struct v3_vm_info * info, void * frontend_data, struct v3_dev_net_ops * ops, @@ -667,42 +941,184 @@ static int connect_fn(struct v3_vm_info * info, struct virtio_dev_state * virtio = (struct virtio_dev_state *)frontend_data; struct virtio_net_state * net_state = (struct virtio_net_state *)V3_Malloc(sizeof(struct virtio_net_state)); + if (!net_state) { + PrintError(info, VCORE_NONE, "Cannot allocate in connect\n"); + return -1; + } + memset(net_state, 0, sizeof(struct virtio_net_state)); register_dev(virtio, net_state); + net_state->vm = info; net_state->net_ops = ops; net_state->backend_data = private_data; + net_state->virtio_dev = virtio; + + switch (virtio->model) { + case GUEST_DRIVEN: + V3_Print(info, VCORE_NONE, "Virtio NIC: Guest-driven operation\n"); + net_state->tx_notify = 1; + net_state->rx_notify = 1; + break; + case VMM_DRIVEN: + V3_Print(info, VCORE_NONE, "Virtio NIC: VMM-driven operation\n"); + net_state->tx_notify = 0; + net_state->rx_notify = 0; + break; + case ADAPTIVE: { + V3_Print(info, VCORE_NONE, "Virtio NIC: Adaptive operation (begins in guest-driven mode)\n"); + net_state->tx_notify = 1; + net_state->rx_notify = 1; + + net_state->timer = v3_add_timer(&(info->cores[0]), &timer_ops,net_state); + + } + break; + + default: + V3_Print(info, VCORE_NONE, "Virtio NIC: Unknown model, using GUEST_DRIVEN\n"); + net_state->tx_notify = 1; + net_state->rx_notify = 1; + break; + } + + + ops->recv = virtio_rx; + ops->poll = virtio_poll; + ops->config.frontend_data = net_state; + ops->config.poll = 1; + ops->config.quote = 64; + ops->config.fnt_mac = V3_Malloc(ETH_ALEN); + + if (!ops->config.fnt_mac) { + PrintError(info, VCORE_NONE, "Cannot allocate in connect\n"); + // should unregister here + return -1; + } + + memcpy(ops->config.fnt_mac, virtio->mac, ETH_ALEN); + + return 0; +} + +static int setup_perf_model(struct virtio_dev_state *virtio_state, v3_cfg_tree_t *t) +{ + char *mode = v3_cfg_val(t,"mode"); + + // defaults + virtio_state->model = GUEST_DRIVEN; + virtio_state->lower_thresh_pps = RATE_LOWER_THRESHOLD_DEFAULT; + virtio_state->upper_thresh_pps = RATE_UPPER_THRESHOLD_DEFAULT; + virtio_state->period_us = PROFILE_PERIOD_DEFAULT; + + + // overrides + if (mode) { + if (!strcasecmp(mode,"guest-driven")) { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: Setting static GUEST_DRIVEN mode of operation (latency optimized)\n"); + virtio_state->model=GUEST_DRIVEN; + } else if (!strcasecmp(mode, "vmm-driven")) { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: Setting static VMM_DRIVEN mode of operation (throughput optimized)\n"); + virtio_state->model=VMM_DRIVEN; + } else if (!strcasecmp(mode, "adaptive")) { + char *s; + + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: Setting dynamic ADAPTIVE mode of operation\n"); + virtio_state->model=ADAPTIVE; + + if (!(s=v3_cfg_val(t,"upper"))) { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: No upper bound given, using default\n"); + } else { + virtio_state->upper_thresh_pps = atoi(s); + } + if (!(s=v3_cfg_val(t,"lower"))) { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: No lower bound given, using default\n"); + } else { + virtio_state->lower_thresh_pps = atoi(s); + } + if (!(s=v3_cfg_val(t,"period"))) { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: No period given, using default\n"); + } else { + virtio_state->period_us = atoi(s); + } + + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: lower_thresh_pps=%llu, upper_thresh_pps=%llu, period_us=%llu\n", + virtio_state->lower_thresh_pps, + virtio_state->upper_thresh_pps, + virtio_state->period_us); + } else { + PrintError(VM_NONE, VCORE_NONE, "Virtio NIC: Unknown mode of operation '%s', using default (guest-driven)\n",mode); + virtio_state->model=GUEST_DRIVEN; + } + } else { + V3_Print(VM_NONE, VCORE_NONE, "Virtio NIC: No model given, using default (guest-driven)\n"); + } return 0; + } +/* + + pci-bus-to-attach-to // required + mac address // if ommited with pic one + + +*/ static int virtio_init(struct v3_vm_info * vm, v3_cfg_tree_t * cfg) { struct vm_device * pci_bus = v3_find_dev(vm, v3_cfg_val(cfg, "bus")); struct virtio_dev_state * virtio_state = NULL; - char * name = v3_cfg_val(cfg, "name"); - - PrintDebug("Virtio NIC: Initializing VIRTIO Network device: %s\n", name); - + char * dev_id = v3_cfg_val(cfg, "ID"); + char * mac = v3_cfg_val(cfg, "mac"); + v3_cfg_tree_t *model = v3_cfg_subtree(cfg,"model"); + if (pci_bus == NULL) { - PrintError("Virtio NIC: VirtIO devices require a PCI Bus"); + PrintError(vm, VCORE_NONE, "Virtio NIC: Virtio device require a PCI Bus"); return -1; } virtio_state = (struct virtio_dev_state *)V3_Malloc(sizeof(struct virtio_dev_state)); + + if (!virtio_state) { + PrintError(vm, VCORE_NONE, "Cannot allocate in init\n"); + return -1; + } + memset(virtio_state, 0, sizeof(struct virtio_dev_state)); INIT_LIST_HEAD(&(virtio_state->dev_list)); virtio_state->pci_bus = pci_bus; - //virtio_state->vm = vm; + virtio_state->vm = vm; + + if (mac) { + if (!str2mac(mac, virtio_state->mac)) { + PrintDebug(vm, VCORE_NONE, "Virtio NIC: Mac specified %s\n", mac); + } else { + PrintError(vm, VCORE_NONE, "Virtio NIC: Mac specified is incorrect, picking a randoom mac\n"); + random_ethaddr(virtio_state->mac); + } + } else { + PrintDebug(vm, VCORE_NONE, "Virtio NIC: no mac specified, so picking a random mac\n"); + random_ethaddr(virtio_state->mac); + } + + if (setup_perf_model(virtio_state,model)<0) { + PrintError(vm, VCORE_NONE, "Cannnot setup performance model\n"); + V3_Free(virtio_state); + return -1; + } + + struct vm_device * dev = v3_add_device(vm, dev_id, &dev_ops, virtio_state); - struct vm_device * dev = v3_allocate_device(name, &dev_ops, virtio_state); - if (v3_attach_device(vm, dev) == -1) { - PrintError("Virtio NIC: Could not attach device %s\n", name); + if (dev == NULL) { + PrintError(vm, VCORE_NONE, "Virtio NIC: Could not attach device %s\n", dev_id); + V3_Free(virtio_state); return -1; } - if (v3_dev_add_net_frontend(vm, name, connect_fn, (void *)virtio_state) == -1) { - PrintError("Virtio NIC: Could not register %s as net frontend\n", name); + if (v3_dev_add_net_frontend(vm, dev_id, connect_fn, (void *)virtio_state) == -1) { + PrintError(vm, VCORE_NONE, "Virtio NIC: Could not register %s as net frontend\n", dev_id); + v3_remove_device(dev); return -1; }