/* sriov sysfs */
int local_pci_num_vf(struct pci_dev *dev)
{
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,34)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 34)
struct iov {
int pos;
int nres;
u16 total;
u16 initial;
u16 nr_virtfn;
- } *iov = (struct iov*)dev->sriov;
+ } *iov = (struct iov *)dev->sriov;
if (!dev->is_physfn)
return 0;
static DEVICE_ATTR(max_vfs, S_IRUGO | S_IWUSR, show_max_vfs, store_max_vfs);
#ifdef RTE_PCI_CONFIG
-static DEVICE_ATTR(extended_tag, S_IRUGO | S_IWUSR, show_extended_tag, \
+static DEVICE_ATTR(extended_tag, S_IRUGO | S_IWUSR, show_extended_tag,
store_extended_tag);
-static DEVICE_ATTR(max_read_request_size, S_IRUGO | S_IWUSR, \
+static DEVICE_ATTR(max_read_request_size, S_IRUGO | S_IWUSR,
show_max_read_request_size, store_max_read_request_size);
#endif
&dev_attr_extended_tag.attr,
&dev_attr_max_read_request_size.attr,
#endif
- NULL,
+ NULL,
};
static const struct attribute_group dev_attr_grp = {
pci_lock(struct pci_dev * pdev)
{
/* Some function names changes between 3.2.0 and 3.3.0... */
-#if LINUX_VERSION_CODE < KERNEL_VERSION(3,3,0)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 3, 0)
pci_block_user_cfg_access(pdev);
return 1;
#else
pci_unlock(struct pci_dev * pdev)
{
/* Some function names changes between 3.2.0 and 3.3.0... */
-#if LINUX_VERSION_CODE < KERNEL_VERSION(3,3,0)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 3, 0)
pci_unblock_user_cfg_access(pdev);
#else
pci_cfg_access_unlock(pdev);
igbuio_dom0_mmap_phys(struct uio_info *info, struct vm_area_struct *vma)
{
int idx;
+
idx = (int)vma->vm_pgoff;
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
vma->vm_page_prot.pgprot |= _PAGE_IOMAP;
if (vma->vm_pgoff >= MAX_UIO_MAPS)
return -EINVAL;
- if(info->mem[vma->vm_pgoff].size == 0)
- return -EINVAL;
+
+ if (info->mem[vma->vm_pgoff].size == 0)
+ return -EINVAL;
idx = (int)vma->vm_pgoff;
switch (info->mem[idx].memtype) {
unsigned long addr, len;
void *internal_addr;
- if (sizeof(info->mem) / sizeof (info->mem[0]) <= n)
- return (EINVAL);
+ if (sizeof(info->mem) / sizeof(info->mem[0]) <= n)
+ return -EINVAL;
addr = pci_resource_start(dev, pci_bar);
len = pci_resource_len(dev, pci_bar);
{
unsigned long addr, len;
- if (sizeof(info->port) / sizeof (info->port[0]) <= n)
- return (EINVAL);
+ if (sizeof(info->port) / sizeof(info->port[0]) <= n)
+ return -EINVAL;
addr = pci_resource_start(dev, pci_bar);
len = pci_resource_len(dev, pci_bar);
if (addr == 0 || len == 0)
- return (-1);
+ return -EINVAL;
info->port[n].name = name;
info->port[n].start = addr;
info->port[n].size = len;
info->port[n].porttype = UIO_PORT_X86;
- return (0);
+ return 0;
}
/* Unmap previously ioremap'd resources */
igbuio_pci_release_iomem(struct uio_info *info)
{
int i;
+
for (i = 0; i < MAX_UIO_MAPS; i++) {
if (info->mem[i].internal_addr)
iounmap(info->mem[i].internal_addr);
pci_resource_start(dev, i) != 0) {
flags = pci_resource_flags(dev, i);
if (flags & IORESOURCE_MEM) {
- if ((ret = igbuio_pci_setup_iomem(dev, info,
- iom, i, bar_names[i])) != 0)
- return (ret);
+ ret = igbuio_pci_setup_iomem(dev, info, iom,
+ i, bar_names[i]);
+ if (ret != 0)
+ return ret;
iom++;
} else if (flags & IORESOURCE_IO) {
- if ((ret = igbuio_pci_setup_ioport(dev, info,
- iop, i, bar_names[i])) != 0)
- return (ret);
+ ret = igbuio_pci_setup_ioport(dev, info, iop,
+ i, bar_names[i]);
+ if (ret != 0)
+ return ret;
iop++;
}
}
return ((iom != 0) ? ret : ENOENT);
}
-#if LINUX_VERSION_CODE < KERNEL_VERSION(3,8,0)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 8, 0)
static int __devinit
#else
static int