ret = device_add(&cdev->dev);
if (ret)
goto add_err;
+ ret = ib_setup_port_attrs(cdev, false);
+ if (ret)
+ goto port_err;
ret = xa_err(xa_store(&device->compat_devs, rnet->id,
cdev, GFP_KERNEL));
return 0;
insert_err:
+ ib_free_port_attrs(cdev);
+port_err:
device_del(&cdev->dev);
add_err:
put_device(&cdev->dev);
cdev = xa_erase(&device->compat_devs, id);
mutex_unlock(&device->compat_devs_mutex);
if (cdev) {
+ ib_free_port_attrs(cdev);
device_del(&cdev->dev);
put_device(&cdev->dev);
}
return;
}
-static int add_port(struct ib_core_device *coredev, int port_num)
+static int add_port(struct ib_core_device *coredev,
+ int port_num, bool alloc_stats)
{
struct ib_device *device = rdma_device_to_ibdev(&coredev->dev);
struct ib_port *p;
goto err_put;
}
- if (device->ops.process_mad) {
+ if (device->ops.process_mad && alloc_stats) {
p->pma_table = get_counter_table(device, port_num);
ret = sysfs_create_group(&p->kobj, p->pma_table);
if (ret)
* port, so holder should be device. Therefore skip per port conunter
* initialization.
*/
- if (device->ops.alloc_hw_stats && port_num)
+ if (device->ops.alloc_hw_stats && port_num && alloc_stats)
setup_hw_stats(device, p, port_num);
list_add_tail(&p->kobj.entry, &coredev->port_list);
.attrs = ib_dev_attrs,
};
-static void ib_free_port_attrs(struct ib_core_device *coredev)
+void ib_free_port_attrs(struct ib_core_device *coredev)
{
struct kobject *p, *t;
kobject_put(coredev->ports_kobj);
}
-static int ib_setup_port_attrs(struct ib_core_device *coredev)
+int ib_setup_port_attrs(struct ib_core_device *coredev, bool alloc_stats)
{
struct ib_device *device = rdma_device_to_ibdev(&coredev->dev);
unsigned int port;
return -ENOMEM;
rdma_for_each_port (device, port) {
- ret = add_port(coredev, port);
+ ret = add_port(coredev, port, alloc_stats);
if (ret)
goto err_put;
}
{
int ret;
- ret = ib_setup_port_attrs(&device->coredev);
+ ret = ib_setup_port_attrs(&device->coredev, true);
if (ret)
return ret;