mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:46:16 +00:00
2f87e9cf0c
Add new UAPI to support the mac address from vdpa tool Function vdpa_nl_cmd_dev_attr_set_doit() will get the new MAC address from the vdpa tool and then set it to the device. The usage is: vdpa dev set name vdpa_name mac **:**:**:**:**:** Here is example: root@L1# vdpa -jp dev config show vdpa0 { "config": { "vdpa0": { "mac": "82:4d:e9:5d:d7:e6", "link ": "up", "link_announce ": false, "mtu": 1500 } } } root@L1# vdpa dev set name vdpa0 mac 00:11:22:33:44:55 root@L1# vdpa -jp dev config show vdpa0 { "config": { "vdpa0": { "mac": "00:11:22:33:44:55", "link ": "up", "link_announce ": false, "mtu": 1500 } } } Signed-off-by: Cindy Lu <lulu@redhat.com> Message-Id: <20240731031653.1047692-2-lulu@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com> Acked-by: Jason Wang <jasowang@redhat.com>
1622 lines
40 KiB
C
1622 lines
40 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* vDPA bus.
|
|
*
|
|
* Copyright (c) 2020, Red Hat. All rights reserved.
|
|
* Author: Jason Wang <jasowang@redhat.com>
|
|
*
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/idr.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/vdpa.h>
|
|
#include <uapi/linux/vdpa.h>
|
|
#include <net/genetlink.h>
|
|
#include <linux/mod_devicetable.h>
|
|
#include <linux/virtio_ids.h>
|
|
|
|
static LIST_HEAD(mdev_head);
|
|
/* A global mutex that protects vdpa management device and device level operations. */
|
|
static DECLARE_RWSEM(vdpa_dev_lock);
|
|
static DEFINE_IDA(vdpa_index_ida);
|
|
|
|
void vdpa_set_status(struct vdpa_device *vdev, u8 status)
|
|
{
|
|
down_write(&vdev->cf_lock);
|
|
vdev->config->set_status(vdev, status);
|
|
up_write(&vdev->cf_lock);
|
|
}
|
|
EXPORT_SYMBOL(vdpa_set_status);
|
|
|
|
static struct genl_family vdpa_nl_family;
|
|
|
|
static int vdpa_dev_probe(struct device *d)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(d);
|
|
struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
|
|
const struct vdpa_config_ops *ops = vdev->config;
|
|
u32 max_num, min_num = 1;
|
|
int ret = 0;
|
|
|
|
d->dma_mask = &d->coherent_dma_mask;
|
|
ret = dma_set_mask_and_coherent(d, DMA_BIT_MASK(64));
|
|
if (ret)
|
|
return ret;
|
|
|
|
max_num = ops->get_vq_num_max(vdev);
|
|
if (ops->get_vq_num_min)
|
|
min_num = ops->get_vq_num_min(vdev);
|
|
if (max_num < min_num)
|
|
return -EINVAL;
|
|
|
|
if (drv && drv->probe)
|
|
ret = drv->probe(vdev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void vdpa_dev_remove(struct device *d)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(d);
|
|
struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
|
|
|
|
if (drv && drv->remove)
|
|
drv->remove(vdev);
|
|
}
|
|
|
|
static int vdpa_dev_match(struct device *dev, const struct device_driver *drv)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(dev);
|
|
|
|
/* Check override first, and if set, only use the named driver */
|
|
if (vdev->driver_override)
|
|
return strcmp(vdev->driver_override, drv->name) == 0;
|
|
|
|
/* Currently devices must be supported by all vDPA bus drivers */
|
|
return 1;
|
|
}
|
|
|
|
static ssize_t driver_override_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(dev);
|
|
int ret;
|
|
|
|
ret = driver_set_override(dev, &vdev->driver_override, buf, count);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return count;
|
|
}
|
|
|
|
static ssize_t driver_override_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(dev);
|
|
ssize_t len;
|
|
|
|
device_lock(dev);
|
|
len = sysfs_emit(buf, "%s\n", vdev->driver_override);
|
|
device_unlock(dev);
|
|
|
|
return len;
|
|
}
|
|
static DEVICE_ATTR_RW(driver_override);
|
|
|
|
static struct attribute *vdpa_dev_attrs[] = {
|
|
&dev_attr_driver_override.attr,
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group vdpa_dev_group = {
|
|
.attrs = vdpa_dev_attrs,
|
|
};
|
|
__ATTRIBUTE_GROUPS(vdpa_dev);
|
|
|
|
static const struct bus_type vdpa_bus = {
|
|
.name = "vdpa",
|
|
.dev_groups = vdpa_dev_groups,
|
|
.match = vdpa_dev_match,
|
|
.probe = vdpa_dev_probe,
|
|
.remove = vdpa_dev_remove,
|
|
};
|
|
|
|
static void vdpa_release_dev(struct device *d)
|
|
{
|
|
struct vdpa_device *vdev = dev_to_vdpa(d);
|
|
const struct vdpa_config_ops *ops = vdev->config;
|
|
|
|
if (ops->free)
|
|
ops->free(vdev);
|
|
|
|
ida_free(&vdpa_index_ida, vdev->index);
|
|
kfree(vdev->driver_override);
|
|
kfree(vdev);
|
|
}
|
|
|
|
/**
|
|
* __vdpa_alloc_device - allocate and initilaize a vDPA device
|
|
* This allows driver to some prepartion after device is
|
|
* initialized but before registered.
|
|
* @parent: the parent device
|
|
* @config: the bus operations that is supported by this device
|
|
* @ngroups: number of groups supported by this device
|
|
* @nas: number of address spaces supported by this device
|
|
* @size: size of the parent structure that contains private data
|
|
* @name: name of the vdpa device; optional.
|
|
* @use_va: indicate whether virtual address must be used by this device
|
|
*
|
|
* Driver should use vdpa_alloc_device() wrapper macro instead of
|
|
* using this directly.
|
|
*
|
|
* Return: Returns an error when parent/config/dma_dev is not set or fail to get
|
|
* ida.
|
|
*/
|
|
struct vdpa_device *__vdpa_alloc_device(struct device *parent,
|
|
const struct vdpa_config_ops *config,
|
|
unsigned int ngroups, unsigned int nas,
|
|
size_t size, const char *name,
|
|
bool use_va)
|
|
{
|
|
struct vdpa_device *vdev;
|
|
int err = -EINVAL;
|
|
|
|
if (!config)
|
|
goto err;
|
|
|
|
if (!!config->dma_map != !!config->dma_unmap)
|
|
goto err;
|
|
|
|
/* It should only work for the device that use on-chip IOMMU */
|
|
if (use_va && !(config->dma_map || config->set_map))
|
|
goto err;
|
|
|
|
err = -ENOMEM;
|
|
vdev = kzalloc(size, GFP_KERNEL);
|
|
if (!vdev)
|
|
goto err;
|
|
|
|
err = ida_alloc(&vdpa_index_ida, GFP_KERNEL);
|
|
if (err < 0)
|
|
goto err_ida;
|
|
|
|
vdev->dev.bus = &vdpa_bus;
|
|
vdev->dev.parent = parent;
|
|
vdev->dev.release = vdpa_release_dev;
|
|
vdev->index = err;
|
|
vdev->config = config;
|
|
vdev->features_valid = false;
|
|
vdev->use_va = use_va;
|
|
vdev->ngroups = ngroups;
|
|
vdev->nas = nas;
|
|
|
|
if (name)
|
|
err = dev_set_name(&vdev->dev, "%s", name);
|
|
else
|
|
err = dev_set_name(&vdev->dev, "vdpa%u", vdev->index);
|
|
if (err)
|
|
goto err_name;
|
|
|
|
init_rwsem(&vdev->cf_lock);
|
|
device_initialize(&vdev->dev);
|
|
|
|
return vdev;
|
|
|
|
err_name:
|
|
ida_free(&vdpa_index_ida, vdev->index);
|
|
err_ida:
|
|
kfree(vdev);
|
|
err:
|
|
return ERR_PTR(err);
|
|
}
|
|
EXPORT_SYMBOL_GPL(__vdpa_alloc_device);
|
|
|
|
static int vdpa_name_match(struct device *dev, const void *data)
|
|
{
|
|
struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
|
|
|
|
return (strcmp(dev_name(&vdev->dev), data) == 0);
|
|
}
|
|
|
|
static int __vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
|
|
{
|
|
struct device *dev;
|
|
|
|
vdev->nvqs = nvqs;
|
|
|
|
lockdep_assert_held(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, dev_name(&vdev->dev), vdpa_name_match);
|
|
if (dev) {
|
|
put_device(dev);
|
|
return -EEXIST;
|
|
}
|
|
return device_add(&vdev->dev);
|
|
}
|
|
|
|
/**
|
|
* _vdpa_register_device - register a vDPA device with vdpa lock held
|
|
* Caller must have a succeed call of vdpa_alloc_device() before.
|
|
* Caller must invoke this routine in the management device dev_add()
|
|
* callback after setting up valid mgmtdev for this vdpa device.
|
|
* @vdev: the vdpa device to be registered to vDPA bus
|
|
* @nvqs: number of virtqueues supported by this device
|
|
*
|
|
* Return: Returns an error when fail to add device to vDPA bus
|
|
*/
|
|
int _vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
|
|
{
|
|
if (!vdev->mdev)
|
|
return -EINVAL;
|
|
|
|
return __vdpa_register_device(vdev, nvqs);
|
|
}
|
|
EXPORT_SYMBOL_GPL(_vdpa_register_device);
|
|
|
|
/**
|
|
* vdpa_register_device - register a vDPA device
|
|
* Callers must have a succeed call of vdpa_alloc_device() before.
|
|
* @vdev: the vdpa device to be registered to vDPA bus
|
|
* @nvqs: number of virtqueues supported by this device
|
|
*
|
|
* Return: Returns an error when fail to add to vDPA bus
|
|
*/
|
|
int vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
|
|
{
|
|
int err;
|
|
|
|
down_write(&vdpa_dev_lock);
|
|
err = __vdpa_register_device(vdev, nvqs);
|
|
up_write(&vdpa_dev_lock);
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_register_device);
|
|
|
|
/**
|
|
* _vdpa_unregister_device - unregister a vDPA device
|
|
* Caller must invoke this routine as part of management device dev_del()
|
|
* callback.
|
|
* @vdev: the vdpa device to be unregisted from vDPA bus
|
|
*/
|
|
void _vdpa_unregister_device(struct vdpa_device *vdev)
|
|
{
|
|
lockdep_assert_held(&vdpa_dev_lock);
|
|
WARN_ON(!vdev->mdev);
|
|
device_unregister(&vdev->dev);
|
|
}
|
|
EXPORT_SYMBOL_GPL(_vdpa_unregister_device);
|
|
|
|
/**
|
|
* vdpa_unregister_device - unregister a vDPA device
|
|
* @vdev: the vdpa device to be unregisted from vDPA bus
|
|
*/
|
|
void vdpa_unregister_device(struct vdpa_device *vdev)
|
|
{
|
|
down_write(&vdpa_dev_lock);
|
|
device_unregister(&vdev->dev);
|
|
up_write(&vdpa_dev_lock);
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_unregister_device);
|
|
|
|
/**
|
|
* __vdpa_register_driver - register a vDPA device driver
|
|
* @drv: the vdpa device driver to be registered
|
|
* @owner: module owner of the driver
|
|
*
|
|
* Return: Returns an err when fail to do the registration
|
|
*/
|
|
int __vdpa_register_driver(struct vdpa_driver *drv, struct module *owner)
|
|
{
|
|
drv->driver.bus = &vdpa_bus;
|
|
drv->driver.owner = owner;
|
|
|
|
return driver_register(&drv->driver);
|
|
}
|
|
EXPORT_SYMBOL_GPL(__vdpa_register_driver);
|
|
|
|
/**
|
|
* vdpa_unregister_driver - unregister a vDPA device driver
|
|
* @drv: the vdpa device driver to be unregistered
|
|
*/
|
|
void vdpa_unregister_driver(struct vdpa_driver *drv)
|
|
{
|
|
driver_unregister(&drv->driver);
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_unregister_driver);
|
|
|
|
/**
|
|
* vdpa_mgmtdev_register - register a vdpa management device
|
|
*
|
|
* @mdev: Pointer to vdpa management device
|
|
* vdpa_mgmtdev_register() register a vdpa management device which supports
|
|
* vdpa device management.
|
|
* Return: Returns 0 on success or failure when required callback ops are not
|
|
* initialized.
|
|
*/
|
|
int vdpa_mgmtdev_register(struct vdpa_mgmt_dev *mdev)
|
|
{
|
|
if (!mdev->device || !mdev->ops || !mdev->ops->dev_add || !mdev->ops->dev_del)
|
|
return -EINVAL;
|
|
|
|
INIT_LIST_HEAD(&mdev->list);
|
|
down_write(&vdpa_dev_lock);
|
|
list_add_tail(&mdev->list, &mdev_head);
|
|
up_write(&vdpa_dev_lock);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_mgmtdev_register);
|
|
|
|
static int vdpa_match_remove(struct device *dev, void *data)
|
|
{
|
|
struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
|
|
struct vdpa_mgmt_dev *mdev = vdev->mdev;
|
|
|
|
if (mdev == data)
|
|
mdev->ops->dev_del(mdev, vdev);
|
|
return 0;
|
|
}
|
|
|
|
void vdpa_mgmtdev_unregister(struct vdpa_mgmt_dev *mdev)
|
|
{
|
|
down_write(&vdpa_dev_lock);
|
|
|
|
list_del(&mdev->list);
|
|
|
|
/* Filter out all the entries belong to this management device and delete it. */
|
|
bus_for_each_dev(&vdpa_bus, NULL, mdev, vdpa_match_remove);
|
|
|
|
up_write(&vdpa_dev_lock);
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_mgmtdev_unregister);
|
|
|
|
static void vdpa_get_config_unlocked(struct vdpa_device *vdev,
|
|
unsigned int offset,
|
|
void *buf, unsigned int len)
|
|
{
|
|
const struct vdpa_config_ops *ops = vdev->config;
|
|
|
|
/*
|
|
* Config accesses aren't supposed to trigger before features are set.
|
|
* If it does happen we assume a legacy guest.
|
|
*/
|
|
if (!vdev->features_valid)
|
|
vdpa_set_features_unlocked(vdev, 0);
|
|
ops->get_config(vdev, offset, buf, len);
|
|
}
|
|
|
|
/**
|
|
* vdpa_get_config - Get one or more device configuration fields.
|
|
* @vdev: vdpa device to operate on
|
|
* @offset: starting byte offset of the field
|
|
* @buf: buffer pointer to read to
|
|
* @len: length of the configuration fields in bytes
|
|
*/
|
|
void vdpa_get_config(struct vdpa_device *vdev, unsigned int offset,
|
|
void *buf, unsigned int len)
|
|
{
|
|
down_read(&vdev->cf_lock);
|
|
vdpa_get_config_unlocked(vdev, offset, buf, len);
|
|
up_read(&vdev->cf_lock);
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_get_config);
|
|
|
|
/**
|
|
* vdpa_set_config - Set one or more device configuration fields.
|
|
* @vdev: vdpa device to operate on
|
|
* @offset: starting byte offset of the field
|
|
* @buf: buffer pointer to read from
|
|
* @length: length of the configuration fields in bytes
|
|
*/
|
|
void vdpa_set_config(struct vdpa_device *vdev, unsigned int offset,
|
|
const void *buf, unsigned int length)
|
|
{
|
|
down_write(&vdev->cf_lock);
|
|
vdev->config->set_config(vdev, offset, buf, length);
|
|
up_write(&vdev->cf_lock);
|
|
}
|
|
EXPORT_SYMBOL_GPL(vdpa_set_config);
|
|
|
|
static bool mgmtdev_handle_match(const struct vdpa_mgmt_dev *mdev,
|
|
const char *busname, const char *devname)
|
|
{
|
|
/* Bus name is optional for simulated management device, so ignore the
|
|
* device with bus if bus attribute is provided.
|
|
*/
|
|
if ((busname && !mdev->device->bus) || (!busname && mdev->device->bus))
|
|
return false;
|
|
|
|
if (!busname && strcmp(dev_name(mdev->device), devname) == 0)
|
|
return true;
|
|
|
|
if (busname && (strcmp(mdev->device->bus->name, busname) == 0) &&
|
|
(strcmp(dev_name(mdev->device), devname) == 0))
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static struct vdpa_mgmt_dev *vdpa_mgmtdev_get_from_attr(struct nlattr **attrs)
|
|
{
|
|
struct vdpa_mgmt_dev *mdev;
|
|
const char *busname = NULL;
|
|
const char *devname;
|
|
|
|
if (!attrs[VDPA_ATTR_MGMTDEV_DEV_NAME])
|
|
return ERR_PTR(-EINVAL);
|
|
devname = nla_data(attrs[VDPA_ATTR_MGMTDEV_DEV_NAME]);
|
|
if (attrs[VDPA_ATTR_MGMTDEV_BUS_NAME])
|
|
busname = nla_data(attrs[VDPA_ATTR_MGMTDEV_BUS_NAME]);
|
|
|
|
list_for_each_entry(mdev, &mdev_head, list) {
|
|
if (mgmtdev_handle_match(mdev, busname, devname))
|
|
return mdev;
|
|
}
|
|
return ERR_PTR(-ENODEV);
|
|
}
|
|
|
|
static int vdpa_nl_mgmtdev_handle_fill(struct sk_buff *msg, const struct vdpa_mgmt_dev *mdev)
|
|
{
|
|
if (mdev->device->bus &&
|
|
nla_put_string(msg, VDPA_ATTR_MGMTDEV_BUS_NAME, mdev->device->bus->name))
|
|
return -EMSGSIZE;
|
|
if (nla_put_string(msg, VDPA_ATTR_MGMTDEV_DEV_NAME, dev_name(mdev->device)))
|
|
return -EMSGSIZE;
|
|
return 0;
|
|
}
|
|
|
|
static u64 vdpa_mgmtdev_get_classes(const struct vdpa_mgmt_dev *mdev,
|
|
unsigned int *nclasses)
|
|
{
|
|
u64 supported_classes = 0;
|
|
unsigned int n = 0;
|
|
|
|
for (int i = 0; mdev->id_table[i].device; i++) {
|
|
if (mdev->id_table[i].device > 63)
|
|
continue;
|
|
supported_classes |= BIT_ULL(mdev->id_table[i].device);
|
|
n++;
|
|
}
|
|
if (nclasses)
|
|
*nclasses = n;
|
|
|
|
return supported_classes;
|
|
}
|
|
|
|
static int vdpa_mgmtdev_fill(const struct vdpa_mgmt_dev *mdev, struct sk_buff *msg,
|
|
u32 portid, u32 seq, int flags)
|
|
{
|
|
void *hdr;
|
|
int err;
|
|
|
|
hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_MGMTDEV_NEW);
|
|
if (!hdr)
|
|
return -EMSGSIZE;
|
|
err = vdpa_nl_mgmtdev_handle_fill(msg, mdev);
|
|
if (err)
|
|
goto msg_err;
|
|
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_MGMTDEV_SUPPORTED_CLASSES,
|
|
vdpa_mgmtdev_get_classes(mdev, NULL),
|
|
VDPA_ATTR_UNSPEC)) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_MGMTDEV_MAX_VQS,
|
|
mdev->max_supported_vqs)) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_SUPPORTED_FEATURES,
|
|
mdev->supported_features, VDPA_ATTR_PAD)) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
|
|
genlmsg_end(msg, hdr);
|
|
return 0;
|
|
|
|
msg_err:
|
|
genlmsg_cancel(msg, hdr);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_mgmtdev_get_doit(struct sk_buff *skb, struct genl_info *info)
|
|
{
|
|
struct vdpa_mgmt_dev *mdev;
|
|
struct sk_buff *msg;
|
|
int err;
|
|
|
|
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
|
|
if (!msg)
|
|
return -ENOMEM;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
|
|
if (IS_ERR(mdev)) {
|
|
up_read(&vdpa_dev_lock);
|
|
NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified mgmt device");
|
|
err = PTR_ERR(mdev);
|
|
goto out;
|
|
}
|
|
|
|
err = vdpa_mgmtdev_fill(mdev, msg, info->snd_portid, info->snd_seq, 0);
|
|
up_read(&vdpa_dev_lock);
|
|
if (err)
|
|
goto out;
|
|
err = genlmsg_reply(msg, info);
|
|
return err;
|
|
|
|
out:
|
|
nlmsg_free(msg);
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
vdpa_nl_cmd_mgmtdev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
|
|
{
|
|
struct vdpa_mgmt_dev *mdev;
|
|
int start = cb->args[0];
|
|
int idx = 0;
|
|
int err;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
list_for_each_entry(mdev, &mdev_head, list) {
|
|
if (idx < start) {
|
|
idx++;
|
|
continue;
|
|
}
|
|
err = vdpa_mgmtdev_fill(mdev, msg, NETLINK_CB(cb->skb).portid,
|
|
cb->nlh->nlmsg_seq, NLM_F_MULTI);
|
|
if (err)
|
|
goto out;
|
|
idx++;
|
|
}
|
|
out:
|
|
up_read(&vdpa_dev_lock);
|
|
cb->args[0] = idx;
|
|
return msg->len;
|
|
}
|
|
|
|
#define VDPA_DEV_NET_ATTRS_MASK (BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR) | \
|
|
BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU) | \
|
|
BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP))
|
|
|
|
/*
|
|
* Bitmask for all per-device features: feature bits VIRTIO_TRANSPORT_F_START
|
|
* through VIRTIO_TRANSPORT_F_END are unset, i.e. 0xfffffc000fffffff for
|
|
* all 64bit features. If the features are extended beyond 64 bits, or new
|
|
* "holes" are reserved for other type of features than per-device, this
|
|
* macro would have to be updated.
|
|
*/
|
|
#define VIRTIO_DEVICE_F_MASK (~0ULL << (VIRTIO_TRANSPORT_F_END + 1) | \
|
|
((1ULL << VIRTIO_TRANSPORT_F_START) - 1))
|
|
|
|
static int vdpa_nl_cmd_dev_add_set_doit(struct sk_buff *skb, struct genl_info *info)
|
|
{
|
|
struct vdpa_dev_set_config config = {};
|
|
struct nlattr **nl_attrs = info->attrs;
|
|
struct vdpa_mgmt_dev *mdev;
|
|
unsigned int ncls = 0;
|
|
const u8 *macaddr;
|
|
const char *name;
|
|
u64 classes;
|
|
int err = 0;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
|
|
name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]) {
|
|
macaddr = nla_data(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]);
|
|
memcpy(config.net.mac, macaddr, sizeof(config.net.mac));
|
|
config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR);
|
|
}
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]) {
|
|
config.net.mtu =
|
|
nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]);
|
|
config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU);
|
|
}
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]) {
|
|
config.net.max_vq_pairs =
|
|
nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]);
|
|
if (!config.net.max_vq_pairs) {
|
|
NL_SET_ERR_MSG_MOD(info->extack,
|
|
"At least one pair of VQs is required");
|
|
return -EINVAL;
|
|
}
|
|
config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP);
|
|
}
|
|
if (nl_attrs[VDPA_ATTR_DEV_FEATURES]) {
|
|
u64 missing = 0x0ULL;
|
|
|
|
config.device_features =
|
|
nla_get_u64(nl_attrs[VDPA_ATTR_DEV_FEATURES]);
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR] &&
|
|
!(config.device_features & BIT_ULL(VIRTIO_NET_F_MAC)))
|
|
missing |= BIT_ULL(VIRTIO_NET_F_MAC);
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU] &&
|
|
!(config.device_features & BIT_ULL(VIRTIO_NET_F_MTU)))
|
|
missing |= BIT_ULL(VIRTIO_NET_F_MTU);
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP] &&
|
|
config.net.max_vq_pairs > 1 &&
|
|
!(config.device_features & BIT_ULL(VIRTIO_NET_F_MQ)))
|
|
missing |= BIT_ULL(VIRTIO_NET_F_MQ);
|
|
if (missing) {
|
|
NL_SET_ERR_MSG_FMT_MOD(info->extack,
|
|
"Missing features 0x%llx for provided attributes",
|
|
missing);
|
|
return -EINVAL;
|
|
}
|
|
config.mask |= BIT_ULL(VDPA_ATTR_DEV_FEATURES);
|
|
}
|
|
|
|
/* Skip checking capability if user didn't prefer to configure any
|
|
* device networking attributes. It is likely that user might have used
|
|
* a device specific method to configure such attributes or using device
|
|
* default attributes.
|
|
*/
|
|
if ((config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
|
|
!netlink_capable(skb, CAP_NET_ADMIN))
|
|
return -EPERM;
|
|
|
|
down_write(&vdpa_dev_lock);
|
|
mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
|
|
if (IS_ERR(mdev)) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified management device");
|
|
err = PTR_ERR(mdev);
|
|
goto err;
|
|
}
|
|
|
|
if ((config.mask & mdev->config_attr_mask) != config.mask) {
|
|
NL_SET_ERR_MSG_FMT_MOD(info->extack,
|
|
"Some provided attributes are not supported: 0x%llx",
|
|
config.mask & ~mdev->config_attr_mask);
|
|
err = -EOPNOTSUPP;
|
|
goto err;
|
|
}
|
|
|
|
classes = vdpa_mgmtdev_get_classes(mdev, &ncls);
|
|
if (config.mask & VDPA_DEV_NET_ATTRS_MASK &&
|
|
!(classes & BIT_ULL(VIRTIO_ID_NET))) {
|
|
NL_SET_ERR_MSG_MOD(info->extack,
|
|
"Network class attributes provided on unsupported management device");
|
|
err = -EINVAL;
|
|
goto err;
|
|
}
|
|
if (!(config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
|
|
config.mask & BIT_ULL(VDPA_ATTR_DEV_FEATURES) &&
|
|
classes & BIT_ULL(VIRTIO_ID_NET) && ncls > 1 &&
|
|
config.device_features & VIRTIO_DEVICE_F_MASK) {
|
|
NL_SET_ERR_MSG_MOD(info->extack,
|
|
"Management device supports multi-class while device features specified are ambiguous");
|
|
err = -EINVAL;
|
|
goto err;
|
|
}
|
|
|
|
err = mdev->ops->dev_add(mdev, name, &config);
|
|
err:
|
|
up_write(&vdpa_dev_lock);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_del_set_doit(struct sk_buff *skb, struct genl_info *info)
|
|
{
|
|
struct vdpa_mgmt_dev *mdev;
|
|
struct vdpa_device *vdev;
|
|
struct device *dev;
|
|
const char *name;
|
|
int err = 0;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
|
|
down_write(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, name, vdpa_name_match);
|
|
if (!dev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "device not found");
|
|
err = -ENODEV;
|
|
goto dev_err;
|
|
}
|
|
vdev = container_of(dev, struct vdpa_device, dev);
|
|
if (!vdev->mdev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "Only user created device can be deleted by user");
|
|
err = -EINVAL;
|
|
goto mdev_err;
|
|
}
|
|
mdev = vdev->mdev;
|
|
mdev->ops->dev_del(mdev, vdev);
|
|
mdev_err:
|
|
put_device(dev);
|
|
dev_err:
|
|
up_write(&vdpa_dev_lock);
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
|
|
int flags, struct netlink_ext_ack *extack)
|
|
{
|
|
u16 max_vq_size;
|
|
u16 min_vq_size = 1;
|
|
u32 device_id;
|
|
u32 vendor_id;
|
|
void *hdr;
|
|
int err;
|
|
|
|
hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_DEV_NEW);
|
|
if (!hdr)
|
|
return -EMSGSIZE;
|
|
|
|
err = vdpa_nl_mgmtdev_handle_fill(msg, vdev->mdev);
|
|
if (err)
|
|
goto msg_err;
|
|
|
|
device_id = vdev->config->get_device_id(vdev);
|
|
vendor_id = vdev->config->get_vendor_id(vdev);
|
|
max_vq_size = vdev->config->get_vq_num_max(vdev);
|
|
if (vdev->config->get_vq_num_min)
|
|
min_vq_size = vdev->config->get_vq_num_min(vdev);
|
|
|
|
err = -EMSGSIZE;
|
|
if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev)))
|
|
goto msg_err;
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id))
|
|
goto msg_err;
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_VENDOR_ID, vendor_id))
|
|
goto msg_err;
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_MAX_VQS, vdev->nvqs))
|
|
goto msg_err;
|
|
if (nla_put_u16(msg, VDPA_ATTR_DEV_MAX_VQ_SIZE, max_vq_size))
|
|
goto msg_err;
|
|
if (nla_put_u16(msg, VDPA_ATTR_DEV_MIN_VQ_SIZE, min_vq_size))
|
|
goto msg_err;
|
|
|
|
genlmsg_end(msg, hdr);
|
|
return 0;
|
|
|
|
msg_err:
|
|
genlmsg_cancel(msg, hdr);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_get_doit(struct sk_buff *skb, struct genl_info *info)
|
|
{
|
|
struct vdpa_device *vdev;
|
|
struct sk_buff *msg;
|
|
const char *devname;
|
|
struct device *dev;
|
|
int err;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
|
|
if (!msg)
|
|
return -ENOMEM;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
|
|
if (!dev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "device not found");
|
|
err = -ENODEV;
|
|
goto err;
|
|
}
|
|
vdev = container_of(dev, struct vdpa_device, dev);
|
|
if (!vdev->mdev) {
|
|
err = -EINVAL;
|
|
goto mdev_err;
|
|
}
|
|
err = vdpa_dev_fill(vdev, msg, info->snd_portid, info->snd_seq, 0, info->extack);
|
|
if (err)
|
|
goto mdev_err;
|
|
|
|
err = genlmsg_reply(msg, info);
|
|
put_device(dev);
|
|
up_read(&vdpa_dev_lock);
|
|
return err;
|
|
|
|
mdev_err:
|
|
put_device(dev);
|
|
err:
|
|
up_read(&vdpa_dev_lock);
|
|
nlmsg_free(msg);
|
|
return err;
|
|
}
|
|
|
|
struct vdpa_dev_dump_info {
|
|
struct sk_buff *msg;
|
|
struct netlink_callback *cb;
|
|
int start_idx;
|
|
int idx;
|
|
};
|
|
|
|
static int vdpa_dev_dump(struct device *dev, void *data)
|
|
{
|
|
struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
|
|
struct vdpa_dev_dump_info *info = data;
|
|
int err;
|
|
|
|
if (!vdev->mdev)
|
|
return 0;
|
|
if (info->idx < info->start_idx) {
|
|
info->idx++;
|
|
return 0;
|
|
}
|
|
err = vdpa_dev_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
|
|
info->cb->nlh->nlmsg_seq, NLM_F_MULTI, info->cb->extack);
|
|
if (err)
|
|
return err;
|
|
|
|
info->idx++;
|
|
return 0;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
|
|
{
|
|
struct vdpa_dev_dump_info info;
|
|
|
|
info.msg = msg;
|
|
info.cb = cb;
|
|
info.start_idx = cb->args[0];
|
|
info.idx = 0;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_dump);
|
|
up_read(&vdpa_dev_lock);
|
|
cb->args[0] = info.idx;
|
|
return msg->len;
|
|
}
|
|
|
|
static int vdpa_dev_net_mq_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_net_config *config)
|
|
{
|
|
u16 val_u16;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_NET_F_MQ)) == 0 &&
|
|
(features & BIT_ULL(VIRTIO_NET_F_RSS)) == 0)
|
|
return 0;
|
|
|
|
val_u16 = __virtio16_to_cpu(true, config->max_virtqueue_pairs);
|
|
|
|
return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MAX_VQP, val_u16);
|
|
}
|
|
|
|
static int vdpa_dev_net_mtu_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_net_config *config)
|
|
{
|
|
u16 val_u16;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_NET_F_MTU)) == 0)
|
|
return 0;
|
|
|
|
val_u16 = __virtio16_to_cpu(true, config->mtu);
|
|
|
|
return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MTU, val_u16);
|
|
}
|
|
|
|
static int vdpa_dev_net_mac_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_net_config *config)
|
|
{
|
|
if ((features & BIT_ULL(VIRTIO_NET_F_MAC)) == 0)
|
|
return 0;
|
|
|
|
return nla_put(msg, VDPA_ATTR_DEV_NET_CFG_MACADDR,
|
|
sizeof(config->mac), config->mac);
|
|
}
|
|
|
|
static int vdpa_dev_net_status_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_net_config *config)
|
|
{
|
|
u16 val_u16;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_NET_F_STATUS)) == 0)
|
|
return 0;
|
|
|
|
val_u16 = __virtio16_to_cpu(true, config->status);
|
|
return nla_put_u16(msg, VDPA_ATTR_DEV_NET_STATUS, val_u16);
|
|
}
|
|
|
|
static int vdpa_dev_net_config_fill(struct vdpa_device *vdev, struct sk_buff *msg)
|
|
{
|
|
struct virtio_net_config config = {};
|
|
u64 features_device;
|
|
|
|
vdev->config->get_config(vdev, 0, &config, sizeof(config));
|
|
|
|
features_device = vdev->config->get_device_features(vdev);
|
|
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_FEATURES, features_device,
|
|
VDPA_ATTR_PAD))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_net_mtu_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_net_mac_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_net_status_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
return vdpa_dev_net_mq_config_fill(msg, features_device, &config);
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_blk_capacity_config_fill(struct sk_buff *msg,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u64 val_u64;
|
|
|
|
val_u64 = __virtio64_to_cpu(true, config->capacity);
|
|
|
|
return nla_put_u64_64bit(msg, VDPA_ATTR_DEV_BLK_CFG_CAPACITY,
|
|
val_u64, VDPA_ATTR_PAD);
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_blk_seg_size_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u32 val_u32;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_SIZE_MAX)) == 0)
|
|
return 0;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->size_max);
|
|
|
|
return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_SIZE_MAX, val_u32);
|
|
}
|
|
|
|
/* fill the block size*/
|
|
static int
|
|
vdpa_dev_blk_block_size_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u32 val_u32;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_BLK_SIZE)) == 0)
|
|
return 0;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->blk_size);
|
|
|
|
return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_BLK_SIZE, val_u32);
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_blk_seg_max_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u32 val_u32;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_SEG_MAX)) == 0)
|
|
return 0;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->seg_max);
|
|
|
|
return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_SEG_MAX, val_u32);
|
|
}
|
|
|
|
static int vdpa_dev_blk_mq_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u16 val_u16;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_MQ)) == 0)
|
|
return 0;
|
|
|
|
val_u16 = __virtio16_to_cpu(true, config->num_queues);
|
|
|
|
return nla_put_u16(msg, VDPA_ATTR_DEV_BLK_CFG_NUM_QUEUES, val_u16);
|
|
}
|
|
|
|
static int vdpa_dev_blk_topology_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u16 min_io_size;
|
|
u32 opt_io_size;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_TOPOLOGY)) == 0)
|
|
return 0;
|
|
|
|
min_io_size = __virtio16_to_cpu(true, config->min_io_size);
|
|
opt_io_size = __virtio32_to_cpu(true, config->opt_io_size);
|
|
|
|
if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_CFG_PHY_BLK_EXP,
|
|
config->physical_block_exp))
|
|
return -EMSGSIZE;
|
|
|
|
if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_CFG_ALIGN_OFFSET,
|
|
config->alignment_offset))
|
|
return -EMSGSIZE;
|
|
|
|
if (nla_put_u16(msg, VDPA_ATTR_DEV_BLK_CFG_MIN_IO_SIZE, min_io_size))
|
|
return -EMSGSIZE;
|
|
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_OPT_IO_SIZE, opt_io_size))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vdpa_dev_blk_discard_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u32 val_u32;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_DISCARD)) == 0)
|
|
return 0;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->max_discard_sectors);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_DISCARD_SEC, val_u32))
|
|
return -EMSGSIZE;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->max_discard_seg);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_DISCARD_SEG, val_u32))
|
|
return -EMSGSIZE;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->discard_sector_alignment);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_DISCARD_SEC_ALIGN, val_u32))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_blk_write_zeroes_config_fill(struct sk_buff *msg, u64 features,
|
|
const struct virtio_blk_config *config)
|
|
{
|
|
u32 val_u32;
|
|
|
|
if ((features & BIT_ULL(VIRTIO_BLK_F_WRITE_ZEROES)) == 0)
|
|
return 0;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->max_write_zeroes_sectors);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_WRITE_ZEROES_SEC, val_u32))
|
|
return -EMSGSIZE;
|
|
|
|
val_u32 = __virtio32_to_cpu(true, config->max_write_zeroes_seg);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_WRITE_ZEROES_SEG, val_u32))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vdpa_dev_blk_ro_config_fill(struct sk_buff *msg, u64 features)
|
|
{
|
|
u8 ro;
|
|
|
|
ro = ((features & BIT_ULL(VIRTIO_BLK_F_RO)) == 0) ? 0 : 1;
|
|
if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_READ_ONLY, ro))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vdpa_dev_blk_flush_config_fill(struct sk_buff *msg, u64 features)
|
|
{
|
|
u8 flush;
|
|
|
|
flush = ((features & BIT_ULL(VIRTIO_BLK_F_FLUSH)) == 0) ? 0 : 1;
|
|
if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_FLUSH, flush))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vdpa_dev_blk_config_fill(struct vdpa_device *vdev,
|
|
struct sk_buff *msg)
|
|
{
|
|
struct virtio_blk_config config = {};
|
|
u64 features_device;
|
|
|
|
vdev->config->get_config(vdev, 0, &config, sizeof(config));
|
|
|
|
features_device = vdev->config->get_device_features(vdev);
|
|
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_FEATURES, features_device,
|
|
VDPA_ATTR_PAD))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_capacity_config_fill(msg, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_seg_size_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_block_size_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_seg_max_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_mq_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_topology_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_discard_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_write_zeroes_config_fill(msg, features_device, &config))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_ro_config_fill(msg, features_device))
|
|
return -EMSGSIZE;
|
|
|
|
if (vdpa_dev_blk_flush_config_fill(msg, features_device))
|
|
return -EMSGSIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
vdpa_dev_config_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
|
|
int flags, struct netlink_ext_ack *extack)
|
|
{
|
|
u64 features_driver;
|
|
u8 status = 0;
|
|
u32 device_id;
|
|
void *hdr;
|
|
int err;
|
|
|
|
down_read(&vdev->cf_lock);
|
|
hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
|
|
VDPA_CMD_DEV_CONFIG_GET);
|
|
if (!hdr) {
|
|
err = -EMSGSIZE;
|
|
goto out;
|
|
}
|
|
|
|
if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
|
|
device_id = vdev->config->get_device_id(vdev);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
|
|
/* only read driver features after the feature negotiation is done */
|
|
status = vdev->config->get_status(vdev);
|
|
if (status & VIRTIO_CONFIG_S_FEATURES_OK) {
|
|
features_driver = vdev->config->get_driver_features(vdev);
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES, features_driver,
|
|
VDPA_ATTR_PAD)) {
|
|
err = -EMSGSIZE;
|
|
goto msg_err;
|
|
}
|
|
}
|
|
|
|
switch (device_id) {
|
|
case VIRTIO_ID_NET:
|
|
err = vdpa_dev_net_config_fill(vdev, msg);
|
|
break;
|
|
case VIRTIO_ID_BLOCK:
|
|
err = vdpa_dev_blk_config_fill(vdev, msg);
|
|
break;
|
|
default:
|
|
err = -EOPNOTSUPP;
|
|
break;
|
|
}
|
|
if (err)
|
|
goto msg_err;
|
|
|
|
up_read(&vdev->cf_lock);
|
|
genlmsg_end(msg, hdr);
|
|
return 0;
|
|
|
|
msg_err:
|
|
genlmsg_cancel(msg, hdr);
|
|
out:
|
|
up_read(&vdev->cf_lock);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_fill_stats_rec(struct vdpa_device *vdev, struct sk_buff *msg,
|
|
struct genl_info *info, u32 index)
|
|
{
|
|
struct virtio_net_config config = {};
|
|
u64 features;
|
|
u8 status;
|
|
int err;
|
|
|
|
status = vdev->config->get_status(vdev);
|
|
if (!(status & VIRTIO_CONFIG_S_FEATURES_OK)) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "feature negotiation not complete");
|
|
return -EAGAIN;
|
|
}
|
|
vdpa_get_config_unlocked(vdev, 0, &config, sizeof(config));
|
|
|
|
features = vdev->config->get_driver_features(vdev);
|
|
if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES,
|
|
features, VDPA_ATTR_PAD))
|
|
return -EMSGSIZE;
|
|
|
|
err = vdpa_dev_net_mq_config_fill(msg, features, &config);
|
|
if (err)
|
|
return err;
|
|
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_QUEUE_INDEX, index))
|
|
return -EMSGSIZE;
|
|
|
|
err = vdev->config->get_vendor_vq_stats(vdev, index, msg, info->extack);
|
|
if (err)
|
|
return err;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vendor_stats_fill(struct vdpa_device *vdev, struct sk_buff *msg,
|
|
struct genl_info *info, u32 index)
|
|
{
|
|
int err;
|
|
|
|
down_read(&vdev->cf_lock);
|
|
if (!vdev->config->get_vendor_vq_stats) {
|
|
err = -EOPNOTSUPP;
|
|
goto out;
|
|
}
|
|
|
|
err = vdpa_fill_stats_rec(vdev, msg, info, index);
|
|
out:
|
|
up_read(&vdev->cf_lock);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_dev_vendor_stats_fill(struct vdpa_device *vdev,
|
|
struct sk_buff *msg,
|
|
struct genl_info *info, u32 index)
|
|
{
|
|
u32 device_id;
|
|
void *hdr;
|
|
int err;
|
|
u32 portid = info->snd_portid;
|
|
u32 seq = info->snd_seq;
|
|
u32 flags = 0;
|
|
|
|
hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
|
|
VDPA_CMD_DEV_VSTATS_GET);
|
|
if (!hdr)
|
|
return -EMSGSIZE;
|
|
|
|
if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
|
|
err = -EMSGSIZE;
|
|
goto undo_msg;
|
|
}
|
|
|
|
device_id = vdev->config->get_device_id(vdev);
|
|
if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
|
|
err = -EMSGSIZE;
|
|
goto undo_msg;
|
|
}
|
|
|
|
switch (device_id) {
|
|
case VIRTIO_ID_NET:
|
|
if (index > VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MAX) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "queue index exceeds max value");
|
|
err = -ERANGE;
|
|
break;
|
|
}
|
|
|
|
err = vendor_stats_fill(vdev, msg, info, index);
|
|
break;
|
|
default:
|
|
err = -EOPNOTSUPP;
|
|
break;
|
|
}
|
|
genlmsg_end(msg, hdr);
|
|
|
|
return err;
|
|
|
|
undo_msg:
|
|
genlmsg_cancel(msg, hdr);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_config_get_doit(struct sk_buff *skb, struct genl_info *info)
|
|
{
|
|
struct vdpa_device *vdev;
|
|
struct sk_buff *msg;
|
|
const char *devname;
|
|
struct device *dev;
|
|
int err;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
|
|
if (!msg)
|
|
return -ENOMEM;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
|
|
if (!dev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "device not found");
|
|
err = -ENODEV;
|
|
goto dev_err;
|
|
}
|
|
vdev = container_of(dev, struct vdpa_device, dev);
|
|
if (!vdev->mdev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
|
|
err = -EINVAL;
|
|
goto mdev_err;
|
|
}
|
|
err = vdpa_dev_config_fill(vdev, msg, info->snd_portid, info->snd_seq,
|
|
0, info->extack);
|
|
if (!err)
|
|
err = genlmsg_reply(msg, info);
|
|
|
|
mdev_err:
|
|
put_device(dev);
|
|
dev_err:
|
|
up_read(&vdpa_dev_lock);
|
|
if (err)
|
|
nlmsg_free(msg);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_dev_net_device_attr_set(struct vdpa_device *vdev,
|
|
struct genl_info *info)
|
|
{
|
|
struct vdpa_dev_set_config set_config = {};
|
|
struct vdpa_mgmt_dev *mdev = vdev->mdev;
|
|
struct nlattr **nl_attrs = info->attrs;
|
|
const u8 *macaddr;
|
|
int err = -EOPNOTSUPP;
|
|
|
|
down_write(&vdev->cf_lock);
|
|
if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]) {
|
|
set_config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR);
|
|
macaddr = nla_data(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]);
|
|
|
|
if (is_valid_ether_addr(macaddr)) {
|
|
ether_addr_copy(set_config.net.mac, macaddr);
|
|
if (mdev->ops->dev_set_attr) {
|
|
err = mdev->ops->dev_set_attr(mdev, vdev,
|
|
&set_config);
|
|
} else {
|
|
NL_SET_ERR_MSG_FMT_MOD(info->extack,
|
|
"Operation not supported by the device.");
|
|
}
|
|
} else {
|
|
NL_SET_ERR_MSG_FMT_MOD(info->extack,
|
|
"Invalid MAC address");
|
|
}
|
|
}
|
|
up_write(&vdev->cf_lock);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_attr_set_doit(struct sk_buff *skb,
|
|
struct genl_info *info)
|
|
{
|
|
struct vdpa_device *vdev;
|
|
struct device *dev;
|
|
const char *name;
|
|
u64 classes;
|
|
int err = 0;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
|
|
name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
|
|
down_write(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, name, vdpa_name_match);
|
|
if (!dev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "device not found");
|
|
err = -ENODEV;
|
|
goto dev_err;
|
|
}
|
|
vdev = container_of(dev, struct vdpa_device, dev);
|
|
if (!vdev->mdev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
|
|
err = -EINVAL;
|
|
goto mdev_err;
|
|
}
|
|
classes = vdpa_mgmtdev_get_classes(vdev->mdev, NULL);
|
|
if (classes & BIT_ULL(VIRTIO_ID_NET)) {
|
|
err = vdpa_dev_net_device_attr_set(vdev, info);
|
|
} else {
|
|
NL_SET_ERR_MSG_FMT_MOD(info->extack, "%s device not supported",
|
|
name);
|
|
}
|
|
|
|
mdev_err:
|
|
put_device(dev);
|
|
dev_err:
|
|
up_write(&vdpa_dev_lock);
|
|
return err;
|
|
}
|
|
|
|
static int vdpa_dev_config_dump(struct device *dev, void *data)
|
|
{
|
|
struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
|
|
struct vdpa_dev_dump_info *info = data;
|
|
int err;
|
|
|
|
if (!vdev->mdev)
|
|
return 0;
|
|
if (info->idx < info->start_idx) {
|
|
info->idx++;
|
|
return 0;
|
|
}
|
|
err = vdpa_dev_config_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
|
|
info->cb->nlh->nlmsg_seq, NLM_F_MULTI,
|
|
info->cb->extack);
|
|
if (err)
|
|
return err;
|
|
|
|
info->idx++;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
vdpa_nl_cmd_dev_config_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
|
|
{
|
|
struct vdpa_dev_dump_info info;
|
|
|
|
info.msg = msg;
|
|
info.cb = cb;
|
|
info.start_idx = cb->args[0];
|
|
info.idx = 0;
|
|
|
|
down_read(&vdpa_dev_lock);
|
|
bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_config_dump);
|
|
up_read(&vdpa_dev_lock);
|
|
cb->args[0] = info.idx;
|
|
return msg->len;
|
|
}
|
|
|
|
static int vdpa_nl_cmd_dev_stats_get_doit(struct sk_buff *skb,
|
|
struct genl_info *info)
|
|
{
|
|
struct vdpa_device *vdev;
|
|
struct sk_buff *msg;
|
|
const char *devname;
|
|
struct device *dev;
|
|
u32 index;
|
|
int err;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_NAME])
|
|
return -EINVAL;
|
|
|
|
if (!info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX])
|
|
return -EINVAL;
|
|
|
|
devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
|
|
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
|
|
if (!msg)
|
|
return -ENOMEM;
|
|
|
|
index = nla_get_u32(info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX]);
|
|
down_read(&vdpa_dev_lock);
|
|
dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
|
|
if (!dev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "device not found");
|
|
err = -ENODEV;
|
|
goto dev_err;
|
|
}
|
|
vdev = container_of(dev, struct vdpa_device, dev);
|
|
if (!vdev->mdev) {
|
|
NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
|
|
err = -EINVAL;
|
|
goto mdev_err;
|
|
}
|
|
err = vdpa_dev_vendor_stats_fill(vdev, msg, info, index);
|
|
if (err)
|
|
goto mdev_err;
|
|
|
|
err = genlmsg_reply(msg, info);
|
|
|
|
put_device(dev);
|
|
up_read(&vdpa_dev_lock);
|
|
|
|
return err;
|
|
|
|
mdev_err:
|
|
put_device(dev);
|
|
dev_err:
|
|
nlmsg_free(msg);
|
|
up_read(&vdpa_dev_lock);
|
|
return err;
|
|
}
|
|
|
|
static const struct nla_policy vdpa_nl_policy[VDPA_ATTR_MAX + 1] = {
|
|
[VDPA_ATTR_MGMTDEV_BUS_NAME] = { .type = NLA_NUL_STRING },
|
|
[VDPA_ATTR_MGMTDEV_DEV_NAME] = { .type = NLA_STRING },
|
|
[VDPA_ATTR_DEV_NAME] = { .type = NLA_STRING },
|
|
[VDPA_ATTR_DEV_NET_CFG_MACADDR] = NLA_POLICY_ETH_ADDR,
|
|
[VDPA_ATTR_DEV_NET_CFG_MAX_VQP] = { .type = NLA_U16 },
|
|
/* virtio spec 1.1 section 5.1.4.1 for valid MTU range */
|
|
[VDPA_ATTR_DEV_NET_CFG_MTU] = NLA_POLICY_MIN(NLA_U16, 68),
|
|
[VDPA_ATTR_DEV_QUEUE_INDEX] = { .type = NLA_U32 },
|
|
[VDPA_ATTR_DEV_FEATURES] = { .type = NLA_U64 },
|
|
};
|
|
|
|
static const struct genl_ops vdpa_nl_ops[] = {
|
|
{
|
|
.cmd = VDPA_CMD_MGMTDEV_GET,
|
|
.doit = vdpa_nl_cmd_mgmtdev_get_doit,
|
|
.dumpit = vdpa_nl_cmd_mgmtdev_get_dumpit,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_NEW,
|
|
.doit = vdpa_nl_cmd_dev_add_set_doit,
|
|
.flags = GENL_ADMIN_PERM,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_DEL,
|
|
.doit = vdpa_nl_cmd_dev_del_set_doit,
|
|
.flags = GENL_ADMIN_PERM,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_GET,
|
|
.doit = vdpa_nl_cmd_dev_get_doit,
|
|
.dumpit = vdpa_nl_cmd_dev_get_dumpit,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_CONFIG_GET,
|
|
.doit = vdpa_nl_cmd_dev_config_get_doit,
|
|
.dumpit = vdpa_nl_cmd_dev_config_get_dumpit,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_VSTATS_GET,
|
|
.doit = vdpa_nl_cmd_dev_stats_get_doit,
|
|
.flags = GENL_ADMIN_PERM,
|
|
},
|
|
{
|
|
.cmd = VDPA_CMD_DEV_ATTR_SET,
|
|
.doit = vdpa_nl_cmd_dev_attr_set_doit,
|
|
.flags = GENL_ADMIN_PERM,
|
|
},
|
|
};
|
|
|
|
static struct genl_family vdpa_nl_family __ro_after_init = {
|
|
.name = VDPA_GENL_NAME,
|
|
.version = VDPA_GENL_VERSION,
|
|
.maxattr = VDPA_ATTR_MAX,
|
|
.policy = vdpa_nl_policy,
|
|
.netnsok = false,
|
|
.module = THIS_MODULE,
|
|
.ops = vdpa_nl_ops,
|
|
.n_ops = ARRAY_SIZE(vdpa_nl_ops),
|
|
.resv_start_op = VDPA_CMD_DEV_VSTATS_GET + 1,
|
|
};
|
|
|
|
static int vdpa_init(void)
|
|
{
|
|
int err;
|
|
|
|
err = bus_register(&vdpa_bus);
|
|
if (err)
|
|
return err;
|
|
err = genl_register_family(&vdpa_nl_family);
|
|
if (err)
|
|
goto err;
|
|
return 0;
|
|
|
|
err:
|
|
bus_unregister(&vdpa_bus);
|
|
return err;
|
|
}
|
|
|
|
static void __exit vdpa_exit(void)
|
|
{
|
|
genl_unregister_family(&vdpa_nl_family);
|
|
bus_unregister(&vdpa_bus);
|
|
ida_destroy(&vdpa_index_ida);
|
|
}
|
|
core_initcall(vdpa_init);
|
|
module_exit(vdpa_exit);
|
|
|
|
MODULE_AUTHOR("Jason Wang <jasowang@redhat.com>");
|
|
MODULE_DESCRIPTION("vDPA bus");
|
|
MODULE_LICENSE("GPL v2");
|