mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-04 01:24:12 +08:00
89e3becd8f
Add device state check before executing command. Without the check the
command can be issued while device is in halt state and causes the driver to
block while waiting for the completion of the command.
Reported-by: Sanjay Kumar <sanjay.k.kumar@intel.com>
Signed-off-by: Dave Jiang <dave.jiang@intel.com>
Tested-by: Sanjay Kumar <sanjay.k.kumar@intel.com>
Fixes: 0d5c10b4c8
("dmaengine: idxd: add work queue drain support")
Link: https://lore.kernel.org/r/161219313921.2976211.12222625226450097465.stgit@djiang5-desk3.ch.intel.com
Signed-off-by: Vinod Koul <vkoul@kernel.org>
825 lines
19 KiB
C
825 lines
19 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright(c) 2019 Intel Corporation. All rights rsvd. */
|
|
#include <linux/init.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/io-64-nonatomic-lo-hi.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/msi.h>
|
|
#include <uapi/linux/idxd.h>
|
|
#include "../dmaengine.h"
|
|
#include "idxd.h"
|
|
#include "registers.h"
|
|
|
|
static void idxd_cmd_exec(struct idxd_device *idxd, int cmd_code, u32 operand,
|
|
u32 *status);
|
|
|
|
/* Interrupt control bits */
|
|
void idxd_mask_msix_vector(struct idxd_device *idxd, int vec_id)
|
|
{
|
|
struct irq_data *data = irq_get_irq_data(idxd->msix_entries[vec_id].vector);
|
|
|
|
pci_msi_mask_irq(data);
|
|
}
|
|
|
|
void idxd_mask_msix_vectors(struct idxd_device *idxd)
|
|
{
|
|
struct pci_dev *pdev = idxd->pdev;
|
|
int msixcnt = pci_msix_vec_count(pdev);
|
|
int i;
|
|
|
|
for (i = 0; i < msixcnt; i++)
|
|
idxd_mask_msix_vector(idxd, i);
|
|
}
|
|
|
|
void idxd_unmask_msix_vector(struct idxd_device *idxd, int vec_id)
|
|
{
|
|
struct irq_data *data = irq_get_irq_data(idxd->msix_entries[vec_id].vector);
|
|
|
|
pci_msi_unmask_irq(data);
|
|
}
|
|
|
|
void idxd_unmask_error_interrupts(struct idxd_device *idxd)
|
|
{
|
|
union genctrl_reg genctrl;
|
|
|
|
genctrl.bits = ioread32(idxd->reg_base + IDXD_GENCTRL_OFFSET);
|
|
genctrl.softerr_int_en = 1;
|
|
iowrite32(genctrl.bits, idxd->reg_base + IDXD_GENCTRL_OFFSET);
|
|
}
|
|
|
|
void idxd_mask_error_interrupts(struct idxd_device *idxd)
|
|
{
|
|
union genctrl_reg genctrl;
|
|
|
|
genctrl.bits = ioread32(idxd->reg_base + IDXD_GENCTRL_OFFSET);
|
|
genctrl.softerr_int_en = 0;
|
|
iowrite32(genctrl.bits, idxd->reg_base + IDXD_GENCTRL_OFFSET);
|
|
}
|
|
|
|
static void free_hw_descs(struct idxd_wq *wq)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < wq->num_descs; i++)
|
|
kfree(wq->hw_descs[i]);
|
|
|
|
kfree(wq->hw_descs);
|
|
}
|
|
|
|
static int alloc_hw_descs(struct idxd_wq *wq, int num)
|
|
{
|
|
struct device *dev = &wq->idxd->pdev->dev;
|
|
int i;
|
|
int node = dev_to_node(dev);
|
|
|
|
wq->hw_descs = kcalloc_node(num, sizeof(struct dsa_hw_desc *),
|
|
GFP_KERNEL, node);
|
|
if (!wq->hw_descs)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < num; i++) {
|
|
wq->hw_descs[i] = kzalloc_node(sizeof(*wq->hw_descs[i]),
|
|
GFP_KERNEL, node);
|
|
if (!wq->hw_descs[i]) {
|
|
free_hw_descs(wq);
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void free_descs(struct idxd_wq *wq)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < wq->num_descs; i++)
|
|
kfree(wq->descs[i]);
|
|
|
|
kfree(wq->descs);
|
|
}
|
|
|
|
static int alloc_descs(struct idxd_wq *wq, int num)
|
|
{
|
|
struct device *dev = &wq->idxd->pdev->dev;
|
|
int i;
|
|
int node = dev_to_node(dev);
|
|
|
|
wq->descs = kcalloc_node(num, sizeof(struct idxd_desc *),
|
|
GFP_KERNEL, node);
|
|
if (!wq->descs)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < num; i++) {
|
|
wq->descs[i] = kzalloc_node(sizeof(*wq->descs[i]),
|
|
GFP_KERNEL, node);
|
|
if (!wq->descs[i]) {
|
|
free_descs(wq);
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* WQ control bits */
|
|
int idxd_wq_alloc_resources(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
int rc, num_descs, i;
|
|
int align;
|
|
u64 tmp;
|
|
|
|
if (wq->type != IDXD_WQT_KERNEL)
|
|
return 0;
|
|
|
|
wq->num_descs = wq->size;
|
|
num_descs = wq->size;
|
|
|
|
rc = alloc_hw_descs(wq, num_descs);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
if (idxd->type == IDXD_TYPE_DSA)
|
|
align = 32;
|
|
else if (idxd->type == IDXD_TYPE_IAX)
|
|
align = 64;
|
|
else
|
|
return -ENODEV;
|
|
|
|
wq->compls_size = num_descs * idxd->compl_size + align;
|
|
wq->compls_raw = dma_alloc_coherent(dev, wq->compls_size,
|
|
&wq->compls_addr_raw, GFP_KERNEL);
|
|
if (!wq->compls_raw) {
|
|
rc = -ENOMEM;
|
|
goto fail_alloc_compls;
|
|
}
|
|
|
|
/* Adjust alignment */
|
|
wq->compls_addr = (wq->compls_addr_raw + (align - 1)) & ~(align - 1);
|
|
tmp = (u64)wq->compls_raw;
|
|
tmp = (tmp + (align - 1)) & ~(align - 1);
|
|
wq->compls = (struct dsa_completion_record *)tmp;
|
|
|
|
rc = alloc_descs(wq, num_descs);
|
|
if (rc < 0)
|
|
goto fail_alloc_descs;
|
|
|
|
rc = sbitmap_queue_init_node(&wq->sbq, num_descs, -1, false, GFP_KERNEL,
|
|
dev_to_node(dev));
|
|
if (rc < 0)
|
|
goto fail_sbitmap_init;
|
|
|
|
for (i = 0; i < num_descs; i++) {
|
|
struct idxd_desc *desc = wq->descs[i];
|
|
|
|
desc->hw = wq->hw_descs[i];
|
|
if (idxd->type == IDXD_TYPE_DSA)
|
|
desc->completion = &wq->compls[i];
|
|
else if (idxd->type == IDXD_TYPE_IAX)
|
|
desc->iax_completion = &wq->iax_compls[i];
|
|
desc->compl_dma = wq->compls_addr + idxd->compl_size * i;
|
|
desc->id = i;
|
|
desc->wq = wq;
|
|
desc->cpu = -1;
|
|
dma_async_tx_descriptor_init(&desc->txd, &wq->dma_chan);
|
|
desc->txd.tx_submit = idxd_dma_tx_submit;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail_sbitmap_init:
|
|
free_descs(wq);
|
|
fail_alloc_descs:
|
|
dma_free_coherent(dev, wq->compls_size, wq->compls_raw,
|
|
wq->compls_addr_raw);
|
|
fail_alloc_compls:
|
|
free_hw_descs(wq);
|
|
return rc;
|
|
}
|
|
|
|
void idxd_wq_free_resources(struct idxd_wq *wq)
|
|
{
|
|
struct device *dev = &wq->idxd->pdev->dev;
|
|
|
|
if (wq->type != IDXD_WQT_KERNEL)
|
|
return;
|
|
|
|
free_hw_descs(wq);
|
|
free_descs(wq);
|
|
dma_free_coherent(dev, wq->compls_size, wq->compls_raw,
|
|
wq->compls_addr_raw);
|
|
sbitmap_queue_free(&wq->sbq);
|
|
}
|
|
|
|
int idxd_wq_enable(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 status;
|
|
|
|
if (wq->state == IDXD_WQ_ENABLED) {
|
|
dev_dbg(dev, "WQ %d already enabled\n", wq->id);
|
|
return -ENXIO;
|
|
}
|
|
|
|
idxd_cmd_exec(idxd, IDXD_CMD_ENABLE_WQ, wq->id, &status);
|
|
|
|
if (status != IDXD_CMDSTS_SUCCESS &&
|
|
status != IDXD_CMDSTS_ERR_WQ_ENABLED) {
|
|
dev_dbg(dev, "WQ enable failed: %#x\n", status);
|
|
return -ENXIO;
|
|
}
|
|
|
|
wq->state = IDXD_WQ_ENABLED;
|
|
dev_dbg(dev, "WQ %d enabled\n", wq->id);
|
|
return 0;
|
|
}
|
|
|
|
int idxd_wq_disable(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 status, operand;
|
|
|
|
dev_dbg(dev, "Disabling WQ %d\n", wq->id);
|
|
|
|
if (wq->state != IDXD_WQ_ENABLED) {
|
|
dev_dbg(dev, "WQ %d in wrong state: %d\n", wq->id, wq->state);
|
|
return 0;
|
|
}
|
|
|
|
operand = BIT(wq->id % 16) | ((wq->id / 16) << 16);
|
|
idxd_cmd_exec(idxd, IDXD_CMD_DISABLE_WQ, operand, &status);
|
|
|
|
if (status != IDXD_CMDSTS_SUCCESS) {
|
|
dev_dbg(dev, "WQ disable failed: %#x\n", status);
|
|
return -ENXIO;
|
|
}
|
|
|
|
wq->state = IDXD_WQ_DISABLED;
|
|
dev_dbg(dev, "WQ %d disabled\n", wq->id);
|
|
return 0;
|
|
}
|
|
|
|
void idxd_wq_drain(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 operand;
|
|
|
|
if (wq->state != IDXD_WQ_ENABLED) {
|
|
dev_dbg(dev, "WQ %d in wrong state: %d\n", wq->id, wq->state);
|
|
return;
|
|
}
|
|
|
|
dev_dbg(dev, "Draining WQ %d\n", wq->id);
|
|
operand = BIT(wq->id % 16) | ((wq->id / 16) << 16);
|
|
idxd_cmd_exec(idxd, IDXD_CMD_DRAIN_WQ, operand, NULL);
|
|
}
|
|
|
|
int idxd_wq_map_portal(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct pci_dev *pdev = idxd->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
resource_size_t start;
|
|
|
|
start = pci_resource_start(pdev, IDXD_WQ_BAR);
|
|
start += idxd_get_wq_portal_full_offset(wq->id, IDXD_PORTAL_LIMITED);
|
|
|
|
wq->portal = devm_ioremap(dev, start, IDXD_PORTAL_SIZE);
|
|
if (!wq->portal)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void idxd_wq_unmap_portal(struct idxd_wq *wq)
|
|
{
|
|
struct device *dev = &wq->idxd->pdev->dev;
|
|
|
|
devm_iounmap(dev, wq->portal);
|
|
}
|
|
|
|
int idxd_wq_set_pasid(struct idxd_wq *wq, int pasid)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
int rc;
|
|
union wqcfg wqcfg;
|
|
unsigned int offset;
|
|
unsigned long flags;
|
|
|
|
rc = idxd_wq_disable(wq);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
offset = WQCFG_OFFSET(idxd, wq->id, WQCFG_PASID_IDX);
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
wqcfg.bits[WQCFG_PASID_IDX] = ioread32(idxd->reg_base + offset);
|
|
wqcfg.pasid_en = 1;
|
|
wqcfg.pasid = pasid;
|
|
iowrite32(wqcfg.bits[WQCFG_PASID_IDX], idxd->reg_base + offset);
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
|
|
rc = idxd_wq_enable(wq);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int idxd_wq_disable_pasid(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
int rc;
|
|
union wqcfg wqcfg;
|
|
unsigned int offset;
|
|
unsigned long flags;
|
|
|
|
rc = idxd_wq_disable(wq);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
offset = WQCFG_OFFSET(idxd, wq->id, WQCFG_PASID_IDX);
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
wqcfg.bits[WQCFG_PASID_IDX] = ioread32(idxd->reg_base + offset);
|
|
wqcfg.pasid_en = 0;
|
|
wqcfg.pasid = 0;
|
|
iowrite32(wqcfg.bits[WQCFG_PASID_IDX], idxd->reg_base + offset);
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
|
|
rc = idxd_wq_enable(wq);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void idxd_wq_disable_cleanup(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
int i, wq_offset;
|
|
|
|
lockdep_assert_held(&idxd->dev_lock);
|
|
memset(wq->wqcfg, 0, idxd->wqcfg_size);
|
|
wq->type = IDXD_WQT_NONE;
|
|
wq->size = 0;
|
|
wq->group = NULL;
|
|
wq->threshold = 0;
|
|
wq->priority = 0;
|
|
wq->ats_dis = 0;
|
|
clear_bit(WQ_FLAG_DEDICATED, &wq->flags);
|
|
memset(wq->name, 0, WQ_NAME_SIZE);
|
|
|
|
for (i = 0; i < WQCFG_STRIDES(idxd); i++) {
|
|
wq_offset = WQCFG_OFFSET(idxd, wq->id, i);
|
|
iowrite32(0, idxd->reg_base + wq_offset);
|
|
dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n",
|
|
wq->id, i, wq_offset,
|
|
ioread32(idxd->reg_base + wq_offset));
|
|
}
|
|
}
|
|
|
|
/* Device control bits */
|
|
static inline bool idxd_is_enabled(struct idxd_device *idxd)
|
|
{
|
|
union gensts_reg gensts;
|
|
|
|
gensts.bits = ioread32(idxd->reg_base + IDXD_GENSTATS_OFFSET);
|
|
|
|
if (gensts.state == IDXD_DEVICE_STATE_ENABLED)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static inline bool idxd_device_is_halted(struct idxd_device *idxd)
|
|
{
|
|
union gensts_reg gensts;
|
|
|
|
gensts.bits = ioread32(idxd->reg_base + IDXD_GENSTATS_OFFSET);
|
|
|
|
return (gensts.state == IDXD_DEVICE_STATE_HALT);
|
|
}
|
|
|
|
/*
|
|
* This is function is only used for reset during probe and will
|
|
* poll for completion. Once the device is setup with interrupts,
|
|
* all commands will be done via interrupt completion.
|
|
*/
|
|
int idxd_device_init_reset(struct idxd_device *idxd)
|
|
{
|
|
struct device *dev = &idxd->pdev->dev;
|
|
union idxd_command_reg cmd;
|
|
unsigned long flags;
|
|
|
|
if (idxd_device_is_halted(idxd)) {
|
|
dev_warn(&idxd->pdev->dev, "Device is HALTED!\n");
|
|
return -ENXIO;
|
|
}
|
|
|
|
memset(&cmd, 0, sizeof(cmd));
|
|
cmd.cmd = IDXD_CMD_RESET_DEVICE;
|
|
dev_dbg(dev, "%s: sending reset for init.\n", __func__);
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
iowrite32(cmd.bits, idxd->reg_base + IDXD_CMD_OFFSET);
|
|
|
|
while (ioread32(idxd->reg_base + IDXD_CMDSTS_OFFSET) &
|
|
IDXD_CMDSTS_ACTIVE)
|
|
cpu_relax();
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
return 0;
|
|
}
|
|
|
|
static void idxd_cmd_exec(struct idxd_device *idxd, int cmd_code, u32 operand,
|
|
u32 *status)
|
|
{
|
|
union idxd_command_reg cmd;
|
|
DECLARE_COMPLETION_ONSTACK(done);
|
|
unsigned long flags;
|
|
|
|
if (idxd_device_is_halted(idxd)) {
|
|
dev_warn(&idxd->pdev->dev, "Device is HALTED!\n");
|
|
*status = IDXD_CMDSTS_HW_ERR;
|
|
return;
|
|
}
|
|
|
|
memset(&cmd, 0, sizeof(cmd));
|
|
cmd.cmd = cmd_code;
|
|
cmd.operand = operand;
|
|
cmd.int_req = 1;
|
|
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
wait_event_lock_irq(idxd->cmd_waitq,
|
|
!test_bit(IDXD_FLAG_CMD_RUNNING, &idxd->flags),
|
|
idxd->dev_lock);
|
|
|
|
dev_dbg(&idxd->pdev->dev, "%s: sending cmd: %#x op: %#x\n",
|
|
__func__, cmd_code, operand);
|
|
|
|
idxd->cmd_status = 0;
|
|
__set_bit(IDXD_FLAG_CMD_RUNNING, &idxd->flags);
|
|
idxd->cmd_done = &done;
|
|
iowrite32(cmd.bits, idxd->reg_base + IDXD_CMD_OFFSET);
|
|
|
|
/*
|
|
* After command submitted, release lock and go to sleep until
|
|
* the command completes via interrupt.
|
|
*/
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
wait_for_completion(&done);
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
if (status) {
|
|
*status = ioread32(idxd->reg_base + IDXD_CMDSTS_OFFSET);
|
|
idxd->cmd_status = *status & GENMASK(7, 0);
|
|
}
|
|
|
|
__clear_bit(IDXD_FLAG_CMD_RUNNING, &idxd->flags);
|
|
/* Wake up other pending commands */
|
|
wake_up(&idxd->cmd_waitq);
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
}
|
|
|
|
int idxd_device_enable(struct idxd_device *idxd)
|
|
{
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 status;
|
|
|
|
if (idxd_is_enabled(idxd)) {
|
|
dev_dbg(dev, "Device already enabled\n");
|
|
return -ENXIO;
|
|
}
|
|
|
|
idxd_cmd_exec(idxd, IDXD_CMD_ENABLE_DEVICE, 0, &status);
|
|
|
|
/* If the command is successful or if the device was enabled */
|
|
if (status != IDXD_CMDSTS_SUCCESS &&
|
|
status != IDXD_CMDSTS_ERR_DEV_ENABLED) {
|
|
dev_dbg(dev, "%s: err_code: %#x\n", __func__, status);
|
|
return -ENXIO;
|
|
}
|
|
|
|
idxd->state = IDXD_DEV_ENABLED;
|
|
return 0;
|
|
}
|
|
|
|
void idxd_device_wqs_clear_state(struct idxd_device *idxd)
|
|
{
|
|
int i;
|
|
|
|
lockdep_assert_held(&idxd->dev_lock);
|
|
|
|
for (i = 0; i < idxd->max_wqs; i++) {
|
|
struct idxd_wq *wq = &idxd->wqs[i];
|
|
|
|
if (wq->state == IDXD_WQ_ENABLED) {
|
|
idxd_wq_disable_cleanup(wq);
|
|
wq->state = IDXD_WQ_DISABLED;
|
|
}
|
|
}
|
|
}
|
|
|
|
int idxd_device_disable(struct idxd_device *idxd)
|
|
{
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 status;
|
|
unsigned long flags;
|
|
|
|
if (!idxd_is_enabled(idxd)) {
|
|
dev_dbg(dev, "Device is not enabled\n");
|
|
return 0;
|
|
}
|
|
|
|
idxd_cmd_exec(idxd, IDXD_CMD_DISABLE_DEVICE, 0, &status);
|
|
|
|
/* If the command is successful or if the device was disabled */
|
|
if (status != IDXD_CMDSTS_SUCCESS &&
|
|
!(status & IDXD_CMDSTS_ERR_DIS_DEV_EN)) {
|
|
dev_dbg(dev, "%s: err_code: %#x\n", __func__, status);
|
|
return -ENXIO;
|
|
}
|
|
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
idxd_device_wqs_clear_state(idxd);
|
|
idxd->state = IDXD_DEV_CONF_READY;
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
return 0;
|
|
}
|
|
|
|
void idxd_device_reset(struct idxd_device *idxd)
|
|
{
|
|
unsigned long flags;
|
|
|
|
idxd_cmd_exec(idxd, IDXD_CMD_RESET_DEVICE, 0, NULL);
|
|
spin_lock_irqsave(&idxd->dev_lock, flags);
|
|
idxd_device_wqs_clear_state(idxd);
|
|
idxd->state = IDXD_DEV_CONF_READY;
|
|
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
|
}
|
|
|
|
void idxd_device_drain_pasid(struct idxd_device *idxd, int pasid)
|
|
{
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 operand;
|
|
|
|
operand = pasid;
|
|
dev_dbg(dev, "cmd: %u operand: %#x\n", IDXD_CMD_DRAIN_PASID, operand);
|
|
idxd_cmd_exec(idxd, IDXD_CMD_DRAIN_PASID, operand, NULL);
|
|
dev_dbg(dev, "pasid %d drained\n", pasid);
|
|
}
|
|
|
|
/* Device configuration bits */
|
|
static void idxd_group_config_write(struct idxd_group *group)
|
|
{
|
|
struct idxd_device *idxd = group->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
int i;
|
|
u32 grpcfg_offset;
|
|
|
|
dev_dbg(dev, "Writing group %d cfg registers\n", group->id);
|
|
|
|
/* setup GRPWQCFG */
|
|
for (i = 0; i < GRPWQCFG_STRIDES; i++) {
|
|
grpcfg_offset = GRPWQCFG_OFFSET(idxd, group->id, i);
|
|
iowrite64(group->grpcfg.wqs[i], idxd->reg_base + grpcfg_offset);
|
|
dev_dbg(dev, "GRPCFG wq[%d:%d: %#x]: %#llx\n",
|
|
group->id, i, grpcfg_offset,
|
|
ioread64(idxd->reg_base + grpcfg_offset));
|
|
}
|
|
|
|
/* setup GRPENGCFG */
|
|
grpcfg_offset = GRPENGCFG_OFFSET(idxd, group->id);
|
|
iowrite64(group->grpcfg.engines, idxd->reg_base + grpcfg_offset);
|
|
dev_dbg(dev, "GRPCFG engs[%d: %#x]: %#llx\n", group->id,
|
|
grpcfg_offset, ioread64(idxd->reg_base + grpcfg_offset));
|
|
|
|
/* setup GRPFLAGS */
|
|
grpcfg_offset = GRPFLGCFG_OFFSET(idxd, group->id);
|
|
iowrite32(group->grpcfg.flags.bits, idxd->reg_base + grpcfg_offset);
|
|
dev_dbg(dev, "GRPFLAGS flags[%d: %#x]: %#x\n",
|
|
group->id, grpcfg_offset,
|
|
ioread32(idxd->reg_base + grpcfg_offset));
|
|
}
|
|
|
|
static int idxd_groups_config_write(struct idxd_device *idxd)
|
|
|
|
{
|
|
union gencfg_reg reg;
|
|
int i;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
|
|
/* Setup bandwidth token limit */
|
|
if (idxd->token_limit) {
|
|
reg.bits = ioread32(idxd->reg_base + IDXD_GENCFG_OFFSET);
|
|
reg.token_limit = idxd->token_limit;
|
|
iowrite32(reg.bits, idxd->reg_base + IDXD_GENCFG_OFFSET);
|
|
}
|
|
|
|
dev_dbg(dev, "GENCFG(%#x): %#x\n", IDXD_GENCFG_OFFSET,
|
|
ioread32(idxd->reg_base + IDXD_GENCFG_OFFSET));
|
|
|
|
for (i = 0; i < idxd->max_groups; i++) {
|
|
struct idxd_group *group = &idxd->groups[i];
|
|
|
|
idxd_group_config_write(group);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int idxd_wq_config_write(struct idxd_wq *wq)
|
|
{
|
|
struct idxd_device *idxd = wq->idxd;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
u32 wq_offset;
|
|
int i;
|
|
|
|
if (!wq->group)
|
|
return 0;
|
|
|
|
memset(wq->wqcfg, 0, idxd->wqcfg_size);
|
|
|
|
/* byte 0-3 */
|
|
wq->wqcfg->wq_size = wq->size;
|
|
|
|
if (wq->size == 0) {
|
|
dev_warn(dev, "Incorrect work queue size: 0\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* bytes 4-7 */
|
|
wq->wqcfg->wq_thresh = wq->threshold;
|
|
|
|
/* byte 8-11 */
|
|
wq->wqcfg->priv = !!(wq->type == IDXD_WQT_KERNEL);
|
|
if (wq_dedicated(wq))
|
|
wq->wqcfg->mode = 1;
|
|
|
|
if (device_pasid_enabled(idxd)) {
|
|
wq->wqcfg->pasid_en = 1;
|
|
if (wq->type == IDXD_WQT_KERNEL && wq_dedicated(wq))
|
|
wq->wqcfg->pasid = idxd->pasid;
|
|
}
|
|
|
|
wq->wqcfg->priority = wq->priority;
|
|
|
|
if (idxd->hw.gen_cap.block_on_fault &&
|
|
test_bit(WQ_FLAG_BLOCK_ON_FAULT, &wq->flags))
|
|
wq->wqcfg->bof = 1;
|
|
|
|
if (idxd->hw.wq_cap.wq_ats_support)
|
|
wq->wqcfg->wq_ats_disable = wq->ats_dis;
|
|
|
|
/* bytes 12-15 */
|
|
wq->wqcfg->max_xfer_shift = ilog2(wq->max_xfer_bytes);
|
|
wq->wqcfg->max_batch_shift = ilog2(wq->max_batch_size);
|
|
|
|
dev_dbg(dev, "WQ %d CFGs\n", wq->id);
|
|
for (i = 0; i < WQCFG_STRIDES(idxd); i++) {
|
|
wq_offset = WQCFG_OFFSET(idxd, wq->id, i);
|
|
iowrite32(wq->wqcfg->bits[i], idxd->reg_base + wq_offset);
|
|
dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n",
|
|
wq->id, i, wq_offset,
|
|
ioread32(idxd->reg_base + wq_offset));
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int idxd_wqs_config_write(struct idxd_device *idxd)
|
|
{
|
|
int i, rc;
|
|
|
|
for (i = 0; i < idxd->max_wqs; i++) {
|
|
struct idxd_wq *wq = &idxd->wqs[i];
|
|
|
|
rc = idxd_wq_config_write(wq);
|
|
if (rc < 0)
|
|
return rc;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void idxd_group_flags_setup(struct idxd_device *idxd)
|
|
{
|
|
int i;
|
|
|
|
/* TC-A 0 and TC-B 1 should be defaults */
|
|
for (i = 0; i < idxd->max_groups; i++) {
|
|
struct idxd_group *group = &idxd->groups[i];
|
|
|
|
if (group->tc_a == -1)
|
|
group->tc_a = group->grpcfg.flags.tc_a = 0;
|
|
else
|
|
group->grpcfg.flags.tc_a = group->tc_a;
|
|
if (group->tc_b == -1)
|
|
group->tc_b = group->grpcfg.flags.tc_b = 1;
|
|
else
|
|
group->grpcfg.flags.tc_b = group->tc_b;
|
|
group->grpcfg.flags.use_token_limit = group->use_token_limit;
|
|
group->grpcfg.flags.tokens_reserved = group->tokens_reserved;
|
|
if (group->tokens_allowed)
|
|
group->grpcfg.flags.tokens_allowed =
|
|
group->tokens_allowed;
|
|
else
|
|
group->grpcfg.flags.tokens_allowed = idxd->max_tokens;
|
|
}
|
|
}
|
|
|
|
static int idxd_engines_setup(struct idxd_device *idxd)
|
|
{
|
|
int i, engines = 0;
|
|
struct idxd_engine *eng;
|
|
struct idxd_group *group;
|
|
|
|
for (i = 0; i < idxd->max_groups; i++) {
|
|
group = &idxd->groups[i];
|
|
group->grpcfg.engines = 0;
|
|
}
|
|
|
|
for (i = 0; i < idxd->max_engines; i++) {
|
|
eng = &idxd->engines[i];
|
|
group = eng->group;
|
|
|
|
if (!group)
|
|
continue;
|
|
|
|
group->grpcfg.engines |= BIT(eng->id);
|
|
engines++;
|
|
}
|
|
|
|
if (!engines)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int idxd_wqs_setup(struct idxd_device *idxd)
|
|
{
|
|
struct idxd_wq *wq;
|
|
struct idxd_group *group;
|
|
int i, j, configured = 0;
|
|
struct device *dev = &idxd->pdev->dev;
|
|
|
|
for (i = 0; i < idxd->max_groups; i++) {
|
|
group = &idxd->groups[i];
|
|
for (j = 0; j < 4; j++)
|
|
group->grpcfg.wqs[j] = 0;
|
|
}
|
|
|
|
for (i = 0; i < idxd->max_wqs; i++) {
|
|
wq = &idxd->wqs[i];
|
|
group = wq->group;
|
|
|
|
if (!wq->group)
|
|
continue;
|
|
if (!wq->size)
|
|
continue;
|
|
|
|
if (wq_shared(wq) && !device_swq_supported(idxd)) {
|
|
dev_warn(dev, "No shared wq support but configured.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
group->grpcfg.wqs[wq->id / 64] |= BIT(wq->id % 64);
|
|
configured++;
|
|
}
|
|
|
|
if (configured == 0)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int idxd_device_config(struct idxd_device *idxd)
|
|
{
|
|
int rc;
|
|
|
|
lockdep_assert_held(&idxd->dev_lock);
|
|
rc = idxd_wqs_setup(idxd);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
rc = idxd_engines_setup(idxd);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
idxd_group_flags_setup(idxd);
|
|
|
|
rc = idxd_wqs_config_write(idxd);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
rc = idxd_groups_config_write(idxd);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
return 0;
|
|
}
|