mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-18 09:44:18 +08:00
IB/hfi1: Search shared contexts on the opened device, not all devices
The search for available shared contexts walks each registered hfi1 device. This search is too broad because other devices may not be on the same fabric, and using its contexts could cause unexpected behavior. Removed walking the list of devices, limiting the search to the opened device. With the device walk removed, the hfi1_devdata (dd) is not available. Added it to the hfi1_filedata for reference. With this change, hfi1_count_units() was rendered obsolete and was removed. Reviewed-by: Mike Marciniszyn <mike.marciniszyn@intel.com> Signed-off-by: Michael J. Ruhl <michael.j.ruhl@intel.com> Signed-off-by: Dennis Dalessandro <dennis.dalessandro@intel.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
This commit is contained in:
parent
780a4c16aa
commit
5fbded483c
@ -210,42 +210,6 @@ int hfi1_count_active_units(void)
|
||||
return nunits_active;
|
||||
}
|
||||
|
||||
/*
|
||||
* Return count of all units, optionally return in arguments
|
||||
* the number of usable (present) units, and the number of
|
||||
* ports that are up.
|
||||
*/
|
||||
int hfi1_count_units(int *npresentp, int *nupp)
|
||||
{
|
||||
int nunits = 0, npresent = 0, nup = 0;
|
||||
struct hfi1_devdata *dd;
|
||||
unsigned long flags;
|
||||
int pidx;
|
||||
struct hfi1_pportdata *ppd;
|
||||
|
||||
spin_lock_irqsave(&hfi1_devs_lock, flags);
|
||||
|
||||
list_for_each_entry(dd, &hfi1_dev_list, list) {
|
||||
nunits++;
|
||||
if ((dd->flags & HFI1_PRESENT) && dd->kregbase)
|
||||
npresent++;
|
||||
for (pidx = 0; pidx < dd->num_pports; ++pidx) {
|
||||
ppd = dd->pport + pidx;
|
||||
if (ppd->lid && ppd->linkup)
|
||||
nup++;
|
||||
}
|
||||
}
|
||||
|
||||
spin_unlock_irqrestore(&hfi1_devs_lock, flags);
|
||||
|
||||
if (npresentp)
|
||||
*npresentp = npresent;
|
||||
if (nupp)
|
||||
*nupp = nup;
|
||||
|
||||
return nunits;
|
||||
}
|
||||
|
||||
/*
|
||||
* Get address of eager buffer from it's index (allocated in chunks, not
|
||||
* contiguous).
|
||||
|
@ -77,7 +77,7 @@ static unsigned int hfi1_poll(struct file *fp, struct poll_table_struct *pt);
|
||||
static int hfi1_file_mmap(struct file *fp, struct vm_area_struct *vma);
|
||||
|
||||
static u64 kvirt_to_phys(void *addr);
|
||||
static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo);
|
||||
static int assign_ctxt(struct hfi1_filedata *fd, struct hfi1_user_info *uinfo);
|
||||
static int init_subctxts(struct hfi1_ctxtdata *uctxt,
|
||||
const struct hfi1_user_info *uinfo);
|
||||
static int user_init(struct hfi1_filedata *fd);
|
||||
@ -87,8 +87,7 @@ static int get_base_info(struct hfi1_filedata *fd, void __user *ubase,
|
||||
__u32 len);
|
||||
static int setup_ctxt(struct hfi1_filedata *fd);
|
||||
static int setup_subctxt(struct hfi1_ctxtdata *uctxt);
|
||||
static int get_user_context(struct hfi1_filedata *fd,
|
||||
struct hfi1_user_info *uinfo, int devno);
|
||||
|
||||
static int find_shared_ctxt(struct hfi1_filedata *fd,
|
||||
const struct hfi1_user_info *uinfo);
|
||||
static int allocate_ctxt(struct hfi1_filedata *fd, struct hfi1_devdata *dd,
|
||||
@ -181,6 +180,9 @@ static int hfi1_file_open(struct inode *inode, struct file *fp)
|
||||
struct hfi1_devdata,
|
||||
user_cdev);
|
||||
|
||||
if (!((dd->flags & HFI1_PRESENT) && dd->kregbase))
|
||||
return -EINVAL;
|
||||
|
||||
if (!atomic_inc_not_zero(&dd->user_refcount))
|
||||
return -ENXIO;
|
||||
|
||||
@ -195,6 +197,7 @@ static int hfi1_file_open(struct inode *inode, struct file *fp)
|
||||
fd->rec_cpu_num = -1; /* no cpu affinity by default */
|
||||
fd->mm = current->mm;
|
||||
mmgrab(fd->mm);
|
||||
fd->dd = dd;
|
||||
fp->private_data = fd;
|
||||
} else {
|
||||
fp->private_data = NULL;
|
||||
@ -237,7 +240,7 @@ static long hfi1_file_ioctl(struct file *fp, unsigned int cmd,
|
||||
sizeof(uinfo)))
|
||||
return -EFAULT;
|
||||
|
||||
ret = assign_ctxt(fp, &uinfo);
|
||||
ret = assign_ctxt(fd, &uinfo);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
ret = setup_ctxt(fd);
|
||||
@ -847,9 +850,9 @@ static u64 kvirt_to_phys(void *addr)
|
||||
return paddr;
|
||||
}
|
||||
|
||||
static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo)
|
||||
static int assign_ctxt(struct hfi1_filedata *fd, struct hfi1_user_info *uinfo)
|
||||
{
|
||||
int i_minor, ret = 0;
|
||||
int ret = 0;
|
||||
unsigned int swmajor, swminor;
|
||||
|
||||
swmajor = uinfo->userversion >> 16;
|
||||
@ -863,8 +866,6 @@ static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo)
|
||||
mutex_lock(&hfi1_mutex);
|
||||
/* First, lets check if we need to setup a shared context? */
|
||||
if (uinfo->subctxt_cnt) {
|
||||
struct hfi1_filedata *fd = fp->private_data;
|
||||
|
||||
ret = find_shared_ctxt(fd, uinfo);
|
||||
if (ret < 0)
|
||||
goto done_unlock;
|
||||
@ -878,94 +879,59 @@ static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo)
|
||||
* We execute the following block if we couldn't find a
|
||||
* shared context or if context sharing is not required.
|
||||
*/
|
||||
if (!ret) {
|
||||
i_minor = iminor(file_inode(fp)) - HFI1_USER_MINOR_BASE;
|
||||
ret = get_user_context(fp->private_data, uinfo, i_minor);
|
||||
}
|
||||
if (!ret)
|
||||
ret = allocate_ctxt(fd, fd->dd, uinfo);
|
||||
|
||||
done_unlock:
|
||||
mutex_unlock(&hfi1_mutex);
|
||||
done:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int get_user_context(struct hfi1_filedata *fd,
|
||||
struct hfi1_user_info *uinfo, int devno)
|
||||
{
|
||||
struct hfi1_devdata *dd = NULL;
|
||||
int devmax, npresent, nup;
|
||||
|
||||
devmax = hfi1_count_units(&npresent, &nup);
|
||||
if (!npresent)
|
||||
return -ENXIO;
|
||||
|
||||
if (!nup)
|
||||
return -ENETDOWN;
|
||||
|
||||
dd = hfi1_lookup(devno);
|
||||
if (!dd)
|
||||
return -ENODEV;
|
||||
else if (!dd->freectxts)
|
||||
return -EBUSY;
|
||||
|
||||
return allocate_ctxt(fd, dd, uinfo);
|
||||
}
|
||||
|
||||
static int find_shared_ctxt(struct hfi1_filedata *fd,
|
||||
const struct hfi1_user_info *uinfo)
|
||||
{
|
||||
int devmax, ndev, i;
|
||||
int ret = 0;
|
||||
int i;
|
||||
struct hfi1_devdata *dd = fd->dd;
|
||||
|
||||
devmax = hfi1_count_units(NULL, NULL);
|
||||
for (i = dd->first_dyn_alloc_ctxt; i < dd->num_rcv_contexts; i++) {
|
||||
struct hfi1_ctxtdata *uctxt = dd->rcd[i];
|
||||
|
||||
for (ndev = 0; ndev < devmax; ndev++) {
|
||||
struct hfi1_devdata *dd = hfi1_lookup(ndev);
|
||||
|
||||
if (!(dd && (dd->flags & HFI1_PRESENT) && dd->kregbase))
|
||||
/* Skip ctxts which are not yet open */
|
||||
if (!uctxt || !uctxt->cnt)
|
||||
continue;
|
||||
for (i = dd->first_dyn_alloc_ctxt;
|
||||
i < dd->num_rcv_contexts; i++) {
|
||||
struct hfi1_ctxtdata *uctxt = dd->rcd[i];
|
||||
|
||||
/* Skip ctxts which are not yet open */
|
||||
if (!uctxt || !uctxt->cnt)
|
||||
continue;
|
||||
/* Skip dynamically allocted kernel contexts */
|
||||
if (uctxt->sc && (uctxt->sc->type == SC_KERNEL))
|
||||
continue;
|
||||
|
||||
/* Skip dynamically allocted kernel contexts */
|
||||
if (uctxt->sc && (uctxt->sc->type == SC_KERNEL))
|
||||
continue;
|
||||
/* Skip ctxt if it doesn't match the requested one */
|
||||
if (memcmp(uctxt->uuid, uinfo->uuid,
|
||||
sizeof(uctxt->uuid)) ||
|
||||
uctxt->jkey != generate_jkey(current_uid()) ||
|
||||
uctxt->subctxt_id != uinfo->subctxt_id ||
|
||||
uctxt->subctxt_cnt != uinfo->subctxt_cnt)
|
||||
continue;
|
||||
|
||||
/* Skip ctxt if it doesn't match the requested one */
|
||||
if (memcmp(uctxt->uuid, uinfo->uuid,
|
||||
sizeof(uctxt->uuid)) ||
|
||||
uctxt->jkey != generate_jkey(current_uid()) ||
|
||||
uctxt->subctxt_id != uinfo->subctxt_id ||
|
||||
uctxt->subctxt_cnt != uinfo->subctxt_cnt)
|
||||
continue;
|
||||
|
||||
/* Verify the sharing process matches the master */
|
||||
if (uctxt->userversion != uinfo->userversion ||
|
||||
uctxt->cnt >= uctxt->subctxt_cnt) {
|
||||
ret = -EINVAL;
|
||||
goto done;
|
||||
}
|
||||
fd->uctxt = uctxt;
|
||||
fd->subctxt = uctxt->cnt++;
|
||||
uctxt->active_slaves |= 1 << fd->subctxt;
|
||||
ret = 1;
|
||||
goto done;
|
||||
/* Verify the sharing process matches the master */
|
||||
if (uctxt->userversion != uinfo->userversion ||
|
||||
uctxt->cnt >= uctxt->subctxt_cnt) {
|
||||
return -EINVAL;
|
||||
}
|
||||
fd->uctxt = uctxt;
|
||||
fd->subctxt = uctxt->cnt++;
|
||||
uctxt->active_slaves |= 1 << fd->subctxt;
|
||||
return 1;
|
||||
}
|
||||
|
||||
done:
|
||||
return ret;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int allocate_ctxt(struct hfi1_filedata *fd, struct hfi1_devdata *dd,
|
||||
struct hfi1_user_info *uinfo)
|
||||
{
|
||||
struct hfi1_ctxtdata *uctxt;
|
||||
unsigned ctxt;
|
||||
unsigned int ctxt;
|
||||
int ret, numa;
|
||||
|
||||
if (dd->flags & HFI1_FROZEN) {
|
||||
@ -979,6 +945,14 @@ static int allocate_ctxt(struct hfi1_filedata *fd, struct hfi1_devdata *dd,
|
||||
return -EIO;
|
||||
}
|
||||
|
||||
/*
|
||||
* This check is sort of redundant to the next EBUSY error. It would
|
||||
* also indicate an inconsistancy in the driver if this value was
|
||||
* zero, but there were still contexts available.
|
||||
*/
|
||||
if (!dd->freectxts)
|
||||
return -EBUSY;
|
||||
|
||||
for (ctxt = dd->first_dyn_alloc_ctxt;
|
||||
ctxt < dd->num_rcv_contexts; ctxt++)
|
||||
if (!dd->rcd[ctxt])
|
||||
|
@ -1238,6 +1238,7 @@ struct mmu_rb_handler;
|
||||
|
||||
/* Private data for file operations */
|
||||
struct hfi1_filedata {
|
||||
struct hfi1_devdata *dd;
|
||||
struct hfi1_ctxtdata *uctxt;
|
||||
unsigned subctxt;
|
||||
struct hfi1_user_sdma_comp_q *cq;
|
||||
@ -1264,7 +1265,6 @@ extern u32 hfi1_cpulist_count;
|
||||
extern unsigned long *hfi1_cpulist;
|
||||
|
||||
int hfi1_init(struct hfi1_devdata *dd, int reinit);
|
||||
int hfi1_count_units(int *npresentp, int *nupp);
|
||||
int hfi1_count_active_units(void);
|
||||
|
||||
int hfi1_diag_add(struct hfi1_devdata *dd);
|
||||
|
Loading…
Reference in New Issue
Block a user