* XEN_DOMCTL_getvcpucontext
* XEN_DOMCTL_max_vcpus
* XEN_DOMCTL_scheduler_op
- * XEN_DOMCTL_irq_permission
* XEN_DOMCTL_iomem_permission
- * XEN_DOMCTL_ioport_permission
* XEN_DOMCTL_gethvmcontext
* XEN_DOMCTL_sethvmcontext
* XEN_DOMCTL_set_address_size
return (iop->remain ? -EFAULT : 0);
}
+#define MAX_IOPORTS 0x10000
+
long arch_do_domctl(
struct xen_domctl *domctl, struct domain *d,
XEN_GUEST_HANDLE_PARAM(xen_domctl_t) u_domctl)
unsigned int np = domctl->u.ioport_permission.nr_ports;
int allow = domctl->u.ioport_permission.allow_access;
- ret = -EINVAL;
- if ( (fp + np) > 65536 )
- break;
-
- if ( np == 0 )
- ret = 0;
- else if ( xsm_ioport_permission(XSM_HOOK, d, fp, fp + np - 1, allow) )
+ if ( (fp + np) <= fp || (fp + np) > MAX_IOPORTS )
+ ret = -EINVAL;
+ else if ( !ioports_access_permitted(current->domain,
+ fp, fp + np - 1) ||
+ xsm_ioport_permission(XSM_HOOK, d, fp, fp + np - 1, allow) )
ret = -EPERM;
else if ( allow )
ret = ioports_permit_access(d, fp, fp + np - 1);
case XEN_DOMCTL_ioport_mapping:
{
-#define MAX_IOPORTS 0x10000
struct hvm_iommu *hd;
unsigned int fgp = domctl->u.ioport_mapping.first_gport;
unsigned int fmp = domctl->u.ioport_mapping.first_mport;
if ( pirq >= d->nr_pirqs )
ret = -EINVAL;
- else if ( xsm_irq_permission(XSM_HOOK, d, pirq, allow) )
+ else if ( !pirq_access_permitted(current->domain, pirq) ||
+ xsm_irq_permission(XSM_HOOK, d, pirq, allow) )
ret = -EPERM;
else if ( allow )
ret = pirq_permit_access(d, pirq);
if ( (mfn + nr_mfns - 1) < mfn ) /* wrap? */
break;
- if ( xsm_iomem_permission(XSM_HOOK, d, mfn, mfn + nr_mfns - 1, allow) )
+ if ( !iomem_access_permitted(current->domain,
+ mfn, mfn + nr_mfns - 1) ||
+ xsm_iomem_permission(XSM_HOOK, d, mfn, mfn + nr_mfns - 1, allow) )
ret = -EPERM;
else if ( allow )
ret = iomem_permit_access(d, mfn, mfn + nr_mfns - 1);