[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PATCH v2 03/10] intel-iommu: add iommu lock
From: |
Peter Xu |
Subject: |
Re: [Qemu-devel] [PATCH v2 03/10] intel-iommu: add iommu lock |
Date: |
Fri, 18 May 2018 13:32:18 +0800 |
User-agent: |
Mutt/1.9.5 (2018-04-13) |
On Thu, May 17, 2018 at 04:32:52PM +0200, Auger Eric wrote:
> Hi Peter,
>
> On 05/04/2018 05:08 AM, Peter Xu wrote:
> > Add a per-iommu big lock to protect IOMMU status. Currently the only
> > thing to be protected is the IOTLB/context cache, since that can be
> > accessed even without BQL, e.g., in IO dataplane.
>
> As discussed together, Peter challenged per device mutex in
> "Re: [PATCH v11 15/17] hw/arm/smmuv3: Cache/invalidate config data"
> https://lists.gnu.org/archive/html/qemu-devel/2018-04/msg02403.html
> but at that time I fail to justify the use :-(
>
> >
> > Note that we don't need to protect device page tables since that's fully
> > controlled by the guest kernel. However there is still possibility that
> > malicious drivers will program the device to not obey the rule. In that
> > case QEMU can't really do anything useful, instead the guest itself will
> > be responsible for all uncertainties.
> >
> > Reported-by: Fam Zheng <address@hidden>
> > Signed-off-by: Peter Xu <address@hidden>
> > ---
> > include/hw/i386/intel_iommu.h | 6 +++++
> > hw/i386/intel_iommu.c | 43 +++++++++++++++++++++++++++++++----
> > 2 files changed, 44 insertions(+), 5 deletions(-)
> >
> > diff --git a/include/hw/i386/intel_iommu.h b/include/hw/i386/intel_iommu.h
> > index 220697253f..ee517704e7 100644
> > --- a/include/hw/i386/intel_iommu.h
> > +++ b/include/hw/i386/intel_iommu.h
> > @@ -300,6 +300,12 @@ struct IntelIOMMUState {
> > OnOffAuto intr_eim; /* Toggle for EIM cabability */
> > bool buggy_eim; /* Force buggy EIM unless eim=off */
> > uint8_t aw_bits; /* Host/IOVA address width (in bits) */
> > +
> > + /*
> > + * Protects IOMMU states in general. Currently it protects the
> > + * per-IOMMU IOTLB cache, and context entry cache in VTDAddressSpace.
> > + */
> > + QemuMutex iommu_lock;
> > };
> >
> > /* Find the VTD Address space associated with the given bus pointer,
> > diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> > index 5987b48d43..112971638d 100644
> > --- a/hw/i386/intel_iommu.c
> > +++ b/hw/i386/intel_iommu.c
> > @@ -128,6 +128,16 @@ static uint64_t
> > vtd_set_clear_mask_quad(IntelIOMMUState *s, hwaddr addr,
> > return new_val;
> > }
> >
> > +static inline void vtd_iommu_lock(IntelIOMMUState *s)
> > +{
> > + qemu_mutex_lock(&s->iommu_lock);
> > +}
> > +
> > +static inline void vtd_iommu_unlock(IntelIOMMUState *s)
> > +{
> > + qemu_mutex_unlock(&s->iommu_lock);
> > +}
> > +
> > /* GHashTable functions */
> > static gboolean vtd_uint64_equal(gconstpointer v1, gconstpointer v2)
> > {
> > @@ -172,7 +182,7 @@ static gboolean vtd_hash_remove_by_page(gpointer key,
> > gpointer value,
> > }
> >
> > /* Reset all the gen of VTDAddressSpace to zero and set the gen of
> > - * IntelIOMMUState to 1.
> > + * IntelIOMMUState to 1. Must be with IOMMU lock held.
> s/Must be/ Must be called ?
Ok.
> not done in vtd_init()
IMHO we can omit that since it's only used during either realization
or system reset. But sure I can add it too.
> > */
> > static void vtd_reset_context_cache(IntelIOMMUState *s)
> > {
> > @@ -197,12 +207,19 @@ static void vtd_reset_context_cache(IntelIOMMUState
> > *s)
> > s->context_cache_gen = 1;
> > }
> >
> > -static void vtd_reset_iotlb(IntelIOMMUState *s)
> > +static void vtd_reset_iotlb_locked(IntelIOMMUState *s)
> add the above comment and keep the original name?
I can add the comment; the original name is defined as another
function below.
> > {
> > assert(s->iotlb);
> > g_hash_table_remove_all(s->iotlb);
> > }
> >
> > +static void vtd_reset_iotlb(IntelIOMMUState *s)
[1]
> > +{
> > + vtd_iommu_lock(s);
> > + vtd_reset_iotlb_locked(s);
> > + vtd_iommu_unlock(s);
> > +}
> > +
> > static uint64_t vtd_get_iotlb_key(uint64_t gfn, uint16_t source_id,
> > uint32_t level)
> > {
> > @@ -215,6 +232,7 @@ static uint64_t vtd_get_iotlb_gfn(hwaddr addr, uint32_t
> > level)
> > return (addr & vtd_slpt_level_page_mask(level)) >> VTD_PAGE_SHIFT_4K;
> > }
> >
> > +/* Must be with IOMMU lock held */
> > static VTDIOTLBEntry *vtd_lookup_iotlb(IntelIOMMUState *s, uint16_t
> > source_id,
> > hwaddr addr)
> > {
> > @@ -235,6 +253,7 @@ out:
> > return entry;
> > }
> >
> > +/* Must be with IOMMU lock held */
> > static void vtd_update_iotlb(IntelIOMMUState *s, uint16_t source_id,
> > uint16_t domain_id, hwaddr addr, uint64_t
> > slpte,
> > uint8_t access_flags, uint32_t level)
> > @@ -246,7 +265,7 @@ static void vtd_update_iotlb(IntelIOMMUState *s,
> > uint16_t source_id,
> > trace_vtd_iotlb_page_update(source_id, addr, slpte, domain_id);
> > if (g_hash_table_size(s->iotlb) >= VTD_IOTLB_MAX_SIZE) {
> > trace_vtd_iotlb_reset("iotlb exceeds size limit");
> > - vtd_reset_iotlb(s);
> > + vtd_reset_iotlb_locked(s);
> > }
> >
> > entry->gfn = gfn;
> > @@ -1106,7 +1125,7 @@ static bool vtd_do_iommu_translate(VTDAddressSpace
> > *vtd_as, PCIBus *bus,
> > IntelIOMMUState *s = vtd_as->iommu_state;
> > VTDContextEntry ce;
> > uint8_t bus_num = pci_bus_num(bus);
> > - VTDContextCacheEntry *cc_entry = &vtd_as->context_cache_entry;
> > + VTDContextCacheEntry *cc_entry;
> > uint64_t slpte, page_mask;
> > uint32_t level;
> > uint16_t source_id = vtd_make_source_id(bus_num, devfn);
> > @@ -1123,6 +1142,10 @@ static bool vtd_do_iommu_translate(VTDAddressSpace
> > *vtd_as, PCIBus *bus,
> > */
> > assert(!vtd_is_interrupt_addr(addr));
> >
> > + vtd_iommu_lock(s);
> > +
> > + cc_entry = &vtd_as->context_cache_entry;
> > +
> > /* Try to fetch slpte form IOTLB */
> > iotlb_entry = vtd_lookup_iotlb(s, source_id, addr);
> > if (iotlb_entry) {
> > @@ -1182,7 +1205,7 @@ static bool vtd_do_iommu_translate(VTDAddressSpace
> > *vtd_as, PCIBus *bus,
> > * IOMMU region can be swapped back.
> > */
> > vtd_pt_enable_fast_path(s, source_id);
> > -
> > + vtd_iommu_unlock(s);
> > return true;
> > }
> >
> > @@ -1203,6 +1226,7 @@ static bool vtd_do_iommu_translate(VTDAddressSpace
> > *vtd_as, PCIBus *bus,
> > vtd_update_iotlb(s, source_id, VTD_CONTEXT_ENTRY_DID(ce.hi), addr,
> > slpte,
> > access_flags, level);
> > out:
> > + vtd_iommu_unlock(s);
> > entry->iova = addr & page_mask;
> > entry->translated_addr = vtd_get_slpte_addr(slpte, s->aw_bits) &
> > page_mask;
> > entry->addr_mask = ~page_mask;
> > @@ -1210,6 +1234,7 @@ out:
> > return true;
> >
> > error:
> > + vtd_iommu_unlock(s);
> > entry->iova = 0;
> > entry->translated_addr = 0;
> > entry->addr_mask = 0;
> > @@ -1258,10 +1283,13 @@ static void vtd_iommu_replay_all(IntelIOMMUState *s)
> > static void vtd_context_global_invalidate(IntelIOMMUState *s)
> > {
> > trace_vtd_inv_desc_cc_global();
> > + /* Protects context cache */
> > + vtd_iommu_lock(s);
> > s->context_cache_gen++;
> > if (s->context_cache_gen == VTD_CONTEXT_CACHE_GEN_MAX) {
> > vtd_reset_context_cache(s);
> > }
> > + vtd_iommu_unlock(s);
> > vtd_switch_address_space_all(s);
> > /*
> > * From VT-d spec 6.5.2.1, a global context entry invalidation
> > @@ -1377,8 +1405,10 @@ static void
> > vtd_iotlb_domain_invalidate(IntelIOMMUState *s, uint16_t domain_id)
> >
> > trace_vtd_inv_desc_iotlb_domain(domain_id);
> >
> > + vtd_iommu_lock(s);
> > g_hash_table_foreach_remove(s->iotlb, vtd_hash_remove_by_domain,
> > &domain_id);
> > + vtd_iommu_unlock(s);
> >
> > QLIST_FOREACH(vtd_as, &s->notifiers_list, next) {
> > if (!vtd_dev_to_context_entry(s, pci_bus_num(vtd_as->bus),
> > @@ -1426,7 +1456,9 @@ static void vtd_iotlb_page_invalidate(IntelIOMMUState
> > *s, uint16_t domain_id,
> > info.domain_id = domain_id;
> > info.addr = addr;
> > info.mask = ~((1 << am) - 1);
> > + vtd_iommu_lock(s);
> > g_hash_table_foreach_remove(s->iotlb, vtd_hash_remove_by_page, &info);
> > + vtd_iommu_unlock(s);
> > vtd_iotlb_page_invalidate_notify(s, domain_id, addr, am);
> > }
> >
> > @@ -3072,6 +3104,7 @@ static void vtd_realize(DeviceState *dev, Error
> > **errp)
> > }
> >
> > QLIST_INIT(&s->notifiers_list);
> > + qemu_mutex_init(&s->iommu_lock);
> > memset(s->vtd_as_by_bus_num, 0, sizeof(s->vtd_as_by_bus_num));
> > memory_region_init_io(&s->csrmem, OBJECT(s), &vtd_mem_ops, s,
> > "intel_iommu", DMAR_REG_SIZE);
> >
> Don't you need to take the lock in vtd_context_device_invalidate() which
> manipulates the vtd_as->context_cache_entry.context_cache_gen?
Yes I think we need that. Thanks,
--
Peter Xu
- [Qemu-devel] [PATCH v2 02/10] intel-iommu: remove IntelIOMMUNotifierNode, (continued)
[Qemu-devel] [PATCH v2 04/10] intel-iommu: only do page walk for MAP notifiers, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 03/10] intel-iommu: add iommu lock, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 05/10] intel-iommu: introduce vtd_page_walk_info, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 06/10] intel-iommu: pass in address space when page walk, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 07/10] util: implement simple interval tree logic, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 08/10] intel-iommu: maintain per-device iova ranges, Peter Xu, 2018/05/03
[Qemu-devel] [PATCH v2 09/10] intel-iommu: don't unmap all for shadow page table, Peter Xu, 2018/05/03