qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH 3/6] hw/acpi: Generic Port Affinity Structure support


From: Markus Armbruster
Subject: Re: [PATCH 3/6] hw/acpi: Generic Port Affinity Structure support
Date: Tue, 04 Jun 2024 14:06:53 +0200
User-agent: Gnus/5.13 (Gnus v5.13)

Sorry for the late reply.  I had to put this aside, and I'm coming back
to it only now.

Jonathan Cameron <Jonathan.Cameron@huawei.com> writes:

> On Tue, 30 Apr 2024 08:55:12 +0200
> Markus Armbruster <armbru@redhat.com> wrote:
>
>> Jonathan Cameron <Jonathan.Cameron@Huawei.com> writes:
>> 
>> > On Tue, 23 Apr 2024 12:56:21 +0200
>> > Markus Armbruster <armbru@redhat.com> wrote:
>> >  
>> >> Jonathan Cameron <Jonathan.Cameron@huawei.com> writes:
>> >>   
>> >> > These are very similar to the recently added Generic Initiators
>> >> > but instead of representing an initiator of memory traffic they
>> >> > represent an edge point beyond which may lie either targets or
>> >> > initiators.  Here we add these ports such that they may
>> >> > be targets of hmat_lb records to describe the latency and
>> >> > bandwidth from host side initiators to the port.  A descoverable
>> >> > mechanism such as UEFI CDAT read from CXL devices and switches
>> >> > is used to discover the remainder fo the path and the OS can build
>> >> > up full latency and bandwidth numbers as need for work and data
>> >> > placement decisions.
>> >> >
>> >> > Signed-off-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>  
>> >
>> > Hi Markus,
>> >
>> > I've again managed a bad job of defining an interface - thanks for
>> > your help!  
>> 
>> Good interfaces are hard!
>> 
>> >> > ---
>> >> >  qapi/qom.json                            |  18 +++
>> >> >  include/hw/acpi/acpi_generic_initiator.h |  18 ++-
>> >> >  include/hw/pci/pci_bridge.h              |   1 +
>> >> >  hw/acpi/acpi_generic_initiator.c         | 141 +++++++++++++++++------
>> >> >  hw/pci-bridge/pci_expander_bridge.c      |   1 -
>> >> >  5 files changed, 141 insertions(+), 38 deletions(-)
>> >> >
>> >> > diff --git a/qapi/qom.json b/qapi/qom.json
>> >> > index 85e6b4f84a..5480d9ca24 100644
>> >> > --- a/qapi/qom.json
>> >> > +++ b/qapi/qom.json
>> >> > @@ -826,6 +826,22 @@
>> >> >    'data': { 'pci-dev': 'str',
>> >> >              'node': 'uint32' } }
>> >> >  
>> >> > +
>> >> > +##
>> >> > +# @AcpiGenericPortProperties:
>> >> > +#
>> >> > +# Properties for acpi-generic-port objects.
>> >> > +#
>> >> > +# @pci-bus: PCI bus of the hostbridge associated with this SRAT entry  
>> >> >   
>> >> 
>> >> What's this exactly?  A QOM path?  A qdev ID?  Something else?  
>> >
>> > QOM path I believe as going to call object_resolve_path_type() on it.  
>> 
>> QOM path then.
>> 
>> > Oddity is it's defined for the bus, not the host bridge that
>> > we care about as the host bridge doesn't have a convenient id to let
>> > us identify it.

Copied from the example below:

    -object acpi-generic-port,id=bob11,pci-bus=cxl.1,node=2

where pci-bus=cxl.1 refers to

    -device pxb-cxl,bus_nr=12,bus=pcie.0,id=cxl.1,hdm_for_passthrough=true

What is "the host bridge that we care about" in that example?

Is pci-bus=cxl.1 a surrogate for a reference to "the host bridge that we
care about"?  I.e. something that enables the C code to find the host
bridge?

>> > e.g. It is specified via --device pxb-cxl,id=XXXX
>> > of TYPE_PXB_CXL_HOST in the command line but ends up on the
>> > TYPE_PCI_BUS with parent set to the PXB_CXL_HOST.
>> > Normally we just want this bus for hanging root ports of it.
>> >
>> > I can clarify it's the QOM path but I'm struggling a bit to explain
>> > the relationship without resorting to an example.
>> > This should also not mention SRAT as at some stage I'd expect DT
>> > bindings to provide similar functionality.  
>> 
>> Let's start with an example.  Not to put it into the doc comment, only
>> to help me understand what you need.  Hopefully I can then assist with
>> improving the interface and/or its documentation.
>
> Stripping out some relevant bits from a test setup and editing it down
> - most of this is about creating relevant SLIT/HMAT tables.
>
> # First a CXL root bridge, root port and direct attached device plus fixed
> # memory window.  Linux currently builds a NUMA node per fixed memory window
> # but that's a simplification that may change over time.
>  -device pxb-cxl,bus_nr=12,bus=pcie.0,id=cxl.1,hdm_for_passthrough=true \
>  -device cxl-rp,port=0,bus=cxl.1,id=cxl_rp_port0,chassis=0,slot=2 \
>  -device 
> cxl-type3,bus=cxl_rp_port0,volatile-memdev=cxl-mem1,persistent-memdev=cxl-mem2,id=cxl-pmem1,lsa=cxl-lsa1,sn=3
>  \
>  -machine 
> cxl-fmw.0.targets.0=cxl.1,cxl-fmw.0.size=4G,cxl-fmw.0.interleave-granularity=1k
>  \
>
> # Next line is the port defintion - see that the pci-bus refers to the one in 
> the id parameter for
> # the PXB CXL, but the ACPI table that is generated refers to the DSDT entry 
> via a ACPI0016 entry.
> # So to get to that we use the PCI bus ID of the root bus that forms part of 
> the root bridge (but
> # is a child object in qemu.
>  -numa node,nodeid=2 \
>  -object acpi-generic-port,id=bob11,pci-bus=cxl.1,node=2 \
>
> # The rest is a the setup for the hmat and slit tables.  I hid most of the 
> config, but
> # left this here as key is that we specify values to and from the generic 
> port 'node'
> # but it's not really a node as such, but just a point along the path to one.
>
>  -numa dist,src=0,dst=0,val=10 -numa dist,src=0,dst=1,val=21 -numa 
> dist,src=0,dst=2,val=21 -numa dist,src=0,dst=3,val=21 -numa 
> dist,src=0,dst=4,val=21 -numa dist,src=0,dst=5,val=21 \
>  -numa dist,src=1,dst=0,val=21 -numa dist,src=1,dst=1,val=10 -numa 
> dist,src=1,dst=2,val=21 -numa dist,src=1,dst=3,val=21 -numa 
> dist,src=1,dst=4,val=21 -numa dist,src=1,dst=5,val=21 \
>  -numa dist,src=2,dst=0,val=21 -numa dist,src=2,dst=1,val=21 -numa 
> dist,src=2,dst=2,val=10 -numa dist,src=2,dst=3,val=21 -numa 
> dist,src=2,dst=4,val=21 -numa dist,src=2,dst=5,val=21 \
>  -numa dist,src=3,dst=0,val=21 -numa dist,src=3,dst=1,val=21 -numa 
> dist,src=3,dst=2,val=21 -numa dist,src=3,dst=3,val=10 -numa 
> dist,src=3,dst=4,val=21 -numa dist,src=3,dst=5,val=21 \
>  -numa dist,src=4,dst=0,val=21 -numa dist,src=4,dst=1,val=21 -numa 
> dist,src=4,dst=2,val=21 -numa dist,src=4,dst=3,val=21 -numa 
> dist,src=4,dst=4,val=10 -numa dist,src=4,dst=5,val=21 \
>  -numa dist,src=5,dst=0,val=21 -numa dist,src=5,dst=1,val=21 -numa 
> dist,src=5,dst=2,val=21 -numa dist,src=5,dst=3,val=21 -numa 
> dist,src=5,dst=4,val=21 -numa dist,src=5,dst=5,val=10 \
>  -numa 
> hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-latency,latency=10
>  \
>  -numa 
> hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=800M
>  \
>  -numa 
> hmat-lb,initiator=0,target=2,hierarchy=memory,data-type=access-latency,latency=100
>  \
>  -numa 
> hmat-lb,initiator=0,target=2,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=0,target=4,hierarchy=memory,data-type=access-latency,latency=100
>  \
>  -numa 
> hmat-lb,initiator=0,target=4,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=0,target=5,hierarchy=memory,data-type=access-latency,latency=200
>  \
>  -numa 
> hmat-lb,initiator=0,target=5,hierarchy=memory,data-type=access-bandwidth,bandwidth=400M
>  \
>  -numa 
> hmat-lb,initiator=1,target=0,hierarchy=memory,data-type=access-latency,latency=500
>  \
>  -numa 
> hmat-lb,initiator=1,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=100M
>  \
>  -numa 
> hmat-lb,initiator=1,target=2,hierarchy=memory,data-type=access-latency,latency=50
>  \
>  -numa 
> hmat-lb,initiator=1,target=2,hierarchy=memory,data-type=access-bandwidth,bandwidth=400M
>  \
>  -numa 
> hmat-lb,initiator=1,target=4,hierarchy=memory,data-type=access-latency,latency=50
>  \
>  -numa 
> hmat-lb,initiator=1,target=4,hierarchy=memory,data-type=access-bandwidth,bandwidth=800M
>  \
>  -numa 
> hmat-lb,initiator=1,target=5,hierarchy=memory,data-type=access-latency,latency=500
>  \
>  -numa 
> hmat-lb,initiator=1,target=5,hierarchy=memory,data-type=access-bandwidth,bandwidth=100M
>  \
>   -numa 
> hmat-lb,initiator=3,target=0,hierarchy=memory,data-type=access-latency,latency=20
>  \
>  -numa 
> hmat-lb,initiator=3,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=400M
>  \
>  -numa 
> hmat-lb,initiator=3,target=2,hierarchy=memory,data-type=access-latency,latency=80
>  \
>  -numa 
> hmat-lb,initiator=3,target=2,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=3,target=4,hierarchy=memory,data-type=access-latency,latency=80
>  \
>  -numa 
> hmat-lb,initiator=3,target=4,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=3,target=5,hierarchy=memory,data-type=access-latency,latency=20
>  \
>  -numa 
> hmat-lb,initiator=3,target=5,hierarchy=memory,data-type=access-bandwidth,bandwidth=400M
>  \
>  -numa 
> hmat-lb,initiator=5,target=0,hierarchy=memory,data-type=access-latency,latency=20
>  \
>  -numa 
> hmat-lb,initiator=5,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=400M
>  \
>  -numa 
> hmat-lb,initiator=5,target=2,hierarchy=memory,data-type=access-latency,latency=80
>  \
>  -numa 
> hmat-lb,initiator=5,target=4,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=5,target=4,hierarchy=memory,data-type=access-latency,latency=80
>  \
>  -numa 
> hmat-lb,initiator=5,target=2,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M
>  \
>  -numa 
> hmat-lb,initiator=5,target=5,hierarchy=memory,data-type=access-latency,latency=10
>  \
>  -numa 
> hmat-lb,initiator=5,target=5,hierarchy=memory,data-type=access-bandwidth,bandwidth=800M

Uff!

Could you throw in output of "info qom-tree /"?

[...]




reply via email to

[Prev in Thread] Current Thread [Next in Thread]