[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
RE: [RFC 22/52] riscv: Replace MachineState.smp access with topology hel
From: |
Mi, Dapeng1 |
Subject: |
RE: [RFC 22/52] riscv: Replace MachineState.smp access with topology helpers |
Date: |
Tue, 14 Feb 2023 02:17:45 +0000 |
> From: Zhao Liu <zhao1.liu@linux.intel.com>
> Sent: Monday, February 13, 2023 5:50 PM
> To: Eduardo Habkost <eduardo@habkost.net>; Marcel Apfelbaum
> <marcel.apfelbaum@gmail.com>; Philippe Mathieu-Daudé <philmd@linaro.org>;
> Yanan Wang <wangyanan55@huawei.com>; Michael S . Tsirkin
> <mst@redhat.com>; Richard Henderson <richard.henderson@linaro.org>; Paolo
> Bonzini <pbonzini@redhat.com>; Eric Blake <eblake@redhat.com>; Markus
> Armbruster <armbru@redhat.com>
> Cc: qemu-devel@nongnu.org; Wang, Zhenyu Z <zhenyu.z.wang@intel.com>; Mi,
> Dapeng1 <dapeng1.mi@intel.com>; Ding, Zhuocheng
> <zhuocheng.ding@intel.com>; Robert Hoo <robert.hu@linux.intel.com>;
> Christopherson,, Sean <seanjc@google.com>; Like Xu
> <like.xu.linux@gmail.com>; Liu, Zhao1 <zhao1.liu@intel.com>; Meng, Bin
> <bin.meng@windriver.com>; Palmer Dabbelt <palmer@dabbelt.com>; Alistair
> Francis <alistair.francis@wdc.com>; Vijai Kumar K <vijai@behindbytes.com>
> Subject: [RFC 22/52] riscv: Replace MachineState.smp access with topology
> helpers
>
> From: Zhao Liu <zhao1.liu@intel.com>
>
> When MachineState.topo is introduced, the topology related structures
> become complicated. So we wrapped the access to topology fields of
> MachineState.topo into some helpers, and we are using these helpers
> to replace the use of MachineState.smp.
>
> In the codes of riscv, it's straightforward to replace topology access
> with wrapped generic interfaces.
>
> Cc: Bin Meng <bin.meng@windriver.com>
> Cc: Palmer Dabbelt <palmer@dabbelt.com>
> Cc: Alistair Francis <alistair.francis@wdc.com>
> CC: Vijai Kumar K <vijai@behindbytes.com>
> Signed-off-by: Zhao Liu <zhao1.liu@intel.com>
> ---
> hw/riscv/microchip_pfsoc.c | 11 ++++++-----
> hw/riscv/numa.c | 21 +++++++++++----------
> hw/riscv/opentitan.c | 8 ++++----
> hw/riscv/shakti_c.c | 2 +-
> hw/riscv/sifive_e.c | 10 ++++++----
> hw/riscv/sifive_u.c | 28 ++++++++++++++--------------
> hw/riscv/virt.c | 24 +++++++++++++-----------
> 7 files changed, 55 insertions(+), 49 deletions(-)
>
> diff --git a/hw/riscv/microchip_pfsoc.c b/hw/riscv/microchip_pfsoc.c
> index 2b91e49561f1..30295cce17e7 100644
> --- a/hw/riscv/microchip_pfsoc.c
> +++ b/hw/riscv/microchip_pfsoc.c
> @@ -164,7 +164,8 @@ static void microchip_pfsoc_soc_instance_init(Object
> *obj)
>
> object_initialize_child(OBJECT(&s->u_cluster), "u-cpus", &s->u_cpus,
> TYPE_RISCV_HART_ARRAY);
> - qdev_prop_set_uint32(DEVICE(&s->u_cpus), "num-harts", ms->smp.cpus - 1);
> + qdev_prop_set_uint32(DEVICE(&s->u_cpus), "num-harts",
> + machine_topo_get_cpus(ms) - 1);
> qdev_prop_set_uint32(DEVICE(&s->u_cpus), "hartid-base", 1);
> qdev_prop_set_string(DEVICE(&s->u_cpus), "cpu-type",
> TYPE_RISCV_CPU_SIFIVE_U54);
> @@ -249,10 +250,10 @@ static void microchip_pfsoc_soc_realize(DeviceState
> *dev, Error **errp)
>
> /* CLINT */
> riscv_aclint_swi_create(memmap[MICROCHIP_PFSOC_CLINT].base,
> - 0, ms->smp.cpus, false);
> + 0, machine_topo_get_cpus(ms), false);
> riscv_aclint_mtimer_create(
> memmap[MICROCHIP_PFSOC_CLINT].base + RISCV_ACLINT_SWI_SIZE,
> - RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, ms->smp.cpus,
> + RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, machine_topo_get_cpus(ms),
> RISCV_ACLINT_DEFAULT_MTIMECMP, RISCV_ACLINT_DEFAULT_MTIME,
> CLINT_TIMEBASE_FREQ, false);
>
> @@ -276,11 +277,11 @@ static void microchip_pfsoc_soc_realize(DeviceState
> *dev, Error **errp)
> l2lim_mem);
>
> /* create PLIC hart topology configuration string */
> - plic_hart_config = riscv_plic_hart_config_string(ms->smp.cpus);
> + plic_hart_config =
> riscv_plic_hart_config_string(machine_topo_get_cpus(ms));
>
> /* PLIC */
> s->plic = sifive_plic_create(memmap[MICROCHIP_PFSOC_PLIC].base,
> - plic_hart_config, ms->smp.cpus, 0,
> + plic_hart_config, machine_topo_get_cpus(ms), 0,
> MICROCHIP_PFSOC_PLIC_NUM_SOURCES,
> MICROCHIP_PFSOC_PLIC_NUM_PRIORITIES,
> MICROCHIP_PFSOC_PLIC_PRIORITY_BASE,
> diff --git a/hw/riscv/numa.c b/hw/riscv/numa.c
> index 472010256183..1fabdc42e767 100644
> --- a/hw/riscv/numa.c
> +++ b/hw/riscv/numa.c
> @@ -37,13 +37,14 @@ int riscv_socket_count(const MachineState *ms)
>
> int riscv_socket_first_hartid(const MachineState *ms, int socket_id)
> {
> - int i, first_hartid = ms->smp.cpus;
> + int i, first_hartid, cpus = machine_topo_get_cpus(ms);
>
> + first_hartid = cpus;
> if (!numa_enabled(ms)) {
> return (!socket_id) ? 0 : -1;
> }
>
> - for (i = 0; i < ms->smp.cpus; i++) {
> + for (i = 0; i < cpus; i++) {
> if (ms->possible_cpus->cpus[i].props.node_id != socket_id) {
> continue;
> }
> @@ -52,18 +53,18 @@ int riscv_socket_first_hartid(const MachineState *ms,
> int socket_id)
> }
> }
>
> - return (first_hartid < ms->smp.cpus) ? first_hartid : -1;
> + return (first_hartid < cpus) ? first_hartid : -1;
> }
>
> int riscv_socket_last_hartid(const MachineState *ms, int socket_id)
> {
> - int i, last_hartid = -1;
> + int i, last_hartid = -1, cpus = machine_topo_get_cpus(ms);
>
> if (!numa_enabled(ms)) {
> - return (!socket_id) ? ms->smp.cpus - 1 : -1;
> + return (!socket_id) ? cpus - 1 : -1;
> }
>
> - for (i = 0; i < ms->smp.cpus; i++) {
> + for (i = 0; i < cpus; i++) {
> if (ms->possible_cpus->cpus[i].props.node_id != socket_id) {
> continue;
> }
> @@ -72,7 +73,7 @@ int riscv_socket_last_hartid(const MachineState *ms, int
> socket_id)
> }
> }
>
> - return (last_hartid < ms->smp.cpus) ? last_hartid : -1;
> + return (last_hartid < cpus) ? last_hartid : -1;
> }
>
> int riscv_socket_hart_count(const MachineState *ms, int socket_id)
> @@ -80,7 +81,7 @@ int riscv_socket_hart_count(const MachineState *ms, int
> socket_id)
> int first_hartid, last_hartid;
>
> if (!numa_enabled(ms)) {
> - return (!socket_id) ? ms->smp.cpus : -1;
> + return (!socket_id) ? machine_topo_get_cpus(ms) : -1;
> }
>
> first_hartid = riscv_socket_first_hartid(ms, socket_id);
> @@ -208,7 +209,7 @@ int64_t riscv_numa_get_default_cpu_node_id(const
> MachineState *ms, int idx)
> int64_t nidx = 0;
>
> if (ms->numa_state->num_nodes) {
> - nidx = idx / (ms->smp.cpus / ms->numa_state->num_nodes);
> + nidx = idx / (machine_topo_get_cpus(ms) / ms->numa_state->num_nodes);
> if (ms->numa_state->num_nodes <= nidx) {
> nidx = ms->numa_state->num_nodes - 1;
> }
> @@ -220,7 +221,7 @@ int64_t riscv_numa_get_default_cpu_node_id(const
> MachineState *ms, int idx)
> const CPUArchIdList *riscv_numa_possible_cpu_arch_ids(MachineState *ms)
> {
> int n;
> - unsigned int max_cpus = ms->smp.max_cpus;
> + unsigned int max_cpus = machine_topo_get_max_cpus(ms);
>
> if (ms->possible_cpus) {
> assert(ms->possible_cpus->len == max_cpus);
> diff --git a/hw/riscv/opentitan.c b/hw/riscv/opentitan.c
> index 353f030d80c1..fb59c43dcc08 100644
> --- a/hw/riscv/opentitan.c
> +++ b/hw/riscv/opentitan.c
> @@ -143,11 +143,11 @@ static void lowrisc_ibex_soc_realize(DeviceState
> *dev_soc, Error **errp)
> MachineState *ms = MACHINE(qdev_get_machine());
> LowRISCIbexSoCState *s = RISCV_IBEX_SOC(dev_soc);
> MemoryRegion *sys_mem = get_system_memory();
> - int i;
> + int i, cpus = machine_topo_get_cpus(ms);
>
> object_property_set_str(OBJECT(&s->cpus), "cpu-type", ms->cpu_type,
> &error_abort);
> - object_property_set_int(OBJECT(&s->cpus), "num-harts", ms->smp.cpus,
> + object_property_set_int(OBJECT(&s->cpus), "num-harts", cpus,
> &error_abort);
> object_property_set_int(OBJECT(&s->cpus), "resetvec", s->resetvec,
> &error_abort);
> @@ -186,10 +186,10 @@ static void lowrisc_ibex_soc_realize(DeviceState
> *dev_soc, Error **errp)
> }
> sysbus_mmio_map(SYS_BUS_DEVICE(&s->plic), 0,
> memmap[IBEX_DEV_PLIC].base);
>
> - for (i = 0; i < ms->smp.cpus; i++) {
> + for (i = 0; i < cpus; i++) {
> CPUState *cpu = qemu_get_cpu(i);
>
> - qdev_connect_gpio_out(DEVICE(&s->plic), ms->smp.cpus + i,
> + qdev_connect_gpio_out(DEVICE(&s->plic), cpus + i,
> qdev_get_gpio_in(DEVICE(cpu), IRQ_M_EXT));
> }
>
> diff --git a/hw/riscv/shakti_c.c b/hw/riscv/shakti_c.c
> index e43cc9445cb1..34499510e2f0 100644
> --- a/hw/riscv/shakti_c.c
> +++ b/hw/riscv/shakti_c.c
> @@ -110,7 +110,7 @@ static void shakti_c_soc_state_realize(DeviceState *dev,
> Error **errp)
> sysbus_realize(SYS_BUS_DEVICE(&sss->cpus), &error_abort);
>
> sss->plic = sifive_plic_create(shakti_c_memmap[SHAKTI_C_PLIC].base,
> - (char *)SHAKTI_C_PLIC_HART_CONFIG, ms->smp.cpus, 0,
> + (char *)SHAKTI_C_PLIC_HART_CONFIG, machine_topo_get_cpus(ms), 0,
> SHAKTI_C_PLIC_NUM_SOURCES,
> SHAKTI_C_PLIC_NUM_PRIORITIES,
> SHAKTI_C_PLIC_PRIORITY_BASE,
> diff --git a/hw/riscv/sifive_e.c b/hw/riscv/sifive_e.c
> index 3e3f4b0088ca..ca3a7a0877cd 100644
> --- a/hw/riscv/sifive_e.c
> +++ b/hw/riscv/sifive_e.c
> @@ -178,7 +178,8 @@ static void sifive_e_soc_init(Object *obj)
> SiFiveESoCState *s = RISCV_E_SOC(obj);
>
> object_initialize_child(obj, "cpus", &s->cpus, TYPE_RISCV_HART_ARRAY);
> - object_property_set_int(OBJECT(&s->cpus), "num-harts", ms->smp.cpus,
> + object_property_set_int(OBJECT(&s->cpus), "num-harts",
> + machine_topo_get_cpus(ms),
> &error_abort);
With this change, we don't need to start a new line for '&error_abort);', just
move
it to previous line.
> object_property_set_int(OBJECT(&s->cpus), "resetvec", 0x1004,
> &error_abort);
> object_initialize_child(obj, "riscv.sifive.e.gpio0", &s->gpio,
> @@ -191,6 +192,7 @@ static void sifive_e_soc_realize(DeviceState *dev, Error
> **errp)
> const MemMapEntry *memmap = sifive_e_memmap;
> SiFiveESoCState *s = RISCV_E_SOC(dev);
> MemoryRegion *sys_mem = get_system_memory();
> + unsigned int cpus = machine_topo_get_cpus(ms);
>
> object_property_set_str(OBJECT(&s->cpus), "cpu-type", ms->cpu_type,
> &error_abort);
> @@ -204,7 +206,7 @@ static void sifive_e_soc_realize(DeviceState *dev, Error
> **errp)
>
> /* MMIO */
> s->plic = sifive_plic_create(memmap[SIFIVE_E_DEV_PLIC].base,
> - (char *)SIFIVE_E_PLIC_HART_CONFIG, ms->smp.cpus, 0,
> + (char *)SIFIVE_E_PLIC_HART_CONFIG, cpus, 0,
> SIFIVE_E_PLIC_NUM_SOURCES,
> SIFIVE_E_PLIC_NUM_PRIORITIES,
> SIFIVE_E_PLIC_PRIORITY_BASE,
> @@ -215,10 +217,10 @@ static void sifive_e_soc_realize(DeviceState *dev,
> Error **errp)
> SIFIVE_E_PLIC_CONTEXT_STRIDE,
> memmap[SIFIVE_E_DEV_PLIC].size);
> riscv_aclint_swi_create(memmap[SIFIVE_E_DEV_CLINT].base,
> - 0, ms->smp.cpus, false);
> + 0, cpus, false);
> riscv_aclint_mtimer_create(memmap[SIFIVE_E_DEV_CLINT].base +
> RISCV_ACLINT_SWI_SIZE,
> - RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, ms->smp.cpus,
> + RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, cpus,
> RISCV_ACLINT_DEFAULT_MTIMECMP, RISCV_ACLINT_DEFAULT_MTIME,
> RISCV_ACLINT_DEFAULT_TIMEBASE_FREQ, false);
> create_unimplemented_device("riscv.sifive.e.aon",
> diff --git a/hw/riscv/sifive_u.c b/hw/riscv/sifive_u.c
> index d3ab7a9cda5d..c7f969510209 100644
> --- a/hw/riscv/sifive_u.c
> +++ b/hw/riscv/sifive_u.c
> @@ -99,7 +99,7 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> MachineState *ms = MACHINE(s);
> uint64_t mem_size = ms->ram_size;
> void *fdt;
> - int cpu, fdt_size;
> + int cpu, fdt_size, num_cpus = machine_topo_get_cpus(ms);
> uint32_t *cells;
> char *nodename;
> uint32_t plic_phandle, prci_phandle, gpio_phandle, phandle = 1;
> @@ -175,7 +175,7 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> qemu_fdt_setprop_cell(fdt, "/cpus", "#size-cells", 0x0);
> qemu_fdt_setprop_cell(fdt, "/cpus", "#address-cells", 0x1);
>
> - for (cpu = ms->smp.cpus - 1; cpu >= 0; cpu--) {
> + for (cpu = num_cpus - 1; cpu >= 0; cpu--) {
> int cpu_phandle = phandle++;
> nodename = g_strdup_printf("/cpus/cpu@%d", cpu);
> char *intc = g_strdup_printf("/cpus/cpu@%d/interrupt-controller",
> cpu);
> @@ -207,8 +207,8 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> g_free(nodename);
> }
>
> - cells = g_new0(uint32_t, ms->smp.cpus * 4);
> - for (cpu = 0; cpu < ms->smp.cpus; cpu++) {
> + cells = g_new0(uint32_t, num_cpus * 4);
> + for (cpu = 0; cpu < num_cpus; cpu++) {
> nodename =
> g_strdup_printf("/cpus/cpu@%d/interrupt-controller", cpu);
> uint32_t intc_phandle = qemu_fdt_get_phandle(fdt, nodename);
> @@ -227,7 +227,7 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> 0x0, memmap[SIFIVE_U_DEV_CLINT].base,
> 0x0, memmap[SIFIVE_U_DEV_CLINT].size);
> qemu_fdt_setprop(fdt, nodename, "interrupts-extended",
> - cells, ms->smp.cpus * sizeof(uint32_t) * 4);
> + cells, num_cpus * sizeof(uint32_t) * 4);
> g_free(cells);
> g_free(nodename);
>
> @@ -258,8 +258,8 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> g_free(nodename);
>
> plic_phandle = phandle++;
> - cells = g_new0(uint32_t, ms->smp.cpus * 4 - 2);
> - for (cpu = 0; cpu < ms->smp.cpus; cpu++) {
> + cells = g_new0(uint32_t, machine_topo_get_cpus(ms) * 4 - 2);
> + for (cpu = 0; cpu < machine_topo_get_cpus(ms); cpu++) {
> nodename =
> g_strdup_printf("/cpus/cpu@%d/interrupt-controller", cpu);
> uint32_t intc_phandle = qemu_fdt_get_phandle(fdt, nodename);
> @@ -283,7 +283,7 @@ static void create_fdt(SiFiveUState *s, const
> MemMapEntry *memmap,
> (char **)&plic_compat, ARRAY_SIZE(plic_compat));
> qemu_fdt_setprop(fdt, nodename, "interrupt-controller", NULL, 0);
> qemu_fdt_setprop(fdt, nodename, "interrupts-extended",
> - cells, (ms->smp.cpus * 4 - 2) * sizeof(uint32_t));
> + cells, (machine_topo_get_cpus(ms) * 4 - 2) * sizeof(uint32_t));
> qemu_fdt_setprop_cells(fdt, nodename, "reg",
> 0x0, memmap[SIFIVE_U_DEV_PLIC].base,
> 0x0, memmap[SIFIVE_U_DEV_PLIC].size);
> @@ -797,10 +797,10 @@ static void sifive_u_soc_realize(DeviceState *dev,
> Error **errp)
> MemoryRegion *mask_rom = g_new(MemoryRegion, 1);
> MemoryRegion *l2lim_mem = g_new(MemoryRegion, 1);
> char *plic_hart_config;
> - int i, j;
> + int i, j, cpus = machine_topo_get_cpus(ms);
> NICInfo *nd = &nd_table[0];
>
> - qdev_prop_set_uint32(DEVICE(&s->u_cpus), "num-harts", ms->smp.cpus - 1);
> + qdev_prop_set_uint32(DEVICE(&s->u_cpus), "num-harts", cpus - 1);
> qdev_prop_set_uint32(DEVICE(&s->u_cpus), "hartid-base", 1);
> qdev_prop_set_string(DEVICE(&s->u_cpus), "cpu-type", s->cpu_type);
> qdev_prop_set_uint64(DEVICE(&s->u_cpus), "resetvec", 0x1004);
> @@ -837,11 +837,11 @@ static void sifive_u_soc_realize(DeviceState *dev,
> Error **errp)
> l2lim_mem);
>
> /* create PLIC hart topology configuration string */
> - plic_hart_config = riscv_plic_hart_config_string(ms->smp.cpus);
> + plic_hart_config = riscv_plic_hart_config_string(cpus);
>
> /* MMIO */
> s->plic = sifive_plic_create(memmap[SIFIVE_U_DEV_PLIC].base,
> - plic_hart_config, ms->smp.cpus, 0,
> + plic_hart_config, cpus, 0,
> SIFIVE_U_PLIC_NUM_SOURCES,
> SIFIVE_U_PLIC_NUM_PRIORITIES,
> SIFIVE_U_PLIC_PRIORITY_BASE,
> @@ -857,10 +857,10 @@ static void sifive_u_soc_realize(DeviceState *dev,
> Error **errp)
> sifive_uart_create(system_memory, memmap[SIFIVE_U_DEV_UART1].base,
> serial_hd(1), qdev_get_gpio_in(DEVICE(s->plic), SIFIVE_U_UART1_IRQ));
> riscv_aclint_swi_create(memmap[SIFIVE_U_DEV_CLINT].base, 0,
> - ms->smp.cpus, false);
> + cpus, false);
> riscv_aclint_mtimer_create(memmap[SIFIVE_U_DEV_CLINT].base +
> RISCV_ACLINT_SWI_SIZE,
> - RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, ms->smp.cpus,
> + RISCV_ACLINT_DEFAULT_MTIMER_SIZE, 0, cpus,
> RISCV_ACLINT_DEFAULT_MTIMECMP, RISCV_ACLINT_DEFAULT_MTIME,
> CLINT_TIMEBASE_FREQ, false);
>
> diff --git a/hw/riscv/virt.c b/hw/riscv/virt.c
> index b81081c70b1c..4b0a3dab5ec1 100644
> --- a/hw/riscv/virt.c
> +++ b/hw/riscv/virt.c
> @@ -506,16 +506,17 @@ static void create_fdt_imsic(RISCVVirtState *s, const
> MemMapEntry *memmap,
> char *imsic_name;
> MachineState *ms = MACHINE(s);
> int socket_count = riscv_socket_count(ms);
> - uint32_t imsic_max_hart_per_socket, imsic_guest_bits;
> + uint32_t imsic_max_hart_per_socket, imsic_guest_bits, cpus;
> uint32_t *imsic_cells, *imsic_regs, imsic_addr, imsic_size;
>
> + cpus = machine_topo_get_cpus(ms);
> *msi_m_phandle = (*phandle)++;
> *msi_s_phandle = (*phandle)++;
> - imsic_cells = g_new0(uint32_t, ms->smp.cpus * 2);
> + imsic_cells = g_new0(uint32_t, cpus * 2);
> imsic_regs = g_new0(uint32_t, socket_count * 4);
>
> /* M-level IMSIC node */
> - for (cpu = 0; cpu < ms->smp.cpus; cpu++) {
> + for (cpu = 0; cpu < cpus; cpu++) {
> imsic_cells[cpu * 2 + 0] = cpu_to_be32(intc_phandles[cpu]);
> imsic_cells[cpu * 2 + 1] = cpu_to_be32(IRQ_M_EXT);
> }
> @@ -544,7 +545,7 @@ static void create_fdt_imsic(RISCVVirtState *s, const
> MemMapEntry *memmap,
> qemu_fdt_setprop(ms->fdt, imsic_name, "msi-controller",
> NULL, 0);
> qemu_fdt_setprop(ms->fdt, imsic_name, "interrupts-extended",
> - imsic_cells, ms->smp.cpus * sizeof(uint32_t) * 2);
> + imsic_cells, cpus * sizeof(uint32_t) * 2);
> qemu_fdt_setprop(ms->fdt, imsic_name, "reg", imsic_regs,
> socket_count * sizeof(uint32_t) * 4);
> qemu_fdt_setprop_cell(ms->fdt, imsic_name, "riscv,num-ids",
> @@ -562,7 +563,7 @@ static void create_fdt_imsic(RISCVVirtState *s, const
> MemMapEntry *memmap,
> g_free(imsic_name);
>
> /* S-level IMSIC node */
> - for (cpu = 0; cpu < ms->smp.cpus; cpu++) {
> + for (cpu = 0; cpu < cpus; cpu++) {
> imsic_cells[cpu * 2 + 0] = cpu_to_be32(intc_phandles[cpu]);
> imsic_cells[cpu * 2 + 1] = cpu_to_be32(IRQ_S_EXT);
> }
> @@ -593,7 +594,7 @@ static void create_fdt_imsic(RISCVVirtState *s, const
> MemMapEntry *memmap,
> qemu_fdt_setprop(ms->fdt, imsic_name, "msi-controller",
> NULL, 0);
> qemu_fdt_setprop(ms->fdt, imsic_name, "interrupts-extended",
> - imsic_cells, ms->smp.cpus * sizeof(uint32_t) * 2);
> + imsic_cells, cpus * sizeof(uint32_t) * 2);
> qemu_fdt_setprop(ms->fdt, imsic_name, "reg", imsic_regs,
> socket_count * sizeof(uint32_t) * 4);
> qemu_fdt_setprop_cell(ms->fdt, imsic_name, "riscv,num-ids",
> @@ -732,7 +733,7 @@ static void create_fdt_sockets(RISCVVirtState *s, const
> MemMapEntry *memmap,
> char *clust_name;
> int socket, phandle_pos;
> MachineState *ms = MACHINE(s);
> - uint32_t msi_m_phandle = 0, msi_s_phandle = 0;
> + uint32_t msi_m_phandle = 0, msi_s_phandle = 0, cpus;
> uint32_t *intc_phandles, xplic_phandles[MAX_NODES];
> int socket_count = riscv_socket_count(ms);
>
> @@ -743,9 +744,10 @@ static void create_fdt_sockets(RISCVVirtState *s, const
> MemMapEntry *memmap,
> qemu_fdt_setprop_cell(ms->fdt, "/cpus", "#address-cells", 0x1);
> qemu_fdt_add_subnode(ms->fdt, "/cpus/cpu-map");
>
> - intc_phandles = g_new0(uint32_t, ms->smp.cpus);
> + cpus = machine_topo_get_cpus(ms);
> + intc_phandles = g_new0(uint32_t, cpus);
>
> - phandle_pos = ms->smp.cpus;
> + phandle_pos = cpus;
> for (socket = (socket_count - 1); socket >= 0; socket--) {
> phandle_pos -= s->soc[socket].num_harts;
>
> @@ -776,7 +778,7 @@ static void create_fdt_sockets(RISCVVirtState *s, const
> MemMapEntry *memmap,
> *msi_pcie_phandle = msi_s_phandle;
> }
>
> - phandle_pos = ms->smp.cpus;
> + phandle_pos = cpus;
> for (socket = (socket_count - 1); socket >= 0; socket--) {
> phandle_pos -= s->soc[socket].num_harts;
>
> @@ -1113,7 +1115,7 @@ static FWCfgState *create_fw_cfg(const
> MachineState *ms)
>
> fw_cfg = fw_cfg_init_mem_wide(base + 8, base, 8, base + 16,
> &address_space_memory);
> - fw_cfg_add_i16(fw_cfg, FW_CFG_NB_CPUS, (uint16_t)ms->smp.cpus);
> + fw_cfg_add_i16(fw_cfg, FW_CFG_NB_CPUS,
> (uint16_t)machine_topo_get_cpus(ms));
>
> return fw_cfg;
> }
> --
> 2.34.1
- Re: [RFC 15/52] migration/postcopy-ram: Use generic topology name and helper, (continued)
- [RFC 16/52] plugins: Use generic topology name and helper, Zhao Liu, 2023/02/13
- [RFC 17/52] softmmu/cpus: Use generic topology helper in vcpus initialization, Zhao Liu, 2023/02/13
- [RFC 18/52] general: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 19/52] i386: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 20/52] s390x: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 21/52] ppc: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 22/52] riscv: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- RE: [RFC 22/52] riscv: Replace MachineState.smp access with topology helpers,
Mi, Dapeng1 <=
- [RFC 23/52] arm: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 25/52] mips: Replace MachineState.smp access with topology helpers, Zhao Liu, 2023/02/13
- [RFC 26/52] hw: Replace MachineState.smp access with topology helpers for all remaining archs, Zhao Liu, 2023/02/13
- [RFC 27/52] test/test-smp-parse: Check fields of MachineState.topo.smp, Zhao Liu, 2023/02/13
- [RFC 28/52] hw/core/machine: Remove support of MachineState.smp, Zhao Liu, 2023/02/13