| // SPDX-License-Identifier: GPL-2.0-only |
| // Copyright(c) 2021 Intel Corporation. All rights reserved. |
| |
| #include <linux/platform_device.h> |
| #include <linux/genalloc.h> |
| #include <linux/module.h> |
| #include <linux/mutex.h> |
| #include <linux/acpi.h> |
| #include <linux/pci.h> |
| #include <linux/mm.h> |
| #include <cxlmem.h> |
| |
| #include "../watermark.h" |
| #include "mock.h" |
| |
| static int interleave_arithmetic; |
| |
| #define FAKE_QTG_ID 42 |
| |
| #define NR_CXL_HOST_BRIDGES 2 |
| #define NR_CXL_SINGLE_HOST 1 |
| #define NR_CXL_RCH 1 |
| #define NR_CXL_ROOT_PORTS 2 |
| #define NR_CXL_SWITCH_PORTS 2 |
| #define NR_CXL_PORT_DECODERS 8 |
| #define NR_BRIDGES (NR_CXL_HOST_BRIDGES + NR_CXL_SINGLE_HOST + NR_CXL_RCH) |
| |
| static struct platform_device *cxl_acpi; |
| static struct platform_device *cxl_host_bridge[NR_CXL_HOST_BRIDGES]; |
| #define NR_MULTI_ROOT (NR_CXL_HOST_BRIDGES * NR_CXL_ROOT_PORTS) |
| static struct platform_device *cxl_root_port[NR_MULTI_ROOT]; |
| static struct platform_device *cxl_switch_uport[NR_MULTI_ROOT]; |
| #define NR_MEM_MULTI \ |
| (NR_CXL_HOST_BRIDGES * NR_CXL_ROOT_PORTS * NR_CXL_SWITCH_PORTS) |
| static struct platform_device *cxl_switch_dport[NR_MEM_MULTI]; |
| |
| static struct platform_device *cxl_hb_single[NR_CXL_SINGLE_HOST]; |
| static struct platform_device *cxl_root_single[NR_CXL_SINGLE_HOST]; |
| static struct platform_device *cxl_swu_single[NR_CXL_SINGLE_HOST]; |
| #define NR_MEM_SINGLE (NR_CXL_SINGLE_HOST * NR_CXL_SWITCH_PORTS) |
| static struct platform_device *cxl_swd_single[NR_MEM_SINGLE]; |
| |
| struct platform_device *cxl_mem[NR_MEM_MULTI]; |
| struct platform_device *cxl_mem_single[NR_MEM_SINGLE]; |
| |
| static struct platform_device *cxl_rch[NR_CXL_RCH]; |
| static struct platform_device *cxl_rcd[NR_CXL_RCH]; |
| |
| static inline bool is_multi_bridge(struct device *dev) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) |
| if (&cxl_host_bridge[i]->dev == dev) |
| return true; |
| return false; |
| } |
| |
| static inline bool is_single_bridge(struct device *dev) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) |
| if (&cxl_hb_single[i]->dev == dev) |
| return true; |
| return false; |
| } |
| |
| static struct acpi_device acpi0017_mock; |
| static struct acpi_device host_bridge[NR_BRIDGES] = { |
| [0] = { |
| .handle = &host_bridge[0], |
| .pnp.unique_id = "0", |
| }, |
| [1] = { |
| .handle = &host_bridge[1], |
| .pnp.unique_id = "1", |
| }, |
| [2] = { |
| .handle = &host_bridge[2], |
| .pnp.unique_id = "2", |
| }, |
| [3] = { |
| .handle = &host_bridge[3], |
| .pnp.unique_id = "3", |
| }, |
| }; |
| |
| static bool is_mock_dev(struct device *dev) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_mem); i++) |
| if (dev == &cxl_mem[i]->dev) |
| return true; |
| for (i = 0; i < ARRAY_SIZE(cxl_mem_single); i++) |
| if (dev == &cxl_mem_single[i]->dev) |
| return true; |
| for (i = 0; i < ARRAY_SIZE(cxl_rcd); i++) |
| if (dev == &cxl_rcd[i]->dev) |
| return true; |
| if (dev == &cxl_acpi->dev) |
| return true; |
| return false; |
| } |
| |
| static bool is_mock_adev(struct acpi_device *adev) |
| { |
| int i; |
| |
| if (adev == &acpi0017_mock) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(host_bridge); i++) |
| if (adev == &host_bridge[i]) |
| return true; |
| |
| return false; |
| } |
| |
| static struct { |
| struct acpi_table_cedt cedt; |
| struct acpi_cedt_chbs chbs[NR_BRIDGES]; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[1]; |
| } cfmws0; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[2]; |
| } cfmws1; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[1]; |
| } cfmws2; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[2]; |
| } cfmws3; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[1]; |
| } cfmws4; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[1]; |
| } cfmws5; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[1]; |
| } cfmws6; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[2]; |
| } cfmws7; |
| struct { |
| struct acpi_cedt_cfmws cfmws; |
| u32 target[4]; |
| } cfmws8; |
| struct { |
| struct acpi_cedt_cxims cxims; |
| u64 xormap_list[2]; |
| } cxims0; |
| } __packed mock_cedt = { |
| .cedt = { |
| .header = { |
| .signature = "CEDT", |
| .length = sizeof(mock_cedt), |
| .revision = 1, |
| }, |
| }, |
| .chbs[0] = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CHBS, |
| .length = sizeof(mock_cedt.chbs[0]), |
| }, |
| .uid = 0, |
| .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, |
| }, |
| .chbs[1] = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CHBS, |
| .length = sizeof(mock_cedt.chbs[0]), |
| }, |
| .uid = 1, |
| .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, |
| }, |
| .chbs[2] = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CHBS, |
| .length = sizeof(mock_cedt.chbs[0]), |
| }, |
| .uid = 2, |
| .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, |
| }, |
| .chbs[3] = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CHBS, |
| .length = sizeof(mock_cedt.chbs[0]), |
| }, |
| .uid = 3, |
| .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL11, |
| }, |
| .cfmws0 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws0), |
| }, |
| .interleave_ways = 0, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 4UL, |
| }, |
| .target = { 0 }, |
| }, |
| .cfmws1 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws1), |
| }, |
| .interleave_ways = 1, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 8UL, |
| }, |
| .target = { 0, 1, }, |
| }, |
| .cfmws2 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws2), |
| }, |
| .interleave_ways = 0, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 4UL, |
| }, |
| .target = { 0 }, |
| }, |
| .cfmws3 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws3), |
| }, |
| .interleave_ways = 1, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 8UL, |
| }, |
| .target = { 0, 1, }, |
| }, |
| .cfmws4 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws4), |
| }, |
| .interleave_ways = 0, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 4UL, |
| }, |
| .target = { 2 }, |
| }, |
| .cfmws5 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws5), |
| }, |
| .interleave_ways = 0, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M, |
| }, |
| .target = { 3 }, |
| }, |
| /* .cfmws6,7,8 use ACPI_CEDT_CFMWS_ARITHMETIC_XOR */ |
| .cfmws6 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws6), |
| }, |
| .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, |
| .interleave_ways = 0, |
| .granularity = 4, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 8UL, |
| }, |
| .target = { 0, }, |
| }, |
| .cfmws7 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws7), |
| }, |
| .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, |
| .interleave_ways = 1, |
| .granularity = 0, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 8UL, |
| }, |
| .target = { 0, 1, }, |
| }, |
| .cfmws8 = { |
| .cfmws = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CFMWS, |
| .length = sizeof(mock_cedt.cfmws8), |
| }, |
| .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, |
| .interleave_ways = 2, |
| .granularity = 0, |
| .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | |
| ACPI_CEDT_CFMWS_RESTRICT_PMEM, |
| .qtg_id = FAKE_QTG_ID, |
| .window_size = SZ_256M * 16UL, |
| }, |
| .target = { 0, 1, 0, 1, }, |
| }, |
| .cxims0 = { |
| .cxims = { |
| .header = { |
| .type = ACPI_CEDT_TYPE_CXIMS, |
| .length = sizeof(mock_cedt.cxims0), |
| }, |
| .hbig = 0, |
| .nr_xormaps = 2, |
| }, |
| .xormap_list = { 0x404100, 0x808200, }, |
| }, |
| }; |
| |
| struct acpi_cedt_cfmws *mock_cfmws[] = { |
| [0] = &mock_cedt.cfmws0.cfmws, |
| [1] = &mock_cedt.cfmws1.cfmws, |
| [2] = &mock_cedt.cfmws2.cfmws, |
| [3] = &mock_cedt.cfmws3.cfmws, |
| [4] = &mock_cedt.cfmws4.cfmws, |
| [5] = &mock_cedt.cfmws5.cfmws, |
| /* Modulo Math above, XOR Math below */ |
| [6] = &mock_cedt.cfmws6.cfmws, |
| [7] = &mock_cedt.cfmws7.cfmws, |
| [8] = &mock_cedt.cfmws8.cfmws, |
| }; |
| |
| static int cfmws_start; |
| static int cfmws_end; |
| #define CFMWS_MOD_ARRAY_START 0 |
| #define CFMWS_MOD_ARRAY_END 5 |
| #define CFMWS_XOR_ARRAY_START 6 |
| #define CFMWS_XOR_ARRAY_END 8 |
| |
| struct acpi_cedt_cxims *mock_cxims[1] = { |
| [0] = &mock_cedt.cxims0.cxims, |
| }; |
| |
| struct cxl_mock_res { |
| struct list_head list; |
| struct range range; |
| }; |
| |
| static LIST_HEAD(mock_res); |
| static DEFINE_MUTEX(mock_res_lock); |
| static struct gen_pool *cxl_mock_pool; |
| |
| static void depopulate_all_mock_resources(void) |
| { |
| struct cxl_mock_res *res, *_res; |
| |
| mutex_lock(&mock_res_lock); |
| list_for_each_entry_safe(res, _res, &mock_res, list) { |
| gen_pool_free(cxl_mock_pool, res->range.start, |
| range_len(&res->range)); |
| list_del(&res->list); |
| kfree(res); |
| } |
| mutex_unlock(&mock_res_lock); |
| } |
| |
| static struct cxl_mock_res *alloc_mock_res(resource_size_t size, int align) |
| { |
| struct cxl_mock_res *res = kzalloc(sizeof(*res), GFP_KERNEL); |
| struct genpool_data_align data = { |
| .align = align, |
| }; |
| unsigned long phys; |
| |
| INIT_LIST_HEAD(&res->list); |
| phys = gen_pool_alloc_algo(cxl_mock_pool, size, |
| gen_pool_first_fit_align, &data); |
| if (!phys) |
| return NULL; |
| |
| res->range = (struct range) { |
| .start = phys, |
| .end = phys + size - 1, |
| }; |
| mutex_lock(&mock_res_lock); |
| list_add(&res->list, &mock_res); |
| mutex_unlock(&mock_res_lock); |
| |
| return res; |
| } |
| |
| static int populate_cedt(void) |
| { |
| struct cxl_mock_res *res; |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(mock_cedt.chbs); i++) { |
| struct acpi_cedt_chbs *chbs = &mock_cedt.chbs[i]; |
| resource_size_t size; |
| |
| if (chbs->cxl_version == ACPI_CEDT_CHBS_VERSION_CXL20) |
| size = ACPI_CEDT_CHBS_LENGTH_CXL20; |
| else |
| size = ACPI_CEDT_CHBS_LENGTH_CXL11; |
| |
| res = alloc_mock_res(size, size); |
| if (!res) |
| return -ENOMEM; |
| chbs->base = res->range.start; |
| chbs->length = size; |
| } |
| |
| for (i = cfmws_start; i <= cfmws_end; i++) { |
| struct acpi_cedt_cfmws *window = mock_cfmws[i]; |
| |
| res = alloc_mock_res(window->window_size, SZ_256M); |
| if (!res) |
| return -ENOMEM; |
| window->base_hpa = res->range.start; |
| } |
| |
| return 0; |
| } |
| |
| static bool is_mock_port(struct device *dev); |
| |
| /* |
| * WARNING, this hack assumes the format of 'struct cxl_cfmws_context' |
| * and 'struct cxl_chbs_context' share the property that the first |
| * struct member is a cxl_test device being probed by the cxl_acpi |
| * driver. |
| */ |
| struct cxl_cedt_context { |
| struct device *dev; |
| }; |
| |
| static int mock_acpi_table_parse_cedt(enum acpi_cedt_type id, |
| acpi_tbl_entry_handler_arg handler_arg, |
| void *arg) |
| { |
| struct cxl_cedt_context *ctx = arg; |
| struct device *dev = ctx->dev; |
| union acpi_subtable_headers *h; |
| unsigned long end; |
| int i; |
| |
| if (!is_mock_port(dev) && !is_mock_dev(dev)) |
| return acpi_table_parse_cedt(id, handler_arg, arg); |
| |
| if (id == ACPI_CEDT_TYPE_CHBS) |
| for (i = 0; i < ARRAY_SIZE(mock_cedt.chbs); i++) { |
| h = (union acpi_subtable_headers *)&mock_cedt.chbs[i]; |
| end = (unsigned long)&mock_cedt.chbs[i + 1]; |
| handler_arg(h, arg, end); |
| } |
| |
| if (id == ACPI_CEDT_TYPE_CFMWS) |
| for (i = cfmws_start; i <= cfmws_end; i++) { |
| h = (union acpi_subtable_headers *) mock_cfmws[i]; |
| end = (unsigned long) h + mock_cfmws[i]->header.length; |
| handler_arg(h, arg, end); |
| } |
| |
| if (id == ACPI_CEDT_TYPE_CXIMS) |
| for (i = 0; i < ARRAY_SIZE(mock_cxims); i++) { |
| h = (union acpi_subtable_headers *)mock_cxims[i]; |
| end = (unsigned long)h + mock_cxims[i]->header.length; |
| handler_arg(h, arg, end); |
| } |
| |
| return 0; |
| } |
| |
| static bool is_mock_bridge(struct device *dev) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) |
| if (dev == &cxl_host_bridge[i]->dev) |
| return true; |
| for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) |
| if (dev == &cxl_hb_single[i]->dev) |
| return true; |
| for (i = 0; i < ARRAY_SIZE(cxl_rch); i++) |
| if (dev == &cxl_rch[i]->dev) |
| return true; |
| |
| return false; |
| } |
| |
| static bool is_mock_port(struct device *dev) |
| { |
| int i; |
| |
| if (is_mock_bridge(dev)) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_root_port); i++) |
| if (dev == &cxl_root_port[i]->dev) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_switch_uport); i++) |
| if (dev == &cxl_switch_uport[i]->dev) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_switch_dport); i++) |
| if (dev == &cxl_switch_dport[i]->dev) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_root_single); i++) |
| if (dev == &cxl_root_single[i]->dev) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_swu_single); i++) |
| if (dev == &cxl_swu_single[i]->dev) |
| return true; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_swd_single); i++) |
| if (dev == &cxl_swd_single[i]->dev) |
| return true; |
| |
| if (is_cxl_memdev(dev)) |
| return is_mock_dev(dev->parent); |
| |
| return false; |
| } |
| |
| static int host_bridge_index(struct acpi_device *adev) |
| { |
| return adev - host_bridge; |
| } |
| |
| static struct acpi_device *find_host_bridge(acpi_handle handle) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(host_bridge); i++) |
| if (handle == host_bridge[i].handle) |
| return &host_bridge[i]; |
| return NULL; |
| } |
| |
| static acpi_status |
| mock_acpi_evaluate_integer(acpi_handle handle, acpi_string pathname, |
| struct acpi_object_list *arguments, |
| unsigned long long *data) |
| { |
| struct acpi_device *adev = find_host_bridge(handle); |
| |
| if (!adev || strcmp(pathname, METHOD_NAME__UID) != 0) |
| return acpi_evaluate_integer(handle, pathname, arguments, data); |
| |
| *data = host_bridge_index(adev); |
| return AE_OK; |
| } |
| |
| static struct pci_bus mock_pci_bus[NR_BRIDGES]; |
| static struct acpi_pci_root mock_pci_root[ARRAY_SIZE(mock_pci_bus)] = { |
| [0] = { |
| .bus = &mock_pci_bus[0], |
| }, |
| [1] = { |
| .bus = &mock_pci_bus[1], |
| }, |
| [2] = { |
| .bus = &mock_pci_bus[2], |
| }, |
| [3] = { |
| .bus = &mock_pci_bus[3], |
| }, |
| |
| }; |
| |
| static bool is_mock_bus(struct pci_bus *bus) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(mock_pci_bus); i++) |
| if (bus == &mock_pci_bus[i]) |
| return true; |
| return false; |
| } |
| |
| static struct acpi_pci_root *mock_acpi_pci_find_root(acpi_handle handle) |
| { |
| struct acpi_device *adev = find_host_bridge(handle); |
| |
| if (!adev) |
| return acpi_pci_find_root(handle); |
| return &mock_pci_root[host_bridge_index(adev)]; |
| } |
| |
| static struct cxl_hdm *mock_cxl_setup_hdm(struct cxl_port *port, |
| struct cxl_endpoint_dvsec_info *info) |
| { |
| struct cxl_hdm *cxlhdm = devm_kzalloc(&port->dev, sizeof(*cxlhdm), GFP_KERNEL); |
| struct device *dev = &port->dev; |
| |
| if (!cxlhdm) |
| return ERR_PTR(-ENOMEM); |
| |
| cxlhdm->port = port; |
| cxlhdm->interleave_mask = ~0U; |
| cxlhdm->iw_cap_mask = ~0UL; |
| dev_set_drvdata(dev, cxlhdm); |
| return cxlhdm; |
| } |
| |
| static int mock_cxl_add_passthrough_decoder(struct cxl_port *port) |
| { |
| dev_err(&port->dev, "unexpected passthrough decoder for cxl_test\n"); |
| return -EOPNOTSUPP; |
| } |
| |
| |
| struct target_map_ctx { |
| int *target_map; |
| int index; |
| int target_count; |
| }; |
| |
| static int map_targets(struct device *dev, void *data) |
| { |
| struct platform_device *pdev = to_platform_device(dev); |
| struct target_map_ctx *ctx = data; |
| |
| ctx->target_map[ctx->index++] = pdev->id; |
| |
| if (ctx->index > ctx->target_count) { |
| dev_WARN_ONCE(dev, 1, "too many targets found?\n"); |
| return -ENXIO; |
| } |
| |
| return 0; |
| } |
| |
| static int mock_decoder_commit(struct cxl_decoder *cxld) |
| { |
| struct cxl_port *port = to_cxl_port(cxld->dev.parent); |
| int id = cxld->id; |
| |
| if (cxld->flags & CXL_DECODER_F_ENABLE) |
| return 0; |
| |
| dev_dbg(&port->dev, "%s commit\n", dev_name(&cxld->dev)); |
| if (cxl_num_decoders_committed(port) != id) { |
| dev_dbg(&port->dev, |
| "%s: out of order commit, expected decoder%d.%d\n", |
| dev_name(&cxld->dev), port->id, |
| cxl_num_decoders_committed(port)); |
| return -EBUSY; |
| } |
| |
| port->commit_end++; |
| cxld->flags |= CXL_DECODER_F_ENABLE; |
| |
| return 0; |
| } |
| |
| static void mock_decoder_reset(struct cxl_decoder *cxld) |
| { |
| struct cxl_port *port = to_cxl_port(cxld->dev.parent); |
| int id = cxld->id; |
| |
| if ((cxld->flags & CXL_DECODER_F_ENABLE) == 0) |
| return; |
| |
| dev_dbg(&port->dev, "%s reset\n", dev_name(&cxld->dev)); |
| if (port->commit_end == id) |
| cxl_port_commit_reap(cxld); |
| else |
| dev_dbg(&port->dev, |
| "%s: out of order reset, expected decoder%d.%d\n", |
| dev_name(&cxld->dev), port->id, port->commit_end); |
| cxld->flags &= ~CXL_DECODER_F_ENABLE; |
| } |
| |
| static void default_mock_decoder(struct cxl_decoder *cxld) |
| { |
| cxld->hpa_range = (struct range){ |
| .start = 0, |
| .end = -1, |
| }; |
| |
| cxld->interleave_ways = 1; |
| cxld->interleave_granularity = 256; |
| cxld->target_type = CXL_DECODER_HOSTONLYMEM; |
| cxld->commit = mock_decoder_commit; |
| cxld->reset = mock_decoder_reset; |
| } |
| |
| static int first_decoder(struct device *dev, void *data) |
| { |
| struct cxl_decoder *cxld; |
| |
| if (!is_switch_decoder(dev)) |
| return 0; |
| cxld = to_cxl_decoder(dev); |
| if (cxld->id == 0) |
| return 1; |
| return 0; |
| } |
| |
| static void mock_init_hdm_decoder(struct cxl_decoder *cxld) |
| { |
| struct acpi_cedt_cfmws *window = mock_cfmws[0]; |
| struct platform_device *pdev = NULL; |
| struct cxl_endpoint_decoder *cxled; |
| struct cxl_switch_decoder *cxlsd; |
| struct cxl_port *port, *iter; |
| const int size = SZ_512M; |
| struct cxl_memdev *cxlmd; |
| struct cxl_dport *dport; |
| struct device *dev; |
| bool hb0 = false; |
| u64 base; |
| int i; |
| |
| if (is_endpoint_decoder(&cxld->dev)) { |
| cxled = to_cxl_endpoint_decoder(&cxld->dev); |
| cxlmd = cxled_to_memdev(cxled); |
| WARN_ON(!dev_is_platform(cxlmd->dev.parent)); |
| pdev = to_platform_device(cxlmd->dev.parent); |
| |
| /* check is endpoint is attach to host-bridge0 */ |
| port = cxled_to_port(cxled); |
| do { |
| if (port->uport_dev == &cxl_host_bridge[0]->dev) { |
| hb0 = true; |
| break; |
| } |
| if (is_cxl_port(port->dev.parent)) |
| port = to_cxl_port(port->dev.parent); |
| else |
| port = NULL; |
| } while (port); |
| port = cxled_to_port(cxled); |
| } |
| |
| /* |
| * The first decoder on the first 2 devices on the first switch |
| * attached to host-bridge0 mock a fake / static RAM region. All |
| * other decoders are default disabled. Given the round robin |
| * assignment those devices are named cxl_mem.0, and cxl_mem.4. |
| * |
| * See 'cxl list -BMPu -m cxl_mem.0,cxl_mem.4' |
| */ |
| if (!hb0 || pdev->id % 4 || pdev->id > 4 || cxld->id > 0) { |
| default_mock_decoder(cxld); |
| return; |
| } |
| |
| base = window->base_hpa; |
| cxld->hpa_range = (struct range) { |
| .start = base, |
| .end = base + size - 1, |
| }; |
| |
| cxld->interleave_ways = 2; |
| eig_to_granularity(window->granularity, &cxld->interleave_granularity); |
| cxld->target_type = CXL_DECODER_HOSTONLYMEM; |
| cxld->flags = CXL_DECODER_F_ENABLE; |
| cxled->state = CXL_DECODER_STATE_AUTO; |
| port->commit_end = cxld->id; |
| devm_cxl_dpa_reserve(cxled, 0, size / cxld->interleave_ways, 0); |
| cxld->commit = mock_decoder_commit; |
| cxld->reset = mock_decoder_reset; |
| |
| /* |
| * Now that endpoint decoder is set up, walk up the hierarchy |
| * and setup the switch and root port decoders targeting @cxlmd. |
| */ |
| iter = port; |
| for (i = 0; i < 2; i++) { |
| dport = iter->parent_dport; |
| iter = dport->port; |
| dev = device_find_child(&iter->dev, NULL, first_decoder); |
| /* |
| * Ancestor ports are guaranteed to be enumerated before |
| * @port, and all ports have at least one decoder. |
| */ |
| if (WARN_ON(!dev)) |
| continue; |
| cxlsd = to_cxl_switch_decoder(dev); |
| if (i == 0) { |
| /* put cxl_mem.4 second in the decode order */ |
| if (pdev->id == 4) |
| cxlsd->target[1] = dport; |
| else |
| cxlsd->target[0] = dport; |
| } else |
| cxlsd->target[0] = dport; |
| cxld = &cxlsd->cxld; |
| cxld->target_type = CXL_DECODER_HOSTONLYMEM; |
| cxld->flags = CXL_DECODER_F_ENABLE; |
| iter->commit_end = 0; |
| /* |
| * Switch targets 2 endpoints, while host bridge targets |
| * one root port |
| */ |
| if (i == 0) |
| cxld->interleave_ways = 2; |
| else |
| cxld->interleave_ways = 1; |
| cxld->interleave_granularity = 4096; |
| cxld->hpa_range = (struct range) { |
| .start = base, |
| .end = base + size - 1, |
| }; |
| put_device(dev); |
| } |
| } |
| |
| static int mock_cxl_enumerate_decoders(struct cxl_hdm *cxlhdm, |
| struct cxl_endpoint_dvsec_info *info) |
| { |
| struct cxl_port *port = cxlhdm->port; |
| struct cxl_port *parent_port = to_cxl_port(port->dev.parent); |
| int target_count, i; |
| |
| if (is_cxl_endpoint(port)) |
| target_count = 0; |
| else if (is_cxl_root(parent_port)) |
| target_count = NR_CXL_ROOT_PORTS; |
| else |
| target_count = NR_CXL_SWITCH_PORTS; |
| |
| for (i = 0; i < NR_CXL_PORT_DECODERS; i++) { |
| int target_map[CXL_DECODER_MAX_INTERLEAVE] = { 0 }; |
| struct target_map_ctx ctx = { |
| .target_map = target_map, |
| .target_count = target_count, |
| }; |
| struct cxl_decoder *cxld; |
| int rc; |
| |
| if (target_count) { |
| struct cxl_switch_decoder *cxlsd; |
| |
| cxlsd = cxl_switch_decoder_alloc(port, target_count); |
| if (IS_ERR(cxlsd)) { |
| dev_warn(&port->dev, |
| "Failed to allocate the decoder\n"); |
| return PTR_ERR(cxlsd); |
| } |
| cxld = &cxlsd->cxld; |
| } else { |
| struct cxl_endpoint_decoder *cxled; |
| |
| cxled = cxl_endpoint_decoder_alloc(port); |
| |
| if (IS_ERR(cxled)) { |
| dev_warn(&port->dev, |
| "Failed to allocate the decoder\n"); |
| return PTR_ERR(cxled); |
| } |
| cxld = &cxled->cxld; |
| } |
| |
| mock_init_hdm_decoder(cxld); |
| |
| if (target_count) { |
| rc = device_for_each_child(port->uport_dev, &ctx, |
| map_targets); |
| if (rc) { |
| put_device(&cxld->dev); |
| return rc; |
| } |
| } |
| |
| rc = cxl_decoder_add_locked(cxld, target_map); |
| if (rc) { |
| put_device(&cxld->dev); |
| dev_err(&port->dev, "Failed to add decoder\n"); |
| return rc; |
| } |
| |
| rc = cxl_decoder_autoremove(&port->dev, cxld); |
| if (rc) |
| return rc; |
| dev_dbg(&cxld->dev, "Added to port %s\n", dev_name(&port->dev)); |
| } |
| |
| return 0; |
| } |
| |
| static int mock_cxl_port_enumerate_dports(struct cxl_port *port) |
| { |
| struct platform_device **array; |
| int i, array_size; |
| |
| if (port->depth == 1) { |
| if (is_multi_bridge(port->uport_dev)) { |
| array_size = ARRAY_SIZE(cxl_root_port); |
| array = cxl_root_port; |
| } else if (is_single_bridge(port->uport_dev)) { |
| array_size = ARRAY_SIZE(cxl_root_single); |
| array = cxl_root_single; |
| } else { |
| dev_dbg(&port->dev, "%s: unknown bridge type\n", |
| dev_name(port->uport_dev)); |
| return -ENXIO; |
| } |
| } else if (port->depth == 2) { |
| struct cxl_port *parent = to_cxl_port(port->dev.parent); |
| |
| if (is_multi_bridge(parent->uport_dev)) { |
| array_size = ARRAY_SIZE(cxl_switch_dport); |
| array = cxl_switch_dport; |
| } else if (is_single_bridge(parent->uport_dev)) { |
| array_size = ARRAY_SIZE(cxl_swd_single); |
| array = cxl_swd_single; |
| } else { |
| dev_dbg(&port->dev, "%s: unknown bridge type\n", |
| dev_name(port->uport_dev)); |
| return -ENXIO; |
| } |
| } else { |
| dev_WARN_ONCE(&port->dev, 1, "unexpected depth %d\n", |
| port->depth); |
| return -ENXIO; |
| } |
| |
| for (i = 0; i < array_size; i++) { |
| struct platform_device *pdev = array[i]; |
| struct cxl_dport *dport; |
| |
| if (pdev->dev.parent != port->uport_dev) { |
| dev_dbg(&port->dev, "%s: mismatch parent %s\n", |
| dev_name(port->uport_dev), |
| dev_name(pdev->dev.parent)); |
| continue; |
| } |
| |
| dport = devm_cxl_add_dport(port, &pdev->dev, pdev->id, |
| CXL_RESOURCE_NONE); |
| |
| if (IS_ERR(dport)) |
| return PTR_ERR(dport); |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * Faking the cxl_dpa_perf for the memdev when appropriate. |
| */ |
| static void dpa_perf_setup(struct cxl_port *endpoint, struct range *range, |
| struct cxl_dpa_perf *dpa_perf) |
| { |
| dpa_perf->qos_class = FAKE_QTG_ID; |
| dpa_perf->dpa_range = *range; |
| for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) { |
| dpa_perf->coord[i].read_latency = 500; |
| dpa_perf->coord[i].write_latency = 500; |
| dpa_perf->coord[i].read_bandwidth = 1000; |
| dpa_perf->coord[i].write_bandwidth = 1000; |
| } |
| } |
| |
| static void mock_cxl_endpoint_parse_cdat(struct cxl_port *port) |
| { |
| struct cxl_root *cxl_root __free(put_cxl_root) = |
| find_cxl_root(port); |
| struct cxl_memdev *cxlmd = to_cxl_memdev(port->uport_dev); |
| struct cxl_dev_state *cxlds = cxlmd->cxlds; |
| struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds); |
| struct access_coordinate ep_c[ACCESS_COORDINATE_MAX]; |
| struct range pmem_range = { |
| .start = cxlds->pmem_res.start, |
| .end = cxlds->pmem_res.end, |
| }; |
| struct range ram_range = { |
| .start = cxlds->ram_res.start, |
| .end = cxlds->ram_res.end, |
| }; |
| |
| if (!cxl_root) |
| return; |
| |
| if (range_len(&ram_range)) |
| dpa_perf_setup(port, &ram_range, &mds->ram_perf); |
| |
| if (range_len(&pmem_range)) |
| dpa_perf_setup(port, &pmem_range, &mds->pmem_perf); |
| |
| cxl_memdev_update_perf(cxlmd); |
| |
| /* |
| * This function is here to only test the topology iterator. It serves |
| * no other purpose. |
| */ |
| cxl_endpoint_get_perf_coordinates(port, ep_c); |
| } |
| |
| static struct cxl_mock_ops cxl_mock_ops = { |
| .is_mock_adev = is_mock_adev, |
| .is_mock_bridge = is_mock_bridge, |
| .is_mock_bus = is_mock_bus, |
| .is_mock_port = is_mock_port, |
| .is_mock_dev = is_mock_dev, |
| .acpi_table_parse_cedt = mock_acpi_table_parse_cedt, |
| .acpi_evaluate_integer = mock_acpi_evaluate_integer, |
| .acpi_pci_find_root = mock_acpi_pci_find_root, |
| .devm_cxl_port_enumerate_dports = mock_cxl_port_enumerate_dports, |
| .devm_cxl_setup_hdm = mock_cxl_setup_hdm, |
| .devm_cxl_add_passthrough_decoder = mock_cxl_add_passthrough_decoder, |
| .devm_cxl_enumerate_decoders = mock_cxl_enumerate_decoders, |
| .cxl_endpoint_parse_cdat = mock_cxl_endpoint_parse_cdat, |
| .list = LIST_HEAD_INIT(cxl_mock_ops.list), |
| }; |
| |
| static void mock_companion(struct acpi_device *adev, struct device *dev) |
| { |
| device_initialize(&adev->dev); |
| fwnode_init(&adev->fwnode, NULL); |
| dev->fwnode = &adev->fwnode; |
| adev->fwnode.dev = dev; |
| } |
| |
| #ifndef SZ_64G |
| #define SZ_64G (SZ_32G * 2) |
| #endif |
| |
| static __init int cxl_rch_topo_init(void) |
| { |
| int rc, i; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_rch); i++) { |
| int idx = NR_CXL_HOST_BRIDGES + NR_CXL_SINGLE_HOST + i; |
| struct acpi_device *adev = &host_bridge[idx]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_host_bridge", idx); |
| if (!pdev) |
| goto err_bridge; |
| |
| mock_companion(adev, &pdev->dev); |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_bridge; |
| } |
| |
| cxl_rch[i] = pdev; |
| mock_pci_bus[idx].bridge = &pdev->dev; |
| rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, |
| "firmware_node"); |
| if (rc) |
| goto err_bridge; |
| } |
| |
| return 0; |
| |
| err_bridge: |
| for (i = ARRAY_SIZE(cxl_rch) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_rch[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "firmware_node"); |
| platform_device_unregister(cxl_rch[i]); |
| } |
| |
| return rc; |
| } |
| |
| static void cxl_rch_topo_exit(void) |
| { |
| int i; |
| |
| for (i = ARRAY_SIZE(cxl_rch) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_rch[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "firmware_node"); |
| platform_device_unregister(cxl_rch[i]); |
| } |
| } |
| |
| static __init int cxl_single_topo_init(void) |
| { |
| int i, rc; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) { |
| struct acpi_device *adev = |
| &host_bridge[NR_CXL_HOST_BRIDGES + i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_host_bridge", |
| NR_CXL_HOST_BRIDGES + i); |
| if (!pdev) |
| goto err_bridge; |
| |
| mock_companion(adev, &pdev->dev); |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_bridge; |
| } |
| |
| cxl_hb_single[i] = pdev; |
| mock_pci_bus[i + NR_CXL_HOST_BRIDGES].bridge = &pdev->dev; |
| rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, |
| "physical_node"); |
| if (rc) |
| goto err_bridge; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_root_single); i++) { |
| struct platform_device *bridge = |
| cxl_hb_single[i % ARRAY_SIZE(cxl_hb_single)]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_root_port", |
| NR_MULTI_ROOT + i); |
| if (!pdev) |
| goto err_port; |
| pdev->dev.parent = &bridge->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_port; |
| } |
| cxl_root_single[i] = pdev; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_swu_single); i++) { |
| struct platform_device *root_port = cxl_root_single[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_switch_uport", |
| NR_MULTI_ROOT + i); |
| if (!pdev) |
| goto err_uport; |
| pdev->dev.parent = &root_port->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_uport; |
| } |
| cxl_swu_single[i] = pdev; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_swd_single); i++) { |
| struct platform_device *uport = |
| cxl_swu_single[i % ARRAY_SIZE(cxl_swu_single)]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_switch_dport", |
| i + NR_MEM_MULTI); |
| if (!pdev) |
| goto err_dport; |
| pdev->dev.parent = &uport->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_dport; |
| } |
| cxl_swd_single[i] = pdev; |
| } |
| |
| return 0; |
| |
| err_dport: |
| for (i = ARRAY_SIZE(cxl_swd_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_swd_single[i]); |
| err_uport: |
| for (i = ARRAY_SIZE(cxl_swu_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_swu_single[i]); |
| err_port: |
| for (i = ARRAY_SIZE(cxl_root_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_root_single[i]); |
| err_bridge: |
| for (i = ARRAY_SIZE(cxl_hb_single) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_hb_single[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "physical_node"); |
| platform_device_unregister(cxl_hb_single[i]); |
| } |
| |
| return rc; |
| } |
| |
| static void cxl_single_topo_exit(void) |
| { |
| int i; |
| |
| for (i = ARRAY_SIZE(cxl_swd_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_swd_single[i]); |
| for (i = ARRAY_SIZE(cxl_swu_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_swu_single[i]); |
| for (i = ARRAY_SIZE(cxl_root_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_root_single[i]); |
| for (i = ARRAY_SIZE(cxl_hb_single) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_hb_single[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "physical_node"); |
| platform_device_unregister(cxl_hb_single[i]); |
| } |
| } |
| |
| static void cxl_mem_exit(void) |
| { |
| int i; |
| |
| for (i = ARRAY_SIZE(cxl_rcd) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_rcd[i]); |
| for (i = ARRAY_SIZE(cxl_mem_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_mem_single[i]); |
| for (i = ARRAY_SIZE(cxl_mem) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_mem[i]); |
| } |
| |
| static int cxl_mem_init(void) |
| { |
| int i, rc; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_mem); i++) { |
| struct platform_device *dport = cxl_switch_dport[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_mem", i); |
| if (!pdev) |
| goto err_mem; |
| pdev->dev.parent = &dport->dev; |
| set_dev_node(&pdev->dev, i % 2); |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_mem; |
| } |
| cxl_mem[i] = pdev; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_mem_single); i++) { |
| struct platform_device *dport = cxl_swd_single[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_mem", NR_MEM_MULTI + i); |
| if (!pdev) |
| goto err_single; |
| pdev->dev.parent = &dport->dev; |
| set_dev_node(&pdev->dev, i % 2); |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_single; |
| } |
| cxl_mem_single[i] = pdev; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_rcd); i++) { |
| int idx = NR_MEM_MULTI + NR_MEM_SINGLE + i; |
| struct platform_device *rch = cxl_rch[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_rcd", idx); |
| if (!pdev) |
| goto err_rcd; |
| pdev->dev.parent = &rch->dev; |
| set_dev_node(&pdev->dev, i % 2); |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_rcd; |
| } |
| cxl_rcd[i] = pdev; |
| } |
| |
| return 0; |
| |
| err_rcd: |
| for (i = ARRAY_SIZE(cxl_rcd) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_rcd[i]); |
| err_single: |
| for (i = ARRAY_SIZE(cxl_mem_single) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_mem_single[i]); |
| err_mem: |
| for (i = ARRAY_SIZE(cxl_mem) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_mem[i]); |
| return rc; |
| } |
| |
| static __init int cxl_test_init(void) |
| { |
| int rc, i; |
| |
| cxl_acpi_test(); |
| cxl_core_test(); |
| cxl_mem_test(); |
| cxl_pmem_test(); |
| cxl_port_test(); |
| |
| register_cxl_mock_ops(&cxl_mock_ops); |
| |
| cxl_mock_pool = gen_pool_create(ilog2(SZ_2M), NUMA_NO_NODE); |
| if (!cxl_mock_pool) { |
| rc = -ENOMEM; |
| goto err_gen_pool_create; |
| } |
| |
| rc = gen_pool_add(cxl_mock_pool, iomem_resource.end + 1 - SZ_64G, |
| SZ_64G, NUMA_NO_NODE); |
| if (rc) |
| goto err_gen_pool_add; |
| |
| if (interleave_arithmetic == 1) { |
| cfmws_start = CFMWS_XOR_ARRAY_START; |
| cfmws_end = CFMWS_XOR_ARRAY_END; |
| } else { |
| cfmws_start = CFMWS_MOD_ARRAY_START; |
| cfmws_end = CFMWS_MOD_ARRAY_END; |
| } |
| |
| rc = populate_cedt(); |
| if (rc) |
| goto err_populate; |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) { |
| struct acpi_device *adev = &host_bridge[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_host_bridge", i); |
| if (!pdev) |
| goto err_bridge; |
| |
| mock_companion(adev, &pdev->dev); |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_bridge; |
| } |
| |
| cxl_host_bridge[i] = pdev; |
| mock_pci_bus[i].bridge = &pdev->dev; |
| rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, |
| "physical_node"); |
| if (rc) |
| goto err_bridge; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_root_port); i++) { |
| struct platform_device *bridge = |
| cxl_host_bridge[i % ARRAY_SIZE(cxl_host_bridge)]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_root_port", i); |
| if (!pdev) |
| goto err_port; |
| pdev->dev.parent = &bridge->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_port; |
| } |
| cxl_root_port[i] = pdev; |
| } |
| |
| BUILD_BUG_ON(ARRAY_SIZE(cxl_switch_uport) != ARRAY_SIZE(cxl_root_port)); |
| for (i = 0; i < ARRAY_SIZE(cxl_switch_uport); i++) { |
| struct platform_device *root_port = cxl_root_port[i]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_switch_uport", i); |
| if (!pdev) |
| goto err_uport; |
| pdev->dev.parent = &root_port->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_uport; |
| } |
| cxl_switch_uport[i] = pdev; |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(cxl_switch_dport); i++) { |
| struct platform_device *uport = |
| cxl_switch_uport[i % ARRAY_SIZE(cxl_switch_uport)]; |
| struct platform_device *pdev; |
| |
| pdev = platform_device_alloc("cxl_switch_dport", i); |
| if (!pdev) |
| goto err_dport; |
| pdev->dev.parent = &uport->dev; |
| |
| rc = platform_device_add(pdev); |
| if (rc) { |
| platform_device_put(pdev); |
| goto err_dport; |
| } |
| cxl_switch_dport[i] = pdev; |
| } |
| |
| rc = cxl_single_topo_init(); |
| if (rc) |
| goto err_dport; |
| |
| rc = cxl_rch_topo_init(); |
| if (rc) |
| goto err_single; |
| |
| cxl_acpi = platform_device_alloc("cxl_acpi", 0); |
| if (!cxl_acpi) |
| goto err_rch; |
| |
| mock_companion(&acpi0017_mock, &cxl_acpi->dev); |
| acpi0017_mock.dev.bus = &platform_bus_type; |
| |
| rc = platform_device_add(cxl_acpi); |
| if (rc) |
| goto err_root; |
| |
| rc = cxl_mem_init(); |
| if (rc) |
| goto err_root; |
| |
| return 0; |
| |
| err_root: |
| platform_device_put(cxl_acpi); |
| err_rch: |
| cxl_rch_topo_exit(); |
| err_single: |
| cxl_single_topo_exit(); |
| err_dport: |
| for (i = ARRAY_SIZE(cxl_switch_dport) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_switch_dport[i]); |
| err_uport: |
| for (i = ARRAY_SIZE(cxl_switch_uport) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_switch_uport[i]); |
| err_port: |
| for (i = ARRAY_SIZE(cxl_root_port) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_root_port[i]); |
| err_bridge: |
| for (i = ARRAY_SIZE(cxl_host_bridge) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_host_bridge[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "physical_node"); |
| platform_device_unregister(cxl_host_bridge[i]); |
| } |
| err_populate: |
| depopulate_all_mock_resources(); |
| err_gen_pool_add: |
| gen_pool_destroy(cxl_mock_pool); |
| err_gen_pool_create: |
| unregister_cxl_mock_ops(&cxl_mock_ops); |
| return rc; |
| } |
| |
| static __exit void cxl_test_exit(void) |
| { |
| int i; |
| |
| cxl_mem_exit(); |
| platform_device_unregister(cxl_acpi); |
| cxl_rch_topo_exit(); |
| cxl_single_topo_exit(); |
| for (i = ARRAY_SIZE(cxl_switch_dport) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_switch_dport[i]); |
| for (i = ARRAY_SIZE(cxl_switch_uport) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_switch_uport[i]); |
| for (i = ARRAY_SIZE(cxl_root_port) - 1; i >= 0; i--) |
| platform_device_unregister(cxl_root_port[i]); |
| for (i = ARRAY_SIZE(cxl_host_bridge) - 1; i >= 0; i--) { |
| struct platform_device *pdev = cxl_host_bridge[i]; |
| |
| if (!pdev) |
| continue; |
| sysfs_remove_link(&pdev->dev.kobj, "physical_node"); |
| platform_device_unregister(cxl_host_bridge[i]); |
| } |
| depopulate_all_mock_resources(); |
| gen_pool_destroy(cxl_mock_pool); |
| unregister_cxl_mock_ops(&cxl_mock_ops); |
| } |
| |
| module_param(interleave_arithmetic, int, 0444); |
| MODULE_PARM_DESC(interleave_arithmetic, "Modulo:0, XOR:1"); |
| module_init(cxl_test_init); |
| module_exit(cxl_test_exit); |
| MODULE_LICENSE("GPL v2"); |
| MODULE_IMPORT_NS(ACPI); |
| MODULE_IMPORT_NS(CXL); |