]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/blob - drivers/cxl/acpi.c
Merge tag 'asoc-fix-v5.15-rc2' of https://git.kernel.org/pub/scm/linux/kernel/git...
[mirror_ubuntu-jammy-kernel.git] / drivers / cxl / acpi.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright(c) 2021 Intel Corporation. All rights reserved. */
3 #include <linux/platform_device.h>
4 #include <linux/module.h>
5 #include <linux/device.h>
6 #include <linux/kernel.h>
7 #include <linux/acpi.h>
8 #include <linux/pci.h>
9 #include "cxl.h"
10
11 static struct acpi_table_header *acpi_cedt;
12
13 /* Encode defined in CXL 2.0 8.2.5.12.7 HDM Decoder Control Register */
14 #define CFMWS_INTERLEAVE_WAYS(x) (1 << (x)->interleave_ways)
15 #define CFMWS_INTERLEAVE_GRANULARITY(x) ((x)->granularity + 8)
16
17 static unsigned long cfmws_to_decoder_flags(int restrictions)
18 {
19 unsigned long flags = 0;
20
21 if (restrictions & ACPI_CEDT_CFMWS_RESTRICT_TYPE2)
22 flags |= CXL_DECODER_F_TYPE2;
23 if (restrictions & ACPI_CEDT_CFMWS_RESTRICT_TYPE3)
24 flags |= CXL_DECODER_F_TYPE3;
25 if (restrictions & ACPI_CEDT_CFMWS_RESTRICT_VOLATILE)
26 flags |= CXL_DECODER_F_RAM;
27 if (restrictions & ACPI_CEDT_CFMWS_RESTRICT_PMEM)
28 flags |= CXL_DECODER_F_PMEM;
29 if (restrictions & ACPI_CEDT_CFMWS_RESTRICT_FIXED)
30 flags |= CXL_DECODER_F_LOCK;
31
32 return flags;
33 }
34
35 static int cxl_acpi_cfmws_verify(struct device *dev,
36 struct acpi_cedt_cfmws *cfmws)
37 {
38 int expected_len;
39
40 if (cfmws->interleave_arithmetic != ACPI_CEDT_CFMWS_ARITHMETIC_MODULO) {
41 dev_err(dev, "CFMWS Unsupported Interleave Arithmetic\n");
42 return -EINVAL;
43 }
44
45 if (!IS_ALIGNED(cfmws->base_hpa, SZ_256M)) {
46 dev_err(dev, "CFMWS Base HPA not 256MB aligned\n");
47 return -EINVAL;
48 }
49
50 if (!IS_ALIGNED(cfmws->window_size, SZ_256M)) {
51 dev_err(dev, "CFMWS Window Size not 256MB aligned\n");
52 return -EINVAL;
53 }
54
55 expected_len = struct_size((cfmws), interleave_targets,
56 CFMWS_INTERLEAVE_WAYS(cfmws));
57
58 if (cfmws->header.length < expected_len) {
59 dev_err(dev, "CFMWS length %d less than expected %d\n",
60 cfmws->header.length, expected_len);
61 return -EINVAL;
62 }
63
64 if (cfmws->header.length > expected_len)
65 dev_dbg(dev, "CFMWS length %d greater than expected %d\n",
66 cfmws->header.length, expected_len);
67
68 return 0;
69 }
70
71 static void cxl_add_cfmws_decoders(struct device *dev,
72 struct cxl_port *root_port)
73 {
74 struct acpi_cedt_cfmws *cfmws;
75 struct cxl_decoder *cxld;
76 acpi_size len, cur = 0;
77 void *cedt_subtable;
78 unsigned long flags;
79 int rc;
80
81 len = acpi_cedt->length - sizeof(*acpi_cedt);
82 cedt_subtable = acpi_cedt + 1;
83
84 while (cur < len) {
85 struct acpi_cedt_header *c = cedt_subtable + cur;
86
87 if (c->type != ACPI_CEDT_TYPE_CFMWS) {
88 cur += c->length;
89 continue;
90 }
91
92 cfmws = cedt_subtable + cur;
93
94 if (cfmws->header.length < sizeof(*cfmws)) {
95 dev_warn_once(dev,
96 "CFMWS entry skipped:invalid length:%u\n",
97 cfmws->header.length);
98 cur += c->length;
99 continue;
100 }
101
102 rc = cxl_acpi_cfmws_verify(dev, cfmws);
103 if (rc) {
104 dev_err(dev, "CFMWS range %#llx-%#llx not registered\n",
105 cfmws->base_hpa, cfmws->base_hpa +
106 cfmws->window_size - 1);
107 cur += c->length;
108 continue;
109 }
110
111 flags = cfmws_to_decoder_flags(cfmws->restrictions);
112 cxld = devm_cxl_add_decoder(dev, root_port,
113 CFMWS_INTERLEAVE_WAYS(cfmws),
114 cfmws->base_hpa, cfmws->window_size,
115 CFMWS_INTERLEAVE_WAYS(cfmws),
116 CFMWS_INTERLEAVE_GRANULARITY(cfmws),
117 CXL_DECODER_EXPANDER,
118 flags);
119
120 if (IS_ERR(cxld)) {
121 dev_err(dev, "Failed to add decoder for %#llx-%#llx\n",
122 cfmws->base_hpa, cfmws->base_hpa +
123 cfmws->window_size - 1);
124 } else {
125 dev_dbg(dev, "add: %s range %#llx-%#llx\n",
126 dev_name(&cxld->dev), cfmws->base_hpa,
127 cfmws->base_hpa + cfmws->window_size - 1);
128 }
129 cur += c->length;
130 }
131 }
132
133 static struct acpi_cedt_chbs *cxl_acpi_match_chbs(struct device *dev, u32 uid)
134 {
135 struct acpi_cedt_chbs *chbs, *chbs_match = NULL;
136 acpi_size len, cur = 0;
137 void *cedt_subtable;
138
139 len = acpi_cedt->length - sizeof(*acpi_cedt);
140 cedt_subtable = acpi_cedt + 1;
141
142 while (cur < len) {
143 struct acpi_cedt_header *c = cedt_subtable + cur;
144
145 if (c->type != ACPI_CEDT_TYPE_CHBS) {
146 cur += c->length;
147 continue;
148 }
149
150 chbs = cedt_subtable + cur;
151
152 if (chbs->header.length < sizeof(*chbs)) {
153 dev_warn_once(dev,
154 "CHBS entry skipped: invalid length:%u\n",
155 chbs->header.length);
156 cur += c->length;
157 continue;
158 }
159
160 if (chbs->uid != uid) {
161 cur += c->length;
162 continue;
163 }
164
165 if (chbs_match) {
166 dev_warn_once(dev,
167 "CHBS entry skipped: duplicate UID:%u\n",
168 uid);
169 cur += c->length;
170 continue;
171 }
172
173 chbs_match = chbs;
174 cur += c->length;
175 }
176
177 return chbs_match ? chbs_match : ERR_PTR(-ENODEV);
178 }
179
180 static resource_size_t get_chbcr(struct acpi_cedt_chbs *chbs)
181 {
182 return IS_ERR(chbs) ? CXL_RESOURCE_NONE : chbs->base;
183 }
184
185 struct cxl_walk_context {
186 struct device *dev;
187 struct pci_bus *root;
188 struct cxl_port *port;
189 int error;
190 int count;
191 };
192
193 static int match_add_root_ports(struct pci_dev *pdev, void *data)
194 {
195 struct cxl_walk_context *ctx = data;
196 struct pci_bus *root_bus = ctx->root;
197 struct cxl_port *port = ctx->port;
198 int type = pci_pcie_type(pdev);
199 struct device *dev = ctx->dev;
200 u32 lnkcap, port_num;
201 int rc;
202
203 if (pdev->bus != root_bus)
204 return 0;
205 if (!pci_is_pcie(pdev))
206 return 0;
207 if (type != PCI_EXP_TYPE_ROOT_PORT)
208 return 0;
209 if (pci_read_config_dword(pdev, pci_pcie_cap(pdev) + PCI_EXP_LNKCAP,
210 &lnkcap) != PCIBIOS_SUCCESSFUL)
211 return 0;
212
213 /* TODO walk DVSEC to find component register base */
214 port_num = FIELD_GET(PCI_EXP_LNKCAP_PN, lnkcap);
215 rc = cxl_add_dport(port, &pdev->dev, port_num, CXL_RESOURCE_NONE);
216 if (rc) {
217 ctx->error = rc;
218 return rc;
219 }
220 ctx->count++;
221
222 dev_dbg(dev, "add dport%d: %s\n", port_num, dev_name(&pdev->dev));
223
224 return 0;
225 }
226
227 static struct cxl_dport *find_dport_by_dev(struct cxl_port *port, struct device *dev)
228 {
229 struct cxl_dport *dport;
230
231 device_lock(&port->dev);
232 list_for_each_entry(dport, &port->dports, list)
233 if (dport->dport == dev) {
234 device_unlock(&port->dev);
235 return dport;
236 }
237
238 device_unlock(&port->dev);
239 return NULL;
240 }
241
242 static struct acpi_device *to_cxl_host_bridge(struct device *dev)
243 {
244 struct acpi_device *adev = to_acpi_device(dev);
245
246 if (!acpi_pci_find_root(adev->handle))
247 return NULL;
248
249 if (strcmp(acpi_device_hid(adev), "ACPI0016") == 0)
250 return adev;
251 return NULL;
252 }
253
254 /*
255 * A host bridge is a dport to a CFMWS decode and it is a uport to the
256 * dport (PCIe Root Ports) in the host bridge.
257 */
258 static int add_host_bridge_uport(struct device *match, void *arg)
259 {
260 struct acpi_device *bridge = to_cxl_host_bridge(match);
261 struct cxl_port *root_port = arg;
262 struct device *host = root_port->dev.parent;
263 struct acpi_pci_root *pci_root;
264 struct cxl_walk_context ctx;
265 struct cxl_decoder *cxld;
266 struct cxl_dport *dport;
267 struct cxl_port *port;
268
269 if (!bridge)
270 return 0;
271
272 dport = find_dport_by_dev(root_port, match);
273 if (!dport) {
274 dev_dbg(host, "host bridge expected and not found\n");
275 return -ENODEV;
276 }
277
278 port = devm_cxl_add_port(host, match, dport->component_reg_phys,
279 root_port);
280 if (IS_ERR(port))
281 return PTR_ERR(port);
282 dev_dbg(host, "%s: add: %s\n", dev_name(match), dev_name(&port->dev));
283
284 /*
285 * Note that this lookup already succeeded in
286 * to_cxl_host_bridge(), so no need to check for failure here
287 */
288 pci_root = acpi_pci_find_root(bridge->handle);
289 ctx = (struct cxl_walk_context){
290 .dev = host,
291 .root = pci_root->bus,
292 .port = port,
293 };
294 pci_walk_bus(pci_root->bus, match_add_root_ports, &ctx);
295
296 if (ctx.count == 0)
297 return -ENODEV;
298 if (ctx.error)
299 return ctx.error;
300
301 /* TODO: Scan CHBCR for HDM Decoder resources */
302
303 /*
304 * In the single-port host-bridge case there are no HDM decoders
305 * in the CHBCR and a 1:1 passthrough decode is implied.
306 */
307 if (ctx.count == 1) {
308 cxld = devm_cxl_add_passthrough_decoder(host, port);
309 if (IS_ERR(cxld))
310 return PTR_ERR(cxld);
311
312 dev_dbg(host, "add: %s\n", dev_name(&cxld->dev));
313 }
314
315 return 0;
316 }
317
318 static int add_host_bridge_dport(struct device *match, void *arg)
319 {
320 int rc;
321 acpi_status status;
322 unsigned long long uid;
323 struct acpi_cedt_chbs *chbs;
324 struct cxl_port *root_port = arg;
325 struct device *host = root_port->dev.parent;
326 struct acpi_device *bridge = to_cxl_host_bridge(match);
327
328 if (!bridge)
329 return 0;
330
331 status = acpi_evaluate_integer(bridge->handle, METHOD_NAME__UID, NULL,
332 &uid);
333 if (status != AE_OK) {
334 dev_err(host, "unable to retrieve _UID of %s\n",
335 dev_name(match));
336 return -ENODEV;
337 }
338
339 chbs = cxl_acpi_match_chbs(host, uid);
340 if (IS_ERR(chbs))
341 dev_dbg(host, "No CHBS found for Host Bridge: %s\n",
342 dev_name(match));
343
344 rc = cxl_add_dport(root_port, match, uid, get_chbcr(chbs));
345 if (rc) {
346 dev_err(host, "failed to add downstream port: %s\n",
347 dev_name(match));
348 return rc;
349 }
350 dev_dbg(host, "add dport%llu: %s\n", uid, dev_name(match));
351 return 0;
352 }
353
354 static int add_root_nvdimm_bridge(struct device *match, void *data)
355 {
356 struct cxl_decoder *cxld;
357 struct cxl_port *root_port = data;
358 struct cxl_nvdimm_bridge *cxl_nvb;
359 struct device *host = root_port->dev.parent;
360
361 if (!is_root_decoder(match))
362 return 0;
363
364 cxld = to_cxl_decoder(match);
365 if (!(cxld->flags & CXL_DECODER_F_PMEM))
366 return 0;
367
368 cxl_nvb = devm_cxl_add_nvdimm_bridge(host, root_port);
369 if (IS_ERR(cxl_nvb)) {
370 dev_dbg(host, "failed to register pmem\n");
371 return PTR_ERR(cxl_nvb);
372 }
373 dev_dbg(host, "%s: add: %s\n", dev_name(&root_port->dev),
374 dev_name(&cxl_nvb->dev));
375 return 1;
376 }
377
378 static int cxl_acpi_probe(struct platform_device *pdev)
379 {
380 int rc;
381 acpi_status status;
382 struct cxl_port *root_port;
383 struct device *host = &pdev->dev;
384 struct acpi_device *adev = ACPI_COMPANION(host);
385
386 root_port = devm_cxl_add_port(host, host, CXL_RESOURCE_NONE, NULL);
387 if (IS_ERR(root_port))
388 return PTR_ERR(root_port);
389 dev_dbg(host, "add: %s\n", dev_name(&root_port->dev));
390
391 status = acpi_get_table(ACPI_SIG_CEDT, 0, &acpi_cedt);
392 if (ACPI_FAILURE(status))
393 return -ENXIO;
394
395 rc = bus_for_each_dev(adev->dev.bus, NULL, root_port,
396 add_host_bridge_dport);
397 if (rc)
398 goto out;
399
400 cxl_add_cfmws_decoders(host, root_port);
401
402 /*
403 * Root level scanned with host-bridge as dports, now scan host-bridges
404 * for their role as CXL uports to their CXL-capable PCIe Root Ports.
405 */
406 rc = bus_for_each_dev(adev->dev.bus, NULL, root_port,
407 add_host_bridge_uport);
408 if (rc)
409 goto out;
410
411 if (IS_ENABLED(CONFIG_CXL_PMEM))
412 rc = device_for_each_child(&root_port->dev, root_port,
413 add_root_nvdimm_bridge);
414
415 out:
416 acpi_put_table(acpi_cedt);
417 if (rc < 0)
418 return rc;
419 return 0;
420 }
421
422 static const struct acpi_device_id cxl_acpi_ids[] = {
423 { "ACPI0017", 0 },
424 { "", 0 },
425 };
426 MODULE_DEVICE_TABLE(acpi, cxl_acpi_ids);
427
428 static struct platform_driver cxl_acpi_driver = {
429 .probe = cxl_acpi_probe,
430 .driver = {
431 .name = KBUILD_MODNAME,
432 .acpi_match_table = cxl_acpi_ids,
433 },
434 };
435
436 module_platform_driver(cxl_acpi_driver);
437 MODULE_LICENSE("GPL v2");
438 MODULE_IMPORT_NS(CXL);