1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright(c) 2020 Intel Corporation. All rights reserved. */
3 #include <linux/io-64-nonatomic-lo-hi.h>
4 #include <linux/security.h>
5 #include <linux/debugfs.h>
6 #include <linux/mutex.h>
7 #include <cxlmem.h>
8 #include <cxl.h>
9
10 #include "core.h"
11
12 static bool cxl_raw_allow_all;
13
14 /**
15 * DOC: cxl mbox
16 *
17 * Core implementation of the CXL 2.0 Type-3 Memory Device Mailbox. The
18 * implementation is used by the cxl_pci driver to initialize the device
19 * and implement the cxl_mem.h IOCTL UAPI. It also implements the
20 * backend of the cxl_pmem_ctl() transport for LIBNVDIMM.
21 */
22
23 #define cxl_for_each_cmd(cmd) \
24 for ((cmd) = &cxl_mem_commands[0]; \
25 ((cmd) - cxl_mem_commands) < ARRAY_SIZE(cxl_mem_commands); (cmd)++)
26
27 #define CXL_CMD(_id, sin, sout, _flags) \
28 [CXL_MEM_COMMAND_ID_##_id] = { \
29 .info = { \
30 .id = CXL_MEM_COMMAND_ID_##_id, \
31 .size_in = sin, \
32 .size_out = sout, \
33 }, \
34 .opcode = CXL_MBOX_OP_##_id, \
35 .flags = _flags, \
36 }
37
38 /*
39 * This table defines the supported mailbox commands for the driver. This table
40 * is made up of a UAPI structure. Non-negative values as parameters in the
41 * table will be validated against the user's input. For example, if size_in is
42 * 0, and the user passed in 1, it is an error.
43 */
44 static struct cxl_mem_command cxl_mem_commands[CXL_MEM_COMMAND_ID_MAX] = {
45 CXL_CMD(IDENTIFY, 0, 0x43, CXL_CMD_FLAG_FORCE_ENABLE),
46 #ifdef CONFIG_CXL_MEM_RAW_COMMANDS
47 CXL_CMD(RAW, ~0, ~0, 0),
48 #endif
49 CXL_CMD(GET_SUPPORTED_LOGS, 0, ~0, CXL_CMD_FLAG_FORCE_ENABLE),
50 CXL_CMD(GET_FW_INFO, 0, 0x50, 0),
51 CXL_CMD(GET_PARTITION_INFO, 0, 0x20, 0),
52 CXL_CMD(GET_LSA, 0x8, ~0, 0),
53 CXL_CMD(GET_HEALTH_INFO, 0, 0x12, 0),
54 CXL_CMD(GET_LOG, 0x18, ~0, CXL_CMD_FLAG_FORCE_ENABLE),
55 CXL_CMD(SET_PARTITION_INFO, 0x0a, 0, 0),
56 CXL_CMD(SET_LSA, ~0, 0, 0),
57 CXL_CMD(GET_ALERT_CONFIG, 0, 0x10, 0),
58 CXL_CMD(SET_ALERT_CONFIG, 0xc, 0, 0),
59 CXL_CMD(GET_SHUTDOWN_STATE, 0, 0x1, 0),
60 CXL_CMD(SET_SHUTDOWN_STATE, 0x1, 0, 0),
61 CXL_CMD(GET_POISON, 0x10, ~0, 0),
62 CXL_CMD(INJECT_POISON, 0x8, 0, 0),
63 CXL_CMD(CLEAR_POISON, 0x48, 0, 0),
64 CXL_CMD(GET_SCAN_MEDIA_CAPS, 0x10, 0x4, 0),
65 CXL_CMD(SCAN_MEDIA, 0x11, 0, 0),
66 CXL_CMD(GET_SCAN_MEDIA, 0, ~0, 0),
67 };
68
69 /*
70 * Commands that RAW doesn't permit. The rationale for each:
71 *
72 * CXL_MBOX_OP_ACTIVATE_FW: Firmware activation requires adjustment /
73 * coordination of transaction timeout values at the root bridge level.
74 *
75 * CXL_MBOX_OP_SET_PARTITION_INFO: The device memory map may change live
76 * and needs to be coordinated with HDM updates.
77 *
78 * CXL_MBOX_OP_SET_LSA: The label storage area may be cached by the
79 * driver and any writes from userspace invalidates those contents.
80 *
81 * CXL_MBOX_OP_SET_SHUTDOWN_STATE: Set shutdown state assumes no writes
82 * to the device after it is marked clean, userspace can not make that
83 * assertion.
84 *
85 * CXL_MBOX_OP_[GET_]SCAN_MEDIA: The kernel provides a native error list that
86 * is kept up to date with patrol notifications and error management.
87 */
88 static u16 cxl_disabled_raw_commands[] = {
89 CXL_MBOX_OP_ACTIVATE_FW,
90 CXL_MBOX_OP_SET_PARTITION_INFO,
91 CXL_MBOX_OP_SET_LSA,
92 CXL_MBOX_OP_SET_SHUTDOWN_STATE,
93 CXL_MBOX_OP_SCAN_MEDIA,
94 CXL_MBOX_OP_GET_SCAN_MEDIA,
95 };
96
97 /*
98 * Command sets that RAW doesn't permit. All opcodes in this set are
99 * disabled because they pass plain text security payloads over the
100 * user/kernel boundary. This functionality is intended to be wrapped
101 * behind the keys ABI which allows for encrypted payloads in the UAPI
102 */
103 static u8 security_command_sets[] = {
104 0x44, /* Sanitize */
105 0x45, /* Persistent Memory Data-at-rest Security */
106 0x46, /* Security Passthrough */
107 };
108
cxl_is_security_command(u16 opcode)109 static bool cxl_is_security_command(u16 opcode)
110 {
111 int i;
112
113 for (i = 0; i < ARRAY_SIZE(security_command_sets); i++)
114 if (security_command_sets[i] == (opcode >> 8))
115 return true;
116 return false;
117 }
118
cxl_mem_find_command(u16 opcode)119 static struct cxl_mem_command *cxl_mem_find_command(u16 opcode)
120 {
121 struct cxl_mem_command *c;
122
123 cxl_for_each_cmd(c)
124 if (c->opcode == opcode)
125 return c;
126
127 return NULL;
128 }
129
130 /**
131 * cxl_mem_mbox_send_cmd() - Send a mailbox command to a memory device.
132 * @cxlm: The CXL memory device to communicate with.
133 * @opcode: Opcode for the mailbox command.
134 * @in: The input payload for the mailbox command.
135 * @in_size: The length of the input payload
136 * @out: Caller allocated buffer for the output.
137 * @out_size: Expected size of output.
138 *
139 * Context: Any context. Will acquire and release mbox_mutex.
140 * Return:
141 * * %>=0 - Number of bytes returned in @out.
142 * * %-E2BIG - Payload is too large for hardware.
143 * * %-EBUSY - Couldn't acquire exclusive mailbox access.
144 * * %-EFAULT - Hardware error occurred.
145 * * %-ENXIO - Command completed, but device reported an error.
146 * * %-EIO - Unexpected output size.
147 *
148 * Mailbox commands may execute successfully yet the device itself reported an
149 * error. While this distinction can be useful for commands from userspace, the
150 * kernel will only be able to use results when both are successful.
151 *
152 * See __cxl_mem_mbox_send_cmd()
153 */
cxl_mem_mbox_send_cmd(struct cxl_mem * cxlm,u16 opcode,void * in,size_t in_size,void * out,size_t out_size)154 int cxl_mem_mbox_send_cmd(struct cxl_mem *cxlm, u16 opcode, void *in,
155 size_t in_size, void *out, size_t out_size)
156 {
157 const struct cxl_mem_command *cmd = cxl_mem_find_command(opcode);
158 struct cxl_mbox_cmd mbox_cmd = {
159 .opcode = opcode,
160 .payload_in = in,
161 .size_in = in_size,
162 .size_out = out_size,
163 .payload_out = out,
164 };
165 int rc;
166
167 if (out_size > cxlm->payload_size)
168 return -E2BIG;
169
170 rc = cxlm->mbox_send(cxlm, &mbox_cmd);
171 if (rc)
172 return rc;
173
174 /* TODO: Map return code to proper kernel style errno */
175 if (mbox_cmd.return_code != CXL_MBOX_SUCCESS)
176 return -ENXIO;
177
178 /*
179 * Variable sized commands can't be validated and so it's up to the
180 * caller to do that if they wish.
181 */
182 if (cmd->info.size_out >= 0 && mbox_cmd.size_out != out_size)
183 return -EIO;
184
185 return 0;
186 }
187 EXPORT_SYMBOL_GPL(cxl_mem_mbox_send_cmd);
188
cxl_mem_raw_command_allowed(u16 opcode)189 static bool cxl_mem_raw_command_allowed(u16 opcode)
190 {
191 int i;
192
193 if (!IS_ENABLED(CONFIG_CXL_MEM_RAW_COMMANDS))
194 return false;
195
196 if (security_locked_down(LOCKDOWN_PCI_ACCESS))
197 return false;
198
199 if (cxl_raw_allow_all)
200 return true;
201
202 if (cxl_is_security_command(opcode))
203 return false;
204
205 for (i = 0; i < ARRAY_SIZE(cxl_disabled_raw_commands); i++)
206 if (cxl_disabled_raw_commands[i] == opcode)
207 return false;
208
209 return true;
210 }
211
212 /**
213 * cxl_validate_cmd_from_user() - Check fields for CXL_MEM_SEND_COMMAND.
214 * @cxlm: &struct cxl_mem device whose mailbox will be used.
215 * @send_cmd: &struct cxl_send_command copied in from userspace.
216 * @out_cmd: Sanitized and populated &struct cxl_mem_command.
217 *
218 * Return:
219 * * %0 - @out_cmd is ready to send.
220 * * %-ENOTTY - Invalid command specified.
221 * * %-EINVAL - Reserved fields or invalid values were used.
222 * * %-ENOMEM - Input or output buffer wasn't sized properly.
223 * * %-EPERM - Attempted to use a protected command.
224 * * %-EBUSY - Kernel has claimed exclusive access to this opcode
225 *
226 * The result of this command is a fully validated command in @out_cmd that is
227 * safe to send to the hardware.
228 *
229 * See handle_mailbox_cmd_from_user()
230 */
cxl_validate_cmd_from_user(struct cxl_mem * cxlm,const struct cxl_send_command * send_cmd,struct cxl_mem_command * out_cmd)231 static int cxl_validate_cmd_from_user(struct cxl_mem *cxlm,
232 const struct cxl_send_command *send_cmd,
233 struct cxl_mem_command *out_cmd)
234 {
235 const struct cxl_command_info *info;
236 struct cxl_mem_command *c;
237
238 if (send_cmd->id == 0 || send_cmd->id >= CXL_MEM_COMMAND_ID_MAX)
239 return -ENOTTY;
240
241 /*
242 * The user can never specify an input payload larger than what hardware
243 * supports, but output can be arbitrarily large (simply write out as
244 * much data as the hardware provides).
245 */
246 if (send_cmd->in.size > cxlm->payload_size)
247 return -EINVAL;
248
249 /*
250 * Checks are bypassed for raw commands but a WARN/taint will occur
251 * later in the callchain
252 */
253 if (send_cmd->id == CXL_MEM_COMMAND_ID_RAW) {
254 const struct cxl_mem_command temp = {
255 .info = {
256 .id = CXL_MEM_COMMAND_ID_RAW,
257 .flags = 0,
258 .size_in = send_cmd->in.size,
259 .size_out = send_cmd->out.size,
260 },
261 .opcode = send_cmd->raw.opcode
262 };
263
264 if (send_cmd->raw.rsvd)
265 return -EINVAL;
266
267 /*
268 * Unlike supported commands, the output size of RAW commands
269 * gets passed along without further checking, so it must be
270 * validated here.
271 */
272 if (send_cmd->out.size > cxlm->payload_size)
273 return -EINVAL;
274
275 if (!cxl_mem_raw_command_allowed(send_cmd->raw.opcode))
276 return -EPERM;
277
278 memcpy(out_cmd, &temp, sizeof(temp));
279
280 return 0;
281 }
282
283 if (send_cmd->flags & ~CXL_MEM_COMMAND_FLAG_MASK)
284 return -EINVAL;
285
286 if (send_cmd->rsvd)
287 return -EINVAL;
288
289 if (send_cmd->in.rsvd || send_cmd->out.rsvd)
290 return -EINVAL;
291
292 /* Convert user's command into the internal representation */
293 c = &cxl_mem_commands[send_cmd->id];
294 info = &c->info;
295
296 /* Check that the command is enabled for hardware */
297 if (!test_bit(info->id, cxlm->enabled_cmds))
298 return -ENOTTY;
299
300 /* Check that the command is not claimed for exclusive kernel use */
301 if (test_bit(info->id, cxlm->exclusive_cmds))
302 return -EBUSY;
303
304 /* Check the input buffer is the expected size */
305 if (info->size_in >= 0 && info->size_in != send_cmd->in.size)
306 return -ENOMEM;
307
308 /* Check the output buffer is at least large enough */
309 if (info->size_out >= 0 && send_cmd->out.size < info->size_out)
310 return -ENOMEM;
311
312 memcpy(out_cmd, c, sizeof(*c));
313 out_cmd->info.size_in = send_cmd->in.size;
314 /*
315 * XXX: out_cmd->info.size_out will be controlled by the driver, and the
316 * specified number of bytes @send_cmd->out.size will be copied back out
317 * to userspace.
318 */
319
320 return 0;
321 }
322
cxl_query_cmd(struct cxl_memdev * cxlmd,struct cxl_mem_query_commands __user * q)323 int cxl_query_cmd(struct cxl_memdev *cxlmd,
324 struct cxl_mem_query_commands __user *q)
325 {
326 struct device *dev = &cxlmd->dev;
327 struct cxl_mem_command *cmd;
328 u32 n_commands;
329 int j = 0;
330
331 dev_dbg(dev, "Query IOCTL\n");
332
333 if (get_user(n_commands, &q->n_commands))
334 return -EFAULT;
335
336 /* returns the total number if 0 elements are requested. */
337 if (n_commands == 0)
338 return put_user(ARRAY_SIZE(cxl_mem_commands), &q->n_commands);
339
340 /*
341 * otherwise, return max(n_commands, total commands) cxl_command_info
342 * structures.
343 */
344 cxl_for_each_cmd(cmd) {
345 const struct cxl_command_info *info = &cmd->info;
346
347 if (copy_to_user(&q->commands[j++], info, sizeof(*info)))
348 return -EFAULT;
349
350 if (j == n_commands)
351 break;
352 }
353
354 return 0;
355 }
356
357 /**
358 * handle_mailbox_cmd_from_user() - Dispatch a mailbox command for userspace.
359 * @cxlm: The CXL memory device to communicate with.
360 * @cmd: The validated command.
361 * @in_payload: Pointer to userspace's input payload.
362 * @out_payload: Pointer to userspace's output payload.
363 * @size_out: (Input) Max payload size to copy out.
364 * (Output) Payload size hardware generated.
365 * @retval: Hardware generated return code from the operation.
366 *
367 * Return:
368 * * %0 - Mailbox transaction succeeded. This implies the mailbox
369 * protocol completed successfully not that the operation itself
370 * was successful.
371 * * %-ENOMEM - Couldn't allocate a bounce buffer.
372 * * %-EFAULT - Something happened with copy_to/from_user.
373 * * %-EINTR - Mailbox acquisition interrupted.
374 * * %-EXXX - Transaction level failures.
375 *
376 * Creates the appropriate mailbox command and dispatches it on behalf of a
377 * userspace request. The input and output payloads are copied between
378 * userspace.
379 *
380 * See cxl_send_cmd().
381 */
handle_mailbox_cmd_from_user(struct cxl_mem * cxlm,const struct cxl_mem_command * cmd,u64 in_payload,u64 out_payload,s32 * size_out,u32 * retval)382 static int handle_mailbox_cmd_from_user(struct cxl_mem *cxlm,
383 const struct cxl_mem_command *cmd,
384 u64 in_payload, u64 out_payload,
385 s32 *size_out, u32 *retval)
386 {
387 struct device *dev = cxlm->dev;
388 struct cxl_mbox_cmd mbox_cmd = {
389 .opcode = cmd->opcode,
390 .size_in = cmd->info.size_in,
391 .size_out = cmd->info.size_out,
392 };
393 int rc;
394
395 if (cmd->info.size_out) {
396 mbox_cmd.payload_out = kvzalloc(cmd->info.size_out, GFP_KERNEL);
397 if (!mbox_cmd.payload_out)
398 return -ENOMEM;
399 }
400
401 if (cmd->info.size_in) {
402 mbox_cmd.payload_in = vmemdup_user(u64_to_user_ptr(in_payload),
403 cmd->info.size_in);
404 if (IS_ERR(mbox_cmd.payload_in)) {
405 kvfree(mbox_cmd.payload_out);
406 return PTR_ERR(mbox_cmd.payload_in);
407 }
408 }
409
410 dev_dbg(dev,
411 "Submitting %s command for user\n"
412 "\topcode: %x\n"
413 "\tsize: %ub\n",
414 cxl_command_names[cmd->info.id].name, mbox_cmd.opcode,
415 cmd->info.size_in);
416
417 dev_WARN_ONCE(dev, cmd->info.id == CXL_MEM_COMMAND_ID_RAW,
418 "raw command path used\n");
419
420 rc = cxlm->mbox_send(cxlm, &mbox_cmd);
421 if (rc)
422 goto out;
423
424 /*
425 * @size_out contains the max size that's allowed to be written back out
426 * to userspace. While the payload may have written more output than
427 * this it will have to be ignored.
428 */
429 if (mbox_cmd.size_out) {
430 dev_WARN_ONCE(dev, mbox_cmd.size_out > *size_out,
431 "Invalid return size\n");
432 if (copy_to_user(u64_to_user_ptr(out_payload),
433 mbox_cmd.payload_out, mbox_cmd.size_out)) {
434 rc = -EFAULT;
435 goto out;
436 }
437 }
438
439 *size_out = mbox_cmd.size_out;
440 *retval = mbox_cmd.return_code;
441
442 out:
443 kvfree(mbox_cmd.payload_in);
444 kvfree(mbox_cmd.payload_out);
445 return rc;
446 }
447
cxl_send_cmd(struct cxl_memdev * cxlmd,struct cxl_send_command __user * s)448 int cxl_send_cmd(struct cxl_memdev *cxlmd, struct cxl_send_command __user *s)
449 {
450 struct cxl_mem *cxlm = cxlmd->cxlm;
451 struct device *dev = &cxlmd->dev;
452 struct cxl_send_command send;
453 struct cxl_mem_command c;
454 int rc;
455
456 dev_dbg(dev, "Send IOCTL\n");
457
458 if (copy_from_user(&send, s, sizeof(send)))
459 return -EFAULT;
460
461 rc = cxl_validate_cmd_from_user(cxlmd->cxlm, &send, &c);
462 if (rc)
463 return rc;
464
465 /* Prepare to handle a full payload for variable sized output */
466 if (c.info.size_out < 0)
467 c.info.size_out = cxlm->payload_size;
468
469 rc = handle_mailbox_cmd_from_user(cxlm, &c, send.in.payload,
470 send.out.payload, &send.out.size,
471 &send.retval);
472 if (rc)
473 return rc;
474
475 if (copy_to_user(s, &send, sizeof(send)))
476 return -EFAULT;
477
478 return 0;
479 }
480
cxl_xfer_log(struct cxl_mem * cxlm,uuid_t * uuid,u32 size,u8 * out)481 static int cxl_xfer_log(struct cxl_mem *cxlm, uuid_t *uuid, u32 size, u8 *out)
482 {
483 u32 remaining = size;
484 u32 offset = 0;
485
486 while (remaining) {
487 u32 xfer_size = min_t(u32, remaining, cxlm->payload_size);
488 struct cxl_mbox_get_log log = {
489 .uuid = *uuid,
490 .offset = cpu_to_le32(offset),
491 .length = cpu_to_le32(xfer_size)
492 };
493 int rc;
494
495 rc = cxl_mem_mbox_send_cmd(cxlm, CXL_MBOX_OP_GET_LOG, &log,
496 sizeof(log), out, xfer_size);
497 if (rc < 0)
498 return rc;
499
500 out += xfer_size;
501 remaining -= xfer_size;
502 offset += xfer_size;
503 }
504
505 return 0;
506 }
507
508 /**
509 * cxl_walk_cel() - Walk through the Command Effects Log.
510 * @cxlm: Device.
511 * @size: Length of the Command Effects Log.
512 * @cel: CEL
513 *
514 * Iterate over each entry in the CEL and determine if the driver supports the
515 * command. If so, the command is enabled for the device and can be used later.
516 */
cxl_walk_cel(struct cxl_mem * cxlm,size_t size,u8 * cel)517 static void cxl_walk_cel(struct cxl_mem *cxlm, size_t size, u8 *cel)
518 {
519 struct cxl_cel_entry *cel_entry;
520 const int cel_entries = size / sizeof(*cel_entry);
521 int i;
522
523 cel_entry = (struct cxl_cel_entry *) cel;
524
525 for (i = 0; i < cel_entries; i++) {
526 u16 opcode = le16_to_cpu(cel_entry[i].opcode);
527 struct cxl_mem_command *cmd = cxl_mem_find_command(opcode);
528
529 if (!cmd) {
530 dev_dbg(cxlm->dev,
531 "Opcode 0x%04x unsupported by driver", opcode);
532 continue;
533 }
534
535 set_bit(cmd->info.id, cxlm->enabled_cmds);
536 }
537 }
538
cxl_get_gsl(struct cxl_mem * cxlm)539 static struct cxl_mbox_get_supported_logs *cxl_get_gsl(struct cxl_mem *cxlm)
540 {
541 struct cxl_mbox_get_supported_logs *ret;
542 int rc;
543
544 ret = kvmalloc(cxlm->payload_size, GFP_KERNEL);
545 if (!ret)
546 return ERR_PTR(-ENOMEM);
547
548 rc = cxl_mem_mbox_send_cmd(cxlm, CXL_MBOX_OP_GET_SUPPORTED_LOGS, NULL,
549 0, ret, cxlm->payload_size);
550 if (rc < 0) {
551 kvfree(ret);
552 return ERR_PTR(rc);
553 }
554
555 return ret;
556 }
557
558 enum {
559 CEL_UUID,
560 VENDOR_DEBUG_UUID,
561 };
562
563 /* See CXL 2.0 Table 170. Get Log Input Payload */
564 static const uuid_t log_uuid[] = {
565 [CEL_UUID] = DEFINE_CXL_CEL_UUID,
566 [VENDOR_DEBUG_UUID] = DEFINE_CXL_VENDOR_DEBUG_UUID,
567 };
568
569 /**
570 * cxl_mem_enumerate_cmds() - Enumerate commands for a device.
571 * @cxlm: The device.
572 *
573 * Returns 0 if enumerate completed successfully.
574 *
575 * CXL devices have optional support for certain commands. This function will
576 * determine the set of supported commands for the hardware and update the
577 * enabled_cmds bitmap in the @cxlm.
578 */
cxl_mem_enumerate_cmds(struct cxl_mem * cxlm)579 int cxl_mem_enumerate_cmds(struct cxl_mem *cxlm)
580 {
581 struct cxl_mbox_get_supported_logs *gsl;
582 struct device *dev = cxlm->dev;
583 struct cxl_mem_command *cmd;
584 int i, rc;
585
586 gsl = cxl_get_gsl(cxlm);
587 if (IS_ERR(gsl))
588 return PTR_ERR(gsl);
589
590 rc = -ENOENT;
591 for (i = 0; i < le16_to_cpu(gsl->entries); i++) {
592 u32 size = le32_to_cpu(gsl->entry[i].size);
593 uuid_t uuid = gsl->entry[i].uuid;
594 u8 *log;
595
596 dev_dbg(dev, "Found LOG type %pU of size %d", &uuid, size);
597
598 if (!uuid_equal(&uuid, &log_uuid[CEL_UUID]))
599 continue;
600
601 log = kvmalloc(size, GFP_KERNEL);
602 if (!log) {
603 rc = -ENOMEM;
604 goto out;
605 }
606
607 rc = cxl_xfer_log(cxlm, &uuid, size, log);
608 if (rc) {
609 kvfree(log);
610 goto out;
611 }
612
613 cxl_walk_cel(cxlm, size, log);
614 kvfree(log);
615
616 /* In case CEL was bogus, enable some default commands. */
617 cxl_for_each_cmd(cmd)
618 if (cmd->flags & CXL_CMD_FLAG_FORCE_ENABLE)
619 set_bit(cmd->info.id, cxlm->enabled_cmds);
620
621 /* Found the required CEL */
622 rc = 0;
623 }
624
625 out:
626 kvfree(gsl);
627 return rc;
628 }
629 EXPORT_SYMBOL_GPL(cxl_mem_enumerate_cmds);
630
631 /**
632 * cxl_mem_get_partition_info - Get partition info
633 * @cxlm: cxl_mem instance to update partition info
634 *
635 * Retrieve the current partition info for the device specified. The active
636 * values are the current capacity in bytes. If not 0, the 'next' values are
637 * the pending values, in bytes, which take affect on next cold reset.
638 *
639 * Return: 0 if no error: or the result of the mailbox command.
640 *
641 * See CXL @8.2.9.5.2.1 Get Partition Info
642 */
cxl_mem_get_partition_info(struct cxl_mem * cxlm)643 static int cxl_mem_get_partition_info(struct cxl_mem *cxlm)
644 {
645 struct cxl_mbox_get_partition_info {
646 __le64 active_volatile_cap;
647 __le64 active_persistent_cap;
648 __le64 next_volatile_cap;
649 __le64 next_persistent_cap;
650 } __packed pi;
651 int rc;
652
653 rc = cxl_mem_mbox_send_cmd(cxlm, CXL_MBOX_OP_GET_PARTITION_INFO,
654 NULL, 0, &pi, sizeof(pi));
655
656 if (rc)
657 return rc;
658
659 cxlm->active_volatile_bytes =
660 le64_to_cpu(pi.active_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
661 cxlm->active_persistent_bytes =
662 le64_to_cpu(pi.active_persistent_cap) * CXL_CAPACITY_MULTIPLIER;
663 cxlm->next_volatile_bytes =
664 le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
665 cxlm->next_persistent_bytes =
666 le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
667
668 return 0;
669 }
670
671 /**
672 * cxl_mem_identify() - Send the IDENTIFY command to the device.
673 * @cxlm: The device to identify.
674 *
675 * Return: 0 if identify was executed successfully.
676 *
677 * This will dispatch the identify command to the device and on success populate
678 * structures to be exported to sysfs.
679 */
cxl_mem_identify(struct cxl_mem * cxlm)680 int cxl_mem_identify(struct cxl_mem *cxlm)
681 {
682 /* See CXL 2.0 Table 175 Identify Memory Device Output Payload */
683 struct cxl_mbox_identify id;
684 int rc;
685
686 rc = cxl_mem_mbox_send_cmd(cxlm, CXL_MBOX_OP_IDENTIFY, NULL, 0, &id,
687 sizeof(id));
688 if (rc < 0)
689 return rc;
690
691 cxlm->total_bytes =
692 le64_to_cpu(id.total_capacity) * CXL_CAPACITY_MULTIPLIER;
693 cxlm->volatile_only_bytes =
694 le64_to_cpu(id.volatile_capacity) * CXL_CAPACITY_MULTIPLIER;
695 cxlm->persistent_only_bytes =
696 le64_to_cpu(id.persistent_capacity) * CXL_CAPACITY_MULTIPLIER;
697 cxlm->partition_align_bytes =
698 le64_to_cpu(id.partition_align) * CXL_CAPACITY_MULTIPLIER;
699
700 dev_dbg(cxlm->dev,
701 "Identify Memory Device\n"
702 " total_bytes = %#llx\n"
703 " volatile_only_bytes = %#llx\n"
704 " persistent_only_bytes = %#llx\n"
705 " partition_align_bytes = %#llx\n",
706 cxlm->total_bytes, cxlm->volatile_only_bytes,
707 cxlm->persistent_only_bytes, cxlm->partition_align_bytes);
708
709 cxlm->lsa_size = le32_to_cpu(id.lsa_size);
710 memcpy(cxlm->firmware_version, id.fw_revision, sizeof(id.fw_revision));
711
712 return 0;
713 }
714 EXPORT_SYMBOL_GPL(cxl_mem_identify);
715
cxl_mem_create_range_info(struct cxl_mem * cxlm)716 int cxl_mem_create_range_info(struct cxl_mem *cxlm)
717 {
718 int rc;
719
720 if (cxlm->partition_align_bytes == 0) {
721 cxlm->ram_range.start = 0;
722 cxlm->ram_range.end = cxlm->volatile_only_bytes - 1;
723 cxlm->pmem_range.start = cxlm->volatile_only_bytes;
724 cxlm->pmem_range.end = cxlm->volatile_only_bytes +
725 cxlm->persistent_only_bytes - 1;
726 return 0;
727 }
728
729 rc = cxl_mem_get_partition_info(cxlm);
730 if (rc) {
731 dev_err(cxlm->dev, "Failed to query partition information\n");
732 return rc;
733 }
734
735 dev_dbg(cxlm->dev,
736 "Get Partition Info\n"
737 " active_volatile_bytes = %#llx\n"
738 " active_persistent_bytes = %#llx\n"
739 " next_volatile_bytes = %#llx\n"
740 " next_persistent_bytes = %#llx\n",
741 cxlm->active_volatile_bytes, cxlm->active_persistent_bytes,
742 cxlm->next_volatile_bytes, cxlm->next_persistent_bytes);
743
744 cxlm->ram_range.start = 0;
745 cxlm->ram_range.end = cxlm->active_volatile_bytes - 1;
746
747 cxlm->pmem_range.start = cxlm->active_volatile_bytes;
748 cxlm->pmem_range.end =
749 cxlm->active_volatile_bytes + cxlm->active_persistent_bytes - 1;
750
751 return 0;
752 }
753 EXPORT_SYMBOL_GPL(cxl_mem_create_range_info);
754
cxl_mem_create(struct device * dev)755 struct cxl_mem *cxl_mem_create(struct device *dev)
756 {
757 struct cxl_mem *cxlm;
758
759 cxlm = devm_kzalloc(dev, sizeof(*cxlm), GFP_KERNEL);
760 if (!cxlm) {
761 dev_err(dev, "No memory available\n");
762 return ERR_PTR(-ENOMEM);
763 }
764
765 mutex_init(&cxlm->mbox_mutex);
766 cxlm->dev = dev;
767
768 return cxlm;
769 }
770 EXPORT_SYMBOL_GPL(cxl_mem_create);
771
772 static struct dentry *cxl_debugfs;
773
cxl_mbox_init(void)774 void __init cxl_mbox_init(void)
775 {
776 struct dentry *mbox_debugfs;
777
778 cxl_debugfs = debugfs_create_dir("cxl", NULL);
779 mbox_debugfs = debugfs_create_dir("mbox", cxl_debugfs);
780 debugfs_create_bool("raw_allow_all", 0600, mbox_debugfs,
781 &cxl_raw_allow_all);
782 }
783
cxl_mbox_exit(void)784 void cxl_mbox_exit(void)
785 {
786 debugfs_remove_recursive(cxl_debugfs);
787 }
788