Lines Matching refs:subsys

117 static u32 nvmet_max_nsid(struct nvmet_subsys *subsys)  in nvmet_max_nsid()  argument
123 xa_for_each(&subsys->namespaces, idx, cur) in nvmet_max_nsid()
235 void nvmet_ns_changed(struct nvmet_subsys *subsys, u32 nsid) in nvmet_ns_changed() argument
239 lockdep_assert_held(&subsys->lock); in nvmet_ns_changed()
241 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_ns_changed()
251 void nvmet_send_ana_event(struct nvmet_subsys *subsys, in nvmet_send_ana_event() argument
256 mutex_lock(&subsys->lock); in nvmet_send_ana_event()
257 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_send_ana_event()
265 mutex_unlock(&subsys->lock); in nvmet_send_ana_event()
274 nvmet_send_ana_event(p->subsys, port); in nvmet_port_send_ana_event()
301 void nvmet_port_del_ctrls(struct nvmet_port *port, struct nvmet_subsys *subsys) in nvmet_port_del_ctrls() argument
305 mutex_lock(&subsys->lock); in nvmet_port_del_ctrls()
306 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_port_del_ctrls()
310 mutex_unlock(&subsys->lock); in nvmet_port_del_ctrls()
544 nvmet_ns_changed(ns->subsys, ns->nsid); in nvmet_ns_revalidate()
549 struct nvmet_subsys *subsys = ns->subsys; in nvmet_ns_enable() local
553 mutex_lock(&subsys->lock); in nvmet_ns_enable()
556 if (nvmet_is_passthru_subsys(subsys)) { in nvmet_ns_enable()
565 if (subsys->nr_namespaces == NVMET_MAX_NAMESPACES) in nvmet_ns_enable()
578 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_enable()
586 if (ns->nsid > subsys->max_nsid) in nvmet_ns_enable()
587 subsys->max_nsid = ns->nsid; in nvmet_ns_enable()
589 ret = xa_insert(&subsys->namespaces, ns->nsid, ns, GFP_KERNEL); in nvmet_ns_enable()
593 subsys->nr_namespaces++; in nvmet_ns_enable()
595 nvmet_ns_changed(subsys, ns->nsid); in nvmet_ns_enable()
599 mutex_unlock(&subsys->lock); in nvmet_ns_enable()
603 subsys->max_nsid = nvmet_max_nsid(subsys); in nvmet_ns_enable()
606 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_enable()
615 struct nvmet_subsys *subsys = ns->subsys; in nvmet_ns_disable() local
618 mutex_lock(&subsys->lock); in nvmet_ns_disable()
623 xa_erase(&ns->subsys->namespaces, ns->nsid); in nvmet_ns_disable()
624 if (ns->nsid == subsys->max_nsid) in nvmet_ns_disable()
625 subsys->max_nsid = nvmet_max_nsid(subsys); in nvmet_ns_disable()
627 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_disable()
630 mutex_unlock(&subsys->lock); in nvmet_ns_disable()
645 mutex_lock(&subsys->lock); in nvmet_ns_disable()
647 subsys->nr_namespaces--; in nvmet_ns_disable()
648 nvmet_ns_changed(subsys, ns->nsid); in nvmet_ns_disable()
651 mutex_unlock(&subsys->lock); in nvmet_ns_disable()
666 struct nvmet_ns *nvmet_ns_alloc(struct nvmet_subsys *subsys, u32 nsid) in nvmet_ns_alloc() argument
677 ns->subsys = subsys; in nvmet_ns_alloc()
1143 if (!nvmet_is_disc_subsys(ctrl->subsys) && in nvmet_start_ctrl()
1213 if (nvmet_is_passthru_subsys(ctrl->subsys)) in nvmet_init_cap()
1222 struct nvmet_subsys *subsys; in nvmet_ctrl_find_get() local
1224 subsys = nvmet_find_get_subsys(req->port, subsysnqn); in nvmet_ctrl_find_get()
1225 if (!subsys) { in nvmet_ctrl_find_get()
1232 mutex_lock(&subsys->lock); in nvmet_ctrl_find_get()
1233 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_ctrl_find_get()
1253 mutex_unlock(&subsys->lock); in nvmet_ctrl_find_get()
1254 nvmet_subsys_put(subsys); in nvmet_ctrl_find_get()
1275 bool nvmet_host_allowed(struct nvmet_subsys *subsys, const char *hostnqn) in nvmet_host_allowed() argument
1281 if (subsys->allow_any_host) in nvmet_host_allowed()
1284 if (nvmet_is_disc_subsys(subsys)) /* allow all access to disc subsys */ in nvmet_host_allowed()
1287 list_for_each_entry(p, &subsys->hosts, entry) { in nvmet_host_allowed()
1309 xa_for_each(&ctrl->subsys->namespaces, idx, ns) in nvmet_setup_p2p_ns_map()
1339 struct nvmet_subsys *subsys; in nvmet_alloc_ctrl() local
1345 subsys = nvmet_find_get_subsys(req->port, subsysnqn); in nvmet_alloc_ctrl()
1346 if (!subsys) { in nvmet_alloc_ctrl()
1355 if (!nvmet_host_allowed(subsys, hostnqn)) { in nvmet_alloc_ctrl()
1385 ctrl->subsys = subsys; in nvmet_alloc_ctrl()
1394 ctrl->sqs = kcalloc(subsys->max_qid + 1, in nvmet_alloc_ctrl()
1400 if (subsys->cntlid_min > subsys->cntlid_max) in nvmet_alloc_ctrl()
1404 subsys->cntlid_min, subsys->cntlid_max, in nvmet_alloc_ctrl()
1416 if (nvmet_is_disc_subsys(ctrl->subsys) && !kato) in nvmet_alloc_ctrl()
1427 mutex_lock(&subsys->lock); in nvmet_alloc_ctrl()
1428 list_add_tail(&ctrl->subsys_entry, &subsys->ctrls); in nvmet_alloc_ctrl()
1430 mutex_unlock(&subsys->lock); in nvmet_alloc_ctrl()
1442 nvmet_subsys_put(subsys); in nvmet_alloc_ctrl()
1450 struct nvmet_subsys *subsys = ctrl->subsys; in nvmet_ctrl_free() local
1452 mutex_lock(&subsys->lock); in nvmet_ctrl_free()
1455 mutex_unlock(&subsys->lock); in nvmet_ctrl_free()
1469 nvmet_subsys_put(subsys); in nvmet_ctrl_free()
1505 if (!strncmp(p->subsys->subsysnqn, subsysnqn, in nvmet_find_get_subsys()
1507 if (!kref_get_unless_zero(&p->subsys->ref)) in nvmet_find_get_subsys()
1510 return p->subsys; in nvmet_find_get_subsys()
1520 struct nvmet_subsys *subsys; in nvmet_subsys_alloc() local
1524 subsys = kzalloc(sizeof(*subsys), GFP_KERNEL); in nvmet_subsys_alloc()
1525 if (!subsys) in nvmet_subsys_alloc()
1528 subsys->ver = NVMET_DEFAULT_VS; in nvmet_subsys_alloc()
1531 bin2hex(subsys->serial, &serial, sizeof(serial)); in nvmet_subsys_alloc()
1533 subsys->model_number = kstrdup(NVMET_DEFAULT_CTRL_MODEL, GFP_KERNEL); in nvmet_subsys_alloc()
1534 if (!subsys->model_number) { in nvmet_subsys_alloc()
1541 subsys->max_qid = NVMET_NR_QUEUES; in nvmet_subsys_alloc()
1545 subsys->max_qid = 0; in nvmet_subsys_alloc()
1552 subsys->type = type; in nvmet_subsys_alloc()
1553 subsys->subsysnqn = kstrndup(subsysnqn, NVMF_NQN_SIZE, in nvmet_subsys_alloc()
1555 if (!subsys->subsysnqn) { in nvmet_subsys_alloc()
1559 subsys->cntlid_min = NVME_CNTLID_MIN; in nvmet_subsys_alloc()
1560 subsys->cntlid_max = NVME_CNTLID_MAX; in nvmet_subsys_alloc()
1561 kref_init(&subsys->ref); in nvmet_subsys_alloc()
1563 mutex_init(&subsys->lock); in nvmet_subsys_alloc()
1564 xa_init(&subsys->namespaces); in nvmet_subsys_alloc()
1565 INIT_LIST_HEAD(&subsys->ctrls); in nvmet_subsys_alloc()
1566 INIT_LIST_HEAD(&subsys->hosts); in nvmet_subsys_alloc()
1568 return subsys; in nvmet_subsys_alloc()
1571 kfree(subsys->model_number); in nvmet_subsys_alloc()
1573 kfree(subsys); in nvmet_subsys_alloc()
1579 struct nvmet_subsys *subsys = in nvmet_subsys_free() local
1582 WARN_ON_ONCE(!xa_empty(&subsys->namespaces)); in nvmet_subsys_free()
1584 xa_destroy(&subsys->namespaces); in nvmet_subsys_free()
1585 nvmet_passthru_subsys_free(subsys); in nvmet_subsys_free()
1587 kfree(subsys->subsysnqn); in nvmet_subsys_free()
1588 kfree(subsys->model_number); in nvmet_subsys_free()
1589 kfree(subsys); in nvmet_subsys_free()
1592 void nvmet_subsys_del_ctrls(struct nvmet_subsys *subsys) in nvmet_subsys_del_ctrls() argument
1596 mutex_lock(&subsys->lock); in nvmet_subsys_del_ctrls()
1597 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_subsys_del_ctrls()
1599 mutex_unlock(&subsys->lock); in nvmet_subsys_del_ctrls()
1602 void nvmet_subsys_put(struct nvmet_subsys *subsys) in nvmet_subsys_put() argument
1604 kref_put(&subsys->ref, nvmet_subsys_free); in nvmet_subsys_put()