Lines Matching refs:keys

115 	if (b->level && b->keys.nsets)  in bch_btree_init_next()
116 bch_btree_sort(&b->keys, &b->c->sort); in bch_btree_init_next()
118 bch_btree_sort_lazy(&b->keys, &b->c->sort); in bch_btree_init_next()
121 bch_bset_init_next(&b->keys, write_block(b), in bch_btree_init_next()
164 iter->b = &b->keys; in bch_btree_node_read_done()
171 b->written < btree_blocks(b) && i->seq == b->keys.set[0].data->seq; in bch_btree_node_read_done()
199 if (i != b->keys.set[0].data && !i->keys) in bch_btree_node_read_done()
209 bset_sector_offset(&b->keys, i) < KEY_SIZE(&b->key); in bch_btree_node_read_done()
211 if (i->seq == b->keys.set[0].data->seq) in bch_btree_node_read_done()
214 bch_btree_sort_and_fix_extents(&b->keys, iter, &b->c->sort); in bch_btree_node_read_done()
216 i = b->keys.set[0].data; in bch_btree_node_read_done()
218 if (b->keys.set[0].size && in bch_btree_node_read_done()
219 bkey_cmp(&b->key, &b->keys.set[0].end) < 0) in bch_btree_node_read_done()
223 bch_bset_init_next(&b->keys, write_block(b), in bch_btree_node_read_done()
232 bset_block_offset(b, i), i->keys); in bch_btree_node_read_done()
259 bch_bio_map(bio, b->keys.set[0].data); in bch_btree_node_read()
373 bset_sector_offset(&b->keys, i)); in do_btree_node_write()
413 BUG_ON(b->written && !i->keys); in __bch_btree_node_write()
415 bch_check_keys(&b->keys, "writing"); in __bch_btree_node_write()
436 unsigned int nsets = b->keys.nsets; in bch_btree_node_write()
446 if (nsets && !b->keys.nsets) in bch_btree_node_write()
483 BUG_ON(!i->keys); in bch_btree_leaf_dirty()
528 bch_btree_keys_free(&b->keys); in mca_data_free()
550 if (!bch_btree_keys_alloc(&b->keys, in mca_data_alloc()
597 BUG_ON(btree_node_dirty(b) && !b->keys.set[0].data); in mca_reap()
599 if (b->keys.page_order < min_order) in mca_reap()
804 c->verify_data->keys.set->data) in bch_btree_cache_alloc()
923 if (!b->keys.set[0].data) in mca_alloc()
934 if (!b->keys.set->data) in mca_alloc()
951 bch_btree_keys_init(&b->keys, &bch_extent_keys_ops, in mca_alloc()
954 bch_btree_keys_init(&b->keys, &bch_btree_keys_ops, in mca_alloc()
1024 for (; i <= b->keys.nsets && b->keys.set[i].size; i++) { in bch_btree_node_get()
1025 prefetch(b->keys.set[i].tree); in bch_btree_node_get()
1026 prefetch(b->keys.set[i].data); in bch_btree_node_get()
1029 for (; i <= b->keys.nsets; i++) in bch_btree_node_get()
1030 prefetch(b->keys.set[i].data); in bch_btree_node_get()
1114 bch_bset_init_next(&b->keys, b->keys.set->data, bset_magic(&b->c->cache->sb)); in __bch_btree_node_alloc()
1143 bch_btree_sort_into(&b->keys, &n->keys, &b->c->sort); in btree_node_alloc_replacement()
1275 unsigned int keys = 0, good_keys = 0; in btree_gc_mark_node() local
1282 for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) { in btree_gc_mark_node()
1284 keys++; in btree_gc_mark_node()
1286 if (bch_ptr_bad(&b->keys, k)) in btree_gc_mark_node()
1296 for (t = b->keys.set; t <= &b->keys.set[b->keys.nsets]; t++) in btree_gc_mark_node()
1298 bset_written(&b->keys, t) && in btree_gc_mark_node()
1308 if ((keys - good_keys) * 2 > keys) in btree_gc_mark_node()
1318 unsigned int keys; member
1329 unsigned int i, nodes = 0, keys = 0, blocks; in btree_gc_coalesce() local
1344 keys += r[nodes++].keys; in btree_gc_coalesce()
1349 __set_blocks(b->keys.set[0].data, keys, in btree_gc_coalesce()
1376 keys = 0; in btree_gc_coalesce()
1382 if (__set_blocks(n1, n1->keys + keys + in btree_gc_coalesce()
1388 keys += bkey_u64s(k); in btree_gc_coalesce()
1399 if (__set_blocks(n1, n1->keys + n2->keys, in btree_gc_coalesce()
1404 keys = n2->keys; in btree_gc_coalesce()
1409 BUG_ON(__set_blocks(n1, n1->keys + keys, block_bytes(b->c->cache)) > in btree_gc_coalesce()
1417 (void *) bset_bkey_idx(n2, keys) - (void *) n2->start); in btree_gc_coalesce()
1419 n1->keys += keys; in btree_gc_coalesce()
1420 r[i].keys = n1->keys; in btree_gc_coalesce()
1423 bset_bkey_idx(n2, keys), in btree_gc_coalesce()
1425 (void *) bset_bkey_idx(n2, keys)); in btree_gc_coalesce()
1427 n2->keys -= keys; in btree_gc_coalesce()
1443 BUG_ON(btree_bset_first(new_nodes[0])->keys); in btree_gc_coalesce()
1500 struct keylist keys; in btree_gc_rewrite_node() local
1517 bch_keylist_init(&keys); in btree_gc_rewrite_node()
1518 bch_keylist_add(&keys, &n->key); in btree_gc_rewrite_node()
1520 make_btree_freeing_key(replace, keys.top); in btree_gc_rewrite_node()
1521 bch_keylist_push(&keys); in btree_gc_rewrite_node()
1523 bch_btree_insert_node(b, op, &keys, NULL, NULL); in btree_gc_rewrite_node()
1524 BUG_ON(!bch_keylist_empty(&keys)); in btree_gc_rewrite_node()
1539 for_each_key_filter(&b->keys, k, &iter, bch_ptr_bad) in btree_gc_count_keys()
1581 bch_btree_iter_init(&b->keys, &iter, &b->c->gc_done); in btree_gc_recurse()
1587 k = bch_btree_iter_next_filter(&iter, &b->keys, bch_ptr_bad); in btree_gc_recurse()
1596 r->keys = btree_gc_count_keys(r->b); in btree_gc_recurse()
1751 &dc->writeback_keys.keys, node) in bch_btree_gc_finish()
1764 for (k = ca->sb.d; k < ca->sb.d + ca->sb.keys; k++) in bch_btree_gc_finish()
1879 for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) in bch_btree_check_recurse()
1885 bch_btree_iter_init(&b->keys, &iter, NULL); in bch_btree_check_recurse()
1888 k = bch_btree_iter_next_filter(&iter, &b->keys, in bch_btree_check_recurse()
1925 bch_btree_iter_init(&c->root->keys, &iter, NULL); in bch_btree_check_thread()
1926 k = bch_btree_iter_next_filter(&iter, &c->root->keys, bch_ptr_bad); in bch_btree_check_thread()
1945 &c->root->keys, in bch_btree_check_thread()
2013 for_each_key_filter(&c->root->keys, k, &iter, bch_ptr_invalid) in bch_btree_check()
2124 status = bch_btree_insert_key(&b->keys, k, replace_key); in btree_insert_key()
2126 bch_check_keys(&b->keys, "%u for %s", status, in btree_insert_key()
2138 long ret = bch_btree_keys_u64s_remaining(&b->keys); in insert_u64s_remaining()
2143 if (b->keys.ops->is_extents) in insert_u64s_remaining()
2154 int oldsize = bch_count_data(&b->keys); in bch_btree_insert_keys()
2157 struct bkey *k = insert_keys->keys; in bch_btree_insert_keys()
2170 bkey_copy(&temp.key, insert_keys->keys); in bch_btree_insert_keys()
2173 bch_cut_front(&b->key, insert_keys->keys); in bch_btree_insert_keys()
2187 BUG_ON(bch_count_data(&b->keys) < oldsize); in bch_btree_insert_keys()
2219 unsigned int keys = 0; in btree_split() local
2221 trace_bcache_btree_node_split(b, btree_bset_first(n1)->keys); in btree_split()
2243 while (keys < (btree_bset_first(n1)->keys * 3) / 5) in btree_split()
2244 keys += bkey_u64s(bset_bkey_idx(btree_bset_first(n1), in btree_split()
2245 keys)); in btree_split()
2248 bset_bkey_idx(btree_bset_first(n1), keys)); in btree_split()
2249 keys += bkey_u64s(bset_bkey_idx(btree_bset_first(n1), keys)); in btree_split()
2251 btree_bset_first(n2)->keys = btree_bset_first(n1)->keys - keys; in btree_split()
2252 btree_bset_first(n1)->keys = keys; in btree_split()
2256 btree_bset_first(n2)->keys * sizeof(uint64_t)); in btree_split()
2265 trace_bcache_btree_node_compact(b, btree_bset_first(n1)->keys); in btree_split()
2339 b->keys.last_set_unwritten) in bch_btree_insert_node()
2421 struct keylist *keys; member
2431 int ret = bch_btree_insert_node(b, &op->op, op->keys, in btree_insert_fn()
2433 if (ret && !bch_keylist_empty(op->keys)) in btree_insert_fn()
2439 int bch_btree_insert(struct cache_set *c, struct keylist *keys, in bch_btree_insert() argument
2446 BUG_ON(bch_keylist_empty(keys)); in bch_btree_insert()
2449 op.keys = keys; in bch_btree_insert()
2453 while (!ret && !bch_keylist_empty(keys)) { in bch_btree_insert()
2456 &START_KEY(keys->keys), in bch_btree_insert()
2465 while ((k = bch_keylist_pop(keys))) in bch_btree_insert()
2509 bch_btree_iter_init(&b->keys, &iter, from); in bch_btree_map_nodes_recurse()
2511 while ((k = bch_btree_iter_next_filter(&iter, &b->keys, in bch_btree_map_nodes_recurse()
2542 bch_btree_iter_init(&b->keys, &iter, from); in bch_btree_map_keys_recurse()
2544 while ((k = bch_btree_iter_next_filter(&iter, &b->keys, bch_ptr_bad))) { in bch_btree_map_keys_recurse()
2623 if (RB_INSERT(&buf->keys, w, node, keybuf_cmp)) in refill_keybuf_fn()
2662 if (!RB_EMPTY_ROOT(&buf->keys)) { in bch_refill_keybuf()
2665 w = RB_FIRST(&buf->keys, struct keybuf_key, node); in bch_refill_keybuf()
2668 w = RB_LAST(&buf->keys, struct keybuf_key, node); in bch_refill_keybuf()
2680 rb_erase(&w->node, &buf->keys); in __bch_keybuf_del()
2704 w = RB_GREATER(&buf->keys, s, node, keybuf_nonoverlapping_cmp); in bch_keybuf_check_overlapping()
2726 w = RB_FIRST(&buf->keys, struct keybuf_key, node); in bch_keybuf_next()
2764 buf->keys = RB_ROOT; in bch_keybuf_init()