/linux/fs/btrfs/ |
A D | extent-io-tree.h | 171 struct extent_state **cached_state, gfp_t mask, 184 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument 187 cached_state, GFP_NOFS, NULL); in clear_extent_uptodate() 207 struct extent_state **cached_state); 211 struct extent_state **cached_state) in set_extent_delalloc() argument 215 0, NULL, cached_state, GFP_NOFS, NULL); in set_extent_delalloc() 219 u64 end, struct extent_state **cached_state) in set_extent_defrag() argument 223 0, NULL, cached_state, GFP_NOFS, NULL); in set_extent_defrag() 237 cached_state, mask, NULL); in set_extent_uptodate() 242 struct extent_state **cached_state); [all …]
|
A D | extent_io.c | 742 if (cached_state) { in __clear_extent_bit() 746 *cached_state = NULL; in __clear_extent_bit() 747 cached_state = NULL; in __clear_extent_bit() 1006 if (cached_state && *cached_state) { in set_extent_bit() 1007 state = *cached_state; in set_extent_bit() 1243 if (cached_state && *cached_state) { in convert_extent_bit() 1244 state = *cached_state; in convert_extent_bit() 1572 if (cached_state && *cached_state) { in find_first_extent_bit() 1584 *cached_state = NULL; in find_first_extent_bit() 2051 cached_state = NULL; in find_lock_delalloc_range() [all …]
|
A D | file.c | 1459 cached_state); in lock_and_cleanup_extent_if_need() 1466 last_pos, cached_state); in lock_and_cleanup_extent_if_need() 1782 &lockend, &cached_state); in btrfs_buffered_write() 2584 cached_state); in btrfs_punch_hole_lock_range() 2605 lockend, cached_state); in btrfs_punch_hole_lock_range() 3059 &cached_state); in btrfs_punch_hole() 3084 &cached_state); in btrfs_punch_hole() 3355 &cached_state); in btrfs_zero_range() 3499 &cached_state); in btrfs_fallocate() 3585 &cached_state); in btrfs_fallocate() [all …]
|
A D | inode.c | 2659 cached_state); in btrfs_set_extent_delalloc() 2665 cached_state); in btrfs_set_extent_delalloc() 2760 &cached_state); in btrfs_writepage_fixup_worker() 2779 &cached_state); in btrfs_writepage_fixup_worker() 3141 &cached_state); in btrfs_finish_ordered_io() 4698 &cached_state); in btrfs_truncate_inode_items() 5101 &cached_state); in btrfs_truncate_block() 5229 &cached_state); in btrfs_cont_expand() 5516 &cached_state); in evict_inode_truncate_pages() 7638 cached_state); in lock_extent_direct() [all …]
|
A D | transaction.c | 1027 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local 1033 mark, &cached_state)) { in btrfs_write_marked_extents() 1038 mark, &cached_state); in btrfs_write_marked_extents() 1062 free_extent_state(cached_state); in btrfs_write_marked_extents() 1063 cached_state = NULL; in btrfs_write_marked_extents() 1083 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local 1088 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents() 1098 EXTENT_NEED_WAIT, 0, 0, &cached_state); in __btrfs_wait_marked_extents() 1105 free_extent_state(cached_state); in __btrfs_wait_marked_extents() 1106 cached_state = NULL; in __btrfs_wait_marked_extents()
|
A D | ordered-data.c | 1019 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument 1025 if (cached_state) in btrfs_lock_and_flush_ordered_range() 1026 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range() 1038 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
|
A D | ordered-data.h | 212 struct extent_state **cached_state);
|
A D | dev-replace.c | 823 struct extent_state *cached_state = NULL; in btrfs_set_target_alloc_state() local 833 CHUNK_ALLOCATED, &cached_state)) { in btrfs_set_target_alloc_state() 841 free_extent_state(cached_state); in btrfs_set_target_alloc_state()
|
A D | ioctl.c | 1066 struct extent_state *cached_state = NULL; in defrag_prepare_one_page() local 1100 lock_extent_bits(&inode->io_tree, page_start, page_end, &cached_state); in defrag_prepare_one_page() 1103 &cached_state); in defrag_prepare_one_page() 1286 struct extent_state **cached_state) in defrag_one_locked_target() argument 1305 EXTENT_DEFRAG, 0, 0, cached_state); in defrag_one_locked_target() 1306 set_extent_defrag(&inode->io_tree, start, start + len - 1, cached_state); in defrag_one_locked_target() 1322 struct extent_state *cached_state = NULL; in defrag_one_range() local 1356 &cached_state); in defrag_one_range() 1372 &cached_state); in defrag_one_range() 1384 &cached_state); in defrag_one_range()
|
A D | free-space-cache.c | 1235 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument 1239 i_size_read(inode) - 1, cached_state); in cleanup_write_cache_enospc() 1329 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local 1365 &cached_state); in __btrfs_write_out_cache() 1407 &cached_state, false); in __btrfs_write_out_cache() 1421 i_size_read(inode) - 1, &cached_state); in __btrfs_write_out_cache() 1443 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
|
A D | disk-io.c | 242 struct extent_state *cached_state = NULL; in verify_parent_transid() local 252 &cached_state); in verify_parent_transid() 266 &cached_state); in verify_parent_transid() 4831 struct extent_state *cached_state = NULL; in btrfs_destroy_pinned_extent() local 4841 EXTENT_DIRTY, &cached_state); in btrfs_destroy_pinned_extent() 4847 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_destroy_pinned_extent() 4848 free_extent_state(cached_state); in btrfs_destroy_pinned_extent()
|
A D | extent-tree.c | 2798 struct extent_state *cached_state = NULL; in btrfs_finish_extent_commit() local 2802 EXTENT_DIRTY, &cached_state); in btrfs_finish_extent_commit() 2812 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_finish_extent_commit() 2815 free_extent_state(cached_state); in btrfs_finish_extent_commit()
|
A D | tree-log.c | 4275 struct extent_state *cached_state = NULL; in log_csums() local 4293 lock_end, &cached_state); in log_csums() 4310 &cached_state); in log_csums()
|
A D | ctree.h | 3221 struct extent_state **cached_state);
|
/linux/drivers/gpu/drm/msm/dsi/phy/ |
A D | dsi_phy_28nm_8960.c | 65 struct pll_28nm_cached_state cached_state; member 346 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local 349 cached_state->postdiv3 = in dsi_28nm_pll_save_state() 351 cached_state->postdiv2 = in dsi_28nm_pll_save_state() 353 cached_state->postdiv1 = in dsi_28nm_pll_save_state() 356 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state() 362 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local 367 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state() 375 cached_state->postdiv3); in dsi_28nm_pll_restore_state() 377 cached_state->postdiv2); in dsi_28nm_pll_restore_state() [all …]
|
A D | dsi_phy_28nm.c | 73 struct pll_28nm_cached_state cached_state; member 481 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local 484 cached_state->postdiv3 = in dsi_28nm_pll_save_state() 486 cached_state->postdiv1 = in dsi_28nm_pll_save_state() 490 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state() 492 cached_state->vco_rate = 0; in dsi_28nm_pll_save_state() 498 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local 503 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state() 511 cached_state->postdiv3); in dsi_28nm_pll_restore_state() 513 cached_state->postdiv1); in dsi_28nm_pll_restore_state() [all …]
|
A D | dsi_phy_14nm.c | 79 struct pll_14nm_cached_state cached_state; member 687 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_save_state() local 693 cached_state->n1postdiv = data & 0xf; in dsi_14nm_pll_save_state() 694 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_14nm_pll_save_state() 697 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_save_state() 699 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_14nm_pll_save_state() 705 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_restore_state() local 711 cached_state->vco_rate, 0); in dsi_14nm_pll_restore_state() 718 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_14nm_pll_restore_state() 721 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_restore_state()
|
A D | dsi_phy_10nm.c | 79 struct pll_10nm_cached_state cached_state; member 472 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state() 495 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_restore_state()
|
A D | dsi_phy_7nm.c | 78 struct pll_7nm_cached_state cached_state; member 498 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() 521 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state()
|
/linux/drivers/iio/multiplexer/ |
A D | iio-mux.c | 29 int cached_state; member 49 mux->cached_state = -1; in iio_mux_select() 53 if (mux->cached_state == chan->channel) in iio_mux_select() 72 mux->cached_state = -1; in iio_mux_select() 77 mux->cached_state = chan->channel; in iio_mux_select() 395 mux->cached_state = -1; in mux_probe()
|
/linux/drivers/mux/ |
A D | core.c | 118 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc() 133 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set() 159 if (mux->idle_state == mux->cached_state) in mux_chip_register() 307 if (mux->cached_state == state) in __mux_control_select() 425 mux->idle_state != mux->cached_state) in mux_control_deselect()
|
/linux/include/linux/mux/ |
A D | driver.h | 48 int cached_state; member
|
/linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
A D | amdgpu_dm.h | 481 struct drm_atomic_state *cached_state; member
|
A D | amdgpu_dm.c | 2365 WARN_ON(adev->dm.cached_state); in dm_suspend() 2366 adev->dm.cached_state = drm_atomic_helper_suspend(adev_to_drm(adev)); in dm_suspend() 2691 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) in dm_resume() 2699 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) { in dm_resume() 2708 for_each_new_plane_in_state(dm->cached_state, plane, new_plane_state, i) { in dm_resume() 2717 drm_atomic_helper_resume(ddev, dm->cached_state); in dm_resume() 2719 dm->cached_state = NULL; in dm_resume()
|