Searched refs:__le64 (Results 1 – 23 of 23) sorted by relevance
419 __le64 type;460 __le64 type;518 __le64 refs;546 __le64 root;563 __le64 root;599 __le64 sec;639 __le64 size;646 __le64 rdev;664 __le64 end;761 __le64 pend;[all …]
422 __le64 sqnum;440 __le64 huge;482 __le64 size;521 __le64 inum;570 __le64 old_size;571 __le64 new_size;636 __le64 rp_size;680 __le64 cmt_no;769 __le64 cmt_no;780 __le64 cmt_no;[all …]
199 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); in avail_orphs()201 if (gap >= UBIFS_ORPH_NODE_SZ + sizeof(__le64)) in avail_orphs()202 avail += (gap - UBIFS_ORPH_NODE_SZ) / sizeof(__le64); in avail_orphs()219 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); in tot_avail_orphs()272 if (gap < UBIFS_ORPH_NODE_SZ + sizeof(__le64)) { in write_orph_node()285 cnt = (gap - UBIFS_ORPH_NODE_SZ) / sizeof(__le64); in write_orph_node()288 len = UBIFS_ORPH_NODE_SZ + cnt * sizeof(__le64); in write_orph_node()
85 __le64 tmp_le64; in create_default_filesystem()
768 UBIFS_ORPH_NODE_SZ + sizeof(__le64); in init_constants_early()
82 __le64 signature;87 __le64 my_lba;88 __le64 alternate_lba;89 __le64 first_usable_lba;90 __le64 last_usable_lba;92 __le64 partition_entry_lba;113 __le64 starting_lba;114 __le64 ending_lba;
129 __le64 mmp_block;
407 return le64_to_cpu((__force __le64)val); in __virtio64_to_cpu()
94 __le64 nsze;95 __le64 ncap;96 __le64 nuse;206 __le64 rkey;233 __le64 prp1;234 __le64 prp2;245 __le64 prp1;246 __le64 prp2;247 __le64 slba;286 __le64 prp1;[all …]
66 __le64 root_inode;67 __le64 bytes_used;68 __le64 id_table_start;69 __le64 xattr_id_table_start;70 __le64 inode_table_start;71 __le64 directory_table_start;72 __le64 fragment_table_start;73 __le64 export_table_start;178 __le64 start_block;179 __le64 file_size;[all …]
84 static int sqfs_calc_n_blks(__le64 start, __le64 end, u64 *offset) in sqfs_calc_n_blks()
20 #define __constant_cpu_to_le64(x) ((__force __le64)___constant_swab64((x)))21 #define __constant_le64_to_cpu(x) ___constant_swab64((__force __u64)(__le64)(x))32 #define __cpu_to_le64(x) ((__force __le64)__swab64((x)))33 #define __le64_to_cpu(x) __swab64((__force __u64)(__le64)(x))45 static inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p()47 return (__force __le64)__swab64p(p); in __cpu_to_le64p()49 static inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup()
20 #define __constant_cpu_to_le64(x) ((__force __le64)(__u64)(x))21 #define __constant_le64_to_cpu(x) ((__force __u64)(__le64)(x))32 #define __cpu_to_le64(x) ((__force __le64)(__u64)(x))33 #define __le64_to_cpu(x) ((__force __u64)(__le64)(x))45 static inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p()47 return (__force __le64)*p; in __cpu_to_le64p()49 static inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup()
154 static inline void le64_add_cpu(__le64 *var, u64 val) in le64_add_cpu()
18 return le64_to_cpup((__le64 *)p); in get_unaligned_le64()48 *((__le64 *)p) = cpu_to_le64(val); in put_unaligned_le64()
394 volatile __le64 erst_base;395 volatile __le64 erst_dequeue;593 __le64 deq;691 __le64 dev_context_ptrs[MAX_HC_SLOTS];702 __le64 buffer;797 volatile __le64 segment_ptr;808 volatile __le64 cmd_trb;1020 __le64 seg_addr;1106 static inline u64 xhci_readq(__le64 volatile *regs) in xhci_readq()1114 static inline void xhci_writeq(__le64 volatile *regs, const u64 val) in xhci_writeq()
120 #define aligned_le64 __le64 __aligned(8)146 typedef __u64 __bitwise __le64; typedef
94 __le64 addr;108 __le64 addr;
47 __le64 addr;
168 #define readq_relaxed(c) ({ u64 __r = le64_to_cpu((__force __le64)__arch_getq(c)); __r; })
635 __le64 dma_adr[DWC3_MAX_HIBER_SCRATCHBUFS];
498 sizeof(__le64)); in xhci_alloc_virt_device()
140 #define readq_relaxed(c) ({ u64 __r = le64_to_cpu((__force __le64) \
Completed in 91 milliseconds