/drivers/mtd/ |
H A D | ftl.c | 158 static int scan_header(partition_t *part) argument 164 part->header.FormattedSize = 0; 165 max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size; 169 offset += part->mbd.mtd->erasesize ? : 0x2000) { 171 err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret, 190 if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) { 192 1 << header.EraseUnitSize,part->mbd.mtd->erasesize); 195 part->header = header; 199 static int build_maps(partition_t *part) argument 334 erase_xfer(partition_t *part, uint16_t xfernum) argument 377 partition_t *part; local 406 prepare_xfer(partition_t *part, int i) argument 464 copy_erase_unit(partition_t *part, uint16_t srcunit, uint16_t xferunit) argument 608 reclaim_block(partition_t *part) argument 715 dump_lists(partition_t *part) argument 727 find_free(partition_t *part) argument 788 ftl_read(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) argument 834 set_bam_entry(partition_t *part, uint32_t log_addr, uint32_t virt_addr) argument 902 ftl_write(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) argument 984 partition_t *part = (void *)dev; local 1012 partition_t *part = (void *)dev; local 1034 ftl_freepart(partition_t *part) argument [all...] |
H A D | rfd_ftl.c | 92 static int build_block_map(struct partition *part, int block_no) argument 94 struct block *block = &part->blocks[block_no]; 97 block->offset = part->block_size * block_no; 99 if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) { 106 for (i=0; i<part->data_sectors_per_block; i++) { 109 entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]); 122 if (entry >= part->sector_count) { 126 part->mbd.mtd->name, block_no, i, entry); 130 if (part->sector_map[entry] != -1) { 133 part 150 scan_header(struct partition *part) argument 243 struct partition *part = (struct partition*)dev; local 271 struct partition *part; local 326 erase_block(struct partition *part, int block) argument 357 move_block_contents(struct partition *part, int block_no, u_long *old_sector) argument 445 reclaim_block(struct partition *part, u_long *old_sector) argument 517 find_free_block(struct partition *part) argument 541 find_writable_block(struct partition *part, u_long *old_sector) argument 583 mark_sector_deleted(struct partition *part, u_long old_addr) argument 621 find_free_sector(const struct partition *part, const struct block *block) argument 642 struct partition *part = (struct partition*)dev; local 710 struct partition *part = (struct partition*)dev; local 751 struct partition *part = (struct partition*)dev; local 762 struct partition *part; local 807 struct partition *part = (struct partition*)dev; local [all...] |
H A D | mtdpart.c | 63 struct mtd_part *part = PART(mtd); local 67 stats = part->master->ecc_stats; 68 res = part->master->_read(part->master, from + part->offset, len, 72 mtd->ecc_stats.corrected += part->master->ecc_stats.corrected - stats.corrected; 74 mtd->ecc_stats.failed += part->master->ecc_stats.failed - stats.failed; 82 struct mtd_part *part = PART(mtd); local 84 return part->master->_point(part 90 struct mtd_part *part = PART(mtd); local 100 struct mtd_part *part = PART(mtd); local 110 struct mtd_part *part = PART(mtd); local 148 struct mtd_part *part = PART(mtd); local 156 struct mtd_part *part = PART(mtd); local 163 struct mtd_part *part = PART(mtd); local 171 struct mtd_part *part = PART(mtd); local 178 struct mtd_part *part = PART(mtd); local 186 struct mtd_part *part = PART(mtd); local 194 struct mtd_part *part = PART(mtd); local 206 struct mtd_part *part = PART(mtd); local 214 struct mtd_part *part = PART(mtd); local 221 struct mtd_part *part = PART(mtd); local 228 struct mtd_part *part = PART(mtd); local 244 struct mtd_part *part = PART(instr->mtd); local 257 struct mtd_part *part = PART(mtd); local 263 struct mtd_part *part = PART(mtd); local 269 struct mtd_part *part = PART(mtd); local 275 struct mtd_part *part = PART(mtd); local 281 struct mtd_part *part = PART(mtd); local 287 struct mtd_part *part = PART(mtd); local 293 struct mtd_part *part = PART(mtd); local 300 struct mtd_part *part = PART(mtd); local 342 allocate_partition(struct mtd_info *master, const struct mtd_partition *part, int partno, uint64_t cur_offset) argument 537 struct mtd_partition part; local 747 struct mtd_part *part; local [all...] |
H A D | cmdlinepart.c | 322 struct cmdline_mtd_partition *part; local 329 for(part = partitions; part; part = part->next) 331 if ((!mtd_id) || (!strcmp(part->mtd_id, mtd_id))) 333 for(i = 0, offset = 0; i < part->num_parts; i++) 335 if (part->parts[i].offset == OFFSET_CONTINUOUS) 336 part->parts[i].offset = offset; 338 offset = part [all...] |
H A D | ofpart.c | 104 } *part; local 114 part = of_get_property(dp, "partitions", &plen); 115 if (!part) 121 nr_parts = plen / sizeof(part[0]); 130 (*pparts)[i].offset = be32_to_cpu(part->offset); 131 (*pparts)[i].size = be32_to_cpu(part->len) & ~1; 133 if (be32_to_cpu(part->len) & 1) 146 part++;
|
/drivers/acpi/acpica/ |
H A D | utmath.c | 68 struct uint64_struct part; member in union:uint64_overlay 112 ACPI_DIV_64_BY_32(0, dividend_ovl.part.hi, divisor, 113 quotient.part.hi, remainder32); 114 ACPI_DIV_64_BY_32(remainder32, dividend_ovl.part.lo, divisor, 115 quotient.part.lo, remainder32); 169 if (divisor.part.hi == 0) { 174 remainder.part.hi = 0; 180 ACPI_DIV_64_BY_32(0, dividend.part.hi, divisor.part.lo, 181 quotient.part [all...] |
/drivers/net/ethernet/sfc/ |
H A D | mtd.c | 57 struct efx_mtd_partition part[0]; member in struct:efx_mtd 60 #define efx_for_each_partition(part, efx_mtd) \ 61 for ((part) = &(efx_mtd)->part[0]; \ 62 (part) != &(efx_mtd)->part[(efx_mtd)->n_parts]; \ 63 (part)++) 74 efx_spi_slow_wait(struct efx_mtd_partition *part, bool uninterruptible) argument 76 struct efx_mtd *efx_mtd = part->mtd.priv; 96 pr_err("%s: timed out waiting for %s\n", part 136 efx_spi_erase(struct efx_mtd_partition *part, loff_t start, size_t len) argument 204 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 214 efx_mtd_remove_partition(struct efx_mtd_partition *part) argument 229 struct efx_mtd_partition *part; local 239 struct efx_mtd_partition *part; local 254 struct efx_mtd_partition *part; local 320 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 338 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 355 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 373 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 463 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 487 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 519 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 550 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 595 struct efx_mtd_partition *part = &efx_mtd->part[part_id]; local 629 struct efx_mtd_partition *part; local [all...] |
/drivers/misc/sgi-xp/ |
H A D | xpc_partition.c | 12 * This is the part of XPC that detects the presence/absence of 270 xpc_partition_disengaged(struct xpc_partition *part) argument 272 short partid = XPC_PARTID(part); 276 if (part->disengage_timeout) { 278 if (time_is_after_jiffies(part->disengage_timeout)) { 294 part->disengage_timeout = 0; 298 del_singleshot_timer_sync(&part->disengage_timer); 300 DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING && 301 part->act_state != XPC_P_AS_INACTIVE); 302 if (part 314 xpc_mark_partition_active(struct xpc_partition *part) argument 338 xpc_deactivate_partition(const int line, struct xpc_partition *part, enum xp_retval reason) argument 386 xpc_mark_partition_inactive(struct xpc_partition *part) argument 528 struct xpc_partition *part; local [all...] |
H A D | xpc_main.c | 15 * partition This part detects the presence/absence of other 19 * channel This part manages the channels and sends/receives 63 .init_name = "", /* set to "part" at xpc_init() time */ 173 struct xpc_partition *part = (struct xpc_partition *)data; local 175 DBUG_ON(time_is_after_jiffies(part->disengage_timeout)); 177 (void)xpc_partition_disengaged(part); 179 DBUG_ON(part->disengage_timeout != 0); 180 DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part))); 223 struct xpc_partition *part; local 235 part 343 xpc_channel_mgr(struct xpc_partition *part) argument 404 xpc_setup_ch_structures(struct xpc_partition *part) argument 495 xpc_teardown_ch_structures(struct xpc_partition *part) argument 537 struct xpc_partition *part = &xpc_partitions[partid]; local 587 xpc_activate_partition(struct xpc_partition *part) argument 651 xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch) argument 688 struct xpc_partition *part = &xpc_partitions[partid]; local 780 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 854 struct xpc_partition *part; local 903 struct xpc_partition *part; local 955 struct xpc_partition *part; local 1089 struct xpc_partition *part; local [all...] |
H A D | xpc_uv.c | 371 struct xpc_partition *part; local 378 part = &xpc_partitions[partid]; 380 if (part->sn.uv.act_state_req == 0) 386 act_state_req = part->sn.uv.act_state_req; 387 part->sn.uv.act_state_req = 0; 391 if (part->act_state == XPC_P_AS_INACTIVE) 392 xpc_activate_partition(part); 393 else if (part->act_state == XPC_P_AS_DEACTIVATING) 394 XPC_DEACTIVATE_PARTITION(part, xpReactivating); 397 if (part 418 xpc_handle_activate_mq_msg_uv(struct xpc_partition *part, struct xpc_activate_mq_msghdr_uv *msg_hdr, int part_setup, int *wakeup_hb_checker) argument 621 struct xpc_partition *part; local 670 xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size, int msg_type) argument 726 xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg, size_t msg_size, int msg_type) argument 740 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 756 xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req) argument 869 xpc_get_remote_heartbeat_uv(struct xpc_partition *part) argument 895 struct xpc_partition *part = &xpc_partitions[partid]; local 922 xpc_request_partition_reactivation_uv(struct xpc_partition *part) argument 928 xpc_request_partition_deactivation_uv(struct xpc_partition *part) argument 946 xpc_cancel_partition_deactivation_request_uv(struct xpc_partition *part) argument 1010 xpc_setup_ch_structures_uv(struct xpc_partition *part) argument 1029 xpc_teardown_ch_structures_uv(struct xpc_partition *part) argument 1036 xpc_make_first_contact_uv(struct xpc_partition *part) argument 1065 xpc_get_chctl_all_flags_uv(struct xpc_partition *part) argument 1254 xpc_send_chctl_local_msgrequest_uv(struct xpc_partition *part, int ch_number) argument 1277 xpc_indicate_partition_engaged_uv(struct xpc_partition *part) argument 1286 xpc_indicate_partition_disengaged_uv(struct xpc_partition *part) argument 1404 xpc_handle_notify_mq_msg_uv(struct xpc_partition *part, struct xpc_notify_mq_msg_uv *msg) argument 1477 struct xpc_partition *part; local 1508 xpc_process_msg_chctl_flags_uv(struct xpc_partition *part, int ch_number) argument [all...] |
H A D | xpc_sn2.c | 268 xpc_check_for_sent_chctl_flags_sn2(struct xpc_partition *part) argument 273 chctl.all_flags = xpc_receive_IRQ_amo_sn2(part->sn.sn2. 278 spin_lock_irqsave(&part->chctl_lock, irq_flags); 279 part->chctl.all_flags |= chctl.all_flags; 280 spin_unlock_irqrestore(&part->chctl_lock, irq_flags); 283 "0x%llx\n", XPC_PARTID(part), chctl.all_flags); 285 xpc_wakeup_channel_mgr(part); 308 struct xpc_partition *part = &xpc_partitions[partid]; local 312 if (xpc_part_ref(part)) { 313 xpc_check_for_sent_chctl_flags_sn2(part); 326 xpc_check_for_dropped_notify_IRQ_sn2(struct xpc_partition *part) argument 348 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 384 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 467 xpc_indicate_partition_engaged_sn2(struct xpc_partition *part) argument 494 xpc_indicate_partition_disengaged_sn2(struct xpc_partition *part) argument 768 xpc_get_remote_heartbeat_sn2(struct xpc_partition *part) argument 837 xpc_request_partition_reactivation_sn2(struct xpc_partition *part) argument 843 xpc_request_partition_deactivation_sn2(struct xpc_partition *part) argument 880 xpc_cancel_partition_deactivation_request_sn2(struct xpc_partition *part) argument 921 xpc_update_partition_info_sn2(struct xpc_partition *part, u8 remote_rp_version, unsigned long *remote_rp_ts_jiffies, unsigned long remote_rp_pa, unsigned long remote_vars_pa, struct xpc_vars_sn2 *remote_vars) argument 995 struct xpc_partition *part; local 1168 xpc_setup_ch_structures_sn2(struct xpc_partition *part) argument 1290 xpc_teardown_ch_structures_sn2(struct xpc_partition *part) argument 1324 xpc_pull_remote_cachelines_sn2(struct xpc_partition *part, void *dst, const unsigned long src_pa, size_t cnt) argument 1349 xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) argument 1446 xpc_make_first_contact_sn2(struct xpc_partition *part) argument 1503 xpc_get_chctl_all_flags_sn2(struct xpc_partition *part) argument 1808 xpc_process_msg_chctl_flags_sn2(struct xpc_partition *part, int ch_number) argument 1912 struct xpc_partition *part = &xpc_partitions[ch->partid]; local [all...] |
H A D | xpc_channel.c | 12 * This is the part of XPC that manages the channels and 82 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 101 if (part->act_state == XPC_P_AS_DEACTIVATING) { 157 atomic_dec(&part->nchannels_active); 168 if (part->act_state != XPC_P_AS_DEACTIVATING) { 170 spin_lock(&part->chctl_lock); 171 part->chctl.flags[ch->number] |= 173 spin_unlock(&part->chctl_lock); 183 xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number, argument 188 &part 546 xpc_process_sent_chctl_flags(struct xpc_partition *part) argument 617 xpc_partition_going_down(struct xpc_partition *part, enum xp_retval reason) argument 658 struct xpc_partition *part; local 714 struct xpc_partition *part; local 875 struct xpc_partition *part = &xpc_partitions[partid]; local 926 struct xpc_partition *part = &xpc_partitions[partid]; local 1000 struct xpc_partition *part = &xpc_partitions[partid]; local [all...] |
H A D | xpc.h | 75 * vars part (ia64-sn2 only) 80 * which are partition specific (vars part). These are setup by XPC. 164 * The vars_part MAGIC numbers play a part in the first contact protocol. 678 int activate_IRQ_phys_cpuid; /* active part's act/deact phys cpuid */ 681 unsigned long remote_vars_part_pa; /* paddr of partition's vars part */ 759 atomic_t nchannels_engaged; /* #of channels engaged with remote part */ 937 xpc_wakeup_channel_mgr(struct xpc_partition *part) argument 939 if (atomic_inc_return(&part->channel_mgr_requests) == 1) 940 wake_up(&part->channel_mgr_wq); 971 xpc_part_deref(struct xpc_partition *part) argument 981 xpc_part_ref(struct xpc_partition *part) argument [all...] |
/drivers/macintosh/ |
H A D | nvram.c | 84 int part, offset; local 85 if (copy_from_user(&part, (void __user*)arg, sizeof(part)) != 0) 87 if (part < pmac_nvram_OF || part > pmac_nvram_NR) 89 offset = pmac_get_partition(part);
|
H A D | smu.c | 1020 const struct smu_sdbp_header *part; local 1037 part = of_get_property(smu->of_node, pname, size); 1038 if (part == NULL) { 1040 part = smu_create_sdb_partition(id); 1041 if (part != NULL && size) 1042 *size = part->len << 2; 1045 return part; 1128 const struct smu_sdbp_header *part; local 1129 part = __smu_get_sdb_partition(hdr.cmd, NULL, 1); 1130 if (part [all...] |
/drivers/gpu/drm/nouveau/ |
H A D | nvc0_vram.c | 114 int ret, part; local 123 for (part = 0; part < parts; part++) { 124 if (!(pmask & (1 << part))) { 125 u32 psize = nv_rd32(dev, 0x11020c + (part * 0x1000)); 132 NV_DEBUG(dev, "%d: mem_amount 0x%08x\n", part, psize);
|
H A D | nouveau_sgdma.c | 222 u32 part = (cnt > max) ? max : cnt; local 223 nv44_sgdma_fill(pgt, list, pte, part); 224 pte += (part << 2); 225 list += part; 226 cnt -= part; 258 u32 part = (cnt > max) ? max : cnt; local 259 nv44_sgdma_fill(pgt, NULL, pte, part); 260 pte += (part << 2); 261 cnt -= part;
|
/drivers/char/ |
H A D | generic_nvram.c | 100 int part, offset; local 104 if (copy_from_user(&part, (void __user*)arg, sizeof(part)) != 0) 106 if (part < pmac_nvram_OF || part > pmac_nvram_NR) 108 offset = pmac_get_partition(part);
|
/drivers/net/wireless/wl12xx/ |
H A D | io.h | 2 * This file is part of wl1271 101 if ((addr >= wl->part.reg.start) && 102 (addr < wl->part.reg.start + wl->part.reg.size)) 103 return addr - wl->part.reg.start + wl->part.mem.size; 105 return addr - wl->part.mem.start;
|
/drivers/isdn/hisax/ |
H A D | hscx.h | 41 extern void inithscxisac(struct IsdnCardState *cs, int part);
|
/drivers/memstick/host/ |
H A D | Kconfig | 40 the Ricoh R5C592 MemoryStick card reader (which is part of 5 in one
|
/drivers/mmc/card/ |
H A D | block.c | 86 struct list_head part; member in struct:mmc_blk_data 219 list_for_each_entry(part_md, &md->part, part) 1518 INIT_LIST_HEAD(&md->part); 1620 list_add(&part_md->part, &md->part); 1644 if (card->part[idx].size) { 1646 card->part[idx].part_cfg, 1647 card->part[idx].size >> 9, 1648 card->part[id [all...] |
/drivers/block/ |
H A D | z2ram.c | 335 static struct kobject *z2_find(dev_t dev, int *part, void *data) argument 337 *part = 0;
|
/drivers/char/ipmi/ |
H A D | Kconfig | 35 string (byte 5 of the event) and part of the string (the rest of the
|
/drivers/cpufreq/ |
H A D | Kconfig.arm | 57 This adds the CPUFreq driver common part for Samsung
|