/drivers/media/video/ |
H A D | aptina-pll.c | 30 const struct aptina_pll_limits *limits, 43 if (pll->ext_clock < limits->ext_clock_min || 44 pll->ext_clock > limits->ext_clock_max) { 49 if (pll->pix_clock == 0 || pll->pix_clock > limits->pix_clock_max) { 69 mf_min = DIV_ROUND_UP(limits->m_min, pll->m); 70 mf_min = max(mf_min, limits->out_clock_min / 71 (pll->ext_clock / limits->n_min * pll->m)); 72 mf_min = max(mf_min, limits->n_min * limits->p1_min / div); 73 mf_max = limits 29 aptina_pll_calculate(struct device *dev, const struct aptina_pll_limits *limits, struct aptina_pll *pll) argument [all...] |
H A D | mt9m032.c | 193 /* hardware limits to 11 bit values */ 246 static const struct aptina_pll_limits limits = { local 270 ret = aptina_pll_calculate(&client->dev, &limits, &pll);
|
H A D | mt9p031.c | 182 static const struct aptina_pll_limits limits = { local 204 return aptina_pll_calculate(&client->dev, &limits, &mt9p031->pll);
|
/drivers/gpu/drm/nouveau/ |
H A D | nva3_pm.c | 112 struct pll_lims limits; local 162 ret = get_pll_limits(dev, pll, &limits); 166 limits.refclk = read_clk(dev, clk - 0x10, true); 167 if (!limits.refclk) 170 ret = nva3_calc_pll(dev, &limits, khz, &N, NULL, &M, &P);
|
H A D | nvc0_pm.c | 213 struct pll_lims limits; local 216 ret = get_pll_limits(dev, 0x137000 + (clk * 0x20), &limits); 220 limits.refclk = read_div(dev, clk, 0x137120, 0x137140); 221 if (!limits.refclk) 224 ret = nva3_calc_pll(dev, &limits, freq, &N, NULL, &M, &P);
|
/drivers/md/ |
H A D | dm-stripe.c | 391 struct queue_limits *limits) 396 blk_limits_io_min(limits, chunk_size); 397 blk_limits_io_opt(limits, chunk_size * sc->stripes); 390 stripe_io_hints(struct dm_target *ti, struct queue_limits *limits) argument
|
H A D | dm-raid.c | 1227 static void raid_io_hints(struct dm_target *ti, struct queue_limits *limits) argument 1233 blk_limits_io_min(limits, chunk_size); 1234 blk_limits_io_opt(limits, chunk_size * (conf->raid_disks - conf->max_degraded));
|
H A D | dm-verity.c | 590 static void verity_io_hints(struct dm_target *ti, struct queue_limits *limits) argument 594 if (limits->logical_block_size < 1 << v->data_dev_block_bits) 595 limits->logical_block_size = 1 << v->data_dev_block_bits; 597 if (limits->physical_block_size < 1 << v->data_dev_block_bits) 598 limits->physical_block_size = 1 << v->data_dev_block_bits; 600 blk_limits_io_min(limits, limits->logical_block_size);
|
H A D | dm-table.c | 366 struct queue_limits *limits = data; local 371 limits->logical_block_size >> SECTOR_SHIFT; 411 limits->logical_block_size, bdevname(bdev, b)); 420 limits->logical_block_size, bdevname(bdev, b)); 519 struct queue_limits *limits = data; local 525 DMWARN("%s: Cannot set limits for nonexistent device %s", 530 if (bdev_stack_limits(limits, bdev, start) < 0) 535 q->limits.physical_block_size, 536 q->limits.logical_block_size, 537 q->limits 671 validate_hardware_logical_block_alignment(struct dm_table *table, struct queue_limits *limits) argument 1218 dm_calculate_queue_limits(struct dm_table *table, struct queue_limits *limits) argument 1371 dm_table_set_restrictions(struct dm_table *t, struct request_queue *q, struct queue_limits *limits) argument [all...] |
H A D | dm-thin.c | 1233 * limits->discard_granularity. It ignores any 2022 * stacking of discard limits (this keeps the pool and 2023 * thin devices' discard limits consistent). 2430 static void set_discard_limits(struct pool *pool, struct queue_limits *limits) argument 2433 * FIXME: these limits may be incompatible with the pool's data device 2435 limits->max_discard_sectors = pool->sectors_per_block; 2441 limits->discard_granularity = pool->sectors_per_block << SECTOR_SHIFT; 2442 limits->discard_zeroes_data = pool->pf.zero_new_blocks; 2445 static void pool_io_hints(struct dm_target *ti, struct queue_limits *limits) argument 2450 blk_limits_io_min(limits, 2720 thin_io_hints(struct dm_target *ti, struct queue_limits *limits) argument [all...] |
H A D | dm.c | 577 DMWARN("Start sector is beyond the geometry limits."); 2054 struct queue_limits *limits) 2093 dm_table_set_restrictions(t, q, limits); 2391 struct queue_limits limits; local 2400 r = dm_calculate_queue_limits(table, &limits); 2406 map = __bind(md, table, &limits); 2053 __bind(struct mapped_device *md, struct dm_table *t, struct queue_limits *limits) argument
|
/drivers/net/wireless/p54/ |
H A D | fwio.c | 454 struct pda_channel_output_limit *limits = local 459 limits->val_bpsk; 461 limits->val_qpsk; 463 limits->val_16qam; 465 limits->val_64qam;
|
/drivers/target/ |
H A D | target_core_file.c | 115 struct queue_limits *limits; local 174 * Setup the local scope queue_limits from struct request_queue->limits 178 limits = &dev_limits.limits; 179 limits->logical_block_size = bdev_logical_block_size(inode->i_bdev); 180 limits->max_hw_sectors = queue_max_hw_sectors(q); 181 limits->max_sectors = queue_max_sectors(q); 202 limits = &dev_limits.limits; 203 limits [all...] |
H A D | target_core_iblock.c | 97 struct queue_limits *limits; local 127 * Setup the local scope queue_limits from struct request_queue->limits 131 limits = &dev_limits.limits; 132 limits->logical_block_size = bdev_logical_block_size(bd); 133 limits->max_hw_sectors = UINT_MAX; 134 limits->max_sectors = UINT_MAX; 153 q->limits.max_discard_sectors; 159 q->limits.discard_granularity >> 9; 161 q->limits [all...] |
H A D | target_core_device.c | 849 struct queue_limits *limits = &dev_limits->limits; local 878 dev->se_sub_dev->se_dev_attrib.hw_block_size = limits->logical_block_size; 879 dev->se_sub_dev->se_dev_attrib.block_size = limits->logical_block_size; 883 dev->se_sub_dev->se_dev_attrib.hw_max_sectors = limits->max_hw_sectors; 887 limits->max_sectors = se_dev_align_max_sectors(limits->max_sectors, 888 limits->logical_block_size); 889 dev->se_sub_dev->se_dev_attrib.max_sectors = limits->max_sectors; 891 * Set fabric_max_sectors, which is reported in block limits [all...] |
H A D | target_core_pscsi.c | 312 struct queue_limits *limits; local 324 * Setup the local scope queue_limits from struct request_queue->limits 328 limits = &dev_limits.limits; 329 limits->logical_block_size = sd->sector_size; 330 limits->max_hw_sectors = min_t(int, sd->host->max_sectors, queue_max_hw_sectors(q)); 331 limits->max_sectors = min_t(int, sd->host->max_sectors, queue_max_sectors(q));
|
/drivers/macintosh/ |
H A D | therm_adt746x.c | 78 u8 limits[3]; member in struct:thermostat 173 th->limits[0], th->limits[1], th->limits[2], 276 " limits: %d,%d,%d;" 279 th->limits[0], th->limits[1], th->limits[2], 297 int var = th->temps[i] - th->limits[i]; 374 th->limits[ [all...] |
/drivers/video/via/ |
H A D | hw.c | 1388 static struct via_pll_config get_pll_config(struct pll_limit *limits, int size, argument 1396 cur.rshift = limits[i].rshift; 1397 cur.divisor = limits[i].divisor; 1408 if (cur.multiplier < limits[i].multiplier_min) 1409 cur.multiplier = limits[i].multiplier_min; 1410 else if (cur.multiplier > limits[i].multiplier_max) 1411 cur.multiplier = limits[i].multiplier_max;
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_dev.h | 337 struct mthca_limits limits; member in struct:mthca_dev
|
/drivers/platform/x86/ |
H A D | intel_ips.c | 258 /* Per-SKU limits */ 332 struct ips_mcp_limits *limits; member in struct:ips_driver 411 * as low as the platform limits will allow (though we could go lower there 601 * mcp_exceeded - check whether we're outside our thermal & power limits 629 * cpu_exceeded - check whether a CPU core is outside its limits 643 if (avg > (ips->limits->core_temp_limit * 100)) 668 if (ips->mch_avg_temp > (ips->limits->mch_temp_limit * 100)) 678 * verify_limits - verify BIOS provided limits 681 * BIOS can optionally provide non-default limits for power and temp. Check 687 if (ips->mcp_power_limit < ips->limits 1370 struct ips_mcp_limits *limits = NULL; local [all...] |
/drivers/video/matrox/ |
H A D | matroxfb_base.h | 490 } limits; member in struct:matrox_fb_info
|