/drivers/crypto/amcc/ |
H A D | crypto4xx_sa.c | 37 u32 get_dynamic_sa_offset_iv_field(struct crypto4xx_ctx *ctx) argument 42 if (ctx->direction == DIR_INBOUND) 43 cts.w = ((struct dynamic_sa_ctl *)(ctx->sa_in))->sa_contents; 45 cts.w = ((struct dynamic_sa_ctl *)(ctx->sa_out))->sa_contents; 60 u32 get_dynamic_sa_offset_state_ptr_field(struct crypto4xx_ctx *ctx) argument 65 if (ctx->direction == DIR_INBOUND) 66 cts.w = ((struct dynamic_sa_ctl *) ctx->sa_in)->sa_contents; 68 cts.w = ((struct dynamic_sa_ctl *) ctx->sa_out)->sa_contents; 87 u32 get_dynamic_sa_iv_size(struct crypto4xx_ctx *ctx) argument 91 if (ctx 98 get_dynamic_sa_offset_key_field(struct crypto4xx_ctx *ctx) argument [all...] |
/drivers/gpu/host1x/ |
H A D | debug.h | 27 void (*fn)(void *ctx, const char *str, size_t len); 28 void *ctx; member in struct:output 32 static inline void write_to_seqfile(void *ctx, const char *str, size_t len) argument 34 seq_write((struct seq_file *)ctx, str, len); 37 static inline void write_to_printk(void *ctx, const char *str, size_t len) argument
|
/drivers/media/platform/s5p-mfc/ |
H A D | s5p_mfc_intr.c | 54 int s5p_mfc_wait_for_done_ctx(struct s5p_mfc_ctx *ctx, argument 60 ret = wait_event_interruptible_timeout(ctx->queue, 61 (ctx->int_cond && (ctx->int_type == command 62 || ctx->int_type == S5P_MFC_R2H_CMD_ERR_RET)), 65 ret = wait_event_timeout(ctx->queue, 66 (ctx->int_cond && (ctx->int_type == command 67 || ctx->int_type == S5P_MFC_R2H_CMD_ERR_RET)), 71 mfc_err("Interrupt (ctx 85 s5p_mfc_clean_ctx_int_flags(struct s5p_mfc_ctx *ctx) argument [all...] |
H A D | s5p_mfc_cmd_v5.c | 75 static int s5p_mfc_open_inst_cmd_v5(struct s5p_mfc_ctx *ctx) argument 77 struct s5p_mfc_dev *dev = ctx->dev; 82 mfc_debug(2, "Getting instance number (codec: %d)\n", ctx->codec_mode); 83 dev->curr_ctx = ctx->num; 85 switch (ctx->codec_mode) { 117 h2r_args.arg[2] = ctx->ctx.ofs; 118 h2r_args.arg[3] = ctx->ctx.size; 123 ctx 128 s5p_mfc_close_inst_cmd_v5(struct s5p_mfc_ctx *ctx) argument [all...] |
H A D | s5p_mfc_cmd_v6.c | 67 static int s5p_mfc_open_inst_cmd_v6(struct s5p_mfc_ctx *ctx) argument 69 struct s5p_mfc_dev *dev = ctx->dev; 73 mfc_debug(2, "Requested codec mode: %d\n", ctx->codec_mode); 74 dev->curr_ctx = ctx->num; 75 switch (ctx->codec_mode) { 119 mfc_write(dev, ctx->ctx.dma, S5P_FIMV_CONTEXT_MEM_ADDR_V6); 120 mfc_write(dev, ctx->ctx.size, S5P_FIMV_CONTEXT_MEM_SIZE_V6); 128 static int s5p_mfc_close_inst_cmd_v6(struct s5p_mfc_ctx *ctx) argument [all...] |
H A D | s5p_mfc_ctrl.c | 379 int s5p_mfc_open_mfc_inst(struct s5p_mfc_dev *dev, struct s5p_mfc_ctx *ctx) argument 383 ret = s5p_mfc_hw_call(dev->mfc_ops, alloc_instance_buffer, ctx); 389 if (ctx->type == MFCINST_DECODER) { 391 alloc_dec_temp_buffers, ctx); 398 set_work_bit_irqsave(ctx); 399 s5p_mfc_clean_ctx_int_flags(ctx); 401 if (s5p_mfc_wait_for_done_ctx(ctx, 409 mfc_debug(2, "Got instance number: %d\n", ctx->inst_no); 413 if (ctx->type == MFCINST_DECODER) 414 s5p_mfc_hw_call_void(dev->mfc_ops, release_dec_desc_buffer, ctx); 421 s5p_mfc_close_mfc_inst(struct s5p_mfc_dev *dev, struct s5p_mfc_ctx *ctx) argument [all...] |
/drivers/clk/samsung/ |
H A D | clk-exynos5440.c | 96 struct samsung_clk_provider *ctx; local 105 ctx = samsung_clk_init(np, reg_base, CLK_NR_CLKS); 106 if (!ctx) 109 samsung_clk_of_register_fixed_ext(ctx, exynos5440_fixed_rate_ext_clks, 115 samsung_clk_register_fixed_rate(ctx, exynos5440_fixed_rate_clks, 117 samsung_clk_register_fixed_factor(ctx, exynos5440_fixed_factor_clks, 119 samsung_clk_register_mux(ctx, exynos5440_mux_clks, 121 samsung_clk_register_div(ctx, exynos5440_div_clks, 123 samsung_clk_register_gate(ctx, exynos5440_gate_clks, 126 samsung_clk_of_add_provider(np, ctx); [all...] |
H A D | clk.c | 54 struct samsung_clk_provider *ctx; local 58 ctx = kzalloc(sizeof(struct samsung_clk_provider), GFP_KERNEL); 59 if (!ctx) 69 ctx->reg_base = base; 70 ctx->clk_data.clks = clk_table; 71 ctx->clk_data.clk_num = nr_clks; 72 spin_lock_init(&ctx->lock); 74 return ctx; 78 struct samsung_clk_provider *ctx) 82 &ctx 77 samsung_clk_of_add_provider(struct device_node *np, struct samsung_clk_provider *ctx) argument 88 samsung_clk_add_lookup(struct samsung_clk_provider *ctx, struct clk *clk, unsigned int id) argument 96 samsung_clk_register_alias(struct samsung_clk_provider *ctx, struct samsung_clock_alias *list, unsigned int nr_clk) argument 130 samsung_clk_register_fixed_rate(struct samsung_clk_provider *ctx, struct samsung_fixed_rate_clock *list, unsigned int nr_clk) argument 159 samsung_clk_register_fixed_factor(struct samsung_clk_provider *ctx, struct samsung_fixed_factor_clock *list, unsigned int nr_clk) argument 179 samsung_clk_register_mux(struct samsung_clk_provider *ctx, struct samsung_mux_clock *list, unsigned int nr_clk) argument 211 samsung_clk_register_div(struct samsung_clk_provider *ctx, struct samsung_div_clock *list, unsigned int nr_clk) argument 250 samsung_clk_register_gate(struct samsung_clk_provider *ctx, struct samsung_gate_clock *list, unsigned int nr_clk) argument 285 samsung_clk_of_register_fixed_ext(struct samsung_clk_provider *ctx, struct samsung_fixed_rate_clock *fixed_rate_clk, unsigned int nr_fixed_rate_clk, const struct of_device_id *clk_matches) argument [all...] |
/drivers/misc/cxl/ |
H A D | context.c | 37 int cxl_context_init(struct cxl_context *ctx, struct cxl_afu *afu, bool master) argument 41 spin_lock_init(&ctx->sste_lock); 42 ctx->afu = afu; 43 ctx->master = master; 44 ctx->pid = NULL; /* Set in start work ioctl */ 53 i = cxl_alloc_sst(ctx); 57 INIT_WORK(&ctx->fault_work, cxl_handle_fault); 59 init_waitqueue_head(&ctx->wq); 60 spin_lock_init(&ctx->lock); 62 ctx 103 cxl_context_iomap(struct cxl_context *ctx, struct vm_area_struct *vma) argument 135 __detach_context(struct cxl_context *ctx) argument 158 cxl_context_detach(struct cxl_context *ctx) argument 168 struct cxl_context *ctx; local 181 cxl_context_free(struct cxl_context *ctx) argument [all...] |
H A D | main.c | 34 static inline void _cxl_slbia(struct cxl_context *ctx, struct mm_struct *mm) argument 38 if (!(task = get_pid_task(ctx->pid, PIDTYPE_PID))) { 40 __func__, pid_nr(ctx->pid)); 48 ctx->afu->adapter->adapter_num, ctx->afu->slice, ctx->pe); 50 spin_lock_irqsave(&ctx->sste_lock, flags); 51 memset(ctx->sstp, 0, ctx->sst_size); 52 spin_unlock_irqrestore(&ctx 63 struct cxl_context *ctx; local 91 cxl_alloc_sst(struct cxl_context *ctx) argument [all...] |
/drivers/power/reset/ |
H A D | syscon-reboot.c | 37 struct syscon_reboot_context *ctx = local 42 regmap_write(ctx->map, ctx->offset, ctx->mask); 52 struct syscon_reboot_context *ctx; local 56 ctx = devm_kzalloc(&pdev->dev, sizeof(*ctx), GFP_KERNEL); 57 if (!ctx) 60 ctx->map = syscon_regmap_lookup_by_phandle(dev->of_node, "regmap"); 61 if (IS_ERR(ctx [all...] |
H A D | xgene-reboot.c | 45 struct xgene_reboot_context *ctx = xgene_restart_ctx; local 49 if (ctx) 50 writel(ctx->mask, ctx->csr); 56 dev_emerg(&ctx->pdev->dev, "Unable to restart system\n"); 61 struct xgene_reboot_context *ctx; local 63 ctx = devm_kzalloc(&pdev->dev, sizeof(*ctx), GFP_KERNEL); 64 if (!ctx) { 69 ctx [all...] |
/drivers/gpu/drm/i915/ |
H A D | intel_lrc.h | 66 struct intel_context *ctx); 67 void intel_lr_context_free(struct intel_context *ctx); 68 int intel_lr_context_deferred_create(struct intel_context *ctx, 75 struct intel_context *ctx, 84 * @ctx: Context to submit to the ELSP. 102 struct intel_context *ctx; member in struct:intel_ctx_submit_request
|
/drivers/crypto/qat/qat_common/ |
H A D | qat_crypto.h | 79 struct qat_alg_session_ctx *ctx; member in struct:qat_crypto_request
|
/drivers/hwmon/ |
H A D | pwm-fan.c | 39 struct pwm_fan_ctx *ctx = dev_get_drvdata(dev); local 46 mutex_lock(&ctx->lock); 48 if (ctx->pwm_value == pwm) 52 pwm_disable(ctx->pwm); 56 duty = DIV_ROUND_UP(pwm * (ctx->pwm->period - 1), MAX_PWM); 57 ret = pwm_config(ctx->pwm, duty, ctx->pwm->period); 61 if (ctx->pwm_value == 0) { 62 ret = pwm_enable(ctx->pwm); 68 ctx 79 struct pwm_fan_ctx *ctx = dev_get_drvdata(dev); local 97 struct pwm_fan_ctx *ctx; local 144 struct pwm_fan_ctx *ctx = platform_get_drvdata(pdev); local 154 struct pwm_fan_ctx *ctx = dev_get_drvdata(dev); local 163 struct pwm_fan_ctx *ctx = dev_get_drvdata(dev); local [all...] |
/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.h | 39 struct usnic_ib_ucontext *ctx; member in struct:usnic_ib_qp_grp
|
/drivers/net/team/ |
H A D | team_mode_activebackup.c | 71 static int ab_active_port_get(struct team *team, struct team_gsetter_ctx *ctx) argument 78 ctx->data.u32_val = active_port->dev->ifindex; 80 ctx->data.u32_val = 0; 84 static int ab_active_port_set(struct team *team, struct team_gsetter_ctx *ctx) argument 89 if (port->dev->ifindex == ctx->data.u32_val) {
|
/drivers/net/usb/ |
H A D | huawei_cdc_ncm.c | 34 struct cdc_ncm_ctx *ctx; member in struct:huawei_cdc_ncm_state 72 struct cdc_ncm_ctx *ctx; local 84 ctx = drvstate->ctx; 91 subdriver = usb_cdc_wdm_register(ctx->control, 114 struct cdc_ncm_ctx *ctx = drvstate->ctx; local 117 drvstate->subdriver->disconnect(ctx->control); 129 struct cdc_ncm_ctx *ctx = drvstate->ctx; local 157 struct cdc_ncm_ctx *ctx = drvstate->ctx; local [all...] |
/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | commonring.c | 37 int (*cr_ring_bell)(void *ctx), 38 int (*cr_update_rptr)(void *ctx), 39 int (*cr_update_wptr)(void *ctx), 40 int (*cr_write_rptr)(void *ctx), 41 int (*cr_write_wptr)(void *ctx), void *ctx) 48 commonring->cr_ctx = ctx; 36 brcmf_commonring_register_cb(struct brcmf_commonring *commonring, int (*cr_ring_bell)(void *ctx), int (*cr_update_rptr)(void *ctx), int (*cr_update_wptr)(void *ctx), int (*cr_write_rptr)(void *ctx), int (*cr_write_wptr)(void *ctx), void *ctx) argument
|
/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | sec_gc.c | 104 void sptlrpc_gc_add_ctx(struct ptlrpc_cli_ctx *ctx) argument 106 LASSERT(list_empty(&ctx->cc_gc_chain)); 108 CDEBUG(D_SEC, "hand over ctx %p(%u->%s)\n", 109 ctx, ctx->cc_vcred.vc_uid, sec2target_str(ctx->cc_sec)); 111 list_add(&ctx->cc_gc_chain, &sec_gc_ctx_list); 121 struct ptlrpc_cli_ctx *ctx; local 126 ctx = list_entry(sec_gc_ctx_list.next, 128 list_del_init(&ctx [all...] |
/drivers/staging/skein/ |
H A D | skein_api.c | 30 int skein_ctx_prepare(struct skein_ctx *ctx, enum skein_size size) argument 32 skein_assert_ret(ctx && size, SKEIN_FAIL); 34 memset(ctx , 0, sizeof(struct skein_ctx)); 35 ctx->skein_size = size; 40 int skein_init(struct skein_ctx *ctx, size_t hash_bit_len) argument 47 skein_assert_ret(ctx, SKEIN_FAIL); 53 x = ctx->m.s256.x; 54 x_len = ctx->skein_size/8; 59 switch (ctx->skein_size) { 61 ret = skein_256_init_ext(&ctx 84 skein_mac_init(struct skein_ctx *ctx, const u8 *key, size_t key_len, size_t hash_bit_len) argument 128 skein_reset(struct skein_ctx *ctx) argument 147 skein_update(struct skein_ctx *ctx, const u8 *msg, size_t msg_byte_cnt) argument 172 skein_update_bits(struct skein_ctx *ctx, const u8 *msg, size_t msg_bit_cnt) argument 221 skein_final(struct skein_ctx *ctx, u8 *hash) argument [all...] |
/drivers/acpi/apei/ |
H A D | apei-internal.h | 14 typedef int (*apei_exec_ins_func_t)(struct apei_exec_context *ctx, 37 void apei_exec_ctx_init(struct apei_exec_context *ctx, 43 static inline void apei_exec_ctx_set_input(struct apei_exec_context *ctx, argument 46 ctx->value = input; 49 static inline u64 apei_exec_ctx_get_output(struct apei_exec_context *ctx) argument 51 return ctx->value; 54 int __apei_exec_run(struct apei_exec_context *ctx, u8 action, bool optional); 56 static inline int apei_exec_run(struct apei_exec_context *ctx, u8 action) argument 58 return __apei_exec_run(ctx, action, 0); 62 static inline int apei_exec_run_optional(struct apei_exec_context *ctx, u argument [all...] |
/drivers/base/regmap/ |
H A D | regmap-mmio.c | 92 struct regmap_mmio_context *ctx = context; local 98 if (!IS_ERR(ctx->clk)) { 99 ret = clk_enable(ctx->clk); 107 switch (ctx->val_bytes) { 109 writeb(*(u8 *)val, ctx->regs + offset); 112 writew(*(u16 *)val, ctx->regs + offset); 115 writel(*(u32 *)val, ctx->regs + offset); 119 writeq(*(u64 *)val, ctx->regs + offset); 126 val_size -= ctx->val_bytes; 127 val += ctx 139 struct regmap_mmio_context *ctx = context; local 152 struct regmap_mmio_context *ctx = context; local 199 struct regmap_mmio_context *ctx = context; local 223 struct regmap_mmio_context *ctx; local 314 struct regmap_mmio_context *ctx; local 340 struct regmap_mmio_context *ctx; local [all...] |
/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-xts.c | 101 struct ccp_ctx *ctx = crypto_tfm_ctx(crypto_ablkcipher_tfm(tfm)); local 108 memcpy(ctx->u.aes.key, key, key_len); 111 ctx->u.aes.key_len = key_len / 2; 112 sg_init_one(&ctx->u.aes.key_sg, ctx->u.aes.key, key_len); 114 return crypto_ablkcipher_setkey(ctx->u.aes.tfm_ablkcipher, key, 123 struct ccp_ctx *ctx = crypto_tfm_ctx(req->base.tfm); local 128 if (!ctx->u.aes.key_len) 142 (ctx->u.aes.key_len != AES_KEYSIZE_128)) { 146 ablkcipher_request_set_tfm(req, ctx 188 struct ccp_ctx *ctx = crypto_tfm_ctx(tfm); local 212 struct ccp_ctx *ctx = crypto_tfm_ctx(tfm); local [all...] |
/drivers/crypto/ux500/cryp/ |
H A D | cryp.c | 287 * @ctx: Crypto device context 290 struct cryp_device_context *ctx, 309 ctx->din = readl_relaxed(&src_reg->din); 311 ctx->cr = readl_relaxed(&src_reg->cr) & CRYP_CR_CONTEXT_SAVE_MASK; 315 ctx->key_4_l = readl_relaxed(&src_reg->key_4_l); 316 ctx->key_4_r = readl_relaxed(&src_reg->key_4_r); 319 ctx->key_3_l = readl_relaxed(&src_reg->key_3_l); 320 ctx->key_3_r = readl_relaxed(&src_reg->key_3_r); 323 ctx->key_2_l = readl_relaxed(&src_reg->key_2_l); 324 ctx 289 cryp_save_device_context(struct cryp_device_data *device_data, struct cryp_device_context *ctx, int cryp_mode) argument 349 cryp_restore_device_context(struct cryp_device_data *device_data, struct cryp_device_context *ctx) argument [all...] |