/* * crypto_mac_init_prov() * * Arguments: * pd: pointer to the descriptor of the provider to use for this * operation. * sid: provider session id. * mech: crypto_mechanism_t pointer. * mech_type is a valid value previously returned by * crypto_mech2id(); * When the mech's parameter is not NULL, its definition depends * on the standard definition of the mechanism. * key: pointer to a crypto_key_t structure. * tmpl: a crypto_ctx_template_t, opaque template of a context of a * MAC with the 'mech' using 'key'. 'tmpl' is created by * a previous call to crypto_create_ctx_template(). * ctxp: Pointer to a crypto_context_t. * cr: crypto_call_req_t calling conditions and call back info. * * Description: * Asynchronously submits a request for, or synchronously performs the * initialization of a MAC operation on the specified provider with * the specified session. * When possible and applicable, will internally use the pre-computed MAC * context from the context template, tmpl. * When complete and successful, 'ctxp' will contain a crypto_context_t * valid for later calls to mac_update() and mac_final(). * The caller should hold a reference on the specified provider * descriptor before calling this function. * * Context: * Process or interrupt, according to the semantics dictated by the 'cr'. * * Returns: * See comment in the beginning of the file. */ int crypto_mac_init_prov(crypto_provider_t provider, crypto_session_id_t sid, crypto_mechanism_t *mech, crypto_key_t *key, crypto_spi_ctx_template_t tmpl, crypto_context_t *ctxp, crypto_call_req_t *crq) { int rv; crypto_ctx_t *ctx; kcf_req_params_t params; kcf_provider_desc_t *pd = provider; kcf_provider_desc_t *real_provider = pd; ASSERT(KCF_PROV_REFHELD(pd)); if (pd->pd_prov_type == CRYPTO_LOGICAL_PROVIDER) { rv = kcf_get_hardware_provider(mech->cm_type, CRYPTO_MECH_INVALID, CHECK_RESTRICT(crq), pd, &real_provider, CRYPTO_FG_MAC); if (rv != CRYPTO_SUCCESS) return (rv); } /* Allocate and initialize the canonical context */ if ((ctx = kcf_new_ctx(crq, real_provider, sid)) == NULL) { if (pd->pd_prov_type == CRYPTO_LOGICAL_PROVIDER) KCF_PROV_REFRELE(real_provider); return (CRYPTO_HOST_MEMORY); } /* The fast path for SW providers. */ if (CHECK_FASTPATH(crq, pd)) { crypto_mechanism_t lmech; lmech = *mech; KCF_SET_PROVIDER_MECHNUM(mech->cm_type, real_provider, &lmech); rv = KCF_PROV_MAC_INIT(real_provider, ctx, &lmech, key, tmpl, KCF_SWFP_RHNDL(crq)); KCF_PROV_INCRSTATS(pd, rv); } else { KCF_WRAP_MAC_OPS_PARAMS(¶ms, KCF_OP_INIT, sid, mech, key, NULL, NULL, tmpl); rv = kcf_submit_request(real_provider, ctx, crq, ¶ms, B_FALSE); } if (pd->pd_prov_type == CRYPTO_LOGICAL_PROVIDER) KCF_PROV_REFRELE(real_provider); if ((rv == CRYPTO_SUCCESS) || (rv == CRYPTO_QUEUED)) *ctxp = (crypto_context_t)ctx; else { /* Release the hold done in kcf_new_ctx(). */ KCF_CONTEXT_REFRELE((kcf_context_t *)ctx->cc_framework_private); } return (rv); }
/* * Same as crypto_digest_prov(), but relies on the KCF scheduler to * choose a provider. See crypto_digest_prov() comments for more information. */ int crypto_digest(crypto_mechanism_t *mech, crypto_data_t *data, crypto_data_t *digest, crypto_call_req_t *crq) { int error; kcf_provider_desc_t *pd; kcf_req_params_t params; kcf_prov_tried_t *list = NULL; retry: /* The pd is returned held */ if ((pd = kcf_get_mech_provider(mech->cm_type, NULL, &error, list, CRYPTO_FG_DIGEST_ATOMIC, CHECK_RESTRICT(crq), data->cd_length)) == NULL) { if (list != NULL) kcf_free_triedlist(list); return (error); } /* The fast path for SW providers. */ if (CHECK_FASTPATH(crq, pd)) { crypto_mechanism_t lmech; lmech = *mech; KCF_SET_PROVIDER_MECHNUM(mech->cm_type, pd, &lmech); error = KCF_PROV_DIGEST_ATOMIC(pd, pd->pd_sid, &lmech, data, digest, KCF_SWFP_RHNDL(crq)); KCF_PROV_INCRSTATS(pd, error); } else { KCF_WRAP_DIGEST_OPS_PARAMS(¶ms, KCF_OP_ATOMIC, pd->pd_sid, mech, NULL, data, digest); /* no crypto context to carry between multiple parts. */ error = kcf_submit_request(pd, NULL, crq, ¶ms, B_FALSE); } if (error != CRYPTO_SUCCESS && error != CRYPTO_QUEUED && IS_RECOVERABLE(error)) { /* Add pd to the linked list of providers tried. */ if (kcf_insert_triedlist(&list, pd, KCF_KMFLAG(crq)) != NULL) goto retry; } if (list != NULL) kcf_free_triedlist(list); KCF_PROV_REFRELE(pd); return (error); }
/* * Same as crypto_mac_verify_prov(), but relies on the KCF scheduler to choose * a provider. See crypto_mac_verify_prov() comments for more information. */ int crypto_mac_verify(crypto_mechanism_t *mech, crypto_data_t *data, crypto_key_t *key, crypto_ctx_template_t tmpl, crypto_data_t *mac, crypto_call_req_t *crq) { int error; kcf_mech_entry_t *me; kcf_req_params_t params; kcf_provider_desc_t *pd; kcf_ctx_template_t *ctx_tmpl; crypto_spi_ctx_template_t spi_ctx_tmpl = NULL; kcf_prov_tried_t *list = NULL; retry: /* The pd is returned held */ if ((pd = kcf_get_mech_provider(mech->cm_type, &me, &error, list, CRYPTO_FG_MAC_ATOMIC, CHECK_RESTRICT(crq), data->cd_length)) == NULL) { if (list != NULL) kcf_free_triedlist(list); return (error); } /* * For SW providers, check the validity of the context template * It is very rare that the generation number mis-matches, so * is acceptable to fail here, and let the consumer recover by * freeing this tmpl and create a new one for the key and new SW * provider */ if ((pd->pd_prov_type == CRYPTO_SW_PROVIDER) && ((ctx_tmpl = (kcf_ctx_template_t *)tmpl) != NULL)) { if (ctx_tmpl->ct_generation != me->me_gen_swprov) { if (list != NULL) kcf_free_triedlist(list); KCF_PROV_REFRELE(pd); return (CRYPTO_OLD_CTX_TEMPLATE); } else { spi_ctx_tmpl = ctx_tmpl->ct_prov_tmpl; } } /* The fast path for SW providers. */ if (CHECK_FASTPATH(crq, pd)) { crypto_mechanism_t lmech; lmech = *mech; KCF_SET_PROVIDER_MECHNUM(mech->cm_type, pd, &lmech); error = KCF_PROV_MAC_VERIFY_ATOMIC(pd, pd->pd_sid, &lmech, key, data, mac, spi_ctx_tmpl, KCF_SWFP_RHNDL(crq)); KCF_PROV_INCRSTATS(pd, error); } else { if (pd->pd_prov_type == CRYPTO_HW_PROVIDER && (pd->pd_flags & CRYPTO_HASH_NO_UPDATE) && (data->cd_length > pd->pd_hash_limit)) { /* see comments in crypto_mac() */ error = CRYPTO_BUFFER_TOO_BIG; } else { KCF_WRAP_MAC_OPS_PARAMS(¶ms, KCF_OP_MAC_VERIFY_ATOMIC, pd->pd_sid, mech, key, data, mac, spi_ctx_tmpl); error = kcf_submit_request(pd, NULL, crq, ¶ms, KCF_ISDUALREQ(crq)); } } if (error != CRYPTO_SUCCESS && error != CRYPTO_QUEUED && IS_RECOVERABLE(error)) { /* Add pd to the linked list of providers tried. */ if (kcf_insert_triedlist(&list, pd, KCF_KMFLAG(crq)) != NULL) goto retry; } if (list != NULL) kcf_free_triedlist(list); KCF_PROV_REFRELE(pd); return (error); }
/* * crypto_cipher_init_prov() * * Arguments: * * pd: provider descriptor * sid: session id * mech: crypto_mechanism_t pointer. * mech_type is a valid value previously returned by * crypto_mech2id(); * When the mech's parameter is not NULL, its definition depends * on the standard definition of the mechanism. * key: pointer to a crypto_key_t structure. * tmpl: a crypto_ctx_template_t, opaque template of a context of an * encryption or decryption with the 'mech' using 'key'. * 'tmpl' is created by a previous call to * crypto_create_ctx_template(). * ctxp: Pointer to a crypto_context_t. * func: CRYPTO_FG_ENCRYPT or CRYPTO_FG_DECRYPT. * cr: crypto_call_req_t calling conditions and call back info. * * Description: * This is a common function invoked internally by both * crypto_encrypt_init() and crypto_decrypt_init(). * Asynchronously submits a request for, or synchronously performs the * initialization of an encryption or a decryption operation. * When possible and applicable, will internally use the pre-expanded key * schedule from the context template, tmpl. * When complete and successful, 'ctxp' will contain a crypto_context_t * valid for later calls to encrypt_update() and encrypt_final(), or * decrypt_update() and decrypt_final(). * The caller should hold a reference on the specified provider * descriptor before calling this function. * * Context: * Process or interrupt, according to the semantics dictated by the 'cr'. * * Returns: * See comment in the beginning of the file. */ static int crypto_cipher_init_prov(crypto_provider_t provider, crypto_session_id_t sid, crypto_mechanism_t *mech, crypto_key_t *key, crypto_spi_ctx_template_t tmpl, crypto_context_t *ctxp, crypto_call_req_t *crq, crypto_func_group_t func) { int error; crypto_ctx_t *ctx; kcf_req_params_t params; kcf_provider_desc_t *pd = provider; kcf_provider_desc_t *real_provider = pd; ASSERT(KCF_PROV_REFHELD(pd)); if (pd->pd_prov_type == CRYPTO_LOGICAL_PROVIDER) { if (func == CRYPTO_FG_ENCRYPT) { error = kcf_get_hardware_provider(mech->cm_type, CRYPTO_MECH_INVALID, CHECK_RESTRICT(crq), pd, &real_provider, CRYPTO_FG_ENCRYPT); } else { error = kcf_get_hardware_provider(mech->cm_type, CRYPTO_MECH_INVALID, CHECK_RESTRICT(crq), pd, &real_provider, CRYPTO_FG_DECRYPT); } if (error != CRYPTO_SUCCESS) return (error); } /* Allocate and initialize the canonical context */ if ((ctx = kcf_new_ctx(crq, real_provider, sid)) == NULL) { if (pd->pd_prov_type == CRYPTO_LOGICAL_PROVIDER) KCF_PROV_REFRELE(real_provider); return (CRYPTO_HOST_MEMORY); } /* The fast path for SW providers. */ if (CHECK_FASTPATH(crq, pd)) { crypto_mechanism_t lmech; lmech = *mech; KCF_SET_PROVIDER_MECHNUM(mech->cm_type, real_provider, &lmech); if (func == CRYPTO_FG_ENCRYPT) error = KCF_PROV_ENCRYPT_INIT(real_provider, ctx, &lmech, key, tmpl, KCF_SWFP_RHNDL(crq)); else { ASSERT(func == CRYPTO_FG_DECRYPT); error = KCF_PROV_DECRYPT_INIT(real_provider, ctx, &lmech, key, tmpl, KCF_SWFP_RHNDL(crq)); } KCF_PROV_INCRSTATS(pd, error); goto done; } /* Check if context sharing is possible */ if (pd->pd_prov_type == CRYPTO_HW_PROVIDER && key->ck_format == CRYPTO_KEY_RAW && KCF_CAN_SHARE_OPSTATE(pd, mech->cm_type)) { kcf_context_t *tctxp = (kcf_context_t *)ctx; kcf_provider_desc_t *tpd = NULL; crypto_mech_info_t *sinfo; if ((kcf_get_sw_prov(mech->cm_type, &tpd, &tctxp->kc_mech, B_FALSE) == CRYPTO_SUCCESS)) { int tlen; sinfo = &(KCF_TO_PROV_MECHINFO(tpd, mech->cm_type)); /* * key->ck_length from the consumer is always in bits. * We convert it to be in the same unit registered by * the provider in order to do a comparison. */ if (sinfo->cm_mech_flags & CRYPTO_KEYSIZE_UNIT_IN_BYTES) tlen = key->ck_length >> 3; else
static int sign_sr_atomic_common(crypto_mechanism_t *mech, crypto_key_t *key, crypto_data_t *data, crypto_ctx_template_t tmpl, crypto_data_t *signature, crypto_call_req_t *crq, crypto_func_group_t fg) { int error; kcf_mech_entry_t *me; kcf_provider_desc_t *pd; kcf_req_params_t params; kcf_prov_tried_t *list = NULL; kcf_ctx_template_t *ctx_tmpl; crypto_spi_ctx_template_t spi_ctx_tmpl = NULL; retry: /* The pd is returned held */ if ((pd = kcf_get_mech_provider(mech->cm_type, &me, &error, list, fg, CHECK_RESTRICT(crq), data->cd_length)) == NULL) { if (list != NULL) kcf_free_triedlist(list); return (error); } /* * For SW providers, check the validity of the context template * It is very rare that the generation number mis-matches, so * it is acceptable to fail here, and let the consumer recover by * freeing this tmpl and create a new one for the key and new SW * provider. */ if ((pd->pd_prov_type == CRYPTO_SW_PROVIDER) && ((ctx_tmpl = (kcf_ctx_template_t *)tmpl) != NULL)) { if (ctx_tmpl->ct_generation != me->me_gen_swprov) { if (list != NULL) kcf_free_triedlist(list); KCF_PROV_REFRELE(pd); return (CRYPTO_OLD_CTX_TEMPLATE); } else { spi_ctx_tmpl = ctx_tmpl->ct_prov_tmpl; } } /* The fast path for SW providers. */ if (CHECK_FASTPATH(crq, pd)) { crypto_mechanism_t lmech; lmech = *mech; KCF_SET_PROVIDER_MECHNUM(mech->cm_type, pd, &lmech); if (fg == CRYPTO_FG_SIGN_ATOMIC) error = KCF_PROV_SIGN_ATOMIC(pd, pd->pd_sid, &lmech, key, data, spi_ctx_tmpl, signature, KCF_SWFP_RHNDL(crq)); else error = KCF_PROV_SIGN_RECOVER_ATOMIC(pd, pd->pd_sid, &lmech, key, data, spi_ctx_tmpl, signature, KCF_SWFP_RHNDL(crq)); KCF_PROV_INCRSTATS(pd, error); } else { kcf_op_type_t op = ((fg == CRYPTO_FG_SIGN_ATOMIC) ? KCF_OP_ATOMIC : KCF_OP_SIGN_RECOVER_ATOMIC); KCF_WRAP_SIGN_OPS_PARAMS(¶ms, op, pd->pd_sid, mech, key, data, signature, spi_ctx_tmpl); /* no crypto context to carry between multiple parts. */ error = kcf_submit_request(pd, NULL, crq, ¶ms, B_FALSE); } if (error != CRYPTO_SUCCESS && error != CRYPTO_QUEUED && IS_RECOVERABLE(error)) { /* Add pd to the linked list of providers tried. */ if (kcf_insert_triedlist(&list, pd, KCF_KMFLAG(crq)) != NULL) goto retry; } if (list != NULL) kcf_free_triedlist(list); KCF_PROV_REFRELE(pd); return (error); }