Print this page
first pass
*** 85,96 ****
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
* OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
* EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
- /* EXPORT DELETE START */
-
#if defined(sun4u)
/* External assembly functions: */
extern void aes_encrypt_impl(const uint32_t rk[], int Nr, const uint32_t pt[4],
uint32_t ct[4]);
extern void aes_decrypt_impl(const uint32_t rk[], int Nr, const uint32_t ct[4],
--- 85,94 ----
*** 1554,1564 ****
(Td4[t0 & 0xff] & 0x000000ff) ^
rk[3];
pt[3] = s3;
}
#endif /* sun4u, __amd64 */
- /* EXPORT DELETE END */
/*
* Initialize AES encryption and decryption key schedules.
*
--- 1552,1561 ----
*** 1569,1579 ****
* Allocated by aes_alloc_keysched().
*/
void
aes_init_keysched(const uint8_t *cipherKey, uint_t keyBits, void *keysched)
{
- /* EXPORT DELETE START */
aes_key_t *newbie = keysched;
uint_t keysize, i, j;
union {
uint64_t ka64[4];
uint32_t ka32[8];
--- 1566,1575 ----
*** 1622,1632 ****
keyarr.ka32[i] = htonl(*(uint32_t *)(void *)&cipherKey[j]);
}
#endif
aes_setupkeys(newbie, keyarr.ka32, keyBits);
- /* EXPORT DELETE END */
}
/*
* Encrypt one block using AES.
--- 1618,1627 ----
*** 1638,1648 ****
* ct Output block (crypto text). Can overlap with pt
*/
int
aes_encrypt_block(const void *ks, const uint8_t *pt, uint8_t *ct)
{
- /* EXPORT DELETE START */
aes_key_t *ksch = (aes_key_t *)ks;
#ifndef AES_BYTE_SWAP
if (IS_P2ALIGNED2(pt, ct, sizeof (uint32_t))) {
/* LINTED: pointer alignment */
--- 1633,1642 ----
*** 1676,1686 ****
*(uint32_t *)(void *)&ct[0] = htonl(buffer[0]);
*(uint32_t *)(void *)&ct[4] = htonl(buffer[1]);
*(uint32_t *)(void *)&ct[8] = htonl(buffer[2]);
*(uint32_t *)(void *)&ct[12] = htonl(buffer[3]);
#endif
- /* EXPORT DELETE END */
return (CRYPTO_SUCCESS);
}
/*
--- 1670,1679 ----
*** 1693,1703 ****
* pt Output block (plain text). Can overlap with pt
*/
int
aes_decrypt_block(const void *ks, const uint8_t *ct, uint8_t *pt)
{
- /* EXPORT DELETE START */
aes_key_t *ksch = (aes_key_t *)ks;
#ifndef AES_BYTE_SWAP
if (IS_P2ALIGNED2(ct, pt, sizeof (uint32_t))) {
/* LINTED: pointer alignment */
--- 1686,1695 ----
*** 1732,1742 ****
*(uint32_t *)(void *)&pt[4] = htonl(buffer[1]);
*(uint32_t *)(void *)&pt[8] = htonl(buffer[2]);
*(uint32_t *)(void *)&pt[12] = htonl(buffer[3]);
#endif
- /* EXPORT DELETE END */
return (CRYPTO_SUCCESS);
}
/*
--- 1724,1733 ----
*** 1751,1761 ****
*/
/* ARGSUSED */
void *
aes_alloc_keysched(size_t *size, int kmflag)
{
- /* EXPORT DELETE START */
aes_key_t *keysched;
#ifdef _KERNEL
keysched = (aes_key_t *)kmem_alloc(sizeof (aes_key_t), kmflag);
#else /* !_KERNEL */
--- 1742,1751 ----
*** 1764,1774 ****
if (keysched != NULL) {
*size = sizeof (aes_key_t);
return (keysched);
}
- /* EXPORT DELETE END */
return (NULL);
}
#ifdef __amd64
--- 1754,1763 ----