Print this page
first pass

*** 85,96 **** * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, * EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ - /* EXPORT DELETE START */ - #if defined(sun4u) /* External assembly functions: */ extern void aes_encrypt_impl(const uint32_t rk[], int Nr, const uint32_t pt[4], uint32_t ct[4]); extern void aes_decrypt_impl(const uint32_t rk[], int Nr, const uint32_t ct[4], --- 85,94 ----
*** 1554,1564 **** (Td4[t0 & 0xff] & 0x000000ff) ^ rk[3]; pt[3] = s3; } #endif /* sun4u, __amd64 */ - /* EXPORT DELETE END */ /* * Initialize AES encryption and decryption key schedules. * --- 1552,1561 ----
*** 1569,1579 **** * Allocated by aes_alloc_keysched(). */ void aes_init_keysched(const uint8_t *cipherKey, uint_t keyBits, void *keysched) { - /* EXPORT DELETE START */ aes_key_t *newbie = keysched; uint_t keysize, i, j; union { uint64_t ka64[4]; uint32_t ka32[8]; --- 1566,1575 ----
*** 1622,1632 **** keyarr.ka32[i] = htonl(*(uint32_t *)(void *)&cipherKey[j]); } #endif aes_setupkeys(newbie, keyarr.ka32, keyBits); - /* EXPORT DELETE END */ } /* * Encrypt one block using AES. --- 1618,1627 ----
*** 1638,1648 **** * ct Output block (crypto text). Can overlap with pt */ int aes_encrypt_block(const void *ks, const uint8_t *pt, uint8_t *ct) { - /* EXPORT DELETE START */ aes_key_t *ksch = (aes_key_t *)ks; #ifndef AES_BYTE_SWAP if (IS_P2ALIGNED2(pt, ct, sizeof (uint32_t))) { /* LINTED: pointer alignment */ --- 1633,1642 ----
*** 1676,1686 **** *(uint32_t *)(void *)&ct[0] = htonl(buffer[0]); *(uint32_t *)(void *)&ct[4] = htonl(buffer[1]); *(uint32_t *)(void *)&ct[8] = htonl(buffer[2]); *(uint32_t *)(void *)&ct[12] = htonl(buffer[3]); #endif - /* EXPORT DELETE END */ return (CRYPTO_SUCCESS); } /* --- 1670,1679 ----
*** 1693,1703 **** * pt Output block (plain text). Can overlap with pt */ int aes_decrypt_block(const void *ks, const uint8_t *ct, uint8_t *pt) { - /* EXPORT DELETE START */ aes_key_t *ksch = (aes_key_t *)ks; #ifndef AES_BYTE_SWAP if (IS_P2ALIGNED2(ct, pt, sizeof (uint32_t))) { /* LINTED: pointer alignment */ --- 1686,1695 ----
*** 1732,1742 **** *(uint32_t *)(void *)&pt[4] = htonl(buffer[1]); *(uint32_t *)(void *)&pt[8] = htonl(buffer[2]); *(uint32_t *)(void *)&pt[12] = htonl(buffer[3]); #endif - /* EXPORT DELETE END */ return (CRYPTO_SUCCESS); } /* --- 1724,1733 ----
*** 1751,1761 **** */ /* ARGSUSED */ void * aes_alloc_keysched(size_t *size, int kmflag) { - /* EXPORT DELETE START */ aes_key_t *keysched; #ifdef _KERNEL keysched = (aes_key_t *)kmem_alloc(sizeof (aes_key_t), kmflag); #else /* !_KERNEL */ --- 1742,1751 ----
*** 1764,1774 **** if (keysched != NULL) { *size = sizeof (aes_key_t); return (keysched); } - /* EXPORT DELETE END */ return (NULL); } #ifdef __amd64 --- 1754,1763 ----