AES.hpp 24 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754
  1. /*
  2. * ZeroTier One - Network Virtualization Everywhere
  3. * Copyright (C) 2011-2019 ZeroTier, Inc. https://www.zerotier.com/
  4. *
  5. * This program is free software: you can redistribute it and/or modify
  6. * it under the terms of the GNU General Public License as published by
  7. * the Free Software Foundation, either version 3 of the License, or
  8. * (at your option) any later version.
  9. *
  10. * This program is distributed in the hope that it will be useful,
  11. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  12. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  13. * GNU General Public License for more details.
  14. *
  15. * You should have received a copy of the GNU General Public License
  16. * along with this program. If not, see <http://www.gnu.org/licenses/>.
  17. *
  18. * --
  19. *
  20. * You can be released from the requirements of the license by purchasing
  21. * a commercial license. Buying such a license is mandatory as soon as you
  22. * develop commercial closed-source software that incorporates or links
  23. * directly against ZeroTier software without disclosing the source code
  24. * of your own application.
  25. */
  26. #ifndef ZT_AES_HPP
  27. #define ZT_AES_HPP
  28. #include "Constants.hpp"
  29. #include "Utils.hpp"
  30. #if (defined(__amd64) || defined(__amd64__) || defined(__x86_64) || defined(__x86_64__) || defined(__AMD64) || defined(__AMD64__) || defined(_M_X64))
  31. #include <wmmintrin.h>
  32. #include <emmintrin.h>
  33. #include <smmintrin.h>
  34. #define ZT_AES_AESNI 1
  35. #endif
  36. namespace ZeroTier {
  37. /**
  38. * AES-256 and GCM AEAD
  39. *
  40. * AES with 128-bit or 192-bit key sizes isn't supported here. This also only
  41. * supports the encrypt operation since we use AES in GCM mode. For HW acceleration
  42. * the code is inlined for maximum performance.
  43. */
  44. class AES
  45. {
  46. public:
  47. /**
  48. * This will be true if your platform's type of AES acceleration is supported on this machine
  49. */
  50. static const bool HW_ACCEL;
  51. inline AES() {}
  52. inline AES(const uint8_t key[32]) { this->init(key); }
  53. inline ~AES()
  54. {
  55. Utils::burn(&_k,sizeof(_k));
  56. }
  57. inline void init(const uint8_t key[32])
  58. {
  59. #ifdef ZT_AES_AESNI
  60. if (HW_ACCEL) {
  61. _init_aesni(key);
  62. return;
  63. }
  64. #endif
  65. _initSW(key);
  66. }
  67. inline void encrypt(const uint8_t in[16],uint8_t out[16]) const
  68. {
  69. #ifdef ZT_AES_AESNI
  70. if (HW_ACCEL) {
  71. _encrypt_aesni(in,out);
  72. return;
  73. }
  74. #endif
  75. _encryptSW(in,out);
  76. }
  77. inline void gcmEncrypt(const uint8_t iv[12],const void *in,unsigned int inlen,const void *assoc,unsigned int assoclen,void *out,uint8_t *tag,unsigned int taglen)
  78. {
  79. #ifdef ZT_AES_AESNI
  80. if (HW_ACCEL) {
  81. _encrypt_gcm256_aesni(inlen,(const uint8_t *)in,(uint8_t *)out,iv,assoclen,(const uint8_t *)assoc,tag,taglen);
  82. return;
  83. }
  84. #endif
  85. abort(); // TODO: software
  86. }
  87. inline bool gcmDecrypt(const uint8_t iv[12],const void *in,unsigned int inlen,const void *assoc,unsigned int assoclen,void *out,const uint8_t *tag,unsigned int taglen)
  88. {
  89. #ifdef ZT_AES_AESNI
  90. if (HW_ACCEL) {
  91. uint8_t tagbuf[16];
  92. _decrypt_gcm256_aesni(inlen,(const uint8_t *)in,(uint8_t *)out,iv,assoclen,(const uint8_t *)assoc,tagbuf,taglen);
  93. return Utils::secureEq(tagbuf,tag,taglen);
  94. }
  95. #endif
  96. abort(); // TODO: software
  97. return false;
  98. }
  99. private:
  100. void _initSW(const uint8_t key[32]);
  101. void _encryptSW(const uint8_t in[16],uint8_t out[16]) const;
  102. union {
  103. #ifdef ZT_AES_AESNI
  104. struct {
  105. __m128i k[15];
  106. __m128i h,hh,hhh,hhhh;
  107. } ni;
  108. #endif
  109. struct {
  110. uint32_t k[60];
  111. } sw;
  112. } _k;
  113. #ifdef ZT_AES_AESNI /********************************************************/
  114. static inline __m128i _init256_1_aesni(__m128i a,__m128i b)
  115. {
  116. __m128i x,y;
  117. b = _mm_shuffle_epi32(b,0xff);
  118. y = _mm_slli_si128(a,0x04);
  119. x = _mm_xor_si128(a,y);
  120. y = _mm_slli_si128(y,0x04);
  121. x = _mm_xor_si128(x,y);
  122. y = _mm_slli_si128(y,0x04);
  123. x = _mm_xor_si128(x,y);
  124. x = _mm_xor_si128(x,b);
  125. return x;
  126. }
  127. static inline __m128i _init256_2_aesni(__m128i a,__m128i b)
  128. {
  129. __m128i x,y,z;
  130. y = _mm_aeskeygenassist_si128(a,0x00);
  131. z = _mm_shuffle_epi32(y,0xaa);
  132. y = _mm_slli_si128(b,0x04);
  133. x = _mm_xor_si128(b,y);
  134. y = _mm_slli_si128(y,0x04);
  135. x = _mm_xor_si128(x,y);
  136. y = _mm_slli_si128(y,0x04);
  137. x = _mm_xor_si128(x,y);
  138. x = _mm_xor_si128(x,z);
  139. return x;
  140. }
  141. inline void _init_aesni(const uint8_t key[32])
  142. {
  143. /* Init AES itself */
  144. __m128i t1,t2;
  145. _k.ni.k[0] = t1 = _mm_loadu_si128((const __m128i *)key);
  146. _k.ni.k[1] = t2 = _mm_loadu_si128((const __m128i *)(key+16));
  147. _k.ni.k[2] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x01));
  148. _k.ni.k[3] = t2 = _init256_2_aesni(t1,t2);
  149. _k.ni.k[4] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x02));
  150. _k.ni.k[5] = t2 = _init256_2_aesni(t1,t2);
  151. _k.ni.k[6] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x04));
  152. _k.ni.k[7] = t2 = _init256_2_aesni(t1,t2);
  153. _k.ni.k[8] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x08));
  154. _k.ni.k[9] = t2 = _init256_2_aesni(t1,t2);
  155. _k.ni.k[10] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x10));
  156. _k.ni.k[11] = t2 = _init256_2_aesni(t1,t2);
  157. _k.ni.k[12] = t1 = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x20));
  158. _k.ni.k[13] = t2 = _init256_2_aesni(t1,t2);
  159. _k.ni.k[14] = _init256_1_aesni(t1,_mm_aeskeygenassist_si128(t2,0x40));
  160. /* Init GCM / GHASH */
  161. __m128i h = _mm_xor_si128(_mm_setzero_si128(),_k.ni.k[0]);
  162. h = _mm_aesenc_si128(h,_k.ni.k[1]);
  163. h = _mm_aesenc_si128(h,_k.ni.k[2]);
  164. h = _mm_aesenc_si128(h,_k.ni.k[3]);
  165. h = _mm_aesenc_si128(h,_k.ni.k[4]);
  166. h = _mm_aesenc_si128(h,_k.ni.k[5]);
  167. h = _mm_aesenc_si128(h,_k.ni.k[6]);
  168. h = _mm_aesenc_si128(h,_k.ni.k[7]);
  169. h = _mm_aesenc_si128(h,_k.ni.k[8]);
  170. h = _mm_aesenc_si128(h,_k.ni.k[9]);
  171. h = _mm_aesenc_si128(h,_k.ni.k[10]);
  172. h = _mm_aesenc_si128(h,_k.ni.k[11]);
  173. h = _mm_aesenc_si128(h,_k.ni.k[12]);
  174. h = _mm_aesenc_si128(h,_k.ni.k[13]);
  175. h = _mm_aesenclast_si128(h,_k.ni.k[14]);
  176. __m128i hswap = _swap128_aesni(h);
  177. __m128i hh = _mult_block_aesni(hswap,h);
  178. __m128i hhh = _mult_block_aesni(hswap,hh);
  179. __m128i hhhh = _mult_block_aesni(hswap,hhh);
  180. _k.ni.h = hswap;
  181. _k.ni.hh = _swap128_aesni(hh);
  182. _k.ni.hhh = _swap128_aesni(hhh);
  183. _k.ni.hhhh = _swap128_aesni(hhhh);
  184. }
  185. inline void _encrypt_aesni(const void *in,void *out) const
  186. {
  187. __m128i tmp;
  188. tmp = _mm_loadu_si128((const __m128i *)in);
  189. tmp = _mm_xor_si128(tmp,_k.ni.k[0]);
  190. tmp = _mm_aesenc_si128(tmp,_k.ni.k[1]);
  191. tmp = _mm_aesenc_si128(tmp,_k.ni.k[2]);
  192. tmp = _mm_aesenc_si128(tmp,_k.ni.k[3]);
  193. tmp = _mm_aesenc_si128(tmp,_k.ni.k[4]);
  194. tmp = _mm_aesenc_si128(tmp,_k.ni.k[5]);
  195. tmp = _mm_aesenc_si128(tmp,_k.ni.k[6]);
  196. tmp = _mm_aesenc_si128(tmp,_k.ni.k[7]);
  197. tmp = _mm_aesenc_si128(tmp,_k.ni.k[8]);
  198. tmp = _mm_aesenc_si128(tmp,_k.ni.k[9]);
  199. tmp = _mm_aesenc_si128(tmp,_k.ni.k[10]);
  200. tmp = _mm_aesenc_si128(tmp,_k.ni.k[11]);
  201. tmp = _mm_aesenc_si128(tmp,_k.ni.k[12]);
  202. tmp = _mm_aesenc_si128(tmp,_k.ni.k[13]);
  203. _mm_storeu_si128((__m128i *)out,_mm_aesenclast_si128(tmp,_k.ni.k[14]));
  204. }
  205. static inline __m128i _swap128_aesni(__m128i x) { return _mm_shuffle_epi8(x,_mm_set_epi8(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15)); }
  206. static inline __m128i _mult_block_aesni(__m128i h,__m128i y)
  207. {
  208. __m128i t1,t2,t3,t4,t5,t6;
  209. y = _swap128_aesni(y);
  210. t1 = _mm_clmulepi64_si128(h,y,0x00);
  211. t2 = _mm_clmulepi64_si128(h,y,0x01);
  212. t3 = _mm_clmulepi64_si128(h,y,0x10);
  213. t4 = _mm_clmulepi64_si128(h,y,0x11);
  214. t2 = _mm_xor_si128(t2,t3);
  215. t3 = _mm_slli_si128(t2,8);
  216. t2 = _mm_srli_si128(t2,8);
  217. t1 = _mm_xor_si128(t1,t3);
  218. t4 = _mm_xor_si128(t4,t2);
  219. t5 = _mm_srli_epi32(t1,31);
  220. t1 = _mm_slli_epi32(t1,1);
  221. t6 = _mm_srli_epi32(t4,31);
  222. t4 = _mm_slli_epi32(t4,1);
  223. t3 = _mm_srli_si128(t5,12);
  224. t6 = _mm_slli_si128(t6,4);
  225. t5 = _mm_slli_si128(t5,4);
  226. t1 = _mm_or_si128(t1,t5);
  227. t4 = _mm_or_si128(t4,t6);
  228. t4 = _mm_or_si128(t4,t3);
  229. t5 = _mm_slli_epi32(t1,31);
  230. t6 = _mm_slli_epi32(t1,30);
  231. t3 = _mm_slli_epi32(t1,25);
  232. t5 = _mm_xor_si128(t5,t6);
  233. t5 = _mm_xor_si128(t5,t3);
  234. t6 = _mm_srli_si128(t5,4);
  235. t4 = _mm_xor_si128(t4,t6);
  236. t5 = _mm_slli_si128(t5,12);
  237. t1 = _mm_xor_si128(t1,t5);
  238. t4 = _mm_xor_si128(t4,t1);
  239. t5 = _mm_srli_epi32(t1,1);
  240. t2 = _mm_srli_epi32(t1,2);
  241. t3 = _mm_srli_epi32(t1,7);
  242. t4 = _mm_xor_si128(t4,t2);
  243. t4 = _mm_xor_si128(t4,t3);
  244. t4 = _mm_xor_si128(t4,t5);
  245. return _swap128_aesni(t4);
  246. }
  247. static inline __m128i _mult4xor_aesni(__m128i h1,__m128i h2,__m128i h3,__m128i h4,__m128i d1,__m128i d2,__m128i d3,__m128i d4)
  248. {
  249. __m128i t0,t1,t2,t3,t4,t5,t6,t7,t8,t9;
  250. d1 = _swap128_aesni(d1);
  251. d2 = _swap128_aesni(d2);
  252. d3 = _swap128_aesni(d3);
  253. d4 = _swap128_aesni(d4);
  254. t0 = _mm_clmulepi64_si128(h1,d1,0x00);
  255. t1 = _mm_clmulepi64_si128(h2,d2,0x00);
  256. t2 = _mm_clmulepi64_si128(h3,d3,0x00);
  257. t3 = _mm_clmulepi64_si128(h4,d4,0x00);
  258. t8 = _mm_xor_si128(t0,t1);
  259. t8 = _mm_xor_si128(t8,t2);
  260. t8 = _mm_xor_si128(t8,t3);
  261. t4 = _mm_clmulepi64_si128(h1,d1,0x11);
  262. t5 = _mm_clmulepi64_si128(h2,d2,0x11);
  263. t6 = _mm_clmulepi64_si128(h3,d3,0x11);
  264. t7 = _mm_clmulepi64_si128(h4,d4,0x11);
  265. t9 = _mm_xor_si128(t4,t5);
  266. t9 = _mm_xor_si128(t9,t6);
  267. t9 = _mm_xor_si128(t9,t7);
  268. t0 = _mm_shuffle_epi32(h1,78);
  269. t4 = _mm_shuffle_epi32(d1,78);
  270. t0 = _mm_xor_si128(t0,h1);
  271. t4 = _mm_xor_si128(t4,d1);
  272. t1 = _mm_shuffle_epi32(h2,78);
  273. t5 = _mm_shuffle_epi32(d2,78);
  274. t1 = _mm_xor_si128(t1,h2);
  275. t5 = _mm_xor_si128(t5,d2);
  276. t2 = _mm_shuffle_epi32(h3,78);
  277. t6 = _mm_shuffle_epi32(d3,78);
  278. t2 = _mm_xor_si128(t2,h3);
  279. t6 = _mm_xor_si128(t6,d3);
  280. t3 = _mm_shuffle_epi32(h4,78);
  281. t7 = _mm_shuffle_epi32(d4,78);
  282. t3 = _mm_xor_si128(t3,h4);
  283. t7 = _mm_xor_si128(t7,d4);
  284. t0 = _mm_clmulepi64_si128(t0,t4,0x00);
  285. t1 = _mm_clmulepi64_si128(t1,t5,0x00);
  286. t2 = _mm_clmulepi64_si128(t2,t6,0x00);
  287. t3 = _mm_clmulepi64_si128(t3,t7,0x00);
  288. t0 = _mm_xor_si128(t0,t8);
  289. t0 = _mm_xor_si128(t0,t9);
  290. t0 = _mm_xor_si128(t1,t0);
  291. t0 = _mm_xor_si128(t2,t0);
  292. t0 = _mm_xor_si128(t3,t0);
  293. t4 = _mm_slli_si128(t0,8);
  294. t0 = _mm_srli_si128(t0,8);
  295. t3 = _mm_xor_si128(t4,t8);
  296. t6 = _mm_xor_si128(t0,t9);
  297. t7 = _mm_srli_epi32(t3,31);
  298. t8 = _mm_srli_epi32(t6,31);
  299. t3 = _mm_slli_epi32(t3,1);
  300. t6 = _mm_slli_epi32(t6,1);
  301. t9 = _mm_srli_si128(t7,12);
  302. t8 = _mm_slli_si128(t8,4);
  303. t7 = _mm_slli_si128(t7,4);
  304. t3 = _mm_or_si128(t3,t7);
  305. t6 = _mm_or_si128(t6,t8);
  306. t6 = _mm_or_si128(t6,t9);
  307. t7 = _mm_slli_epi32(t3,31);
  308. t8 = _mm_slli_epi32(t3,30);
  309. t9 = _mm_slli_epi32(t3,25);
  310. t7 = _mm_xor_si128(t7,t8);
  311. t7 = _mm_xor_si128(t7,t9);
  312. t8 = _mm_srli_si128(t7,4);
  313. t7 = _mm_slli_si128(t7,12);
  314. t3 = _mm_xor_si128(t3,t7);
  315. t2 = _mm_srli_epi32(t3,1);
  316. t4 = _mm_srli_epi32(t3,2);
  317. t5 = _mm_srli_epi32(t3,7);
  318. t2 = _mm_xor_si128(t2,t4);
  319. t2 = _mm_xor_si128(t2,t5);
  320. t2 = _mm_xor_si128(t2,t8);
  321. t3 = _mm_xor_si128(t3,t2);
  322. t6 = _mm_xor_si128(t6,t3);
  323. return _swap128_aesni(t6);
  324. }
  325. static inline __m128i _ghash_aesni(__m128i h,__m128i y,__m128i x) { return _mult_block_aesni(h,_mm_xor_si128(y,x)); }
  326. static inline __m128i _increment_be_aesni(__m128i x)
  327. {
  328. x = _swap128_aesni(x);
  329. x = _mm_add_epi64(x,_mm_set_epi32(0,0,0,1));
  330. x = _swap128_aesni(x);
  331. return x;
  332. }
  333. static inline void _htoun64_aesni(void *network,const uint64_t host) { *((uint64_t *)network) = Utils::hton(host); }
  334. inline __m128i _create_j_aesni(const uint8_t *iv) const
  335. {
  336. uint8_t j[16];
  337. *((uint64_t *)j) = *((const uint64_t *)iv);
  338. *((uint32_t *)(j+8)) = *((const uint32_t *)(iv+8));
  339. j[12] = 0;
  340. j[13] = 0;
  341. j[14] = 0;
  342. j[15] = 1;
  343. return _mm_loadu_si128((__m128i *)j);
  344. }
  345. inline __m128i _icv_header_aesni(const void *assoc,unsigned int alen) const
  346. {
  347. unsigned int blocks,pblocks,rem,i;
  348. __m128i h1,h2,h3,h4,d1,d2,d3,d4;
  349. __m128i y,last;
  350. const __m128i *ab;
  351. h1 = _k.ni.hhhh;
  352. h2 = _k.ni.hhh;
  353. h3 = _k.ni.hh;
  354. h4 = _k.ni.h;
  355. y = _mm_setzero_si128();
  356. ab = (const __m128i *)assoc;
  357. blocks = alen / 16;
  358. pblocks = blocks - (blocks % 4);
  359. rem = alen % 16;
  360. for (i=0;i<pblocks;i+=4) {
  361. d1 = _mm_loadu_si128(ab + i + 0);
  362. d2 = _mm_loadu_si128(ab + i + 1);
  363. d3 = _mm_loadu_si128(ab + i + 2);
  364. d4 = _mm_loadu_si128(ab + i + 3);
  365. y = _mm_xor_si128(y, d1);
  366. y = _mult4xor_aesni(h1,h2,h3,h4,y,d2,d3,d4);
  367. }
  368. for (i = pblocks; i < blocks; i++)
  369. y = _ghash_aesni(_k.ni.h,y,_mm_loadu_si128(ab + i));
  370. if (rem) {
  371. last = _mm_setzero_si128();
  372. memcpy(&last,ab + blocks,rem);
  373. y = _ghash_aesni(_k.ni.h,y,last);
  374. }
  375. return y;
  376. }
  377. inline __m128i _icv_tailer_aesni(__m128i y,size_t alen,size_t dlen) const
  378. {
  379. __m128i b;
  380. _htoun64_aesni(&b, alen * 8);
  381. _htoun64_aesni((uint8_t *)&b + sizeof(uint64_t), dlen * 8);
  382. return _ghash_aesni(_k.ni.h, y, b);
  383. }
  384. inline void _icv_crypt_aesni(__m128i y,__m128i j,uint8_t *icv,unsigned int icvsize) const
  385. {
  386. __m128i t,b;
  387. t = _mm_xor_si128(j,_k.ni.k[0]);
  388. t = _mm_aesenc_si128(t,_k.ni.k[1]);
  389. t = _mm_aesenc_si128(t,_k.ni.k[2]);
  390. t = _mm_aesenc_si128(t,_k.ni.k[3]);
  391. t = _mm_aesenc_si128(t,_k.ni.k[4]);
  392. t = _mm_aesenc_si128(t,_k.ni.k[5]);
  393. t = _mm_aesenc_si128(t,_k.ni.k[6]);
  394. t = _mm_aesenc_si128(t,_k.ni.k[7]);
  395. t = _mm_aesenc_si128(t,_k.ni.k[8]);
  396. t = _mm_aesenc_si128(t,_k.ni.k[9]);
  397. t = _mm_aesenc_si128(t,_k.ni.k[10]);
  398. t = _mm_aesenc_si128(t,_k.ni.k[11]);
  399. t = _mm_aesenc_si128(t,_k.ni.k[12]);
  400. t = _mm_aesenc_si128(t,_k.ni.k[13]);
  401. t = _mm_aesenclast_si128(t,_k.ni.k[14]);
  402. t = _mm_xor_si128(y, t);
  403. _mm_storeu_si128(&b, t);
  404. memcpy(icv,&b,icvsize);
  405. }
  406. inline __m128i _encrypt_gcm_rem_aesni(unsigned int rem,const void *in,void *out,__m128i cb,__m128i y) const
  407. {
  408. __m128i t,b;
  409. memset(&b,0,sizeof(b));
  410. memcpy(&b,in,rem);
  411. t = _mm_xor_si128(cb,_k.ni.k[0]);
  412. t = _mm_aesenc_si128(t,_k.ni.k[1]);
  413. t = _mm_aesenc_si128(t,_k.ni.k[2]);
  414. t = _mm_aesenc_si128(t,_k.ni.k[3]);
  415. t = _mm_aesenc_si128(t,_k.ni.k[4]);
  416. t = _mm_aesenc_si128(t,_k.ni.k[5]);
  417. t = _mm_aesenc_si128(t,_k.ni.k[6]);
  418. t = _mm_aesenc_si128(t,_k.ni.k[7]);
  419. t = _mm_aesenc_si128(t,_k.ni.k[8]);
  420. t = _mm_aesenc_si128(t,_k.ni.k[9]);
  421. t = _mm_aesenc_si128(t,_k.ni.k[10]);
  422. t = _mm_aesenc_si128(t,_k.ni.k[11]);
  423. t = _mm_aesenc_si128(t,_k.ni.k[12]);
  424. t = _mm_aesenc_si128(t,_k.ni.k[13]);
  425. t = _mm_aesenclast_si128(t,_k.ni.k[14]);
  426. b = _mm_xor_si128(t,b);
  427. memcpy(out,&b,rem);
  428. memset((u_char*)&b + rem,0,16 - rem);
  429. return _ghash_aesni(_k.ni.h,y,b);
  430. }
  431. inline void _encrypt_gcm256_aesni(unsigned int len,const uint8_t *in,uint8_t *out,const uint8_t *iv,unsigned int alen,const uint8_t *assoc,uint8_t *icv,unsigned int icvsize) const
  432. {
  433. __m128i j = _create_j_aesni(iv);
  434. __m128i cb = _increment_be_aesni(j);
  435. __m128i y = _icv_header_aesni(assoc,alen);
  436. unsigned int blocks = len / 16;
  437. unsigned int pblocks = blocks - (blocks % 4);
  438. unsigned int rem = len % 16;
  439. __m128i *bi = (__m128i *)in;
  440. __m128i *bo = (__m128i *)out;
  441. __m128i k0 = _k.ni.k[0];
  442. __m128i k1 = _k.ni.k[1];
  443. __m128i k2 = _k.ni.k[2];
  444. __m128i k3 = _k.ni.k[3];
  445. __m128i k4 = _k.ni.k[4];
  446. __m128i k5 = _k.ni.k[5];
  447. __m128i k6 = _k.ni.k[6];
  448. __m128i k7 = _k.ni.k[7];
  449. __m128i k8 = _k.ni.k[8];
  450. __m128i k9 = _k.ni.k[9];
  451. __m128i k10 = _k.ni.k[10];
  452. __m128i k11 = _k.ni.k[11];
  453. __m128i k12 = _k.ni.k[12];
  454. __m128i k13 = _k.ni.k[13];
  455. __m128i k14 = _k.ni.k[14];
  456. unsigned int i;
  457. for (i=0;i<pblocks;i+=4) {
  458. __m128i d1 = _mm_loadu_si128(bi + i + 0);
  459. __m128i d2 = _mm_loadu_si128(bi + i + 1);
  460. __m128i d3 = _mm_loadu_si128(bi + i + 2);
  461. __m128i d4 = _mm_loadu_si128(bi + i + 3);
  462. __m128i t1 = _mm_xor_si128(cb,k0);
  463. cb = _increment_be_aesni(cb);
  464. __m128i t2 = _mm_xor_si128(cb,k0);
  465. cb = _increment_be_aesni(cb);
  466. __m128i t3 = _mm_xor_si128(cb,k0);
  467. cb = _increment_be_aesni(cb);
  468. __m128i t4 = _mm_xor_si128(cb,k0);
  469. cb = _increment_be_aesni(cb);
  470. t1 = _mm_aesenc_si128(t1,k1);
  471. t2 = _mm_aesenc_si128(t2,k1);
  472. t3 = _mm_aesenc_si128(t3,k1);
  473. t4 = _mm_aesenc_si128(t4,k1);
  474. t1 = _mm_aesenc_si128(t1,k2);
  475. t2 = _mm_aesenc_si128(t2,k2);
  476. t3 = _mm_aesenc_si128(t3,k2);
  477. t4 = _mm_aesenc_si128(t4,k2);
  478. t1 = _mm_aesenc_si128(t1,k3);
  479. t2 = _mm_aesenc_si128(t2,k3);
  480. t3 = _mm_aesenc_si128(t3,k3);
  481. t4 = _mm_aesenc_si128(t4,k3);
  482. t1 = _mm_aesenc_si128(t1,k4);
  483. t2 = _mm_aesenc_si128(t2,k4);
  484. t3 = _mm_aesenc_si128(t3,k4);
  485. t4 = _mm_aesenc_si128(t4,k4);
  486. t1 = _mm_aesenc_si128(t1,k5);
  487. t2 = _mm_aesenc_si128(t2,k5);
  488. t3 = _mm_aesenc_si128(t3,k5);
  489. t4 = _mm_aesenc_si128(t4,k5);
  490. t1 = _mm_aesenc_si128(t1,k6);
  491. t2 = _mm_aesenc_si128(t2,k6);
  492. t3 = _mm_aesenc_si128(t3,k6);
  493. t4 = _mm_aesenc_si128(t4,k6);
  494. t1 = _mm_aesenc_si128(t1,k7);
  495. t2 = _mm_aesenc_si128(t2,k7);
  496. t3 = _mm_aesenc_si128(t3,k7);
  497. t4 = _mm_aesenc_si128(t4,k7);
  498. t1 = _mm_aesenc_si128(t1,k8);
  499. t2 = _mm_aesenc_si128(t2,k8);
  500. t3 = _mm_aesenc_si128(t3,k8);
  501. t4 = _mm_aesenc_si128(t4,k8);
  502. t1 = _mm_aesenc_si128(t1,k9);
  503. t2 = _mm_aesenc_si128(t2,k9);
  504. t3 = _mm_aesenc_si128(t3,k9);
  505. t4 = _mm_aesenc_si128(t4,k9);
  506. t1 = _mm_aesenc_si128(t1,k10);
  507. t2 = _mm_aesenc_si128(t2,k10);
  508. t3 = _mm_aesenc_si128(t3,k10);
  509. t4 = _mm_aesenc_si128(t4,k10);
  510. t1 = _mm_aesenc_si128(t1,k11);
  511. t2 = _mm_aesenc_si128(t2,k11);
  512. t3 = _mm_aesenc_si128(t3,k11);
  513. t4 = _mm_aesenc_si128(t4,k11);
  514. t1 = _mm_aesenc_si128(t1,k12);
  515. t2 = _mm_aesenc_si128(t2,k12);
  516. t3 = _mm_aesenc_si128(t3,k12);
  517. t4 = _mm_aesenc_si128(t4,k12);
  518. t1 = _mm_aesenc_si128(t1,k13);
  519. t2 = _mm_aesenc_si128(t2,k13);
  520. t3 = _mm_aesenc_si128(t3,k13);
  521. t4 = _mm_aesenc_si128(t4,k13);
  522. t1 = _mm_aesenclast_si128(t1,k14);
  523. t2 = _mm_aesenclast_si128(t2,k14);
  524. t3 = _mm_aesenclast_si128(t3,k14);
  525. t4 = _mm_aesenclast_si128(t4,k14);
  526. t1 = _mm_xor_si128(t1,d1);
  527. t2 = _mm_xor_si128(t2,d2);
  528. t3 = _mm_xor_si128(t3,d3);
  529. t4 = _mm_xor_si128(t4,d4);
  530. y = _mm_xor_si128(y,t1);
  531. y = _mult4xor_aesni(_k.ni.hhhh,_k.ni.hhh,_k.ni.hh,_k.ni.h,y,t2,t3,t4);
  532. _mm_storeu_si128(bo + i + 0,t1);
  533. _mm_storeu_si128(bo + i + 1,t2);
  534. _mm_storeu_si128(bo + i + 2,t3);
  535. _mm_storeu_si128(bo + i + 3,t4);
  536. }
  537. for (i=pblocks;i<blocks;++i) {
  538. __m128i d1 = _mm_loadu_si128(bi + i);
  539. __m128i t1 = _mm_xor_si128(cb,k0);
  540. t1 = _mm_aesenc_si128(t1,k1);
  541. t1 = _mm_aesenc_si128(t1,k2);
  542. t1 = _mm_aesenc_si128(t1,k3);
  543. t1 = _mm_aesenc_si128(t1,k4);
  544. t1 = _mm_aesenc_si128(t1,k5);
  545. t1 = _mm_aesenc_si128(t1,k6);
  546. t1 = _mm_aesenc_si128(t1,k7);
  547. t1 = _mm_aesenc_si128(t1,k8);
  548. t1 = _mm_aesenc_si128(t1,k9);
  549. t1 = _mm_aesenc_si128(t1,k10);
  550. t1 = _mm_aesenc_si128(t1,k11);
  551. t1 = _mm_aesenc_si128(t1,k12);
  552. t1 = _mm_aesenc_si128(t1,k13);
  553. t1 = _mm_aesenclast_si128(t1,k14);
  554. t1 = _mm_xor_si128(t1,d1);
  555. _mm_storeu_si128(bo + i,t1);
  556. y = _ghash_aesni(_k.ni.h,y,t1);
  557. cb = _increment_be_aesni(cb);
  558. }
  559. if (rem)
  560. y = _encrypt_gcm_rem_aesni(rem,bi + blocks,bo + blocks,cb,y);
  561. y = _icv_tailer_aesni(y,alen,len);
  562. _icv_crypt_aesni(y,j,icv,icvsize);
  563. }
  564. inline __m128i _decrypt_gcm_rem_aesni(unsigned int rem,const void *in,void *out,__m128i cb,__m128i y)
  565. {
  566. __m128i t,b;
  567. memset(&b,0,sizeof(b));
  568. memcpy(&b,in,rem);
  569. y = _ghash_aesni(_k.ni.h,y,b);
  570. t = _mm_xor_si128(cb,_k.ni.k[0]);
  571. t = _mm_aesenc_si128(t,_k.ni.k[1]);
  572. t = _mm_aesenc_si128(t,_k.ni.k[2]);
  573. t = _mm_aesenc_si128(t,_k.ni.k[3]);
  574. t = _mm_aesenc_si128(t,_k.ni.k[4]);
  575. t = _mm_aesenc_si128(t,_k.ni.k[5]);
  576. t = _mm_aesenc_si128(t,_k.ni.k[6]);
  577. t = _mm_aesenc_si128(t,_k.ni.k[7]);
  578. t = _mm_aesenc_si128(t,_k.ni.k[8]);
  579. t = _mm_aesenc_si128(t,_k.ni.k[9]);
  580. t = _mm_aesenc_si128(t,_k.ni.k[10]);
  581. t = _mm_aesenc_si128(t,_k.ni.k[11]);
  582. t = _mm_aesenc_si128(t,_k.ni.k[12]);
  583. t = _mm_aesenc_si128(t,_k.ni.k[13]);
  584. t = _mm_aesenclast_si128(t,_k.ni.k[14]);
  585. b = _mm_xor_si128(t,b);
  586. memcpy(out,&b,rem);
  587. return y;
  588. }
  589. inline void _decrypt_gcm256_aesni(unsigned int len,const uint8_t *in,uint8_t *out,const uint8_t *iv,unsigned int alen,const uint8_t *assoc,uint8_t *icv,unsigned int icvsize)
  590. {
  591. __m128i j = _create_j_aesni(iv);
  592. __m128i cb = _increment_be_aesni(j);
  593. __m128i y = _icv_header_aesni(assoc,alen);
  594. unsigned int blocks = len / 16;
  595. unsigned int pblocks = blocks - (blocks % 4);
  596. unsigned int rem = len % 16;
  597. __m128i *bi = (__m128i *)in;
  598. __m128i *bo = (__m128i *)out;
  599. __m128i k0 = _k.ni.k[0];
  600. __m128i k1 = _k.ni.k[1];
  601. __m128i k2 = _k.ni.k[2];
  602. __m128i k3 = _k.ni.k[3];
  603. __m128i k4 = _k.ni.k[4];
  604. __m128i k5 = _k.ni.k[5];
  605. __m128i k6 = _k.ni.k[6];
  606. __m128i k7 = _k.ni.k[7];
  607. __m128i k8 = _k.ni.k[8];
  608. __m128i k9 = _k.ni.k[9];
  609. __m128i k10 = _k.ni.k[10];
  610. __m128i k11 = _k.ni.k[11];
  611. __m128i k12 = _k.ni.k[12];
  612. __m128i k13 = _k.ni.k[13];
  613. __m128i k14 = _k.ni.k[14];
  614. unsigned int i;
  615. for (i=0;i<pblocks;i+=4) {
  616. __m128i d1 = _mm_loadu_si128(bi + i + 0);
  617. __m128i d2 = _mm_loadu_si128(bi + i + 1);
  618. __m128i d3 = _mm_loadu_si128(bi + i + 2);
  619. __m128i d4 = _mm_loadu_si128(bi + i + 3);
  620. y = _mm_xor_si128(y,d1);
  621. y = _mult4xor_aesni(_k.ni.hhhh,_k.ni.hhh,_k.ni.hh,_k.ni.h,y,d2,d3,d4);
  622. __m128i t1 = _mm_xor_si128(cb,k0);
  623. cb = _increment_be_aesni(cb);
  624. __m128i t2 = _mm_xor_si128(cb,k0);
  625. cb = _increment_be_aesni(cb);
  626. __m128i t3 = _mm_xor_si128(cb,k0);
  627. cb = _increment_be_aesni(cb);
  628. __m128i t4 = _mm_xor_si128(cb,k0);
  629. cb = _increment_be_aesni(cb);
  630. t1 = _mm_aesenc_si128(t1,k1);
  631. t2 = _mm_aesenc_si128(t2,k1);
  632. t3 = _mm_aesenc_si128(t3,k1);
  633. t4 = _mm_aesenc_si128(t4,k1);
  634. t1 = _mm_aesenc_si128(t1,k2);
  635. t2 = _mm_aesenc_si128(t2,k2);
  636. t3 = _mm_aesenc_si128(t3,k2);
  637. t4 = _mm_aesenc_si128(t4,k2);
  638. t1 = _mm_aesenc_si128(t1,k3);
  639. t2 = _mm_aesenc_si128(t2,k3);
  640. t3 = _mm_aesenc_si128(t3,k3);
  641. t4 = _mm_aesenc_si128(t4,k3);
  642. t1 = _mm_aesenc_si128(t1,k4);
  643. t2 = _mm_aesenc_si128(t2,k4);
  644. t3 = _mm_aesenc_si128(t3,k4);
  645. t4 = _mm_aesenc_si128(t4,k4);
  646. t1 = _mm_aesenc_si128(t1,k5);
  647. t2 = _mm_aesenc_si128(t2,k5);
  648. t3 = _mm_aesenc_si128(t3,k5);
  649. t4 = _mm_aesenc_si128(t4,k5);
  650. t1 = _mm_aesenc_si128(t1,k6);
  651. t2 = _mm_aesenc_si128(t2,k6);
  652. t3 = _mm_aesenc_si128(t3,k6);
  653. t4 = _mm_aesenc_si128(t4,k6);
  654. t1 = _mm_aesenc_si128(t1,k7);
  655. t2 = _mm_aesenc_si128(t2,k7);
  656. t3 = _mm_aesenc_si128(t3,k7);
  657. t4 = _mm_aesenc_si128(t4,k7);
  658. t1 = _mm_aesenc_si128(t1,k8);
  659. t2 = _mm_aesenc_si128(t2,k8);
  660. t3 = _mm_aesenc_si128(t3,k8);
  661. t4 = _mm_aesenc_si128(t4,k8);
  662. t1 = _mm_aesenc_si128(t1,k9);
  663. t2 = _mm_aesenc_si128(t2,k9);
  664. t3 = _mm_aesenc_si128(t3,k9);
  665. t4 = _mm_aesenc_si128(t4,k9);
  666. t1 = _mm_aesenc_si128(t1,k10);
  667. t2 = _mm_aesenc_si128(t2,k10);
  668. t3 = _mm_aesenc_si128(t3,k10);
  669. t4 = _mm_aesenc_si128(t4,k10);
  670. t1 = _mm_aesenc_si128(t1,k11);
  671. t2 = _mm_aesenc_si128(t2,k11);
  672. t3 = _mm_aesenc_si128(t3,k11);
  673. t4 = _mm_aesenc_si128(t4,k11);
  674. t1 = _mm_aesenc_si128(t1,k12);
  675. t2 = _mm_aesenc_si128(t2,k12);
  676. t3 = _mm_aesenc_si128(t3,k12);
  677. t4 = _mm_aesenc_si128(t4,k12);
  678. t1 = _mm_aesenc_si128(t1,k13);
  679. t2 = _mm_aesenc_si128(t2,k13);
  680. t3 = _mm_aesenc_si128(t3,k13);
  681. t4 = _mm_aesenc_si128(t4,k13);
  682. t1 = _mm_aesenclast_si128(t1,k14);
  683. t2 = _mm_aesenclast_si128(t2,k14);
  684. t3 = _mm_aesenclast_si128(t3,k14);
  685. t4 = _mm_aesenclast_si128(t4,k14);
  686. t1 = _mm_xor_si128(t1,d1);
  687. t2 = _mm_xor_si128(t2,d2);
  688. t3 = _mm_xor_si128(t3,d3);
  689. t4 = _mm_xor_si128(t4,d4);
  690. _mm_storeu_si128(bo + i + 0,t1);
  691. _mm_storeu_si128(bo + i + 1,t2);
  692. _mm_storeu_si128(bo + i + 2,t3);
  693. _mm_storeu_si128(bo + i + 3,t4);
  694. }
  695. for (i=pblocks;i<blocks;i++) {
  696. __m128i d1 = _mm_loadu_si128(bi + i);
  697. y = _ghash_aesni(_k.ni.h,y,d1);
  698. __m128i t1 = _mm_xor_si128(cb,k0);
  699. t1 = _mm_aesenc_si128(t1,k1);
  700. t1 = _mm_aesenc_si128(t1,k2);
  701. t1 = _mm_aesenc_si128(t1,k3);
  702. t1 = _mm_aesenc_si128(t1,k4);
  703. t1 = _mm_aesenc_si128(t1,k5);
  704. t1 = _mm_aesenc_si128(t1,k6);
  705. t1 = _mm_aesenc_si128(t1,k7);
  706. t1 = _mm_aesenc_si128(t1,k8);
  707. t1 = _mm_aesenc_si128(t1,k9);
  708. t1 = _mm_aesenc_si128(t1,k10);
  709. t1 = _mm_aesenc_si128(t1,k11);
  710. t1 = _mm_aesenc_si128(t1,k12);
  711. t1 = _mm_aesenc_si128(t1,k13);
  712. t1 = _mm_aesenclast_si128(t1,k14);
  713. t1 = _mm_xor_si128(t1,d1);
  714. _mm_storeu_si128(bo + i,t1);
  715. cb = _increment_be_aesni(cb);
  716. }
  717. if (rem)
  718. y = _decrypt_gcm_rem_aesni(rem,bi + blocks,bo + blocks,cb,y);
  719. y = _icv_tailer_aesni(y,alen,len);
  720. _icv_crypt_aesni(y,j,icv,icvsize);
  721. }
  722. #endif /* ZT_AES_AESNI ******************************************************/
  723. };
  724. } // namespace ZeroTier
  725. #endif