ecp_nistp224.c 59 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752
  1. /*
  2. * Copyright 2010-2020 The OpenSSL Project Authors. All Rights Reserved.
  3. *
  4. * Licensed under the Apache License 2.0 (the "License"). You may not use
  5. * this file except in compliance with the License. You can obtain a copy
  6. * in the file LICENSE in the source distribution or at
  7. * https://www.openssl.org/source/license.html
  8. */
  9. /* Copyright 2011 Google Inc.
  10. *
  11. * Licensed under the Apache License, Version 2.0 (the "License");
  12. *
  13. * you may not use this file except in compliance with the License.
  14. * You may obtain a copy of the License at
  15. *
  16. * http://www.apache.org/licenses/LICENSE-2.0
  17. *
  18. * Unless required by applicable law or agreed to in writing, software
  19. * distributed under the License is distributed on an "AS IS" BASIS,
  20. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  21. * See the License for the specific language governing permissions and
  22. * limitations under the License.
  23. */
  24. /*
  25. * ECDSA low level APIs are deprecated for public use, but still ok for
  26. * internal use.
  27. */
  28. #include "internal/deprecated.h"
  29. /*
  30. * A 64-bit implementation of the NIST P-224 elliptic curve point multiplication
  31. *
  32. * Inspired by Daniel J. Bernstein's public domain nistp224 implementation
  33. * and Adam Langley's public domain 64-bit C implementation of curve25519
  34. */
  35. #include <openssl/opensslconf.h>
  36. #include <stdint.h>
  37. #include <string.h>
  38. #include <openssl/err.h>
  39. #include "ec_local.h"
  40. #if defined(__SIZEOF_INT128__) && __SIZEOF_INT128__==16
  41. /* even with gcc, the typedef won't work for 32-bit platforms */
  42. typedef __uint128_t uint128_t; /* nonstandard; implemented by gcc on 64-bit
  43. * platforms */
  44. #else
  45. # error "Your compiler doesn't appear to support 128-bit integer types"
  46. #endif
  47. typedef uint8_t u8;
  48. typedef uint64_t u64;
  49. /******************************************************************************/
  50. /*-
  51. * INTERNAL REPRESENTATION OF FIELD ELEMENTS
  52. *
  53. * Field elements are represented as a_0 + 2^56*a_1 + 2^112*a_2 + 2^168*a_3
  54. * using 64-bit coefficients called 'limbs',
  55. * and sometimes (for multiplication results) as
  56. * b_0 + 2^56*b_1 + 2^112*b_2 + 2^168*b_3 + 2^224*b_4 + 2^280*b_5 + 2^336*b_6
  57. * using 128-bit coefficients called 'widelimbs'.
  58. * A 4-limb representation is an 'felem';
  59. * a 7-widelimb representation is a 'widefelem'.
  60. * Even within felems, bits of adjacent limbs overlap, and we don't always
  61. * reduce the representations: we ensure that inputs to each felem
  62. * multiplication satisfy a_i < 2^60, so outputs satisfy b_i < 4*2^60*2^60,
  63. * and fit into a 128-bit word without overflow. The coefficients are then
  64. * again partially reduced to obtain an felem satisfying a_i < 2^57.
  65. * We only reduce to the unique minimal representation at the end of the
  66. * computation.
  67. */
  68. typedef uint64_t limb;
  69. typedef uint64_t limb_aX __attribute((__aligned__(1)));
  70. typedef uint128_t widelimb;
  71. typedef limb felem[4];
  72. typedef widelimb widefelem[7];
  73. /*
  74. * Field element represented as a byte array. 28*8 = 224 bits is also the
  75. * group order size for the elliptic curve, and we also use this type for
  76. * scalars for point multiplication.
  77. */
  78. typedef u8 felem_bytearray[28];
  79. static const felem_bytearray nistp224_curve_params[5] = {
  80. {0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, /* p */
  81. 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0x00, 0x00, 0x00, 0x00,
  82. 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01},
  83. {0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, /* a */
  84. 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFE, 0xFF, 0xFF, 0xFF, 0xFF,
  85. 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFE},
  86. {0xB4, 0x05, 0x0A, 0x85, 0x0C, 0x04, 0xB3, 0xAB, 0xF5, 0x41, /* b */
  87. 0x32, 0x56, 0x50, 0x44, 0xB0, 0xB7, 0xD7, 0xBF, 0xD8, 0xBA,
  88. 0x27, 0x0B, 0x39, 0x43, 0x23, 0x55, 0xFF, 0xB4},
  89. {0xB7, 0x0E, 0x0C, 0xBD, 0x6B, 0xB4, 0xBF, 0x7F, 0x32, 0x13, /* x */
  90. 0x90, 0xB9, 0x4A, 0x03, 0xC1, 0xD3, 0x56, 0xC2, 0x11, 0x22,
  91. 0x34, 0x32, 0x80, 0xD6, 0x11, 0x5C, 0x1D, 0x21},
  92. {0xbd, 0x37, 0x63, 0x88, 0xb5, 0xf7, 0x23, 0xfb, 0x4c, 0x22, /* y */
  93. 0xdf, 0xe6, 0xcd, 0x43, 0x75, 0xa0, 0x5a, 0x07, 0x47, 0x64,
  94. 0x44, 0xd5, 0x81, 0x99, 0x85, 0x00, 0x7e, 0x34}
  95. };
  96. /*-
  97. * Precomputed multiples of the standard generator
  98. * Points are given in coordinates (X, Y, Z) where Z normally is 1
  99. * (0 for the point at infinity).
  100. * For each field element, slice a_0 is word 0, etc.
  101. *
  102. * The table has 2 * 16 elements, starting with the following:
  103. * index | bits | point
  104. * ------+---------+------------------------------
  105. * 0 | 0 0 0 0 | 0G
  106. * 1 | 0 0 0 1 | 1G
  107. * 2 | 0 0 1 0 | 2^56G
  108. * 3 | 0 0 1 1 | (2^56 + 1)G
  109. * 4 | 0 1 0 0 | 2^112G
  110. * 5 | 0 1 0 1 | (2^112 + 1)G
  111. * 6 | 0 1 1 0 | (2^112 + 2^56)G
  112. * 7 | 0 1 1 1 | (2^112 + 2^56 + 1)G
  113. * 8 | 1 0 0 0 | 2^168G
  114. * 9 | 1 0 0 1 | (2^168 + 1)G
  115. * 10 | 1 0 1 0 | (2^168 + 2^56)G
  116. * 11 | 1 0 1 1 | (2^168 + 2^56 + 1)G
  117. * 12 | 1 1 0 0 | (2^168 + 2^112)G
  118. * 13 | 1 1 0 1 | (2^168 + 2^112 + 1)G
  119. * 14 | 1 1 1 0 | (2^168 + 2^112 + 2^56)G
  120. * 15 | 1 1 1 1 | (2^168 + 2^112 + 2^56 + 1)G
  121. * followed by a copy of this with each element multiplied by 2^28.
  122. *
  123. * The reason for this is so that we can clock bits into four different
  124. * locations when doing simple scalar multiplies against the base point,
  125. * and then another four locations using the second 16 elements.
  126. */
  127. static const felem gmul[2][16][3] = {
  128. {{{0, 0, 0, 0},
  129. {0, 0, 0, 0},
  130. {0, 0, 0, 0}},
  131. {{0x3280d6115c1d21, 0xc1d356c2112234, 0x7f321390b94a03, 0xb70e0cbd6bb4bf},
  132. {0xd5819985007e34, 0x75a05a07476444, 0xfb4c22dfe6cd43, 0xbd376388b5f723},
  133. {1, 0, 0, 0}},
  134. {{0xfd9675666ebbe9, 0xbca7664d40ce5e, 0x2242df8d8a2a43, 0x1f49bbb0f99bc5},
  135. {0x29e0b892dc9c43, 0xece8608436e662, 0xdc858f185310d0, 0x9812dd4eb8d321},
  136. {1, 0, 0, 0}},
  137. {{0x6d3e678d5d8eb8, 0x559eed1cb362f1, 0x16e9a3bbce8a3f, 0xeedcccd8c2a748},
  138. {0xf19f90ed50266d, 0xabf2b4bf65f9df, 0x313865468fafec, 0x5cb379ba910a17},
  139. {1, 0, 0, 0}},
  140. {{0x0641966cab26e3, 0x91fb2991fab0a0, 0xefec27a4e13a0b, 0x0499aa8a5f8ebe},
  141. {0x7510407766af5d, 0x84d929610d5450, 0x81d77aae82f706, 0x6916f6d4338c5b},
  142. {1, 0, 0, 0}},
  143. {{0xea95ac3b1f15c6, 0x086000905e82d4, 0xdd323ae4d1c8b1, 0x932b56be7685a3},
  144. {0x9ef93dea25dbbf, 0x41665960f390f0, 0xfdec76dbe2a8a7, 0x523e80f019062a},
  145. {1, 0, 0, 0}},
  146. {{0x822fdd26732c73, 0xa01c83531b5d0f, 0x363f37347c1ba4, 0xc391b45c84725c},
  147. {0xbbd5e1b2d6ad24, 0xddfbcde19dfaec, 0xc393da7e222a7f, 0x1efb7890ede244},
  148. {1, 0, 0, 0}},
  149. {{0x4c9e90ca217da1, 0xd11beca79159bb, 0xff8d33c2c98b7c, 0x2610b39409f849},
  150. {0x44d1352ac64da0, 0xcdbb7b2c46b4fb, 0x966c079b753c89, 0xfe67e4e820b112},
  151. {1, 0, 0, 0}},
  152. {{0xe28cae2df5312d, 0xc71b61d16f5c6e, 0x79b7619a3e7c4c, 0x05c73240899b47},
  153. {0x9f7f6382c73e3a, 0x18615165c56bda, 0x641fab2116fd56, 0x72855882b08394},
  154. {1, 0, 0, 0}},
  155. {{0x0469182f161c09, 0x74a98ca8d00fb5, 0xb89da93489a3e0, 0x41c98768fb0c1d},
  156. {0xe5ea05fb32da81, 0x3dce9ffbca6855, 0x1cfe2d3fbf59e6, 0x0e5e03408738a7},
  157. {1, 0, 0, 0}},
  158. {{0xdab22b2333e87f, 0x4430137a5dd2f6, 0xe03ab9f738beb8, 0xcb0c5d0dc34f24},
  159. {0x764a7df0c8fda5, 0x185ba5c3fa2044, 0x9281d688bcbe50, 0xc40331df893881},
  160. {1, 0, 0, 0}},
  161. {{0xb89530796f0f60, 0xade92bd26909a3, 0x1a0c83fb4884da, 0x1765bf22a5a984},
  162. {0x772a9ee75db09e, 0x23bc6c67cec16f, 0x4c1edba8b14e2f, 0xe2a215d9611369},
  163. {1, 0, 0, 0}},
  164. {{0x571e509fb5efb3, 0xade88696410552, 0xc8ae85fada74fe, 0x6c7e4be83bbde3},
  165. {0xff9f51160f4652, 0xb47ce2495a6539, 0xa2946c53b582f4, 0x286d2db3ee9a60},
  166. {1, 0, 0, 0}},
  167. {{0x40bbd5081a44af, 0x0995183b13926c, 0xbcefba6f47f6d0, 0x215619e9cc0057},
  168. {0x8bc94d3b0df45e, 0xf11c54a3694f6f, 0x8631b93cdfe8b5, 0xe7e3f4b0982db9},
  169. {1, 0, 0, 0}},
  170. {{0xb17048ab3e1c7b, 0xac38f36ff8a1d8, 0x1c29819435d2c6, 0xc813132f4c07e9},
  171. {0x2891425503b11f, 0x08781030579fea, 0xf5426ba5cc9674, 0x1e28ebf18562bc},
  172. {1, 0, 0, 0}},
  173. {{0x9f31997cc864eb, 0x06cd91d28b5e4c, 0xff17036691a973, 0xf1aef351497c58},
  174. {0xdd1f2d600564ff, 0xdead073b1402db, 0x74a684435bd693, 0xeea7471f962558},
  175. {1, 0, 0, 0}}},
  176. {{{0, 0, 0, 0},
  177. {0, 0, 0, 0},
  178. {0, 0, 0, 0}},
  179. {{0x9665266dddf554, 0x9613d78b60ef2d, 0xce27a34cdba417, 0xd35ab74d6afc31},
  180. {0x85ccdd22deb15e, 0x2137e5783a6aab, 0xa141cffd8c93c6, 0x355a1830e90f2d},
  181. {1, 0, 0, 0}},
  182. {{0x1a494eadaade65, 0xd6da4da77fe53c, 0xe7992996abec86, 0x65c3553c6090e3},
  183. {0xfa610b1fb09346, 0xf1c6540b8a4aaf, 0xc51a13ccd3cbab, 0x02995b1b18c28a},
  184. {1, 0, 0, 0}},
  185. {{0x7874568e7295ef, 0x86b419fbe38d04, 0xdc0690a7550d9a, 0xd3966a44beac33},
  186. {0x2b7280ec29132f, 0xbeaa3b6a032df3, 0xdc7dd88ae41200, 0xd25e2513e3a100},
  187. {1, 0, 0, 0}},
  188. {{0x924857eb2efafd, 0xac2bce41223190, 0x8edaa1445553fc, 0x825800fd3562d5},
  189. {0x8d79148ea96621, 0x23a01c3dd9ed8d, 0xaf8b219f9416b5, 0xd8db0cc277daea},
  190. {1, 0, 0, 0}},
  191. {{0x76a9c3b1a700f0, 0xe9acd29bc7e691, 0x69212d1a6b0327, 0x6322e97fe154be},
  192. {0x469fc5465d62aa, 0x8d41ed18883b05, 0x1f8eae66c52b88, 0xe4fcbe9325be51},
  193. {1, 0, 0, 0}},
  194. {{0x825fdf583cac16, 0x020b857c7b023a, 0x683c17744b0165, 0x14ffd0a2daf2f1},
  195. {0x323b36184218f9, 0x4944ec4e3b47d4, 0xc15b3080841acf, 0x0bced4b01a28bb},
  196. {1, 0, 0, 0}},
  197. {{0x92ac22230df5c4, 0x52f33b4063eda8, 0xcb3f19870c0c93, 0x40064f2ba65233},
  198. {0xfe16f0924f8992, 0x012da25af5b517, 0x1a57bb24f723a6, 0x06f8bc76760def},
  199. {1, 0, 0, 0}},
  200. {{0x4a7084f7817cb9, 0xbcab0738ee9a78, 0x3ec11e11d9c326, 0xdc0fe90e0f1aae},
  201. {0xcf639ea5f98390, 0x5c350aa22ffb74, 0x9afae98a4047b7, 0x956ec2d617fc45},
  202. {1, 0, 0, 0}},
  203. {{0x4306d648c1be6a, 0x9247cd8bc9a462, 0xf5595e377d2f2e, 0xbd1c3caff1a52e},
  204. {0x045e14472409d0, 0x29f3e17078f773, 0x745a602b2d4f7d, 0x191837685cdfbb},
  205. {1, 0, 0, 0}},
  206. {{0x5b6ee254a8cb79, 0x4953433f5e7026, 0xe21faeb1d1def4, 0xc4c225785c09de},
  207. {0x307ce7bba1e518, 0x31b125b1036db8, 0x47e91868839e8f, 0xc765866e33b9f3},
  208. {1, 0, 0, 0}},
  209. {{0x3bfece24f96906, 0x4794da641e5093, 0xde5df64f95db26, 0x297ecd89714b05},
  210. {0x701bd3ebb2c3aa, 0x7073b4f53cb1d5, 0x13c5665658af16, 0x9895089d66fe58},
  211. {1, 0, 0, 0}},
  212. {{0x0fef05f78c4790, 0x2d773633b05d2e, 0x94229c3a951c94, 0xbbbd70df4911bb},
  213. {0xb2c6963d2c1168, 0x105f47a72b0d73, 0x9fdf6111614080, 0x7b7e94b39e67b0},
  214. {1, 0, 0, 0}},
  215. {{0xad1a7d6efbe2b3, 0xf012482c0da69d, 0x6b3bdf12438345, 0x40d7558d7aa4d9},
  216. {0x8a09fffb5c6d3d, 0x9a356e5d9ffd38, 0x5973f15f4f9b1c, 0xdcd5f59f63c3ea},
  217. {1, 0, 0, 0}},
  218. {{0xacf39f4c5ca7ab, 0x4c8071cc5fd737, 0xc64e3602cd1184, 0x0acd4644c9abba},
  219. {0x6c011a36d8bf6e, 0xfecd87ba24e32a, 0x19f6f56574fad8, 0x050b204ced9405},
  220. {1, 0, 0, 0}},
  221. {{0xed4f1cae7d9a96, 0x5ceef7ad94c40a, 0x778e4a3bf3ef9b, 0x7405783dc3b55e},
  222. {0x32477c61b6e8c6, 0xb46a97570f018b, 0x91176d0a7e95d1, 0x3df90fbc4c7d0e},
  223. {1, 0, 0, 0}}}
  224. };
  225. /* Precomputation for the group generator. */
  226. struct nistp224_pre_comp_st {
  227. felem g_pre_comp[2][16][3];
  228. CRYPTO_REF_COUNT references;
  229. CRYPTO_RWLOCK *lock;
  230. };
  231. const EC_METHOD *EC_GFp_nistp224_method(void)
  232. {
  233. static const EC_METHOD ret = {
  234. EC_FLAGS_DEFAULT_OCT,
  235. NID_X9_62_prime_field,
  236. ec_GFp_nistp224_group_init,
  237. ec_GFp_simple_group_finish,
  238. ec_GFp_simple_group_clear_finish,
  239. ec_GFp_nist_group_copy,
  240. ec_GFp_nistp224_group_set_curve,
  241. ec_GFp_simple_group_get_curve,
  242. ec_GFp_simple_group_get_degree,
  243. ec_group_simple_order_bits,
  244. ec_GFp_simple_group_check_discriminant,
  245. ec_GFp_simple_point_init,
  246. ec_GFp_simple_point_finish,
  247. ec_GFp_simple_point_clear_finish,
  248. ec_GFp_simple_point_copy,
  249. ec_GFp_simple_point_set_to_infinity,
  250. ec_GFp_simple_point_set_affine_coordinates,
  251. ec_GFp_nistp224_point_get_affine_coordinates,
  252. 0 /* point_set_compressed_coordinates */ ,
  253. 0 /* point2oct */ ,
  254. 0 /* oct2point */ ,
  255. ec_GFp_simple_add,
  256. ec_GFp_simple_dbl,
  257. ec_GFp_simple_invert,
  258. ec_GFp_simple_is_at_infinity,
  259. ec_GFp_simple_is_on_curve,
  260. ec_GFp_simple_cmp,
  261. ec_GFp_simple_make_affine,
  262. ec_GFp_simple_points_make_affine,
  263. ec_GFp_nistp224_points_mul,
  264. ec_GFp_nistp224_precompute_mult,
  265. ec_GFp_nistp224_have_precompute_mult,
  266. ec_GFp_nist_field_mul,
  267. ec_GFp_nist_field_sqr,
  268. 0 /* field_div */ ,
  269. ec_GFp_simple_field_inv,
  270. 0 /* field_encode */ ,
  271. 0 /* field_decode */ ,
  272. 0, /* field_set_to_one */
  273. ec_key_simple_priv2oct,
  274. ec_key_simple_oct2priv,
  275. 0, /* set private */
  276. ec_key_simple_generate_key,
  277. ec_key_simple_check_key,
  278. ec_key_simple_generate_public_key,
  279. 0, /* keycopy */
  280. 0, /* keyfinish */
  281. ecdh_simple_compute_key,
  282. ecdsa_simple_sign_setup,
  283. ecdsa_simple_sign_sig,
  284. ecdsa_simple_verify_sig,
  285. 0, /* field_inverse_mod_ord */
  286. 0, /* blind_coordinates */
  287. 0, /* ladder_pre */
  288. 0, /* ladder_step */
  289. 0 /* ladder_post */
  290. };
  291. return &ret;
  292. }
  293. /*
  294. * Helper functions to convert field elements to/from internal representation
  295. */
  296. static void bin28_to_felem(felem out, const u8 in[28])
  297. {
  298. out[0] = *((const limb *)(in)) & 0x00ffffffffffffff;
  299. out[1] = (*((const limb_aX *)(in + 7))) & 0x00ffffffffffffff;
  300. out[2] = (*((const limb_aX *)(in + 14))) & 0x00ffffffffffffff;
  301. out[3] = (*((const limb_aX *)(in + 20))) >> 8;
  302. }
  303. static void felem_to_bin28(u8 out[28], const felem in)
  304. {
  305. unsigned i;
  306. for (i = 0; i < 7; ++i) {
  307. out[i] = in[0] >> (8 * i);
  308. out[i + 7] = in[1] >> (8 * i);
  309. out[i + 14] = in[2] >> (8 * i);
  310. out[i + 21] = in[3] >> (8 * i);
  311. }
  312. }
  313. /* From OpenSSL BIGNUM to internal representation */
  314. static int BN_to_felem(felem out, const BIGNUM *bn)
  315. {
  316. felem_bytearray b_out;
  317. int num_bytes;
  318. if (BN_is_negative(bn)) {
  319. ECerr(EC_F_BN_TO_FELEM, EC_R_BIGNUM_OUT_OF_RANGE);
  320. return 0;
  321. }
  322. num_bytes = BN_bn2lebinpad(bn, b_out, sizeof(b_out));
  323. if (num_bytes < 0) {
  324. ECerr(EC_F_BN_TO_FELEM, EC_R_BIGNUM_OUT_OF_RANGE);
  325. return 0;
  326. }
  327. bin28_to_felem(out, b_out);
  328. return 1;
  329. }
  330. /* From internal representation to OpenSSL BIGNUM */
  331. static BIGNUM *felem_to_BN(BIGNUM *out, const felem in)
  332. {
  333. felem_bytearray b_out;
  334. felem_to_bin28(b_out, in);
  335. return BN_lebin2bn(b_out, sizeof(b_out), out);
  336. }
  337. /******************************************************************************/
  338. /*-
  339. * FIELD OPERATIONS
  340. *
  341. * Field operations, using the internal representation of field elements.
  342. * NB! These operations are specific to our point multiplication and cannot be
  343. * expected to be correct in general - e.g., multiplication with a large scalar
  344. * will cause an overflow.
  345. *
  346. */
  347. static void felem_one(felem out)
  348. {
  349. out[0] = 1;
  350. out[1] = 0;
  351. out[2] = 0;
  352. out[3] = 0;
  353. }
  354. static void felem_assign(felem out, const felem in)
  355. {
  356. out[0] = in[0];
  357. out[1] = in[1];
  358. out[2] = in[2];
  359. out[3] = in[3];
  360. }
  361. /* Sum two field elements: out += in */
  362. static void felem_sum(felem out, const felem in)
  363. {
  364. out[0] += in[0];
  365. out[1] += in[1];
  366. out[2] += in[2];
  367. out[3] += in[3];
  368. }
  369. /* Subtract field elements: out -= in */
  370. /* Assumes in[i] < 2^57 */
  371. static void felem_diff(felem out, const felem in)
  372. {
  373. static const limb two58p2 = (((limb) 1) << 58) + (((limb) 1) << 2);
  374. static const limb two58m2 = (((limb) 1) << 58) - (((limb) 1) << 2);
  375. static const limb two58m42m2 = (((limb) 1) << 58) -
  376. (((limb) 1) << 42) - (((limb) 1) << 2);
  377. /* Add 0 mod 2^224-2^96+1 to ensure out > in */
  378. out[0] += two58p2;
  379. out[1] += two58m42m2;
  380. out[2] += two58m2;
  381. out[3] += two58m2;
  382. out[0] -= in[0];
  383. out[1] -= in[1];
  384. out[2] -= in[2];
  385. out[3] -= in[3];
  386. }
  387. /* Subtract in unreduced 128-bit mode: out -= in */
  388. /* Assumes in[i] < 2^119 */
  389. static void widefelem_diff(widefelem out, const widefelem in)
  390. {
  391. static const widelimb two120 = ((widelimb) 1) << 120;
  392. static const widelimb two120m64 = (((widelimb) 1) << 120) -
  393. (((widelimb) 1) << 64);
  394. static const widelimb two120m104m64 = (((widelimb) 1) << 120) -
  395. (((widelimb) 1) << 104) - (((widelimb) 1) << 64);
  396. /* Add 0 mod 2^224-2^96+1 to ensure out > in */
  397. out[0] += two120;
  398. out[1] += two120m64;
  399. out[2] += two120m64;
  400. out[3] += two120;
  401. out[4] += two120m104m64;
  402. out[5] += two120m64;
  403. out[6] += two120m64;
  404. out[0] -= in[0];
  405. out[1] -= in[1];
  406. out[2] -= in[2];
  407. out[3] -= in[3];
  408. out[4] -= in[4];
  409. out[5] -= in[5];
  410. out[6] -= in[6];
  411. }
  412. /* Subtract in mixed mode: out128 -= in64 */
  413. /* in[i] < 2^63 */
  414. static void felem_diff_128_64(widefelem out, const felem in)
  415. {
  416. static const widelimb two64p8 = (((widelimb) 1) << 64) +
  417. (((widelimb) 1) << 8);
  418. static const widelimb two64m8 = (((widelimb) 1) << 64) -
  419. (((widelimb) 1) << 8);
  420. static const widelimb two64m48m8 = (((widelimb) 1) << 64) -
  421. (((widelimb) 1) << 48) - (((widelimb) 1) << 8);
  422. /* Add 0 mod 2^224-2^96+1 to ensure out > in */
  423. out[0] += two64p8;
  424. out[1] += two64m48m8;
  425. out[2] += two64m8;
  426. out[3] += two64m8;
  427. out[0] -= in[0];
  428. out[1] -= in[1];
  429. out[2] -= in[2];
  430. out[3] -= in[3];
  431. }
  432. /*
  433. * Multiply a field element by a scalar: out = out * scalar The scalars we
  434. * actually use are small, so results fit without overflow
  435. */
  436. static void felem_scalar(felem out, const limb scalar)
  437. {
  438. out[0] *= scalar;
  439. out[1] *= scalar;
  440. out[2] *= scalar;
  441. out[3] *= scalar;
  442. }
  443. /*
  444. * Multiply an unreduced field element by a scalar: out = out * scalar The
  445. * scalars we actually use are small, so results fit without overflow
  446. */
  447. static void widefelem_scalar(widefelem out, const widelimb scalar)
  448. {
  449. out[0] *= scalar;
  450. out[1] *= scalar;
  451. out[2] *= scalar;
  452. out[3] *= scalar;
  453. out[4] *= scalar;
  454. out[5] *= scalar;
  455. out[6] *= scalar;
  456. }
  457. /* Square a field element: out = in^2 */
  458. static void felem_square(widefelem out, const felem in)
  459. {
  460. limb tmp0, tmp1, tmp2;
  461. tmp0 = 2 * in[0];
  462. tmp1 = 2 * in[1];
  463. tmp2 = 2 * in[2];
  464. out[0] = ((widelimb) in[0]) * in[0];
  465. out[1] = ((widelimb) in[0]) * tmp1;
  466. out[2] = ((widelimb) in[0]) * tmp2 + ((widelimb) in[1]) * in[1];
  467. out[3] = ((widelimb) in[3]) * tmp0 + ((widelimb) in[1]) * tmp2;
  468. out[4] = ((widelimb) in[3]) * tmp1 + ((widelimb) in[2]) * in[2];
  469. out[5] = ((widelimb) in[3]) * tmp2;
  470. out[6] = ((widelimb) in[3]) * in[3];
  471. }
  472. /* Multiply two field elements: out = in1 * in2 */
  473. static void felem_mul(widefelem out, const felem in1, const felem in2)
  474. {
  475. out[0] = ((widelimb) in1[0]) * in2[0];
  476. out[1] = ((widelimb) in1[0]) * in2[1] + ((widelimb) in1[1]) * in2[0];
  477. out[2] = ((widelimb) in1[0]) * in2[2] + ((widelimb) in1[1]) * in2[1] +
  478. ((widelimb) in1[2]) * in2[0];
  479. out[3] = ((widelimb) in1[0]) * in2[3] + ((widelimb) in1[1]) * in2[2] +
  480. ((widelimb) in1[2]) * in2[1] + ((widelimb) in1[3]) * in2[0];
  481. out[4] = ((widelimb) in1[1]) * in2[3] + ((widelimb) in1[2]) * in2[2] +
  482. ((widelimb) in1[3]) * in2[1];
  483. out[5] = ((widelimb) in1[2]) * in2[3] + ((widelimb) in1[3]) * in2[2];
  484. out[6] = ((widelimb) in1[3]) * in2[3];
  485. }
  486. /*-
  487. * Reduce seven 128-bit coefficients to four 64-bit coefficients.
  488. * Requires in[i] < 2^126,
  489. * ensures out[0] < 2^56, out[1] < 2^56, out[2] < 2^56, out[3] <= 2^56 + 2^16 */
  490. static void felem_reduce(felem out, const widefelem in)
  491. {
  492. static const widelimb two127p15 = (((widelimb) 1) << 127) +
  493. (((widelimb) 1) << 15);
  494. static const widelimb two127m71 = (((widelimb) 1) << 127) -
  495. (((widelimb) 1) << 71);
  496. static const widelimb two127m71m55 = (((widelimb) 1) << 127) -
  497. (((widelimb) 1) << 71) - (((widelimb) 1) << 55);
  498. widelimb output[5];
  499. /* Add 0 mod 2^224-2^96+1 to ensure all differences are positive */
  500. output[0] = in[0] + two127p15;
  501. output[1] = in[1] + two127m71m55;
  502. output[2] = in[2] + two127m71;
  503. output[3] = in[3];
  504. output[4] = in[4];
  505. /* Eliminate in[4], in[5], in[6] */
  506. output[4] += in[6] >> 16;
  507. output[3] += (in[6] & 0xffff) << 40;
  508. output[2] -= in[6];
  509. output[3] += in[5] >> 16;
  510. output[2] += (in[5] & 0xffff) << 40;
  511. output[1] -= in[5];
  512. output[2] += output[4] >> 16;
  513. output[1] += (output[4] & 0xffff) << 40;
  514. output[0] -= output[4];
  515. /* Carry 2 -> 3 -> 4 */
  516. output[3] += output[2] >> 56;
  517. output[2] &= 0x00ffffffffffffff;
  518. output[4] = output[3] >> 56;
  519. output[3] &= 0x00ffffffffffffff;
  520. /* Now output[2] < 2^56, output[3] < 2^56, output[4] < 2^72 */
  521. /* Eliminate output[4] */
  522. output[2] += output[4] >> 16;
  523. /* output[2] < 2^56 + 2^56 = 2^57 */
  524. output[1] += (output[4] & 0xffff) << 40;
  525. output[0] -= output[4];
  526. /* Carry 0 -> 1 -> 2 -> 3 */
  527. output[1] += output[0] >> 56;
  528. out[0] = output[0] & 0x00ffffffffffffff;
  529. output[2] += output[1] >> 56;
  530. /* output[2] < 2^57 + 2^72 */
  531. out[1] = output[1] & 0x00ffffffffffffff;
  532. output[3] += output[2] >> 56;
  533. /* output[3] <= 2^56 + 2^16 */
  534. out[2] = output[2] & 0x00ffffffffffffff;
  535. /*-
  536. * out[0] < 2^56, out[1] < 2^56, out[2] < 2^56,
  537. * out[3] <= 2^56 + 2^16 (due to final carry),
  538. * so out < 2*p
  539. */
  540. out[3] = output[3];
  541. }
  542. static void felem_square_reduce(felem out, const felem in)
  543. {
  544. widefelem tmp;
  545. felem_square(tmp, in);
  546. felem_reduce(out, tmp);
  547. }
  548. static void felem_mul_reduce(felem out, const felem in1, const felem in2)
  549. {
  550. widefelem tmp;
  551. felem_mul(tmp, in1, in2);
  552. felem_reduce(out, tmp);
  553. }
  554. /*
  555. * Reduce to unique minimal representation. Requires 0 <= in < 2*p (always
  556. * call felem_reduce first)
  557. */
  558. static void felem_contract(felem out, const felem in)
  559. {
  560. static const int64_t two56 = ((limb) 1) << 56;
  561. /* 0 <= in < 2*p, p = 2^224 - 2^96 + 1 */
  562. /* if in > p , reduce in = in - 2^224 + 2^96 - 1 */
  563. int64_t tmp[4], a;
  564. tmp[0] = in[0];
  565. tmp[1] = in[1];
  566. tmp[2] = in[2];
  567. tmp[3] = in[3];
  568. /* Case 1: a = 1 iff in >= 2^224 */
  569. a = (in[3] >> 56);
  570. tmp[0] -= a;
  571. tmp[1] += a << 40;
  572. tmp[3] &= 0x00ffffffffffffff;
  573. /*
  574. * Case 2: a = 0 iff p <= in < 2^224, i.e., the high 128 bits are all 1
  575. * and the lower part is non-zero
  576. */
  577. a = ((in[3] & in[2] & (in[1] | 0x000000ffffffffff)) + 1) |
  578. (((int64_t) (in[0] + (in[1] & 0x000000ffffffffff)) - 1) >> 63);
  579. a &= 0x00ffffffffffffff;
  580. /* turn a into an all-one mask (if a = 0) or an all-zero mask */
  581. a = (a - 1) >> 63;
  582. /* subtract 2^224 - 2^96 + 1 if a is all-one */
  583. tmp[3] &= a ^ 0xffffffffffffffff;
  584. tmp[2] &= a ^ 0xffffffffffffffff;
  585. tmp[1] &= (a ^ 0xffffffffffffffff) | 0x000000ffffffffff;
  586. tmp[0] -= 1 & a;
  587. /*
  588. * eliminate negative coefficients: if tmp[0] is negative, tmp[1] must be
  589. * non-zero, so we only need one step
  590. */
  591. a = tmp[0] >> 63;
  592. tmp[0] += two56 & a;
  593. tmp[1] -= 1 & a;
  594. /* carry 1 -> 2 -> 3 */
  595. tmp[2] += tmp[1] >> 56;
  596. tmp[1] &= 0x00ffffffffffffff;
  597. tmp[3] += tmp[2] >> 56;
  598. tmp[2] &= 0x00ffffffffffffff;
  599. /* Now 0 <= out < p */
  600. out[0] = tmp[0];
  601. out[1] = tmp[1];
  602. out[2] = tmp[2];
  603. out[3] = tmp[3];
  604. }
  605. /*
  606. * Get negative value: out = -in
  607. * Requires in[i] < 2^63,
  608. * ensures out[0] < 2^56, out[1] < 2^56, out[2] < 2^56, out[3] <= 2^56 + 2^16
  609. */
  610. static void felem_neg(felem out, const felem in)
  611. {
  612. widefelem tmp;
  613. memset(tmp, 0, sizeof(tmp));
  614. felem_diff_128_64(tmp, in);
  615. felem_reduce(out, tmp);
  616. }
  617. /*
  618. * Zero-check: returns 1 if input is 0, and 0 otherwise. We know that field
  619. * elements are reduced to in < 2^225, so we only need to check three cases:
  620. * 0, 2^224 - 2^96 + 1, and 2^225 - 2^97 + 2
  621. */
  622. static limb felem_is_zero(const felem in)
  623. {
  624. limb zero, two224m96p1, two225m97p2;
  625. zero = in[0] | in[1] | in[2] | in[3];
  626. zero = (((int64_t) (zero) - 1) >> 63) & 1;
  627. two224m96p1 = (in[0] ^ 1) | (in[1] ^ 0x00ffff0000000000)
  628. | (in[2] ^ 0x00ffffffffffffff) | (in[3] ^ 0x00ffffffffffffff);
  629. two224m96p1 = (((int64_t) (two224m96p1) - 1) >> 63) & 1;
  630. two225m97p2 = (in[0] ^ 2) | (in[1] ^ 0x00fffe0000000000)
  631. | (in[2] ^ 0x00ffffffffffffff) | (in[3] ^ 0x01ffffffffffffff);
  632. two225m97p2 = (((int64_t) (two225m97p2) - 1) >> 63) & 1;
  633. return (zero | two224m96p1 | two225m97p2);
  634. }
  635. static int felem_is_zero_int(const void *in)
  636. {
  637. return (int)(felem_is_zero(in) & ((limb) 1));
  638. }
  639. /* Invert a field element */
  640. /* Computation chain copied from djb's code */
  641. static void felem_inv(felem out, const felem in)
  642. {
  643. felem ftmp, ftmp2, ftmp3, ftmp4;
  644. widefelem tmp;
  645. unsigned i;
  646. felem_square(tmp, in);
  647. felem_reduce(ftmp, tmp); /* 2 */
  648. felem_mul(tmp, in, ftmp);
  649. felem_reduce(ftmp, tmp); /* 2^2 - 1 */
  650. felem_square(tmp, ftmp);
  651. felem_reduce(ftmp, tmp); /* 2^3 - 2 */
  652. felem_mul(tmp, in, ftmp);
  653. felem_reduce(ftmp, tmp); /* 2^3 - 1 */
  654. felem_square(tmp, ftmp);
  655. felem_reduce(ftmp2, tmp); /* 2^4 - 2 */
  656. felem_square(tmp, ftmp2);
  657. felem_reduce(ftmp2, tmp); /* 2^5 - 4 */
  658. felem_square(tmp, ftmp2);
  659. felem_reduce(ftmp2, tmp); /* 2^6 - 8 */
  660. felem_mul(tmp, ftmp2, ftmp);
  661. felem_reduce(ftmp, tmp); /* 2^6 - 1 */
  662. felem_square(tmp, ftmp);
  663. felem_reduce(ftmp2, tmp); /* 2^7 - 2 */
  664. for (i = 0; i < 5; ++i) { /* 2^12 - 2^6 */
  665. felem_square(tmp, ftmp2);
  666. felem_reduce(ftmp2, tmp);
  667. }
  668. felem_mul(tmp, ftmp2, ftmp);
  669. felem_reduce(ftmp2, tmp); /* 2^12 - 1 */
  670. felem_square(tmp, ftmp2);
  671. felem_reduce(ftmp3, tmp); /* 2^13 - 2 */
  672. for (i = 0; i < 11; ++i) { /* 2^24 - 2^12 */
  673. felem_square(tmp, ftmp3);
  674. felem_reduce(ftmp3, tmp);
  675. }
  676. felem_mul(tmp, ftmp3, ftmp2);
  677. felem_reduce(ftmp2, tmp); /* 2^24 - 1 */
  678. felem_square(tmp, ftmp2);
  679. felem_reduce(ftmp3, tmp); /* 2^25 - 2 */
  680. for (i = 0; i < 23; ++i) { /* 2^48 - 2^24 */
  681. felem_square(tmp, ftmp3);
  682. felem_reduce(ftmp3, tmp);
  683. }
  684. felem_mul(tmp, ftmp3, ftmp2);
  685. felem_reduce(ftmp3, tmp); /* 2^48 - 1 */
  686. felem_square(tmp, ftmp3);
  687. felem_reduce(ftmp4, tmp); /* 2^49 - 2 */
  688. for (i = 0; i < 47; ++i) { /* 2^96 - 2^48 */
  689. felem_square(tmp, ftmp4);
  690. felem_reduce(ftmp4, tmp);
  691. }
  692. felem_mul(tmp, ftmp3, ftmp4);
  693. felem_reduce(ftmp3, tmp); /* 2^96 - 1 */
  694. felem_square(tmp, ftmp3);
  695. felem_reduce(ftmp4, tmp); /* 2^97 - 2 */
  696. for (i = 0; i < 23; ++i) { /* 2^120 - 2^24 */
  697. felem_square(tmp, ftmp4);
  698. felem_reduce(ftmp4, tmp);
  699. }
  700. felem_mul(tmp, ftmp2, ftmp4);
  701. felem_reduce(ftmp2, tmp); /* 2^120 - 1 */
  702. for (i = 0; i < 6; ++i) { /* 2^126 - 2^6 */
  703. felem_square(tmp, ftmp2);
  704. felem_reduce(ftmp2, tmp);
  705. }
  706. felem_mul(tmp, ftmp2, ftmp);
  707. felem_reduce(ftmp, tmp); /* 2^126 - 1 */
  708. felem_square(tmp, ftmp);
  709. felem_reduce(ftmp, tmp); /* 2^127 - 2 */
  710. felem_mul(tmp, ftmp, in);
  711. felem_reduce(ftmp, tmp); /* 2^127 - 1 */
  712. for (i = 0; i < 97; ++i) { /* 2^224 - 2^97 */
  713. felem_square(tmp, ftmp);
  714. felem_reduce(ftmp, tmp);
  715. }
  716. felem_mul(tmp, ftmp, ftmp3);
  717. felem_reduce(out, tmp); /* 2^224 - 2^96 - 1 */
  718. }
  719. /*
  720. * Copy in constant time: if icopy == 1, copy in to out, if icopy == 0, copy
  721. * out to itself.
  722. */
  723. static void copy_conditional(felem out, const felem in, limb icopy)
  724. {
  725. unsigned i;
  726. /*
  727. * icopy is a (64-bit) 0 or 1, so copy is either all-zero or all-one
  728. */
  729. const limb copy = -icopy;
  730. for (i = 0; i < 4; ++i) {
  731. const limb tmp = copy & (in[i] ^ out[i]);
  732. out[i] ^= tmp;
  733. }
  734. }
  735. /******************************************************************************/
  736. /*-
  737. * ELLIPTIC CURVE POINT OPERATIONS
  738. *
  739. * Points are represented in Jacobian projective coordinates:
  740. * (X, Y, Z) corresponds to the affine point (X/Z^2, Y/Z^3),
  741. * or to the point at infinity if Z == 0.
  742. *
  743. */
  744. /*-
  745. * Double an elliptic curve point:
  746. * (X', Y', Z') = 2 * (X, Y, Z), where
  747. * X' = (3 * (X - Z^2) * (X + Z^2))^2 - 8 * X * Y^2
  748. * Y' = 3 * (X - Z^2) * (X + Z^2) * (4 * X * Y^2 - X') - 8 * Y^4
  749. * Z' = (Y + Z)^2 - Y^2 - Z^2 = 2 * Y * Z
  750. * Outputs can equal corresponding inputs, i.e., x_out == x_in is allowed,
  751. * while x_out == y_in is not (maybe this works, but it's not tested).
  752. */
  753. static void
  754. point_double(felem x_out, felem y_out, felem z_out,
  755. const felem x_in, const felem y_in, const felem z_in)
  756. {
  757. widefelem tmp, tmp2;
  758. felem delta, gamma, beta, alpha, ftmp, ftmp2;
  759. felem_assign(ftmp, x_in);
  760. felem_assign(ftmp2, x_in);
  761. /* delta = z^2 */
  762. felem_square(tmp, z_in);
  763. felem_reduce(delta, tmp);
  764. /* gamma = y^2 */
  765. felem_square(tmp, y_in);
  766. felem_reduce(gamma, tmp);
  767. /* beta = x*gamma */
  768. felem_mul(tmp, x_in, gamma);
  769. felem_reduce(beta, tmp);
  770. /* alpha = 3*(x-delta)*(x+delta) */
  771. felem_diff(ftmp, delta);
  772. /* ftmp[i] < 2^57 + 2^58 + 2 < 2^59 */
  773. felem_sum(ftmp2, delta);
  774. /* ftmp2[i] < 2^57 + 2^57 = 2^58 */
  775. felem_scalar(ftmp2, 3);
  776. /* ftmp2[i] < 3 * 2^58 < 2^60 */
  777. felem_mul(tmp, ftmp, ftmp2);
  778. /* tmp[i] < 2^60 * 2^59 * 4 = 2^121 */
  779. felem_reduce(alpha, tmp);
  780. /* x' = alpha^2 - 8*beta */
  781. felem_square(tmp, alpha);
  782. /* tmp[i] < 4 * 2^57 * 2^57 = 2^116 */
  783. felem_assign(ftmp, beta);
  784. felem_scalar(ftmp, 8);
  785. /* ftmp[i] < 8 * 2^57 = 2^60 */
  786. felem_diff_128_64(tmp, ftmp);
  787. /* tmp[i] < 2^116 + 2^64 + 8 < 2^117 */
  788. felem_reduce(x_out, tmp);
  789. /* z' = (y + z)^2 - gamma - delta */
  790. felem_sum(delta, gamma);
  791. /* delta[i] < 2^57 + 2^57 = 2^58 */
  792. felem_assign(ftmp, y_in);
  793. felem_sum(ftmp, z_in);
  794. /* ftmp[i] < 2^57 + 2^57 = 2^58 */
  795. felem_square(tmp, ftmp);
  796. /* tmp[i] < 4 * 2^58 * 2^58 = 2^118 */
  797. felem_diff_128_64(tmp, delta);
  798. /* tmp[i] < 2^118 + 2^64 + 8 < 2^119 */
  799. felem_reduce(z_out, tmp);
  800. /* y' = alpha*(4*beta - x') - 8*gamma^2 */
  801. felem_scalar(beta, 4);
  802. /* beta[i] < 4 * 2^57 = 2^59 */
  803. felem_diff(beta, x_out);
  804. /* beta[i] < 2^59 + 2^58 + 2 < 2^60 */
  805. felem_mul(tmp, alpha, beta);
  806. /* tmp[i] < 4 * 2^57 * 2^60 = 2^119 */
  807. felem_square(tmp2, gamma);
  808. /* tmp2[i] < 4 * 2^57 * 2^57 = 2^116 */
  809. widefelem_scalar(tmp2, 8);
  810. /* tmp2[i] < 8 * 2^116 = 2^119 */
  811. widefelem_diff(tmp, tmp2);
  812. /* tmp[i] < 2^119 + 2^120 < 2^121 */
  813. felem_reduce(y_out, tmp);
  814. }
  815. /*-
  816. * Add two elliptic curve points:
  817. * (X_1, Y_1, Z_1) + (X_2, Y_2, Z_2) = (X_3, Y_3, Z_3), where
  818. * X_3 = (Z_1^3 * Y_2 - Z_2^3 * Y_1)^2 - (Z_1^2 * X_2 - Z_2^2 * X_1)^3 -
  819. * 2 * Z_2^2 * X_1 * (Z_1^2 * X_2 - Z_2^2 * X_1)^2
  820. * Y_3 = (Z_1^3 * Y_2 - Z_2^3 * Y_1) * (Z_2^2 * X_1 * (Z_1^2 * X_2 - Z_2^2 * X_1)^2 - X_3) -
  821. * Z_2^3 * Y_1 * (Z_1^2 * X_2 - Z_2^2 * X_1)^3
  822. * Z_3 = (Z_1^2 * X_2 - Z_2^2 * X_1) * (Z_1 * Z_2)
  823. *
  824. * This runs faster if 'mixed' is set, which requires Z_2 = 1 or Z_2 = 0.
  825. */
  826. /*
  827. * This function is not entirely constant-time: it includes a branch for
  828. * checking whether the two input points are equal, (while not equal to the
  829. * point at infinity). This case never happens during single point
  830. * multiplication, so there is no timing leak for ECDH or ECDSA signing.
  831. */
  832. static void point_add(felem x3, felem y3, felem z3,
  833. const felem x1, const felem y1, const felem z1,
  834. const int mixed, const felem x2, const felem y2,
  835. const felem z2)
  836. {
  837. felem ftmp, ftmp2, ftmp3, ftmp4, ftmp5, x_out, y_out, z_out;
  838. widefelem tmp, tmp2;
  839. limb z1_is_zero, z2_is_zero, x_equal, y_equal;
  840. limb points_equal;
  841. if (!mixed) {
  842. /* ftmp2 = z2^2 */
  843. felem_square(tmp, z2);
  844. felem_reduce(ftmp2, tmp);
  845. /* ftmp4 = z2^3 */
  846. felem_mul(tmp, ftmp2, z2);
  847. felem_reduce(ftmp4, tmp);
  848. /* ftmp4 = z2^3*y1 */
  849. felem_mul(tmp2, ftmp4, y1);
  850. felem_reduce(ftmp4, tmp2);
  851. /* ftmp2 = z2^2*x1 */
  852. felem_mul(tmp2, ftmp2, x1);
  853. felem_reduce(ftmp2, tmp2);
  854. } else {
  855. /*
  856. * We'll assume z2 = 1 (special case z2 = 0 is handled later)
  857. */
  858. /* ftmp4 = z2^3*y1 */
  859. felem_assign(ftmp4, y1);
  860. /* ftmp2 = z2^2*x1 */
  861. felem_assign(ftmp2, x1);
  862. }
  863. /* ftmp = z1^2 */
  864. felem_square(tmp, z1);
  865. felem_reduce(ftmp, tmp);
  866. /* ftmp3 = z1^3 */
  867. felem_mul(tmp, ftmp, z1);
  868. felem_reduce(ftmp3, tmp);
  869. /* tmp = z1^3*y2 */
  870. felem_mul(tmp, ftmp3, y2);
  871. /* tmp[i] < 4 * 2^57 * 2^57 = 2^116 */
  872. /* ftmp3 = z1^3*y2 - z2^3*y1 */
  873. felem_diff_128_64(tmp, ftmp4);
  874. /* tmp[i] < 2^116 + 2^64 + 8 < 2^117 */
  875. felem_reduce(ftmp3, tmp);
  876. /* tmp = z1^2*x2 */
  877. felem_mul(tmp, ftmp, x2);
  878. /* tmp[i] < 4 * 2^57 * 2^57 = 2^116 */
  879. /* ftmp = z1^2*x2 - z2^2*x1 */
  880. felem_diff_128_64(tmp, ftmp2);
  881. /* tmp[i] < 2^116 + 2^64 + 8 < 2^117 */
  882. felem_reduce(ftmp, tmp);
  883. /*
  884. * The formulae are incorrect if the points are equal, in affine coordinates
  885. * (X_1, Y_1) == (X_2, Y_2), so we check for this and do doubling if this
  886. * happens.
  887. *
  888. * We use bitwise operations to avoid potential side-channels introduced by
  889. * the short-circuiting behaviour of boolean operators.
  890. */
  891. x_equal = felem_is_zero(ftmp);
  892. y_equal = felem_is_zero(ftmp3);
  893. /*
  894. * The special case of either point being the point at infinity (z1 and/or
  895. * z2 are zero), is handled separately later on in this function, so we
  896. * avoid jumping to point_double here in those special cases.
  897. */
  898. z1_is_zero = felem_is_zero(z1);
  899. z2_is_zero = felem_is_zero(z2);
  900. /*
  901. * Compared to `ecp_nistp256.c` and `ecp_nistp521.c`, in this
  902. * specific implementation `felem_is_zero()` returns truth as `0x1`
  903. * (rather than `0xff..ff`).
  904. *
  905. * This implies that `~true` in this implementation becomes
  906. * `0xff..fe` (rather than `0x0`): for this reason, to be used in
  907. * the if expression, we mask out only the last bit in the next
  908. * line.
  909. */
  910. points_equal = (x_equal & y_equal & (~z1_is_zero) & (~z2_is_zero)) & 1;
  911. if (points_equal) {
  912. /*
  913. * This is obviously not constant-time but, as mentioned before, this
  914. * case never happens during single point multiplication, so there is no
  915. * timing leak for ECDH or ECDSA signing.
  916. */
  917. point_double(x3, y3, z3, x1, y1, z1);
  918. return;
  919. }
  920. /* ftmp5 = z1*z2 */
  921. if (!mixed) {
  922. felem_mul(tmp, z1, z2);
  923. felem_reduce(ftmp5, tmp);
  924. } else {
  925. /* special case z2 = 0 is handled later */
  926. felem_assign(ftmp5, z1);
  927. }
  928. /* z_out = (z1^2*x2 - z2^2*x1)*(z1*z2) */
  929. felem_mul(tmp, ftmp, ftmp5);
  930. felem_reduce(z_out, tmp);
  931. /* ftmp = (z1^2*x2 - z2^2*x1)^2 */
  932. felem_assign(ftmp5, ftmp);
  933. felem_square(tmp, ftmp);
  934. felem_reduce(ftmp, tmp);
  935. /* ftmp5 = (z1^2*x2 - z2^2*x1)^3 */
  936. felem_mul(tmp, ftmp, ftmp5);
  937. felem_reduce(ftmp5, tmp);
  938. /* ftmp2 = z2^2*x1*(z1^2*x2 - z2^2*x1)^2 */
  939. felem_mul(tmp, ftmp2, ftmp);
  940. felem_reduce(ftmp2, tmp);
  941. /* tmp = z2^3*y1*(z1^2*x2 - z2^2*x1)^3 */
  942. felem_mul(tmp, ftmp4, ftmp5);
  943. /* tmp[i] < 4 * 2^57 * 2^57 = 2^116 */
  944. /* tmp2 = (z1^3*y2 - z2^3*y1)^2 */
  945. felem_square(tmp2, ftmp3);
  946. /* tmp2[i] < 4 * 2^57 * 2^57 < 2^116 */
  947. /* tmp2 = (z1^3*y2 - z2^3*y1)^2 - (z1^2*x2 - z2^2*x1)^3 */
  948. felem_diff_128_64(tmp2, ftmp5);
  949. /* tmp2[i] < 2^116 + 2^64 + 8 < 2^117 */
  950. /* ftmp5 = 2*z2^2*x1*(z1^2*x2 - z2^2*x1)^2 */
  951. felem_assign(ftmp5, ftmp2);
  952. felem_scalar(ftmp5, 2);
  953. /* ftmp5[i] < 2 * 2^57 = 2^58 */
  954. /*-
  955. * x_out = (z1^3*y2 - z2^3*y1)^2 - (z1^2*x2 - z2^2*x1)^3 -
  956. * 2*z2^2*x1*(z1^2*x2 - z2^2*x1)^2
  957. */
  958. felem_diff_128_64(tmp2, ftmp5);
  959. /* tmp2[i] < 2^117 + 2^64 + 8 < 2^118 */
  960. felem_reduce(x_out, tmp2);
  961. /* ftmp2 = z2^2*x1*(z1^2*x2 - z2^2*x1)^2 - x_out */
  962. felem_diff(ftmp2, x_out);
  963. /* ftmp2[i] < 2^57 + 2^58 + 2 < 2^59 */
  964. /*
  965. * tmp2 = (z1^3*y2 - z2^3*y1)*(z2^2*x1*(z1^2*x2 - z2^2*x1)^2 - x_out)
  966. */
  967. felem_mul(tmp2, ftmp3, ftmp2);
  968. /* tmp2[i] < 4 * 2^57 * 2^59 = 2^118 */
  969. /*-
  970. * y_out = (z1^3*y2 - z2^3*y1)*(z2^2*x1*(z1^2*x2 - z2^2*x1)^2 - x_out) -
  971. * z2^3*y1*(z1^2*x2 - z2^2*x1)^3
  972. */
  973. widefelem_diff(tmp2, tmp);
  974. /* tmp2[i] < 2^118 + 2^120 < 2^121 */
  975. felem_reduce(y_out, tmp2);
  976. /*
  977. * the result (x_out, y_out, z_out) is incorrect if one of the inputs is
  978. * the point at infinity, so we need to check for this separately
  979. */
  980. /*
  981. * if point 1 is at infinity, copy point 2 to output, and vice versa
  982. */
  983. copy_conditional(x_out, x2, z1_is_zero);
  984. copy_conditional(x_out, x1, z2_is_zero);
  985. copy_conditional(y_out, y2, z1_is_zero);
  986. copy_conditional(y_out, y1, z2_is_zero);
  987. copy_conditional(z_out, z2, z1_is_zero);
  988. copy_conditional(z_out, z1, z2_is_zero);
  989. felem_assign(x3, x_out);
  990. felem_assign(y3, y_out);
  991. felem_assign(z3, z_out);
  992. }
  993. /*
  994. * select_point selects the |idx|th point from a precomputation table and
  995. * copies it to out.
  996. * The pre_comp array argument should be size of |size| argument
  997. */
  998. static void select_point(const u64 idx, unsigned int size,
  999. const felem pre_comp[][3], felem out[3])
  1000. {
  1001. unsigned i, j;
  1002. limb *outlimbs = &out[0][0];
  1003. memset(out, 0, sizeof(*out) * 3);
  1004. for (i = 0; i < size; i++) {
  1005. const limb *inlimbs = &pre_comp[i][0][0];
  1006. u64 mask = i ^ idx;
  1007. mask |= mask >> 4;
  1008. mask |= mask >> 2;
  1009. mask |= mask >> 1;
  1010. mask &= 1;
  1011. mask--;
  1012. for (j = 0; j < 4 * 3; j++)
  1013. outlimbs[j] |= inlimbs[j] & mask;
  1014. }
  1015. }
  1016. /* get_bit returns the |i|th bit in |in| */
  1017. static char get_bit(const felem_bytearray in, unsigned i)
  1018. {
  1019. if (i >= 224)
  1020. return 0;
  1021. return (in[i >> 3] >> (i & 7)) & 1;
  1022. }
  1023. /*
  1024. * Interleaved point multiplication using precomputed point multiples: The
  1025. * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars
  1026. * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
  1027. * generator, using certain (large) precomputed multiples in g_pre_comp.
  1028. * Output point (X, Y, Z) is stored in x_out, y_out, z_out
  1029. */
  1030. static void batch_mul(felem x_out, felem y_out, felem z_out,
  1031. const felem_bytearray scalars[],
  1032. const unsigned num_points, const u8 *g_scalar,
  1033. const int mixed, const felem pre_comp[][17][3],
  1034. const felem g_pre_comp[2][16][3])
  1035. {
  1036. int i, skip;
  1037. unsigned num;
  1038. unsigned gen_mul = (g_scalar != NULL);
  1039. felem nq[3], tmp[4];
  1040. u64 bits;
  1041. u8 sign, digit;
  1042. /* set nq to the point at infinity */
  1043. memset(nq, 0, sizeof(nq));
  1044. /*
  1045. * Loop over all scalars msb-to-lsb, interleaving additions of multiples
  1046. * of the generator (two in each of the last 28 rounds) and additions of
  1047. * other points multiples (every 5th round).
  1048. */
  1049. skip = 1; /* save two point operations in the first
  1050. * round */
  1051. for (i = (num_points ? 220 : 27); i >= 0; --i) {
  1052. /* double */
  1053. if (!skip)
  1054. point_double(nq[0], nq[1], nq[2], nq[0], nq[1], nq[2]);
  1055. /* add multiples of the generator */
  1056. if (gen_mul && (i <= 27)) {
  1057. /* first, look 28 bits upwards */
  1058. bits = get_bit(g_scalar, i + 196) << 3;
  1059. bits |= get_bit(g_scalar, i + 140) << 2;
  1060. bits |= get_bit(g_scalar, i + 84) << 1;
  1061. bits |= get_bit(g_scalar, i + 28);
  1062. /* select the point to add, in constant time */
  1063. select_point(bits, 16, g_pre_comp[1], tmp);
  1064. if (!skip) {
  1065. /* value 1 below is argument for "mixed" */
  1066. point_add(nq[0], nq[1], nq[2],
  1067. nq[0], nq[1], nq[2], 1, tmp[0], tmp[1], tmp[2]);
  1068. } else {
  1069. memcpy(nq, tmp, 3 * sizeof(felem));
  1070. skip = 0;
  1071. }
  1072. /* second, look at the current position */
  1073. bits = get_bit(g_scalar, i + 168) << 3;
  1074. bits |= get_bit(g_scalar, i + 112) << 2;
  1075. bits |= get_bit(g_scalar, i + 56) << 1;
  1076. bits |= get_bit(g_scalar, i);
  1077. /* select the point to add, in constant time */
  1078. select_point(bits, 16, g_pre_comp[0], tmp);
  1079. point_add(nq[0], nq[1], nq[2],
  1080. nq[0], nq[1], nq[2],
  1081. 1 /* mixed */ , tmp[0], tmp[1], tmp[2]);
  1082. }
  1083. /* do other additions every 5 doublings */
  1084. if (num_points && (i % 5 == 0)) {
  1085. /* loop over all scalars */
  1086. for (num = 0; num < num_points; ++num) {
  1087. bits = get_bit(scalars[num], i + 4) << 5;
  1088. bits |= get_bit(scalars[num], i + 3) << 4;
  1089. bits |= get_bit(scalars[num], i + 2) << 3;
  1090. bits |= get_bit(scalars[num], i + 1) << 2;
  1091. bits |= get_bit(scalars[num], i) << 1;
  1092. bits |= get_bit(scalars[num], i - 1);
  1093. ec_GFp_nistp_recode_scalar_bits(&sign, &digit, bits);
  1094. /* select the point to add or subtract */
  1095. select_point(digit, 17, pre_comp[num], tmp);
  1096. felem_neg(tmp[3], tmp[1]); /* (X, -Y, Z) is the negative
  1097. * point */
  1098. copy_conditional(tmp[1], tmp[3], sign);
  1099. if (!skip) {
  1100. point_add(nq[0], nq[1], nq[2],
  1101. nq[0], nq[1], nq[2],
  1102. mixed, tmp[0], tmp[1], tmp[2]);
  1103. } else {
  1104. memcpy(nq, tmp, 3 * sizeof(felem));
  1105. skip = 0;
  1106. }
  1107. }
  1108. }
  1109. }
  1110. felem_assign(x_out, nq[0]);
  1111. felem_assign(y_out, nq[1]);
  1112. felem_assign(z_out, nq[2]);
  1113. }
  1114. /******************************************************************************/
  1115. /*
  1116. * FUNCTIONS TO MANAGE PRECOMPUTATION
  1117. */
  1118. static NISTP224_PRE_COMP *nistp224_pre_comp_new(void)
  1119. {
  1120. NISTP224_PRE_COMP *ret = OPENSSL_zalloc(sizeof(*ret));
  1121. if (!ret) {
  1122. ECerr(EC_F_NISTP224_PRE_COMP_NEW, ERR_R_MALLOC_FAILURE);
  1123. return ret;
  1124. }
  1125. ret->references = 1;
  1126. ret->lock = CRYPTO_THREAD_lock_new();
  1127. if (ret->lock == NULL) {
  1128. ECerr(EC_F_NISTP224_PRE_COMP_NEW, ERR_R_MALLOC_FAILURE);
  1129. OPENSSL_free(ret);
  1130. return NULL;
  1131. }
  1132. return ret;
  1133. }
  1134. NISTP224_PRE_COMP *EC_nistp224_pre_comp_dup(NISTP224_PRE_COMP *p)
  1135. {
  1136. int i;
  1137. if (p != NULL)
  1138. CRYPTO_UP_REF(&p->references, &i, p->lock);
  1139. return p;
  1140. }
  1141. void EC_nistp224_pre_comp_free(NISTP224_PRE_COMP *p)
  1142. {
  1143. int i;
  1144. if (p == NULL)
  1145. return;
  1146. CRYPTO_DOWN_REF(&p->references, &i, p->lock);
  1147. REF_PRINT_COUNT("EC_nistp224", x);
  1148. if (i > 0)
  1149. return;
  1150. REF_ASSERT_ISNT(i < 0);
  1151. CRYPTO_THREAD_lock_free(p->lock);
  1152. OPENSSL_free(p);
  1153. }
  1154. /******************************************************************************/
  1155. /*
  1156. * OPENSSL EC_METHOD FUNCTIONS
  1157. */
  1158. int ec_GFp_nistp224_group_init(EC_GROUP *group)
  1159. {
  1160. int ret;
  1161. ret = ec_GFp_simple_group_init(group);
  1162. group->a_is_minus3 = 1;
  1163. return ret;
  1164. }
  1165. int ec_GFp_nistp224_group_set_curve(EC_GROUP *group, const BIGNUM *p,
  1166. const BIGNUM *a, const BIGNUM *b,
  1167. BN_CTX *ctx)
  1168. {
  1169. int ret = 0;
  1170. BIGNUM *curve_p, *curve_a, *curve_b;
  1171. #ifndef FIPS_MODULE
  1172. BN_CTX *new_ctx = NULL;
  1173. if (ctx == NULL)
  1174. ctx = new_ctx = BN_CTX_new();
  1175. #endif
  1176. if (ctx == NULL)
  1177. return 0;
  1178. BN_CTX_start(ctx);
  1179. curve_p = BN_CTX_get(ctx);
  1180. curve_a = BN_CTX_get(ctx);
  1181. curve_b = BN_CTX_get(ctx);
  1182. if (curve_b == NULL)
  1183. goto err;
  1184. BN_bin2bn(nistp224_curve_params[0], sizeof(felem_bytearray), curve_p);
  1185. BN_bin2bn(nistp224_curve_params[1], sizeof(felem_bytearray), curve_a);
  1186. BN_bin2bn(nistp224_curve_params[2], sizeof(felem_bytearray), curve_b);
  1187. if ((BN_cmp(curve_p, p)) || (BN_cmp(curve_a, a)) || (BN_cmp(curve_b, b))) {
  1188. ECerr(EC_F_EC_GFP_NISTP224_GROUP_SET_CURVE,
  1189. EC_R_WRONG_CURVE_PARAMETERS);
  1190. goto err;
  1191. }
  1192. group->field_mod_func = BN_nist_mod_224;
  1193. ret = ec_GFp_simple_group_set_curve(group, p, a, b, ctx);
  1194. err:
  1195. BN_CTX_end(ctx);
  1196. #ifndef FIPS_MODULE
  1197. BN_CTX_free(new_ctx);
  1198. #endif
  1199. return ret;
  1200. }
  1201. /*
  1202. * Takes the Jacobian coordinates (X, Y, Z) of a point and returns (X', Y') =
  1203. * (X/Z^2, Y/Z^3)
  1204. */
  1205. int ec_GFp_nistp224_point_get_affine_coordinates(const EC_GROUP *group,
  1206. const EC_POINT *point,
  1207. BIGNUM *x, BIGNUM *y,
  1208. BN_CTX *ctx)
  1209. {
  1210. felem z1, z2, x_in, y_in, x_out, y_out;
  1211. widefelem tmp;
  1212. if (EC_POINT_is_at_infinity(group, point)) {
  1213. ECerr(EC_F_EC_GFP_NISTP224_POINT_GET_AFFINE_COORDINATES,
  1214. EC_R_POINT_AT_INFINITY);
  1215. return 0;
  1216. }
  1217. if ((!BN_to_felem(x_in, point->X)) || (!BN_to_felem(y_in, point->Y)) ||
  1218. (!BN_to_felem(z1, point->Z)))
  1219. return 0;
  1220. felem_inv(z2, z1);
  1221. felem_square(tmp, z2);
  1222. felem_reduce(z1, tmp);
  1223. felem_mul(tmp, x_in, z1);
  1224. felem_reduce(x_in, tmp);
  1225. felem_contract(x_out, x_in);
  1226. if (x != NULL) {
  1227. if (!felem_to_BN(x, x_out)) {
  1228. ECerr(EC_F_EC_GFP_NISTP224_POINT_GET_AFFINE_COORDINATES,
  1229. ERR_R_BN_LIB);
  1230. return 0;
  1231. }
  1232. }
  1233. felem_mul(tmp, z1, z2);
  1234. felem_reduce(z1, tmp);
  1235. felem_mul(tmp, y_in, z1);
  1236. felem_reduce(y_in, tmp);
  1237. felem_contract(y_out, y_in);
  1238. if (y != NULL) {
  1239. if (!felem_to_BN(y, y_out)) {
  1240. ECerr(EC_F_EC_GFP_NISTP224_POINT_GET_AFFINE_COORDINATES,
  1241. ERR_R_BN_LIB);
  1242. return 0;
  1243. }
  1244. }
  1245. return 1;
  1246. }
  1247. static void make_points_affine(size_t num, felem points[ /* num */ ][3],
  1248. felem tmp_felems[ /* num+1 */ ])
  1249. {
  1250. /*
  1251. * Runs in constant time, unless an input is the point at infinity (which
  1252. * normally shouldn't happen).
  1253. */
  1254. ec_GFp_nistp_points_make_affine_internal(num,
  1255. points,
  1256. sizeof(felem),
  1257. tmp_felems,
  1258. (void (*)(void *))felem_one,
  1259. felem_is_zero_int,
  1260. (void (*)(void *, const void *))
  1261. felem_assign,
  1262. (void (*)(void *, const void *))
  1263. felem_square_reduce, (void (*)
  1264. (void *,
  1265. const void
  1266. *,
  1267. const void
  1268. *))
  1269. felem_mul_reduce,
  1270. (void (*)(void *, const void *))
  1271. felem_inv,
  1272. (void (*)(void *, const void *))
  1273. felem_contract);
  1274. }
  1275. /*
  1276. * Computes scalar*generator + \sum scalars[i]*points[i], ignoring NULL
  1277. * values Result is stored in r (r can equal one of the inputs).
  1278. */
  1279. int ec_GFp_nistp224_points_mul(const EC_GROUP *group, EC_POINT *r,
  1280. const BIGNUM *scalar, size_t num,
  1281. const EC_POINT *points[],
  1282. const BIGNUM *scalars[], BN_CTX *ctx)
  1283. {
  1284. int ret = 0;
  1285. int j;
  1286. unsigned i;
  1287. int mixed = 0;
  1288. BIGNUM *x, *y, *z, *tmp_scalar;
  1289. felem_bytearray g_secret;
  1290. felem_bytearray *secrets = NULL;
  1291. felem (*pre_comp)[17][3] = NULL;
  1292. felem *tmp_felems = NULL;
  1293. int num_bytes;
  1294. int have_pre_comp = 0;
  1295. size_t num_points = num;
  1296. felem x_in, y_in, z_in, x_out, y_out, z_out;
  1297. NISTP224_PRE_COMP *pre = NULL;
  1298. const felem(*g_pre_comp)[16][3] = NULL;
  1299. EC_POINT *generator = NULL;
  1300. const EC_POINT *p = NULL;
  1301. const BIGNUM *p_scalar = NULL;
  1302. BN_CTX_start(ctx);
  1303. x = BN_CTX_get(ctx);
  1304. y = BN_CTX_get(ctx);
  1305. z = BN_CTX_get(ctx);
  1306. tmp_scalar = BN_CTX_get(ctx);
  1307. if (tmp_scalar == NULL)
  1308. goto err;
  1309. if (scalar != NULL) {
  1310. pre = group->pre_comp.nistp224;
  1311. if (pre)
  1312. /* we have precomputation, try to use it */
  1313. g_pre_comp = (const felem(*)[16][3])pre->g_pre_comp;
  1314. else
  1315. /* try to use the standard precomputation */
  1316. g_pre_comp = &gmul[0];
  1317. generator = EC_POINT_new(group);
  1318. if (generator == NULL)
  1319. goto err;
  1320. /* get the generator from precomputation */
  1321. if (!felem_to_BN(x, g_pre_comp[0][1][0]) ||
  1322. !felem_to_BN(y, g_pre_comp[0][1][1]) ||
  1323. !felem_to_BN(z, g_pre_comp[0][1][2])) {
  1324. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_BN_LIB);
  1325. goto err;
  1326. }
  1327. if (!ec_GFp_simple_set_Jprojective_coordinates_GFp(group, generator, x,
  1328. y, z, ctx))
  1329. goto err;
  1330. if (0 == EC_POINT_cmp(group, generator, group->generator, ctx))
  1331. /* precomputation matches generator */
  1332. have_pre_comp = 1;
  1333. else
  1334. /*
  1335. * we don't have valid precomputation: treat the generator as a
  1336. * random point
  1337. */
  1338. num_points = num_points + 1;
  1339. }
  1340. if (num_points > 0) {
  1341. if (num_points >= 3) {
  1342. /*
  1343. * unless we precompute multiples for just one or two points,
  1344. * converting those into affine form is time well spent
  1345. */
  1346. mixed = 1;
  1347. }
  1348. secrets = OPENSSL_zalloc(sizeof(*secrets) * num_points);
  1349. pre_comp = OPENSSL_zalloc(sizeof(*pre_comp) * num_points);
  1350. if (mixed)
  1351. tmp_felems =
  1352. OPENSSL_malloc(sizeof(felem) * (num_points * 17 + 1));
  1353. if ((secrets == NULL) || (pre_comp == NULL)
  1354. || (mixed && (tmp_felems == NULL))) {
  1355. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_MALLOC_FAILURE);
  1356. goto err;
  1357. }
  1358. /*
  1359. * we treat NULL scalars as 0, and NULL points as points at infinity,
  1360. * i.e., they contribute nothing to the linear combination
  1361. */
  1362. for (i = 0; i < num_points; ++i) {
  1363. if (i == num) {
  1364. /* the generator */
  1365. p = EC_GROUP_get0_generator(group);
  1366. p_scalar = scalar;
  1367. } else {
  1368. /* the i^th point */
  1369. p = points[i];
  1370. p_scalar = scalars[i];
  1371. }
  1372. if ((p_scalar != NULL) && (p != NULL)) {
  1373. /* reduce scalar to 0 <= scalar < 2^224 */
  1374. if ((BN_num_bits(p_scalar) > 224)
  1375. || (BN_is_negative(p_scalar))) {
  1376. /*
  1377. * this is an unusual input, and we don't guarantee
  1378. * constant-timeness
  1379. */
  1380. if (!BN_nnmod(tmp_scalar, p_scalar, group->order, ctx)) {
  1381. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_BN_LIB);
  1382. goto err;
  1383. }
  1384. num_bytes = BN_bn2lebinpad(tmp_scalar,
  1385. secrets[i], sizeof(secrets[i]));
  1386. } else {
  1387. num_bytes = BN_bn2lebinpad(p_scalar,
  1388. secrets[i], sizeof(secrets[i]));
  1389. }
  1390. if (num_bytes < 0) {
  1391. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_BN_LIB);
  1392. goto err;
  1393. }
  1394. /* precompute multiples */
  1395. if ((!BN_to_felem(x_out, p->X)) ||
  1396. (!BN_to_felem(y_out, p->Y)) ||
  1397. (!BN_to_felem(z_out, p->Z)))
  1398. goto err;
  1399. felem_assign(pre_comp[i][1][0], x_out);
  1400. felem_assign(pre_comp[i][1][1], y_out);
  1401. felem_assign(pre_comp[i][1][2], z_out);
  1402. for (j = 2; j <= 16; ++j) {
  1403. if (j & 1) {
  1404. point_add(pre_comp[i][j][0], pre_comp[i][j][1],
  1405. pre_comp[i][j][2], pre_comp[i][1][0],
  1406. pre_comp[i][1][1], pre_comp[i][1][2], 0,
  1407. pre_comp[i][j - 1][0],
  1408. pre_comp[i][j - 1][1],
  1409. pre_comp[i][j - 1][2]);
  1410. } else {
  1411. point_double(pre_comp[i][j][0], pre_comp[i][j][1],
  1412. pre_comp[i][j][2], pre_comp[i][j / 2][0],
  1413. pre_comp[i][j / 2][1],
  1414. pre_comp[i][j / 2][2]);
  1415. }
  1416. }
  1417. }
  1418. }
  1419. if (mixed)
  1420. make_points_affine(num_points * 17, pre_comp[0], tmp_felems);
  1421. }
  1422. /* the scalar for the generator */
  1423. if ((scalar != NULL) && (have_pre_comp)) {
  1424. memset(g_secret, 0, sizeof(g_secret));
  1425. /* reduce scalar to 0 <= scalar < 2^224 */
  1426. if ((BN_num_bits(scalar) > 224) || (BN_is_negative(scalar))) {
  1427. /*
  1428. * this is an unusual input, and we don't guarantee
  1429. * constant-timeness
  1430. */
  1431. if (!BN_nnmod(tmp_scalar, scalar, group->order, ctx)) {
  1432. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_BN_LIB);
  1433. goto err;
  1434. }
  1435. num_bytes = BN_bn2lebinpad(tmp_scalar, g_secret, sizeof(g_secret));
  1436. } else {
  1437. num_bytes = BN_bn2lebinpad(scalar, g_secret, sizeof(g_secret));
  1438. }
  1439. /* do the multiplication with generator precomputation */
  1440. batch_mul(x_out, y_out, z_out,
  1441. (const felem_bytearray(*))secrets, num_points,
  1442. g_secret,
  1443. mixed, (const felem(*)[17][3])pre_comp, g_pre_comp);
  1444. } else {
  1445. /* do the multiplication without generator precomputation */
  1446. batch_mul(x_out, y_out, z_out,
  1447. (const felem_bytearray(*))secrets, num_points,
  1448. NULL, mixed, (const felem(*)[17][3])pre_comp, NULL);
  1449. }
  1450. /* reduce the output to its unique minimal representation */
  1451. felem_contract(x_in, x_out);
  1452. felem_contract(y_in, y_out);
  1453. felem_contract(z_in, z_out);
  1454. if ((!felem_to_BN(x, x_in)) || (!felem_to_BN(y, y_in)) ||
  1455. (!felem_to_BN(z, z_in))) {
  1456. ECerr(EC_F_EC_GFP_NISTP224_POINTS_MUL, ERR_R_BN_LIB);
  1457. goto err;
  1458. }
  1459. ret = ec_GFp_simple_set_Jprojective_coordinates_GFp(group, r, x, y, z, ctx);
  1460. err:
  1461. BN_CTX_end(ctx);
  1462. EC_POINT_free(generator);
  1463. OPENSSL_free(secrets);
  1464. OPENSSL_free(pre_comp);
  1465. OPENSSL_free(tmp_felems);
  1466. return ret;
  1467. }
  1468. int ec_GFp_nistp224_precompute_mult(EC_GROUP *group, BN_CTX *ctx)
  1469. {
  1470. int ret = 0;
  1471. NISTP224_PRE_COMP *pre = NULL;
  1472. int i, j;
  1473. BIGNUM *x, *y;
  1474. EC_POINT *generator = NULL;
  1475. felem tmp_felems[32];
  1476. #ifndef FIPS_MODULE
  1477. BN_CTX *new_ctx = NULL;
  1478. #endif
  1479. /* throw away old precomputation */
  1480. EC_pre_comp_free(group);
  1481. #ifndef FIPS_MODULE
  1482. if (ctx == NULL)
  1483. ctx = new_ctx = BN_CTX_new();
  1484. #endif
  1485. if (ctx == NULL)
  1486. return 0;
  1487. BN_CTX_start(ctx);
  1488. x = BN_CTX_get(ctx);
  1489. y = BN_CTX_get(ctx);
  1490. if (y == NULL)
  1491. goto err;
  1492. /* get the generator */
  1493. if (group->generator == NULL)
  1494. goto err;
  1495. generator = EC_POINT_new(group);
  1496. if (generator == NULL)
  1497. goto err;
  1498. BN_bin2bn(nistp224_curve_params[3], sizeof(felem_bytearray), x);
  1499. BN_bin2bn(nistp224_curve_params[4], sizeof(felem_bytearray), y);
  1500. if (!EC_POINT_set_affine_coordinates(group, generator, x, y, ctx))
  1501. goto err;
  1502. if ((pre = nistp224_pre_comp_new()) == NULL)
  1503. goto err;
  1504. /*
  1505. * if the generator is the standard one, use built-in precomputation
  1506. */
  1507. if (0 == EC_POINT_cmp(group, generator, group->generator, ctx)) {
  1508. memcpy(pre->g_pre_comp, gmul, sizeof(pre->g_pre_comp));
  1509. goto done;
  1510. }
  1511. if ((!BN_to_felem(pre->g_pre_comp[0][1][0], group->generator->X)) ||
  1512. (!BN_to_felem(pre->g_pre_comp[0][1][1], group->generator->Y)) ||
  1513. (!BN_to_felem(pre->g_pre_comp[0][1][2], group->generator->Z)))
  1514. goto err;
  1515. /*
  1516. * compute 2^56*G, 2^112*G, 2^168*G for the first table, 2^28*G, 2^84*G,
  1517. * 2^140*G, 2^196*G for the second one
  1518. */
  1519. for (i = 1; i <= 8; i <<= 1) {
  1520. point_double(pre->g_pre_comp[1][i][0], pre->g_pre_comp[1][i][1],
  1521. pre->g_pre_comp[1][i][2], pre->g_pre_comp[0][i][0],
  1522. pre->g_pre_comp[0][i][1], pre->g_pre_comp[0][i][2]);
  1523. for (j = 0; j < 27; ++j) {
  1524. point_double(pre->g_pre_comp[1][i][0], pre->g_pre_comp[1][i][1],
  1525. pre->g_pre_comp[1][i][2], pre->g_pre_comp[1][i][0],
  1526. pre->g_pre_comp[1][i][1], pre->g_pre_comp[1][i][2]);
  1527. }
  1528. if (i == 8)
  1529. break;
  1530. point_double(pre->g_pre_comp[0][2 * i][0],
  1531. pre->g_pre_comp[0][2 * i][1],
  1532. pre->g_pre_comp[0][2 * i][2], pre->g_pre_comp[1][i][0],
  1533. pre->g_pre_comp[1][i][1], pre->g_pre_comp[1][i][2]);
  1534. for (j = 0; j < 27; ++j) {
  1535. point_double(pre->g_pre_comp[0][2 * i][0],
  1536. pre->g_pre_comp[0][2 * i][1],
  1537. pre->g_pre_comp[0][2 * i][2],
  1538. pre->g_pre_comp[0][2 * i][0],
  1539. pre->g_pre_comp[0][2 * i][1],
  1540. pre->g_pre_comp[0][2 * i][2]);
  1541. }
  1542. }
  1543. for (i = 0; i < 2; i++) {
  1544. /* g_pre_comp[i][0] is the point at infinity */
  1545. memset(pre->g_pre_comp[i][0], 0, sizeof(pre->g_pre_comp[i][0]));
  1546. /* the remaining multiples */
  1547. /* 2^56*G + 2^112*G resp. 2^84*G + 2^140*G */
  1548. point_add(pre->g_pre_comp[i][6][0], pre->g_pre_comp[i][6][1],
  1549. pre->g_pre_comp[i][6][2], pre->g_pre_comp[i][4][0],
  1550. pre->g_pre_comp[i][4][1], pre->g_pre_comp[i][4][2],
  1551. 0, pre->g_pre_comp[i][2][0], pre->g_pre_comp[i][2][1],
  1552. pre->g_pre_comp[i][2][2]);
  1553. /* 2^56*G + 2^168*G resp. 2^84*G + 2^196*G */
  1554. point_add(pre->g_pre_comp[i][10][0], pre->g_pre_comp[i][10][1],
  1555. pre->g_pre_comp[i][10][2], pre->g_pre_comp[i][8][0],
  1556. pre->g_pre_comp[i][8][1], pre->g_pre_comp[i][8][2],
  1557. 0, pre->g_pre_comp[i][2][0], pre->g_pre_comp[i][2][1],
  1558. pre->g_pre_comp[i][2][2]);
  1559. /* 2^112*G + 2^168*G resp. 2^140*G + 2^196*G */
  1560. point_add(pre->g_pre_comp[i][12][0], pre->g_pre_comp[i][12][1],
  1561. pre->g_pre_comp[i][12][2], pre->g_pre_comp[i][8][0],
  1562. pre->g_pre_comp[i][8][1], pre->g_pre_comp[i][8][2],
  1563. 0, pre->g_pre_comp[i][4][0], pre->g_pre_comp[i][4][1],
  1564. pre->g_pre_comp[i][4][2]);
  1565. /*
  1566. * 2^56*G + 2^112*G + 2^168*G resp. 2^84*G + 2^140*G + 2^196*G
  1567. */
  1568. point_add(pre->g_pre_comp[i][14][0], pre->g_pre_comp[i][14][1],
  1569. pre->g_pre_comp[i][14][2], pre->g_pre_comp[i][12][0],
  1570. pre->g_pre_comp[i][12][1], pre->g_pre_comp[i][12][2],
  1571. 0, pre->g_pre_comp[i][2][0], pre->g_pre_comp[i][2][1],
  1572. pre->g_pre_comp[i][2][2]);
  1573. for (j = 1; j < 8; ++j) {
  1574. /* odd multiples: add G resp. 2^28*G */
  1575. point_add(pre->g_pre_comp[i][2 * j + 1][0],
  1576. pre->g_pre_comp[i][2 * j + 1][1],
  1577. pre->g_pre_comp[i][2 * j + 1][2],
  1578. pre->g_pre_comp[i][2 * j][0],
  1579. pre->g_pre_comp[i][2 * j][1],
  1580. pre->g_pre_comp[i][2 * j][2], 0,
  1581. pre->g_pre_comp[i][1][0], pre->g_pre_comp[i][1][1],
  1582. pre->g_pre_comp[i][1][2]);
  1583. }
  1584. }
  1585. make_points_affine(31, &(pre->g_pre_comp[0][1]), tmp_felems);
  1586. done:
  1587. SETPRECOMP(group, nistp224, pre);
  1588. pre = NULL;
  1589. ret = 1;
  1590. err:
  1591. BN_CTX_end(ctx);
  1592. EC_POINT_free(generator);
  1593. #ifndef FIPS_MODULE
  1594. BN_CTX_free(new_ctx);
  1595. #endif
  1596. EC_nistp224_pre_comp_free(pre);
  1597. return ret;
  1598. }
  1599. int ec_GFp_nistp224_have_precompute_mult(const EC_GROUP *group)
  1600. {
  1601. return HAVEPRECOMP(group, nistp224);
  1602. }