sp_armthumb.c 629 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900490149024903490449054906490749084909491049114912491349144915491649174918491949204921492249234924492549264927492849294930493149324933493449354936493749384939494049414942494349444945494649474948494949504951495249534954495549564957495849594960496149624963496449654966496749684969497049714972497349744975497649774978497949804981498249834984498549864987498849894990499149924993499449954996499749984999500050015002500350045005500650075008500950105011501250135014501550165017501850195020502150225023502450255026502750285029503050315032503350345035503650375038503950405041504250435044504550465047504850495050505150525053505450555056505750585059506050615062506350645065506650675068506950705071507250735074507550765077507850795080508150825083508450855086508750885089509050915092509350945095509650975098509951005101510251035104510551065107510851095110511151125113511451155116511751185119512051215122512351245125512651275128512951305131513251335134513551365137513851395140514151425143514451455146514751485149515051515152515351545155515651575158515951605161516251635164516551665167516851695170517151725173517451755176517751785179518051815182518351845185518651875188518951905191519251935194519551965197519851995200520152025203520452055206520752085209521052115212521352145215521652175218521952205221522252235224522552265227522852295230523152325233523452355236523752385239524052415242524352445245524652475248524952505251525252535254525552565257525852595260526152625263526452655266526752685269527052715272527352745275527652775278527952805281528252835284528552865287528852895290529152925293529452955296529752985299530053015302530353045305530653075308530953105311531253135314531553165317531853195320532153225323532453255326532753285329533053315332533353345335533653375338533953405341534253435344534553465347534853495350535153525353535453555356535753585359536053615362536353645365536653675368536953705371537253735374537553765377537853795380538153825383538453855386538753885389539053915392539353945395539653975398539954005401540254035404540554065407540854095410541154125413541454155416541754185419542054215422542354245425542654275428542954305431543254335434543554365437543854395440544154425443544454455446544754485449545054515452545354545455545654575458545954605461546254635464546554665467546854695470547154725473547454755476547754785479548054815482548354845485548654875488548954905491549254935494549554965497549854995500550155025503550455055506550755085509551055115512551355145515551655175518551955205521552255235524552555265527552855295530553155325533553455355536553755385539554055415542554355445545554655475548554955505551555255535554555555565557555855595560556155625563556455655566556755685569557055715572557355745575557655775578557955805581558255835584558555865587558855895590559155925593559455955596559755985599560056015602560356045605560656075608560956105611561256135614561556165617561856195620562156225623562456255626562756285629563056315632563356345635563656375638563956405641564256435644564556465647564856495650565156525653565456555656565756585659566056615662566356645665566656675668566956705671567256735674567556765677567856795680568156825683568456855686568756885689569056915692569356945695569656975698569957005701570257035704570557065707570857095710571157125713571457155716571757185719572057215722572357245725572657275728572957305731573257335734573557365737573857395740574157425743574457455746574757485749575057515752575357545755575657575758575957605761576257635764576557665767576857695770577157725773577457755776577757785779578057815782578357845785578657875788578957905791579257935794579557965797579857995800580158025803580458055806580758085809581058115812581358145815581658175818581958205821582258235824582558265827582858295830583158325833583458355836583758385839584058415842584358445845584658475848584958505851585258535854585558565857585858595860586158625863586458655866586758685869587058715872587358745875587658775878587958805881588258835884588558865887588858895890589158925893589458955896589758985899590059015902590359045905590659075908590959105911591259135914591559165917591859195920592159225923592459255926592759285929593059315932593359345935593659375938593959405941594259435944594559465947594859495950595159525953595459555956595759585959596059615962596359645965596659675968596959705971597259735974597559765977597859795980598159825983598459855986598759885989599059915992599359945995599659975998599960006001600260036004600560066007600860096010601160126013601460156016601760186019602060216022602360246025602660276028602960306031603260336034603560366037603860396040604160426043604460456046604760486049605060516052605360546055605660576058605960606061606260636064606560666067606860696070607160726073607460756076607760786079608060816082608360846085608660876088608960906091609260936094609560966097609860996100610161026103610461056106610761086109611061116112611361146115611661176118611961206121612261236124612561266127612861296130613161326133613461356136613761386139614061416142614361446145614661476148614961506151615261536154615561566157615861596160616161626163616461656166616761686169617061716172617361746175617661776178617961806181618261836184618561866187618861896190619161926193619461956196619761986199620062016202620362046205620662076208620962106211621262136214621562166217621862196220622162226223622462256226622762286229623062316232623362346235623662376238623962406241624262436244624562466247624862496250625162526253625462556256625762586259626062616262626362646265626662676268626962706271627262736274627562766277627862796280628162826283628462856286628762886289629062916292629362946295629662976298629963006301630263036304630563066307630863096310631163126313631463156316631763186319632063216322632363246325632663276328632963306331633263336334633563366337633863396340634163426343634463456346634763486349635063516352635363546355635663576358635963606361636263636364636563666367636863696370637163726373637463756376637763786379638063816382638363846385638663876388638963906391639263936394639563966397639863996400640164026403640464056406640764086409641064116412641364146415641664176418641964206421642264236424642564266427642864296430643164326433643464356436643764386439644064416442644364446445644664476448644964506451645264536454645564566457645864596460646164626463646464656466646764686469647064716472647364746475647664776478647964806481648264836484648564866487648864896490649164926493649464956496649764986499650065016502650365046505650665076508650965106511651265136514651565166517651865196520652165226523652465256526652765286529653065316532653365346535653665376538653965406541654265436544654565466547654865496550655165526553655465556556655765586559656065616562656365646565656665676568656965706571657265736574657565766577657865796580658165826583658465856586658765886589659065916592659365946595659665976598659966006601660266036604660566066607660866096610661166126613661466156616661766186619662066216622662366246625662666276628662966306631663266336634663566366637663866396640664166426643664466456646664766486649665066516652665366546655665666576658665966606661666266636664666566666667666866696670667166726673667466756676667766786679668066816682668366846685668666876688668966906691669266936694669566966697669866996700670167026703670467056706670767086709671067116712671367146715671667176718671967206721672267236724672567266727672867296730673167326733673467356736673767386739674067416742674367446745674667476748674967506751675267536754675567566757675867596760676167626763676467656766676767686769677067716772677367746775677667776778677967806781678267836784678567866787678867896790679167926793679467956796679767986799680068016802680368046805680668076808680968106811681268136814681568166817681868196820682168226823682468256826682768286829683068316832683368346835683668376838683968406841684268436844684568466847684868496850685168526853685468556856685768586859686068616862686368646865686668676868686968706871687268736874687568766877687868796880688168826883688468856886688768886889689068916892689368946895689668976898689969006901690269036904690569066907690869096910691169126913691469156916691769186919692069216922692369246925692669276928692969306931693269336934693569366937693869396940694169426943694469456946694769486949695069516952695369546955695669576958695969606961696269636964696569666967696869696970697169726973697469756976697769786979698069816982698369846985698669876988698969906991699269936994699569966997699869997000700170027003700470057006700770087009701070117012701370147015701670177018701970207021702270237024702570267027702870297030703170327033703470357036703770387039704070417042704370447045704670477048704970507051705270537054705570567057705870597060706170627063706470657066706770687069707070717072707370747075707670777078707970807081708270837084708570867087708870897090709170927093709470957096709770987099710071017102710371047105710671077108710971107111711271137114711571167117711871197120712171227123712471257126712771287129713071317132713371347135713671377138713971407141714271437144714571467147714871497150715171527153715471557156715771587159716071617162716371647165716671677168716971707171717271737174717571767177717871797180718171827183718471857186718771887189719071917192719371947195719671977198719972007201720272037204720572067207720872097210721172127213721472157216721772187219722072217222722372247225722672277228722972307231723272337234723572367237723872397240724172427243724472457246724772487249725072517252725372547255725672577258725972607261726272637264726572667267726872697270727172727273727472757276727772787279728072817282728372847285728672877288728972907291729272937294729572967297729872997300730173027303730473057306730773087309731073117312731373147315731673177318731973207321732273237324732573267327732873297330733173327333733473357336733773387339734073417342734373447345734673477348734973507351735273537354735573567357735873597360736173627363736473657366736773687369737073717372737373747375737673777378737973807381738273837384738573867387738873897390739173927393739473957396739773987399740074017402740374047405740674077408740974107411741274137414741574167417741874197420742174227423742474257426742774287429743074317432743374347435743674377438743974407441744274437444744574467447744874497450745174527453745474557456745774587459746074617462746374647465746674677468746974707471747274737474747574767477747874797480748174827483748474857486748774887489749074917492749374947495749674977498749975007501750275037504750575067507750875097510751175127513751475157516751775187519752075217522752375247525752675277528752975307531753275337534753575367537753875397540754175427543754475457546754775487549755075517552755375547555755675577558755975607561756275637564756575667567756875697570757175727573757475757576757775787579758075817582758375847585758675877588758975907591759275937594759575967597759875997600760176027603760476057606760776087609761076117612761376147615761676177618761976207621762276237624762576267627762876297630763176327633763476357636763776387639764076417642764376447645764676477648764976507651765276537654765576567657765876597660766176627663766476657666766776687669767076717672767376747675767676777678767976807681768276837684768576867687768876897690769176927693769476957696769776987699770077017702770377047705770677077708770977107711771277137714771577167717771877197720772177227723772477257726772777287729773077317732773377347735773677377738773977407741774277437744774577467747774877497750775177527753775477557756775777587759776077617762776377647765776677677768776977707771777277737774777577767777777877797780778177827783778477857786778777887789779077917792779377947795779677977798779978007801780278037804780578067807780878097810781178127813781478157816781778187819782078217822782378247825782678277828782978307831783278337834783578367837783878397840784178427843784478457846784778487849785078517852785378547855785678577858785978607861786278637864786578667867786878697870787178727873787478757876787778787879788078817882788378847885788678877888788978907891789278937894789578967897789878997900790179027903790479057906790779087909791079117912791379147915791679177918791979207921792279237924792579267927792879297930793179327933793479357936793779387939794079417942794379447945794679477948794979507951795279537954795579567957795879597960796179627963796479657966796779687969797079717972797379747975797679777978797979807981798279837984798579867987798879897990799179927993799479957996799779987999800080018002800380048005800680078008800980108011801280138014801580168017801880198020802180228023802480258026802780288029803080318032803380348035803680378038803980408041804280438044804580468047804880498050805180528053805480558056805780588059806080618062806380648065806680678068806980708071807280738074807580768077807880798080808180828083808480858086808780888089809080918092809380948095809680978098809981008101810281038104810581068107810881098110811181128113811481158116811781188119812081218122812381248125812681278128812981308131813281338134813581368137813881398140814181428143814481458146814781488149815081518152815381548155815681578158815981608161816281638164816581668167816881698170817181728173817481758176817781788179818081818182818381848185818681878188818981908191819281938194819581968197819881998200820182028203820482058206820782088209821082118212821382148215821682178218821982208221822282238224822582268227822882298230823182328233823482358236823782388239824082418242824382448245824682478248824982508251825282538254825582568257825882598260826182628263826482658266826782688269827082718272827382748275827682778278827982808281828282838284828582868287828882898290829182928293829482958296829782988299830083018302830383048305830683078308830983108311831283138314831583168317831883198320832183228323832483258326832783288329833083318332833383348335833683378338833983408341834283438344834583468347834883498350835183528353835483558356835783588359836083618362836383648365836683678368836983708371837283738374837583768377837883798380838183828383838483858386838783888389839083918392839383948395839683978398839984008401840284038404840584068407840884098410841184128413841484158416841784188419842084218422842384248425842684278428842984308431843284338434843584368437843884398440844184428443844484458446844784488449845084518452845384548455845684578458845984608461846284638464846584668467846884698470847184728473847484758476847784788479848084818482848384848485848684878488848984908491849284938494849584968497849884998500850185028503850485058506850785088509851085118512851385148515851685178518851985208521852285238524852585268527852885298530853185328533853485358536853785388539854085418542854385448545854685478548854985508551855285538554855585568557855885598560856185628563856485658566856785688569857085718572857385748575857685778578857985808581858285838584858585868587858885898590859185928593859485958596859785988599860086018602860386048605860686078608860986108611861286138614861586168617861886198620862186228623862486258626862786288629863086318632863386348635863686378638863986408641864286438644864586468647864886498650865186528653865486558656865786588659866086618662866386648665866686678668866986708671867286738674867586768677867886798680868186828683868486858686868786888689869086918692869386948695869686978698869987008701870287038704870587068707870887098710871187128713871487158716871787188719872087218722872387248725872687278728872987308731873287338734873587368737873887398740874187428743874487458746874787488749875087518752875387548755875687578758875987608761876287638764876587668767876887698770877187728773877487758776877787788779878087818782878387848785878687878788878987908791879287938794879587968797879887998800880188028803880488058806880788088809881088118812881388148815881688178818881988208821882288238824882588268827882888298830883188328833883488358836883788388839884088418842884388448845884688478848884988508851885288538854885588568857885888598860886188628863886488658866886788688869887088718872887388748875887688778878887988808881888288838884888588868887888888898890889188928893889488958896889788988899890089018902890389048905890689078908890989108911891289138914891589168917891889198920892189228923892489258926892789288929893089318932893389348935893689378938893989408941894289438944894589468947894889498950895189528953895489558956895789588959896089618962896389648965896689678968896989708971897289738974897589768977897889798980898189828983898489858986898789888989899089918992899389948995899689978998899990009001900290039004900590069007900890099010901190129013901490159016901790189019902090219022902390249025902690279028902990309031903290339034903590369037903890399040904190429043904490459046904790489049905090519052905390549055905690579058905990609061906290639064906590669067906890699070907190729073907490759076907790789079908090819082908390849085908690879088908990909091909290939094909590969097909890999100910191029103910491059106910791089109911091119112911391149115911691179118911991209121912291239124912591269127912891299130913191329133913491359136913791389139914091419142914391449145914691479148914991509151915291539154915591569157915891599160916191629163916491659166916791689169917091719172917391749175917691779178917991809181918291839184918591869187918891899190919191929193919491959196919791989199920092019202920392049205920692079208920992109211921292139214921592169217921892199220922192229223922492259226922792289229923092319232923392349235923692379238923992409241924292439244924592469247924892499250925192529253925492559256925792589259926092619262926392649265926692679268926992709271927292739274927592769277927892799280928192829283928492859286928792889289929092919292929392949295929692979298929993009301930293039304930593069307930893099310931193129313931493159316931793189319932093219322932393249325932693279328932993309331933293339334933593369337933893399340934193429343934493459346934793489349935093519352935393549355935693579358935993609361936293639364936593669367936893699370937193729373937493759376937793789379938093819382938393849385938693879388938993909391939293939394939593969397939893999400940194029403940494059406940794089409941094119412941394149415941694179418941994209421942294239424942594269427942894299430943194329433943494359436943794389439944094419442944394449445944694479448944994509451945294539454945594569457945894599460946194629463946494659466946794689469947094719472947394749475947694779478947994809481948294839484948594869487948894899490949194929493949494959496949794989499950095019502950395049505950695079508950995109511951295139514951595169517951895199520952195229523952495259526952795289529953095319532953395349535953695379538953995409541954295439544954595469547954895499550955195529553955495559556955795589559956095619562956395649565956695679568956995709571957295739574957595769577957895799580958195829583958495859586958795889589959095919592959395949595959695979598959996009601960296039604960596069607960896099610961196129613961496159616961796189619962096219622962396249625962696279628962996309631963296339634963596369637963896399640964196429643964496459646964796489649965096519652965396549655965696579658965996609661966296639664966596669667966896699670967196729673967496759676967796789679968096819682968396849685968696879688968996909691969296939694969596969697969896999700970197029703970497059706970797089709971097119712971397149715971697179718971997209721972297239724972597269727972897299730973197329733973497359736973797389739974097419742974397449745974697479748974997509751975297539754975597569757975897599760976197629763976497659766976797689769977097719772977397749775977697779778977997809781978297839784978597869787978897899790979197929793979497959796979797989799980098019802980398049805980698079808980998109811981298139814981598169817981898199820982198229823982498259826982798289829983098319832983398349835983698379838983998409841984298439844984598469847984898499850985198529853985498559856985798589859986098619862986398649865986698679868986998709871987298739874987598769877987898799880988198829883988498859886988798889889989098919892989398949895989698979898989999009901990299039904990599069907990899099910991199129913991499159916991799189919992099219922992399249925992699279928992999309931993299339934993599369937993899399940994199429943994499459946994799489949995099519952995399549955995699579958995999609961996299639964996599669967996899699970997199729973997499759976997799789979998099819982998399849985998699879988998999909991999299939994999599969997999899991000010001100021000310004100051000610007100081000910010100111001210013100141001510016100171001810019100201002110022100231002410025100261002710028100291003010031100321003310034100351003610037100381003910040100411004210043100441004510046100471004810049100501005110052100531005410055100561005710058100591006010061100621006310064100651006610067100681006910070100711007210073100741007510076100771007810079100801008110082100831008410085100861008710088100891009010091100921009310094100951009610097100981009910100101011010210103101041010510106101071010810109101101011110112101131011410115101161011710118101191012010121101221012310124101251012610127101281012910130101311013210133101341013510136101371013810139101401014110142101431014410145101461014710148101491015010151101521015310154101551015610157101581015910160101611016210163101641016510166101671016810169101701017110172101731017410175101761017710178101791018010181101821018310184101851018610187101881018910190101911019210193101941019510196101971019810199102001020110202102031020410205102061020710208102091021010211102121021310214102151021610217102181021910220102211022210223102241022510226102271022810229102301023110232102331023410235102361023710238102391024010241102421024310244102451024610247102481024910250102511025210253102541025510256102571025810259102601026110262102631026410265102661026710268102691027010271102721027310274102751027610277102781027910280102811028210283102841028510286102871028810289102901029110292102931029410295102961029710298102991030010301103021030310304103051030610307103081030910310103111031210313103141031510316103171031810319103201032110322103231032410325103261032710328103291033010331103321033310334103351033610337103381033910340103411034210343103441034510346103471034810349103501035110352103531035410355103561035710358103591036010361103621036310364103651036610367103681036910370103711037210373103741037510376103771037810379103801038110382103831038410385103861038710388103891039010391103921039310394103951039610397103981039910400104011040210403104041040510406104071040810409104101041110412104131041410415104161041710418104191042010421104221042310424104251042610427104281042910430104311043210433104341043510436104371043810439104401044110442104431044410445104461044710448104491045010451104521045310454104551045610457104581045910460104611046210463104641046510466104671046810469104701047110472104731047410475104761047710478104791048010481104821048310484104851048610487104881048910490104911049210493104941049510496104971049810499105001050110502105031050410505105061050710508105091051010511105121051310514105151051610517105181051910520105211052210523105241052510526105271052810529105301053110532105331053410535105361053710538105391054010541105421054310544105451054610547105481054910550105511055210553105541055510556105571055810559105601056110562105631056410565105661056710568105691057010571105721057310574105751057610577105781057910580105811058210583105841058510586105871058810589105901059110592105931059410595105961059710598105991060010601106021060310604106051060610607106081060910610106111061210613106141061510616106171061810619106201062110622106231062410625106261062710628106291063010631106321063310634106351063610637106381063910640106411064210643106441064510646106471064810649106501065110652106531065410655106561065710658106591066010661106621066310664106651066610667106681066910670106711067210673106741067510676106771067810679106801068110682106831068410685106861068710688106891069010691106921069310694106951069610697106981069910700107011070210703107041070510706107071070810709107101071110712107131071410715107161071710718107191072010721107221072310724107251072610727107281072910730107311073210733107341073510736107371073810739107401074110742107431074410745107461074710748107491075010751107521075310754107551075610757107581075910760107611076210763107641076510766107671076810769107701077110772107731077410775107761077710778107791078010781107821078310784107851078610787107881078910790107911079210793107941079510796107971079810799108001080110802108031080410805108061080710808108091081010811108121081310814108151081610817108181081910820108211082210823108241082510826108271082810829108301083110832108331083410835108361083710838108391084010841108421084310844108451084610847108481084910850108511085210853108541085510856108571085810859108601086110862108631086410865108661086710868108691087010871108721087310874108751087610877108781087910880108811088210883108841088510886108871088810889108901089110892108931089410895108961089710898108991090010901109021090310904109051090610907109081090910910109111091210913109141091510916109171091810919109201092110922109231092410925109261092710928109291093010931109321093310934109351093610937109381093910940109411094210943109441094510946109471094810949109501095110952109531095410955109561095710958109591096010961109621096310964109651096610967109681096910970109711097210973109741097510976109771097810979109801098110982109831098410985109861098710988109891099010991109921099310994109951099610997109981099911000110011100211003110041100511006110071100811009110101101111012110131101411015110161101711018110191102011021110221102311024110251102611027110281102911030110311103211033110341103511036110371103811039110401104111042110431104411045110461104711048110491105011051110521105311054110551105611057110581105911060110611106211063110641106511066110671106811069110701107111072110731107411075110761107711078110791108011081110821108311084110851108611087110881108911090110911109211093110941109511096110971109811099111001110111102111031110411105111061110711108111091111011111111121111311114111151111611117111181111911120111211112211123111241112511126111271112811129111301113111132111331113411135111361113711138111391114011141111421114311144111451114611147111481114911150111511115211153111541115511156111571115811159111601116111162111631116411165111661116711168111691117011171111721117311174111751117611177111781117911180111811118211183111841118511186111871118811189111901119111192111931119411195111961119711198111991120011201112021120311204112051120611207112081120911210112111121211213112141121511216112171121811219112201122111222112231122411225112261122711228112291123011231112321123311234112351123611237112381123911240112411124211243112441124511246112471124811249112501125111252112531125411255112561125711258112591126011261112621126311264112651126611267112681126911270112711127211273112741127511276112771127811279112801128111282112831128411285112861128711288112891129011291112921129311294112951129611297112981129911300113011130211303113041130511306113071130811309113101131111312113131131411315113161131711318113191132011321113221132311324113251132611327113281132911330113311133211333113341133511336113371133811339113401134111342113431134411345113461134711348113491135011351113521135311354113551135611357113581135911360113611136211363113641136511366113671136811369113701137111372113731137411375113761137711378113791138011381113821138311384113851138611387113881138911390113911139211393113941139511396113971139811399114001140111402114031140411405114061140711408114091141011411114121141311414114151141611417114181141911420114211142211423114241142511426114271142811429114301143111432114331143411435114361143711438114391144011441114421144311444114451144611447114481144911450114511145211453114541145511456114571145811459114601146111462114631146411465114661146711468114691147011471114721147311474114751147611477114781147911480114811148211483114841148511486114871148811489114901149111492114931149411495114961149711498114991150011501115021150311504115051150611507115081150911510115111151211513115141151511516115171151811519115201152111522115231152411525115261152711528115291153011531115321153311534115351153611537115381153911540115411154211543115441154511546115471154811549115501155111552115531155411555115561155711558115591156011561115621156311564115651156611567115681156911570115711157211573115741157511576115771157811579115801158111582115831158411585115861158711588115891159011591115921159311594115951159611597115981159911600116011160211603116041160511606116071160811609116101161111612116131161411615116161161711618116191162011621116221162311624116251162611627116281162911630116311163211633116341163511636116371163811639116401164111642116431164411645116461164711648116491165011651116521165311654116551165611657116581165911660116611166211663116641166511666116671166811669116701167111672116731167411675116761167711678116791168011681116821168311684116851168611687116881168911690116911169211693116941169511696116971169811699117001170111702117031170411705117061170711708117091171011711117121171311714117151171611717117181171911720117211172211723117241172511726117271172811729117301173111732117331173411735117361173711738117391174011741117421174311744117451174611747117481174911750117511175211753117541175511756117571175811759117601176111762117631176411765117661176711768117691177011771117721177311774117751177611777117781177911780117811178211783117841178511786117871178811789117901179111792117931179411795117961179711798117991180011801118021180311804118051180611807118081180911810118111181211813118141181511816118171181811819118201182111822118231182411825118261182711828118291183011831118321183311834118351183611837118381183911840118411184211843118441184511846118471184811849118501185111852118531185411855118561185711858118591186011861118621186311864118651186611867118681186911870118711187211873118741187511876118771187811879118801188111882118831188411885118861188711888118891189011891118921189311894118951189611897118981189911900119011190211903119041190511906119071190811909119101191111912119131191411915119161191711918119191192011921119221192311924119251192611927119281192911930119311193211933119341193511936119371193811939119401194111942119431194411945119461194711948119491195011951119521195311954119551195611957119581195911960119611196211963119641196511966119671196811969119701197111972119731197411975119761197711978119791198011981119821198311984119851198611987119881198911990119911199211993119941199511996119971199811999120001200112002120031200412005120061200712008120091201012011120121201312014120151201612017120181201912020120211202212023120241202512026120271202812029120301203112032120331203412035120361203712038120391204012041120421204312044120451204612047120481204912050120511205212053120541205512056120571205812059120601206112062120631206412065120661206712068120691207012071120721207312074120751207612077120781207912080120811208212083120841208512086120871208812089120901209112092120931209412095120961209712098120991210012101121021210312104121051210612107121081210912110121111211212113121141211512116121171211812119121201212112122121231212412125121261212712128121291213012131121321213312134121351213612137121381213912140121411214212143121441214512146121471214812149121501215112152121531215412155121561215712158121591216012161121621216312164121651216612167121681216912170121711217212173121741217512176121771217812179121801218112182121831218412185121861218712188121891219012191121921219312194121951219612197121981219912200122011220212203122041220512206122071220812209122101221112212122131221412215122161221712218122191222012221122221222312224122251222612227122281222912230122311223212233122341223512236122371223812239122401224112242122431224412245122461224712248122491225012251122521225312254122551225612257122581225912260122611226212263122641226512266122671226812269122701227112272122731227412275122761227712278122791228012281122821228312284122851228612287122881228912290122911229212293122941229512296122971229812299123001230112302123031230412305123061230712308123091231012311123121231312314123151231612317123181231912320123211232212323123241232512326123271232812329123301233112332123331233412335123361233712338123391234012341123421234312344123451234612347123481234912350123511235212353123541235512356123571235812359123601236112362123631236412365123661236712368123691237012371123721237312374123751237612377123781237912380123811238212383123841238512386123871238812389123901239112392123931239412395123961239712398123991240012401124021240312404124051240612407124081240912410124111241212413124141241512416124171241812419124201242112422124231242412425124261242712428124291243012431124321243312434124351243612437124381243912440124411244212443124441244512446124471244812449124501245112452124531245412455124561245712458124591246012461124621246312464124651246612467124681246912470124711247212473124741247512476124771247812479124801248112482124831248412485124861248712488124891249012491124921249312494124951249612497124981249912500125011250212503125041250512506125071250812509125101251112512125131251412515125161251712518125191252012521125221252312524125251252612527125281252912530125311253212533125341253512536125371253812539125401254112542125431254412545125461254712548125491255012551125521255312554125551255612557125581255912560125611256212563125641256512566125671256812569125701257112572125731257412575125761257712578125791258012581125821258312584125851258612587125881258912590125911259212593125941259512596125971259812599126001260112602126031260412605126061260712608126091261012611126121261312614126151261612617126181261912620126211262212623126241262512626126271262812629126301263112632126331263412635126361263712638126391264012641126421264312644126451264612647126481264912650126511265212653126541265512656126571265812659126601266112662126631266412665126661266712668126691267012671126721267312674126751267612677126781267912680126811268212683126841268512686126871268812689126901269112692126931269412695126961269712698126991270012701127021270312704127051270612707127081270912710127111271212713127141271512716127171271812719127201272112722127231272412725127261272712728127291273012731127321273312734127351273612737127381273912740127411274212743127441274512746127471274812749127501275112752127531275412755127561275712758127591276012761127621276312764127651276612767127681276912770127711277212773127741277512776127771277812779127801278112782127831278412785127861278712788127891279012791127921279312794127951279612797127981279912800128011280212803128041280512806128071280812809128101281112812128131281412815128161281712818128191282012821128221282312824128251282612827128281282912830128311283212833128341283512836128371283812839128401284112842128431284412845128461284712848128491285012851128521285312854128551285612857128581285912860128611286212863128641286512866128671286812869128701287112872128731287412875128761287712878128791288012881128821288312884128851288612887128881288912890128911289212893128941289512896128971289812899129001290112902129031290412905129061290712908129091291012911129121291312914129151291612917129181291912920129211292212923129241292512926129271292812929129301293112932129331293412935129361293712938129391294012941129421294312944129451294612947129481294912950129511295212953129541295512956129571295812959129601296112962129631296412965129661296712968129691297012971129721297312974129751297612977129781297912980129811298212983129841298512986129871298812989129901299112992129931299412995129961299712998129991300013001130021300313004130051300613007130081300913010130111301213013130141301513016130171301813019130201302113022130231302413025130261302713028130291303013031130321303313034130351303613037130381303913040130411304213043130441304513046130471304813049130501305113052130531305413055130561305713058130591306013061130621306313064130651306613067130681306913070130711307213073130741307513076130771307813079130801308113082130831308413085130861308713088130891309013091130921309313094130951309613097130981309913100131011310213103131041310513106131071310813109131101311113112131131311413115131161311713118131191312013121131221312313124131251312613127131281312913130131311313213133131341313513136131371313813139131401314113142131431314413145131461314713148131491315013151131521315313154131551315613157131581315913160131611316213163131641316513166131671316813169131701317113172131731317413175131761317713178131791318013181131821318313184131851318613187131881318913190131911319213193131941319513196131971319813199132001320113202132031320413205132061320713208132091321013211132121321313214132151321613217132181321913220132211322213223132241322513226132271322813229132301323113232132331323413235132361323713238132391324013241132421324313244132451324613247132481324913250132511325213253132541325513256132571325813259132601326113262132631326413265132661326713268132691327013271132721327313274132751327613277132781327913280132811328213283132841328513286132871328813289132901329113292132931329413295132961329713298132991330013301133021330313304133051330613307133081330913310133111331213313133141331513316133171331813319133201332113322133231332413325133261332713328133291333013331133321333313334133351333613337133381333913340133411334213343133441334513346133471334813349133501335113352133531335413355133561335713358133591336013361133621336313364133651336613367133681336913370133711337213373133741337513376133771337813379133801338113382133831338413385133861338713388133891339013391133921339313394133951339613397133981339913400134011340213403134041340513406134071340813409134101341113412134131341413415134161341713418134191342013421134221342313424134251342613427134281342913430134311343213433134341343513436134371343813439134401344113442134431344413445134461344713448134491345013451134521345313454134551345613457134581345913460134611346213463134641346513466134671346813469134701347113472134731347413475134761347713478134791348013481134821348313484134851348613487134881348913490134911349213493134941349513496134971349813499135001350113502135031350413505135061350713508135091351013511135121351313514135151351613517135181351913520135211352213523135241352513526135271352813529135301353113532135331353413535135361353713538135391354013541135421354313544135451354613547135481354913550135511355213553135541355513556135571355813559135601356113562135631356413565135661356713568135691357013571135721357313574135751357613577135781357913580135811358213583135841358513586135871358813589135901359113592135931359413595135961359713598135991360013601136021360313604136051360613607136081360913610136111361213613136141361513616136171361813619136201362113622136231362413625136261362713628136291363013631136321363313634136351363613637136381363913640136411364213643136441364513646136471364813649136501365113652136531365413655136561365713658136591366013661136621366313664136651366613667136681366913670136711367213673136741367513676136771367813679136801368113682136831368413685136861368713688136891369013691136921369313694136951369613697136981369913700137011370213703137041370513706137071370813709137101371113712137131371413715137161371713718137191372013721137221372313724137251372613727137281372913730137311373213733137341373513736137371373813739137401374113742137431374413745137461374713748137491375013751137521375313754137551375613757137581375913760137611376213763137641376513766137671376813769137701377113772137731377413775137761377713778137791378013781137821378313784137851378613787137881378913790137911379213793137941379513796137971379813799138001380113802138031380413805138061380713808138091381013811138121381313814138151381613817138181381913820138211382213823138241382513826138271382813829138301383113832138331383413835138361383713838138391384013841138421384313844138451384613847138481384913850138511385213853138541385513856138571385813859138601386113862138631386413865138661386713868138691387013871138721387313874138751387613877138781387913880138811388213883138841388513886138871388813889138901389113892138931389413895138961389713898138991390013901139021390313904139051390613907139081390913910139111391213913139141391513916139171391813919139201392113922139231392413925139261392713928139291393013931139321393313934139351393613937139381393913940139411394213943139441394513946139471394813949139501395113952139531395413955139561395713958139591396013961139621396313964139651396613967139681396913970139711397213973139741397513976139771397813979139801398113982139831398413985139861398713988139891399013991139921399313994139951399613997139981399914000140011400214003140041400514006140071400814009140101401114012140131401414015140161401714018140191402014021140221402314024140251402614027140281402914030140311403214033140341403514036140371403814039140401404114042140431404414045140461404714048140491405014051140521405314054140551405614057140581405914060140611406214063140641406514066140671406814069140701407114072140731407414075140761407714078140791408014081140821408314084140851408614087140881408914090140911409214093140941409514096140971409814099141001410114102141031410414105141061410714108141091411014111141121411314114141151411614117141181411914120141211412214123141241412514126141271412814129141301413114132141331413414135141361413714138141391414014141141421414314144141451414614147141481414914150141511415214153141541415514156141571415814159141601416114162141631416414165141661416714168141691417014171141721417314174141751417614177141781417914180141811418214183141841418514186141871418814189141901419114192141931419414195141961419714198141991420014201142021420314204142051420614207142081420914210142111421214213142141421514216142171421814219142201422114222142231422414225142261422714228142291423014231142321423314234142351423614237142381423914240142411424214243142441424514246142471424814249142501425114252142531425414255142561425714258142591426014261142621426314264142651426614267142681426914270142711427214273142741427514276142771427814279142801428114282142831428414285142861428714288142891429014291142921429314294142951429614297142981429914300143011430214303143041430514306143071430814309143101431114312143131431414315143161431714318143191432014321143221432314324143251432614327143281432914330143311433214333143341433514336143371433814339143401434114342143431434414345143461434714348143491435014351143521435314354143551435614357143581435914360143611436214363143641436514366143671436814369143701437114372143731437414375143761437714378143791438014381143821438314384143851438614387143881438914390143911439214393143941439514396143971439814399144001440114402144031440414405144061440714408144091441014411144121441314414144151441614417144181441914420144211442214423144241442514426144271442814429144301443114432144331443414435144361443714438144391444014441144421444314444144451444614447144481444914450144511445214453144541445514456144571445814459144601446114462144631446414465144661446714468144691447014471144721447314474144751447614477144781447914480144811448214483144841448514486144871448814489144901449114492144931449414495144961449714498144991450014501145021450314504145051450614507145081450914510145111451214513145141451514516145171451814519145201452114522145231452414525145261452714528145291453014531145321453314534145351453614537145381453914540145411454214543145441454514546145471454814549145501455114552145531455414555145561455714558145591456014561145621456314564145651456614567145681456914570145711457214573145741457514576145771457814579145801458114582145831458414585145861458714588145891459014591145921459314594145951459614597145981459914600146011460214603146041460514606146071460814609146101461114612146131461414615146161461714618146191462014621146221462314624146251462614627146281462914630146311463214633146341463514636146371463814639146401464114642146431464414645146461464714648146491465014651146521465314654146551465614657146581465914660146611466214663146641466514666146671466814669146701467114672146731467414675146761467714678146791468014681146821468314684146851468614687146881468914690146911469214693146941469514696146971469814699147001470114702147031470414705147061470714708147091471014711147121471314714147151471614717147181471914720147211472214723147241472514726147271472814729147301473114732147331473414735147361473714738147391474014741147421474314744147451474614747147481474914750147511475214753147541475514756147571475814759147601476114762147631476414765147661476714768147691477014771147721477314774147751477614777147781477914780147811478214783147841478514786147871478814789147901479114792147931479414795147961479714798147991480014801148021480314804148051480614807148081480914810148111481214813148141481514816148171481814819148201482114822148231482414825148261482714828148291483014831148321483314834148351483614837148381483914840148411484214843148441484514846148471484814849148501485114852148531485414855148561485714858148591486014861148621486314864148651486614867148681486914870148711487214873148741487514876148771487814879148801488114882148831488414885148861488714888148891489014891148921489314894148951489614897148981489914900149011490214903149041490514906149071490814909149101491114912149131491414915149161491714918149191492014921149221492314924149251492614927149281492914930149311493214933149341493514936149371493814939149401494114942149431494414945149461494714948149491495014951149521495314954149551495614957149581495914960149611496214963149641496514966149671496814969149701497114972149731497414975149761497714978149791498014981149821498314984149851498614987149881498914990149911499214993149941499514996149971499814999150001500115002150031500415005150061500715008150091501015011150121501315014150151501615017150181501915020150211502215023150241502515026150271502815029150301503115032150331503415035150361503715038150391504015041150421504315044150451504615047150481504915050150511505215053150541505515056150571505815059150601506115062150631506415065150661506715068150691507015071150721507315074150751507615077150781507915080150811508215083150841508515086150871508815089150901509115092150931509415095150961509715098150991510015101151021510315104151051510615107151081510915110151111511215113151141511515116151171511815119151201512115122151231512415125151261512715128151291513015131151321513315134151351513615137151381513915140151411514215143151441514515146151471514815149151501515115152151531515415155151561515715158151591516015161151621516315164151651516615167151681516915170151711517215173151741517515176151771517815179151801518115182151831518415185151861518715188151891519015191151921519315194151951519615197151981519915200152011520215203152041520515206152071520815209152101521115212152131521415215152161521715218152191522015221152221522315224152251522615227152281522915230152311523215233152341523515236152371523815239152401524115242152431524415245152461524715248152491525015251152521525315254152551525615257152581525915260152611526215263152641526515266152671526815269152701527115272152731527415275152761527715278152791528015281152821528315284152851528615287152881528915290152911529215293152941529515296152971529815299153001530115302153031530415305153061530715308153091531015311153121531315314153151531615317153181531915320153211532215323153241532515326153271532815329153301533115332153331533415335153361533715338153391534015341153421534315344153451534615347153481534915350153511535215353153541535515356153571535815359153601536115362153631536415365153661536715368153691537015371153721537315374153751537615377153781537915380153811538215383153841538515386153871538815389153901539115392153931539415395153961539715398153991540015401154021540315404154051540615407154081540915410154111541215413154141541515416154171541815419154201542115422154231542415425154261542715428154291543015431154321543315434154351543615437154381543915440154411544215443154441544515446154471544815449154501545115452154531545415455154561545715458154591546015461154621546315464154651546615467154681546915470154711547215473154741547515476154771547815479154801548115482154831548415485154861548715488154891549015491154921549315494154951549615497154981549915500155011550215503155041550515506155071550815509155101551115512155131551415515155161551715518155191552015521155221552315524155251552615527155281552915530155311553215533155341553515536155371553815539155401554115542155431554415545155461554715548155491555015551155521555315554155551555615557155581555915560155611556215563155641556515566155671556815569155701557115572155731557415575155761557715578155791558015581155821558315584155851558615587155881558915590155911559215593155941559515596155971559815599156001560115602156031560415605156061560715608156091561015611156121561315614156151561615617156181561915620156211562215623156241562515626156271562815629156301563115632156331563415635156361563715638156391564015641156421564315644156451564615647156481564915650156511565215653156541565515656156571565815659156601566115662156631566415665156661566715668156691567015671156721567315674156751567615677156781567915680156811568215683156841568515686156871568815689156901569115692156931569415695156961569715698156991570015701157021570315704157051570615707157081570915710157111571215713157141571515716157171571815719157201572115722157231572415725157261572715728157291573015731157321573315734157351573615737157381573915740157411574215743157441574515746157471574815749157501575115752157531575415755157561575715758157591576015761157621576315764157651576615767157681576915770157711577215773157741577515776157771577815779157801578115782157831578415785157861578715788157891579015791157921579315794157951579615797157981579915800158011580215803158041580515806158071580815809158101581115812158131581415815158161581715818158191582015821158221582315824158251582615827158281582915830158311583215833158341583515836158371583815839158401584115842158431584415845158461584715848158491585015851158521585315854158551585615857158581585915860158611586215863158641586515866158671586815869158701587115872158731587415875158761587715878158791588015881158821588315884158851588615887158881588915890158911589215893158941589515896158971589815899159001590115902159031590415905159061590715908159091591015911159121591315914159151591615917159181591915920159211592215923159241592515926159271592815929159301593115932159331593415935159361593715938159391594015941159421594315944159451594615947159481594915950159511595215953159541595515956159571595815959159601596115962159631596415965159661596715968159691597015971159721597315974159751597615977159781597915980159811598215983159841598515986159871598815989159901599115992159931599415995159961599715998159991600016001160021600316004160051600616007160081600916010160111601216013160141601516016160171601816019160201602116022160231602416025160261602716028160291603016031160321603316034160351603616037160381603916040160411604216043160441604516046160471604816049160501605116052160531605416055160561605716058160591606016061160621606316064160651606616067160681606916070160711607216073160741607516076160771607816079160801608116082160831608416085160861608716088160891609016091160921609316094160951609616097160981609916100161011610216103161041610516106161071610816109161101611116112161131611416115161161611716118161191612016121161221612316124161251612616127161281612916130161311613216133161341613516136161371613816139161401614116142161431614416145161461614716148161491615016151161521615316154161551615616157161581615916160161611616216163161641616516166161671616816169161701617116172161731617416175161761617716178161791618016181161821618316184161851618616187161881618916190161911619216193161941619516196161971619816199162001620116202162031620416205162061620716208162091621016211162121621316214162151621616217162181621916220162211622216223162241622516226162271622816229162301623116232162331623416235162361623716238162391624016241162421624316244162451624616247162481624916250162511625216253162541625516256162571625816259162601626116262162631626416265162661626716268162691627016271162721627316274162751627616277162781627916280162811628216283162841628516286162871628816289162901629116292162931629416295162961629716298162991630016301163021630316304163051630616307163081630916310163111631216313163141631516316163171631816319163201632116322163231632416325163261632716328163291633016331163321633316334163351633616337163381633916340163411634216343163441634516346163471634816349163501635116352163531635416355163561635716358163591636016361163621636316364163651636616367163681636916370163711637216373163741637516376163771637816379163801638116382163831638416385163861638716388163891639016391163921639316394163951639616397163981639916400164011640216403164041640516406164071640816409164101641116412164131641416415164161641716418164191642016421164221642316424164251642616427164281642916430164311643216433164341643516436164371643816439164401644116442164431644416445164461644716448164491645016451164521645316454164551645616457164581645916460164611646216463164641646516466164671646816469164701647116472164731647416475164761647716478164791648016481164821648316484164851648616487164881648916490164911649216493164941649516496164971649816499165001650116502165031650416505165061650716508165091651016511165121651316514165151651616517165181651916520165211652216523165241652516526165271652816529165301653116532165331653416535165361653716538165391654016541165421654316544165451654616547165481654916550165511655216553165541655516556165571655816559165601656116562165631656416565165661656716568165691657016571165721657316574165751657616577165781657916580165811658216583165841658516586165871658816589165901659116592165931659416595165961659716598165991660016601166021660316604166051660616607166081660916610166111661216613166141661516616166171661816619166201662116622166231662416625166261662716628166291663016631166321663316634166351663616637166381663916640166411664216643166441664516646166471664816649166501665116652166531665416655166561665716658166591666016661166621666316664166651666616667166681666916670166711667216673166741667516676166771667816679166801668116682166831668416685166861668716688166891669016691166921669316694166951669616697166981669916700167011670216703167041670516706167071670816709167101671116712167131671416715167161671716718167191672016721167221672316724167251672616727167281672916730167311673216733167341673516736167371673816739167401674116742167431674416745167461674716748167491675016751167521675316754167551675616757167581675916760167611676216763167641676516766167671676816769167701677116772167731677416775167761677716778167791678016781167821678316784167851678616787167881678916790167911679216793167941679516796167971679816799168001680116802168031680416805168061680716808168091681016811168121681316814168151681616817168181681916820168211682216823168241682516826168271682816829168301683116832168331683416835168361683716838168391684016841168421684316844168451684616847168481684916850168511685216853168541685516856168571685816859168601686116862168631686416865168661686716868168691687016871168721687316874168751687616877168781687916880168811688216883168841688516886168871688816889168901689116892168931689416895168961689716898168991690016901169021690316904169051690616907169081690916910169111691216913169141691516916169171691816919169201692116922169231692416925169261692716928169291693016931169321693316934169351693616937169381693916940169411694216943169441694516946169471694816949169501695116952169531695416955169561695716958169591696016961169621696316964169651696616967169681696916970169711697216973169741697516976169771697816979169801698116982169831698416985169861698716988169891699016991169921699316994169951699616997169981699917000170011700217003170041700517006170071700817009170101701117012170131701417015170161701717018170191702017021170221702317024170251702617027170281702917030170311703217033170341703517036170371703817039170401704117042170431704417045170461704717048170491705017051170521705317054170551705617057170581705917060170611706217063170641706517066170671706817069170701707117072170731707417075170761707717078170791708017081170821708317084170851708617087170881708917090170911709217093170941709517096170971709817099171001710117102171031710417105171061710717108171091711017111171121711317114171151711617117171181711917120171211712217123171241712517126171271712817129171301713117132171331713417135171361713717138171391714017141171421714317144171451714617147171481714917150171511715217153171541715517156171571715817159171601716117162171631716417165171661716717168171691717017171171721717317174171751717617177171781717917180171811718217183171841718517186171871718817189171901719117192171931719417195171961719717198171991720017201172021720317204172051720617207172081720917210172111721217213172141721517216172171721817219172201722117222172231722417225172261722717228172291723017231172321723317234172351723617237172381723917240172411724217243172441724517246172471724817249172501725117252172531725417255172561725717258172591726017261172621726317264172651726617267172681726917270172711727217273172741727517276172771727817279172801728117282172831728417285172861728717288172891729017291172921729317294172951729617297172981729917300173011730217303173041730517306173071730817309173101731117312173131731417315173161731717318173191732017321173221732317324173251732617327173281732917330173311733217333173341733517336173371733817339173401734117342173431734417345173461734717348173491735017351173521735317354173551735617357173581735917360173611736217363173641736517366173671736817369173701737117372173731737417375173761737717378173791738017381173821738317384173851738617387173881738917390173911739217393173941739517396173971739817399174001740117402174031740417405174061740717408174091741017411174121741317414174151741617417174181741917420174211742217423174241742517426174271742817429174301743117432174331743417435174361743717438174391744017441174421744317444174451744617447174481744917450174511745217453174541745517456174571745817459174601746117462174631746417465174661746717468174691747017471174721747317474174751747617477174781747917480174811748217483174841748517486174871748817489174901749117492174931749417495174961749717498174991750017501175021750317504175051750617507175081750917510175111751217513175141751517516175171751817519175201752117522175231752417525175261752717528175291753017531175321753317534175351753617537175381753917540175411754217543175441754517546175471754817549175501755117552175531755417555175561755717558175591756017561175621756317564175651756617567175681756917570175711757217573175741757517576175771757817579175801758117582175831758417585175861758717588175891759017591175921759317594175951759617597175981759917600176011760217603176041760517606176071760817609176101761117612176131761417615176161761717618176191762017621176221762317624176251762617627176281762917630176311763217633176341763517636176371763817639176401764117642176431764417645176461764717648176491765017651176521765317654176551765617657176581765917660176611766217663176641766517666176671766817669176701767117672176731767417675176761767717678176791768017681176821768317684176851768617687176881768917690176911769217693176941769517696176971769817699177001770117702177031770417705177061770717708177091771017711177121771317714177151771617717177181771917720177211772217723177241772517726177271772817729177301773117732177331773417735177361773717738177391774017741177421774317744177451774617747177481774917750177511775217753177541775517756177571775817759177601776117762177631776417765177661776717768177691777017771177721777317774177751777617777177781777917780177811778217783177841778517786177871778817789177901779117792177931779417795177961779717798177991780017801178021780317804178051780617807178081780917810178111781217813178141781517816178171781817819178201782117822178231782417825178261782717828178291783017831178321783317834178351783617837178381783917840178411784217843178441784517846178471784817849178501785117852178531785417855178561785717858178591786017861178621786317864178651786617867178681786917870178711787217873178741787517876178771787817879178801788117882178831788417885178861788717888178891789017891178921789317894178951789617897178981789917900179011790217903179041790517906179071790817909179101791117912179131791417915179161791717918179191792017921179221792317924179251792617927179281792917930179311793217933179341793517936179371793817939179401794117942179431794417945179461794717948179491795017951179521795317954179551795617957179581795917960179611796217963179641796517966179671796817969179701797117972179731797417975179761797717978179791798017981179821798317984179851798617987179881798917990179911799217993179941799517996179971799817999180001800118002180031800418005180061800718008180091801018011180121801318014180151801618017180181801918020180211802218023180241802518026180271802818029180301803118032180331803418035180361803718038180391804018041180421804318044180451804618047180481804918050180511805218053180541805518056180571805818059180601806118062180631806418065180661806718068180691807018071180721807318074180751807618077180781807918080180811808218083180841808518086180871808818089180901809118092180931809418095180961809718098180991810018101181021810318104181051810618107181081810918110181111811218113181141811518116181171811818119181201812118122181231812418125181261812718128181291813018131181321813318134181351813618137181381813918140181411814218143181441814518146181471814818149181501815118152181531815418155181561815718158181591816018161181621816318164181651816618167181681816918170181711817218173181741817518176181771817818179181801818118182181831818418185181861818718188181891819018191181921819318194181951819618197181981819918200182011820218203182041820518206182071820818209182101821118212182131821418215182161821718218182191822018221182221822318224182251822618227182281822918230182311823218233182341823518236182371823818239182401824118242182431824418245182461824718248182491825018251182521825318254182551825618257182581825918260182611826218263182641826518266182671826818269182701827118272182731827418275182761827718278182791828018281182821828318284182851828618287182881828918290182911829218293182941829518296182971829818299183001830118302183031830418305183061830718308183091831018311183121831318314183151831618317183181831918320183211832218323183241832518326183271832818329183301833118332183331833418335183361833718338183391834018341183421834318344183451834618347183481834918350183511835218353183541835518356183571835818359183601836118362183631836418365183661836718368183691837018371183721837318374183751837618377183781837918380183811838218383183841838518386183871838818389183901839118392183931839418395183961839718398183991840018401184021840318404184051840618407184081840918410184111841218413184141841518416184171841818419184201842118422184231842418425184261842718428184291843018431184321843318434184351843618437184381843918440184411844218443184441844518446184471844818449184501845118452184531845418455184561845718458184591846018461184621846318464184651846618467184681846918470184711847218473184741847518476184771847818479184801848118482184831848418485184861848718488184891849018491184921849318494184951849618497184981849918500185011850218503185041850518506185071850818509185101851118512185131851418515185161851718518185191852018521185221852318524185251852618527185281852918530185311853218533185341853518536185371853818539185401854118542185431854418545185461854718548185491855018551185521855318554185551855618557185581855918560185611856218563185641856518566185671856818569185701857118572185731857418575185761857718578185791858018581185821858318584185851858618587185881858918590185911859218593185941859518596185971859818599186001860118602186031860418605186061860718608186091861018611186121861318614186151861618617186181861918620186211862218623186241862518626186271862818629186301863118632186331863418635186361863718638186391864018641186421864318644186451864618647186481864918650186511865218653186541865518656186571865818659186601866118662186631866418665186661866718668186691867018671186721867318674186751867618677186781867918680186811868218683186841868518686186871868818689186901869118692186931869418695186961869718698186991870018701187021870318704187051870618707187081870918710187111871218713187141871518716187171871818719187201872118722187231872418725187261872718728187291873018731187321873318734187351873618737187381873918740187411874218743187441874518746187471874818749187501875118752187531875418755187561875718758187591876018761187621876318764187651876618767187681876918770187711877218773187741877518776187771877818779187801878118782187831878418785187861878718788187891879018791187921879318794187951879618797187981879918800188011880218803188041880518806188071880818809188101881118812188131881418815188161881718818188191882018821188221882318824188251882618827188281882918830188311883218833188341883518836188371883818839188401884118842188431884418845188461884718848188491885018851188521885318854188551885618857188581885918860188611886218863188641886518866188671886818869188701887118872188731887418875188761887718878188791888018881188821888318884188851888618887188881888918890188911889218893188941889518896188971889818899189001890118902189031890418905189061890718908189091891018911189121891318914189151891618917189181891918920189211892218923189241892518926189271892818929189301893118932189331893418935189361893718938189391894018941189421894318944189451894618947189481894918950189511895218953189541895518956189571895818959189601896118962189631896418965189661896718968189691897018971189721897318974189751897618977189781897918980189811898218983189841898518986189871898818989189901899118992189931899418995189961899718998189991900019001190021900319004190051900619007190081900919010190111901219013190141901519016190171901819019190201902119022190231902419025190261902719028190291903019031190321903319034190351903619037190381903919040190411904219043190441904519046190471904819049190501905119052190531905419055190561905719058190591906019061190621906319064190651906619067190681906919070190711907219073190741907519076190771907819079190801908119082190831908419085190861908719088190891909019091190921909319094190951909619097190981909919100191011910219103191041910519106191071910819109191101911119112191131911419115191161911719118191191912019121191221912319124191251912619127191281912919130191311913219133191341913519136191371913819139191401914119142191431914419145191461914719148191491915019151191521915319154191551915619157191581915919160191611916219163191641916519166191671916819169191701917119172191731917419175191761917719178191791918019181191821918319184191851918619187191881918919190191911919219193191941919519196191971919819199192001920119202192031920419205192061920719208192091921019211192121921319214192151921619217192181921919220192211922219223192241922519226192271922819229192301923119232192331923419235192361923719238192391924019241192421924319244192451924619247192481924919250192511925219253192541925519256192571925819259192601926119262192631926419265192661926719268192691927019271192721927319274192751927619277192781927919280192811928219283192841928519286192871928819289192901929119292192931929419295192961929719298192991930019301193021930319304193051930619307193081930919310193111931219313193141931519316193171931819319193201932119322193231932419325193261932719328193291933019331193321933319334193351933619337193381933919340193411934219343193441934519346193471934819349193501935119352193531935419355193561935719358193591936019361193621936319364193651936619367193681936919370193711937219373193741937519376193771937819379193801938119382193831938419385193861938719388193891939019391193921939319394193951939619397193981939919400194011940219403194041940519406194071940819409194101941119412194131941419415194161941719418194191942019421194221942319424194251942619427194281942919430194311943219433194341943519436194371943819439194401944119442194431944419445194461944719448194491945019451194521945319454194551945619457194581945919460194611946219463194641946519466194671946819469194701947119472194731947419475194761947719478194791948019481194821948319484194851948619487194881948919490194911949219493194941949519496194971949819499195001950119502195031950419505195061950719508195091951019511195121951319514195151951619517195181951919520195211952219523195241952519526195271952819529195301953119532195331953419535195361953719538195391954019541195421954319544195451954619547195481954919550195511955219553195541955519556195571955819559195601956119562195631956419565195661956719568195691957019571195721957319574195751957619577195781957919580195811958219583195841958519586195871958819589195901959119592195931959419595195961959719598195991960019601196021960319604196051960619607196081960919610196111961219613196141961519616196171961819619196201962119622196231962419625196261962719628196291963019631196321963319634196351963619637196381963919640196411964219643196441964519646196471964819649196501965119652196531965419655196561965719658196591966019661196621966319664196651966619667196681966919670196711967219673196741967519676196771967819679196801968119682196831968419685196861968719688196891969019691196921969319694196951969619697196981969919700197011970219703197041970519706197071970819709197101971119712197131971419715197161971719718197191972019721197221972319724197251972619727197281972919730197311973219733197341973519736197371973819739197401974119742197431974419745197461974719748197491975019751197521975319754197551975619757197581975919760197611976219763197641976519766197671976819769197701977119772197731977419775197761977719778197791978019781197821978319784197851978619787197881978919790197911979219793197941979519796197971979819799198001980119802198031980419805198061980719808198091981019811198121981319814198151981619817198181981919820198211982219823198241982519826198271982819829198301983119832198331983419835198361983719838198391984019841198421984319844198451984619847198481984919850198511985219853198541985519856198571985819859198601986119862198631986419865198661986719868198691987019871198721987319874198751987619877198781987919880198811988219883198841988519886198871988819889198901989119892198931989419895198961989719898198991990019901199021990319904199051990619907199081990919910199111991219913199141991519916199171991819919199201992119922199231992419925199261992719928199291993019931199321993319934199351993619937199381993919940199411994219943199441994519946199471994819949199501995119952199531995419955199561995719958199591996019961199621996319964199651996619967199681996919970199711997219973199741997519976199771997819979199801998119982199831998419985199861998719988199891999019991199921999319994199951999619997199981999920000200012000220003200042000520006200072000820009200102001120012200132001420015200162001720018200192002020021200222002320024200252002620027200282002920030200312003220033200342003520036200372003820039200402004120042200432004420045200462004720048200492005020051200522005320054200552005620057200582005920060200612006220063200642006520066200672006820069200702007120072200732007420075200762007720078200792008020081200822008320084200852008620087200882008920090200912009220093200942009520096200972009820099201002010120102201032010420105201062010720108201092011020111201122011320114201152011620117201182011920120201212012220123201242012520126201272012820129201302013120132201332013420135201362013720138201392014020141201422014320144201452014620147201482014920150201512015220153201542015520156201572015820159201602016120162201632016420165201662016720168201692017020171201722017320174201752017620177201782017920180201812018220183201842018520186201872018820189201902019120192201932019420195201962019720198201992020020201202022020320204202052020620207202082020920210202112021220213202142021520216202172021820219202202022120222202232022420225202262022720228202292023020231202322023320234202352023620237202382023920240202412024220243202442024520246202472024820249202502025120252202532025420255202562025720258202592026020261202622026320264202652026620267202682026920270202712027220273202742027520276202772027820279202802028120282202832028420285202862028720288202892029020291202922029320294202952029620297202982029920300203012030220303203042030520306203072030820309203102031120312203132031420315203162031720318203192032020321203222032320324203252032620327203282032920330203312033220333203342033520336203372033820339203402034120342203432034420345203462034720348203492035020351203522035320354203552035620357203582035920360203612036220363203642036520366203672036820369203702037120372203732037420375203762037720378203792038020381203822038320384203852038620387203882038920390203912039220393203942039520396203972039820399204002040120402204032040420405204062040720408204092041020411204122041320414204152041620417204182041920420204212042220423204242042520426204272042820429204302043120432204332043420435204362043720438204392044020441204422044320444204452044620447204482044920450204512045220453204542045520456204572045820459204602046120462204632046420465204662046720468204692047020471204722047320474204752047620477204782047920480204812048220483204842048520486204872048820489204902049120492204932049420495204962049720498204992050020501205022050320504205052050620507205082050920510205112051220513205142051520516205172051820519205202052120522205232052420525205262052720528205292053020531205322053320534205352053620537205382053920540205412054220543205442054520546205472054820549205502055120552205532055420555205562055720558205592056020561205622056320564205652056620567205682056920570205712057220573205742057520576205772057820579205802058120582205832058420585205862058720588205892059020591205922059320594205952059620597205982059920600206012060220603206042060520606206072060820609206102061120612206132061420615206162061720618206192062020621206222062320624206252062620627206282062920630206312063220633206342063520636206372063820639206402064120642206432064420645206462064720648206492065020651206522065320654206552065620657206582065920660206612066220663206642066520666206672066820669206702067120672206732067420675206762067720678206792068020681206822068320684206852068620687206882068920690206912069220693206942069520696206972069820699207002070120702207032070420705207062070720708207092071020711207122071320714207152071620717207182071920720207212072220723207242072520726207272072820729207302073120732207332073420735207362073720738207392074020741207422074320744207452074620747207482074920750207512075220753207542075520756207572075820759207602076120762207632076420765207662076720768207692077020771207722077320774207752077620777207782077920780207812078220783207842078520786207872078820789207902079120792207932079420795207962079720798207992080020801208022080320804208052080620807208082080920810208112081220813208142081520816208172081820819208202082120822208232082420825208262082720828208292083020831208322083320834208352083620837208382083920840208412084220843208442084520846208472084820849208502085120852208532085420855208562085720858208592086020861208622086320864208652086620867208682086920870208712087220873208742087520876208772087820879208802088120882208832088420885208862088720888208892089020891208922089320894208952089620897208982089920900209012090220903209042090520906209072090820909209102091120912209132091420915209162091720918209192092020921209222092320924209252092620927209282092920930209312093220933209342093520936209372093820939209402094120942209432094420945209462094720948209492095020951209522095320954209552095620957209582095920960209612096220963209642096520966209672096820969209702097120972209732097420975209762097720978209792098020981209822098320984209852098620987209882098920990209912099220993209942099520996209972099820999210002100121002210032100421005210062100721008210092101021011210122101321014210152101621017210182101921020210212102221023210242102521026210272102821029210302103121032210332103421035210362103721038210392104021041210422104321044210452104621047210482104921050210512105221053210542105521056210572105821059210602106121062210632106421065210662106721068210692107021071210722107321074210752107621077210782107921080210812108221083210842108521086210872108821089210902109121092210932109421095210962109721098210992110021101211022110321104211052110621107211082110921110211112111221113211142111521116211172111821119211202112121122211232112421125211262112721128211292113021131211322113321134211352113621137211382113921140211412114221143211442114521146211472114821149211502115121152211532115421155211562115721158211592116021161211622116321164211652116621167211682116921170211712117221173211742117521176211772117821179211802118121182211832118421185211862118721188211892119021191211922119321194211952119621197211982119921200212012120221203212042120521206212072120821209212102121121212212132121421215212162121721218212192122021221212222122321224212252122621227212282122921230212312123221233212342123521236212372123821239212402124121242212432124421245212462124721248212492125021251212522125321254212552125621257212582125921260212612126221263212642126521266212672126821269212702127121272212732127421275212762127721278212792128021281212822128321284212852128621287212882128921290212912129221293212942129521296212972129821299213002130121302213032130421305213062130721308213092131021311213122131321314213152131621317213182131921320213212132221323213242132521326213272132821329213302133121332213332133421335213362133721338213392134021341213422134321344213452134621347213482134921350213512135221353213542135521356213572135821359213602136121362213632136421365213662136721368213692137021371213722137321374213752137621377213782137921380213812138221383213842138521386213872138821389213902139121392213932139421395213962139721398213992140021401214022140321404214052140621407214082140921410214112141221413214142141521416214172141821419214202142121422214232142421425214262142721428214292143021431214322143321434214352143621437214382143921440214412144221443214442144521446214472144821449214502145121452214532145421455214562145721458214592146021461214622146321464214652146621467214682146921470214712147221473214742147521476214772147821479214802148121482214832148421485214862148721488214892149021491214922149321494214952149621497214982149921500215012150221503215042150521506215072150821509215102151121512215132151421515215162151721518215192152021521215222152321524215252152621527215282152921530215312153221533215342153521536215372153821539215402154121542215432154421545215462154721548215492155021551215522155321554215552155621557215582155921560215612156221563215642156521566215672156821569215702157121572215732157421575215762157721578215792158021581215822158321584215852158621587215882158921590215912159221593215942159521596215972159821599216002160121602216032160421605216062160721608216092161021611216122161321614216152161621617216182161921620216212162221623216242162521626216272162821629216302163121632216332163421635216362163721638216392164021641216422164321644216452164621647216482164921650216512165221653216542165521656216572165821659216602166121662216632166421665216662166721668216692167021671216722167321674216752167621677216782167921680216812168221683216842168521686216872168821689216902169121692216932169421695216962169721698216992170021701217022170321704217052170621707217082170921710217112171221713217142171521716217172171821719217202172121722217232172421725217262172721728217292173021731
  1. /* sp.c
  2. *
  3. * Copyright (C) 2006-2020 wolfSSL Inc.
  4. *
  5. * This file is part of wolfSSL.
  6. *
  7. * wolfSSL is free software; you can redistribute it and/or modify
  8. * it under the terms of the GNU General Public License as published by
  9. * the Free Software Foundation; either version 2 of the License, or
  10. * (at your option) any later version.
  11. *
  12. * wolfSSL is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  15. * GNU General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU General Public License
  18. * along with this program; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335, USA
  20. */
  21. /* Implementation by Sean Parkinson. */
  22. #ifdef HAVE_CONFIG_H
  23. #include <config.h>
  24. #endif
  25. #include <wolfssl/wolfcrypt/settings.h>
  26. #include <wolfssl/wolfcrypt/error-crypt.h>
  27. #include <wolfssl/wolfcrypt/cpuid.h>
  28. #ifdef NO_INLINE
  29. #include <wolfssl/wolfcrypt/misc.h>
  30. #else
  31. #define WOLFSSL_MISC_INCLUDED
  32. #include <wolfcrypt/src/misc.c>
  33. #endif
  34. #if defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH) || \
  35. defined(WOLFSSL_HAVE_SP_ECC)
  36. #ifdef RSA_LOW_MEM
  37. #ifndef WOLFSSL_SP_SMALL
  38. #define WOLFSSL_SP_SMALL
  39. #endif
  40. #endif
  41. #include <wolfssl/wolfcrypt/sp.h>
  42. #ifdef WOLFSSL_SP_ARM_THUMB_ASM
  43. #if defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)
  44. #ifndef WOLFSSL_SP_NO_2048
  45. /* Read big endian unsigned byte array into r.
  46. *
  47. * r A single precision integer.
  48. * size Maximum number of bytes to convert
  49. * a Byte array.
  50. * n Number of bytes in array to read.
  51. */
  52. static void sp_2048_from_bin(sp_digit* r, int size, const byte* a, int n)
  53. {
  54. int i, j = 0;
  55. word32 s = 0;
  56. r[0] = 0;
  57. for (i = n-1; i >= 0; i--) {
  58. r[j] |= (((sp_digit)a[i]) << s);
  59. if (s >= 24U) {
  60. r[j] &= 0xffffffff;
  61. s = 32U - s;
  62. if (j + 1 >= size) {
  63. break;
  64. }
  65. r[++j] = (sp_digit)a[i] >> s;
  66. s = 8U - s;
  67. }
  68. else {
  69. s += 8U;
  70. }
  71. }
  72. for (j++; j < size; j++) {
  73. r[j] = 0;
  74. }
  75. }
  76. /* Convert an mp_int to an array of sp_digit.
  77. *
  78. * r A single precision integer.
  79. * size Maximum number of bytes to convert
  80. * a A multi-precision integer.
  81. */
  82. static void sp_2048_from_mp(sp_digit* r, int size, const mp_int* a)
  83. {
  84. #if DIGIT_BIT == 32
  85. int j;
  86. XMEMCPY(r, a->dp, sizeof(sp_digit) * a->used);
  87. for (j = a->used; j < size; j++) {
  88. r[j] = 0;
  89. }
  90. #elif DIGIT_BIT > 32
  91. int i, j = 0;
  92. word32 s = 0;
  93. r[0] = 0;
  94. for (i = 0; i < a->used && j < size; i++) {
  95. r[j] |= ((sp_digit)a->dp[i] << s);
  96. r[j] &= 0xffffffff;
  97. s = 32U - s;
  98. if (j + 1 >= size) {
  99. break;
  100. }
  101. /* lint allow cast of mismatch word32 and mp_digit */
  102. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  103. while ((s + 32U) <= (word32)DIGIT_BIT) {
  104. s += 32U;
  105. r[j] &= 0xffffffff;
  106. if (j + 1 >= size) {
  107. break;
  108. }
  109. if (s < (word32)DIGIT_BIT) {
  110. /* lint allow cast of mismatch word32 and mp_digit */
  111. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  112. }
  113. else {
  114. r[++j] = 0L;
  115. }
  116. }
  117. s = (word32)DIGIT_BIT - s;
  118. }
  119. for (j++; j < size; j++) {
  120. r[j] = 0;
  121. }
  122. #else
  123. int i, j = 0, s = 0;
  124. r[0] = 0;
  125. for (i = 0; i < a->used && j < size; i++) {
  126. r[j] |= ((sp_digit)a->dp[i]) << s;
  127. if (s + DIGIT_BIT >= 32) {
  128. r[j] &= 0xffffffff;
  129. if (j + 1 >= size) {
  130. break;
  131. }
  132. s = 32 - s;
  133. if (s == DIGIT_BIT) {
  134. r[++j] = 0;
  135. s = 0;
  136. }
  137. else {
  138. r[++j] = a->dp[i] >> s;
  139. s = DIGIT_BIT - s;
  140. }
  141. }
  142. else {
  143. s += DIGIT_BIT;
  144. }
  145. }
  146. for (j++; j < size; j++) {
  147. r[j] = 0;
  148. }
  149. #endif
  150. }
  151. /* Write r as big endian to byte array.
  152. * Fixed length number of bytes written: 256
  153. *
  154. * r A single precision integer.
  155. * a Byte array.
  156. */
  157. static void sp_2048_to_bin(sp_digit* r, byte* a)
  158. {
  159. int i, j, s = 0, b;
  160. j = 2048 / 8 - 1;
  161. a[j] = 0;
  162. for (i=0; i<64 && j>=0; i++) {
  163. b = 0;
  164. /* lint allow cast of mismatch sp_digit and int */
  165. a[j--] |= (byte)(r[i] << s); b += 8 - s; /*lint !e9033*/
  166. if (j < 0) {
  167. break;
  168. }
  169. while (b < 32) {
  170. a[j--] = r[i] >> b; b += 8;
  171. if (j < 0) {
  172. break;
  173. }
  174. }
  175. s = 8 - (b - 32);
  176. if (j >= 0) {
  177. a[j] = 0;
  178. }
  179. if (s != 0) {
  180. j++;
  181. }
  182. }
  183. }
  184. #ifndef WOLFSSL_SP_SMALL
  185. /* Multiply a and b into r. (r = a * b)
  186. *
  187. * r A single precision integer.
  188. * a A single precision integer.
  189. * b A single precision integer.
  190. */
  191. SP_NOINLINE static void sp_2048_mul_8(sp_digit* r, const sp_digit* a,
  192. const sp_digit* b)
  193. {
  194. sp_digit tmp[8 * 2];
  195. __asm__ __volatile__ (
  196. "mov r3, #0\n\t"
  197. "mov r4, #0\n\t"
  198. "mov r8, r3\n\t"
  199. "mov r11, %[r]\n\t"
  200. "mov r9, %[a]\n\t"
  201. "mov r10, %[b]\n\t"
  202. "mov r6, #32\n\t"
  203. "add r6, r9\n\t"
  204. "mov r12, r6\n\t"
  205. "\n1:\n\t"
  206. "mov %[r], #0\n\t"
  207. "mov r5, #0\n\t"
  208. "mov r6, #28\n\t"
  209. "mov %[a], r8\n\t"
  210. "sub %[a], r6\n\t"
  211. "sbc r6, r6\n\t"
  212. "mvn r6, r6\n\t"
  213. "and %[a], r6\n\t"
  214. "mov %[b], r8\n\t"
  215. "sub %[b], %[a]\n\t"
  216. "add %[a], r9\n\t"
  217. "add %[b], r10\n\t"
  218. "\n2:\n\t"
  219. "# Multiply Start\n\t"
  220. "ldr r6, [%[a]]\n\t"
  221. "ldr r7, [%[b]]\n\t"
  222. "lsl r6, r6, #16\n\t"
  223. "lsl r7, r7, #16\n\t"
  224. "lsr r6, r6, #16\n\t"
  225. "lsr r7, r7, #16\n\t"
  226. "mul r7, r6\n\t"
  227. "add r3, r7\n\t"
  228. "adc r4, %[r]\n\t"
  229. "adc r5, %[r]\n\t"
  230. "ldr r7, [%[b]]\n\t"
  231. "lsr r7, r7, #16\n\t"
  232. "mul r6, r7\n\t"
  233. "lsr r7, r6, #16\n\t"
  234. "lsl r6, r6, #16\n\t"
  235. "add r3, r6\n\t"
  236. "adc r4, r7\n\t"
  237. "adc r5, %[r]\n\t"
  238. "ldr r6, [%[a]]\n\t"
  239. "ldr r7, [%[b]]\n\t"
  240. "lsr r6, r6, #16\n\t"
  241. "lsr r7, r7, #16\n\t"
  242. "mul r7, r6\n\t"
  243. "add r4, r7\n\t"
  244. "adc r5, %[r]\n\t"
  245. "ldr r7, [%[b]]\n\t"
  246. "lsl r7, r7, #16\n\t"
  247. "lsr r7, r7, #16\n\t"
  248. "mul r6, r7\n\t"
  249. "lsr r7, r6, #16\n\t"
  250. "lsl r6, r6, #16\n\t"
  251. "add r3, r6\n\t"
  252. "adc r4, r7\n\t"
  253. "adc r5, %[r]\n\t"
  254. "# Multiply Done\n\t"
  255. "add %[a], #4\n\t"
  256. "sub %[b], #4\n\t"
  257. "cmp %[a], r12\n\t"
  258. "beq 3f\n\t"
  259. "mov r6, r8\n\t"
  260. "add r6, r9\n\t"
  261. "cmp %[a], r6\n\t"
  262. "ble 2b\n\t"
  263. "\n3:\n\t"
  264. "mov %[r], r11\n\t"
  265. "mov r7, r8\n\t"
  266. "str r3, [%[r], r7]\n\t"
  267. "mov r3, r4\n\t"
  268. "mov r4, r5\n\t"
  269. "add r7, #4\n\t"
  270. "mov r8, r7\n\t"
  271. "mov r6, #56\n\t"
  272. "cmp r7, r6\n\t"
  273. "ble 1b\n\t"
  274. "str r3, [%[r], r7]\n\t"
  275. "mov %[a], r9\n\t"
  276. "mov %[b], r10\n\t"
  277. :
  278. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  279. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  280. );
  281. XMEMCPY(r, tmp, sizeof(tmp));
  282. }
  283. /* Square a and put result in r. (r = a * a)
  284. *
  285. * r A single precision integer.
  286. * a A single precision integer.
  287. */
  288. SP_NOINLINE static void sp_2048_sqr_8(sp_digit* r, const sp_digit* a)
  289. {
  290. __asm__ __volatile__ (
  291. "mov r3, #0\n\t"
  292. "mov r4, #0\n\t"
  293. "mov r5, #0\n\t"
  294. "mov r8, r3\n\t"
  295. "mov r11, %[r]\n\t"
  296. "mov r6, #64\n\t"
  297. "neg r6, r6\n\t"
  298. "add sp, r6\n\t"
  299. "mov r10, sp\n\t"
  300. "mov r9, %[a]\n\t"
  301. "\n1:\n\t"
  302. "mov %[r], #0\n\t"
  303. "mov r6, #28\n\t"
  304. "mov %[a], r8\n\t"
  305. "sub %[a], r6\n\t"
  306. "sbc r6, r6\n\t"
  307. "mvn r6, r6\n\t"
  308. "and %[a], r6\n\t"
  309. "mov r2, r8\n\t"
  310. "sub r2, %[a]\n\t"
  311. "add %[a], r9\n\t"
  312. "add r2, r9\n\t"
  313. "\n2:\n\t"
  314. "cmp r2, %[a]\n\t"
  315. "beq 4f\n\t"
  316. "# Multiply * 2: Start\n\t"
  317. "ldr r6, [%[a]]\n\t"
  318. "ldr r7, [r2]\n\t"
  319. "lsl r6, r6, #16\n\t"
  320. "lsl r7, r7, #16\n\t"
  321. "lsr r6, r6, #16\n\t"
  322. "lsr r7, r7, #16\n\t"
  323. "mul r7, r6\n\t"
  324. "add r3, r7\n\t"
  325. "adc r4, %[r]\n\t"
  326. "adc r5, %[r]\n\t"
  327. "add r3, r7\n\t"
  328. "adc r4, %[r]\n\t"
  329. "adc r5, %[r]\n\t"
  330. "ldr r7, [r2]\n\t"
  331. "lsr r7, r7, #16\n\t"
  332. "mul r6, r7\n\t"
  333. "lsr r7, r6, #16\n\t"
  334. "lsl r6, r6, #16\n\t"
  335. "add r3, r6\n\t"
  336. "adc r4, r7\n\t"
  337. "adc r5, %[r]\n\t"
  338. "add r3, r6\n\t"
  339. "adc r4, r7\n\t"
  340. "adc r5, %[r]\n\t"
  341. "ldr r6, [%[a]]\n\t"
  342. "ldr r7, [r2]\n\t"
  343. "lsr r6, r6, #16\n\t"
  344. "lsr r7, r7, #16\n\t"
  345. "mul r7, r6\n\t"
  346. "add r4, r7\n\t"
  347. "adc r5, %[r]\n\t"
  348. "add r4, r7\n\t"
  349. "adc r5, %[r]\n\t"
  350. "ldr r7, [r2]\n\t"
  351. "lsl r7, r7, #16\n\t"
  352. "lsr r7, r7, #16\n\t"
  353. "mul r6, r7\n\t"
  354. "lsr r7, r6, #16\n\t"
  355. "lsl r6, r6, #16\n\t"
  356. "add r3, r6\n\t"
  357. "adc r4, r7\n\t"
  358. "adc r5, %[r]\n\t"
  359. "add r3, r6\n\t"
  360. "adc r4, r7\n\t"
  361. "adc r5, %[r]\n\t"
  362. "# Multiply * 2: Done\n\t"
  363. "bal 5f\n\t"
  364. "\n4:\n\t"
  365. "# Square: Start\n\t"
  366. "ldr r6, [%[a]]\n\t"
  367. "lsr r7, r6, #16\n\t"
  368. "lsl r6, r6, #16\n\t"
  369. "lsr r6, r6, #16\n\t"
  370. "mul r6, r6\n\t"
  371. "add r3, r6\n\t"
  372. "adc r4, %[r]\n\t"
  373. "adc r5, %[r]\n\t"
  374. "mul r7, r7\n\t"
  375. "add r4, r7\n\t"
  376. "adc r5, %[r]\n\t"
  377. "ldr r6, [%[a]]\n\t"
  378. "lsr r7, r6, #16\n\t"
  379. "lsl r6, r6, #16\n\t"
  380. "lsr r6, r6, #16\n\t"
  381. "mul r6, r7\n\t"
  382. "lsr r7, r6, #15\n\t"
  383. "lsl r6, r6, #17\n\t"
  384. "add r3, r6\n\t"
  385. "adc r4, r7\n\t"
  386. "adc r5, %[r]\n\t"
  387. "# Square: Done\n\t"
  388. "\n5:\n\t"
  389. "add %[a], #4\n\t"
  390. "sub r2, #4\n\t"
  391. "mov r6, #32\n\t"
  392. "add r6, r9\n\t"
  393. "cmp %[a], r6\n\t"
  394. "beq 3f\n\t"
  395. "cmp %[a], r2\n\t"
  396. "bgt 3f\n\t"
  397. "mov r7, r8\n\t"
  398. "add r7, r9\n\t"
  399. "cmp %[a], r7\n\t"
  400. "ble 2b\n\t"
  401. "\n3:\n\t"
  402. "mov %[r], r10\n\t"
  403. "mov r7, r8\n\t"
  404. "str r3, [%[r], r7]\n\t"
  405. "mov r3, r4\n\t"
  406. "mov r4, r5\n\t"
  407. "mov r5, #0\n\t"
  408. "add r7, #4\n\t"
  409. "mov r8, r7\n\t"
  410. "mov r6, #56\n\t"
  411. "cmp r7, r6\n\t"
  412. "ble 1b\n\t"
  413. "mov %[a], r9\n\t"
  414. "str r3, [%[r], r7]\n\t"
  415. "mov %[r], r11\n\t"
  416. "mov %[a], r10\n\t"
  417. "mov r3, #60\n\t"
  418. "\n4:\n\t"
  419. "ldr r6, [%[a], r3]\n\t"
  420. "str r6, [%[r], r3]\n\t"
  421. "sub r3, #4\n\t"
  422. "bge 4b\n\t"
  423. "mov r6, #64\n\t"
  424. "add sp, r6\n\t"
  425. :
  426. : [r] "r" (r), [a] "r" (a)
  427. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  428. );
  429. }
  430. /* Add b to a into r. (r = a + b)
  431. *
  432. * r A single precision integer.
  433. * a A single precision integer.
  434. * b A single precision integer.
  435. */
  436. SP_NOINLINE static sp_digit sp_2048_add_8(sp_digit* r, const sp_digit* a,
  437. const sp_digit* b)
  438. {
  439. sp_digit c = 0;
  440. __asm__ __volatile__ (
  441. "ldr r4, [%[a], #0]\n\t"
  442. "ldr r5, [%[b], #0]\n\t"
  443. "add r4, r5\n\t"
  444. "str r4, [%[r], #0]\n\t"
  445. "ldr r4, [%[a], #4]\n\t"
  446. "ldr r5, [%[b], #4]\n\t"
  447. "adc r4, r5\n\t"
  448. "str r4, [%[r], #4]\n\t"
  449. "ldr r4, [%[a], #8]\n\t"
  450. "ldr r5, [%[b], #8]\n\t"
  451. "adc r4, r5\n\t"
  452. "str r4, [%[r], #8]\n\t"
  453. "ldr r4, [%[a], #12]\n\t"
  454. "ldr r5, [%[b], #12]\n\t"
  455. "adc r4, r5\n\t"
  456. "str r4, [%[r], #12]\n\t"
  457. "ldr r4, [%[a], #16]\n\t"
  458. "ldr r5, [%[b], #16]\n\t"
  459. "adc r4, r5\n\t"
  460. "str r4, [%[r], #16]\n\t"
  461. "ldr r4, [%[a], #20]\n\t"
  462. "ldr r5, [%[b], #20]\n\t"
  463. "adc r4, r5\n\t"
  464. "str r4, [%[r], #20]\n\t"
  465. "ldr r4, [%[a], #24]\n\t"
  466. "ldr r5, [%[b], #24]\n\t"
  467. "adc r4, r5\n\t"
  468. "str r4, [%[r], #24]\n\t"
  469. "ldr r4, [%[a], #28]\n\t"
  470. "ldr r5, [%[b], #28]\n\t"
  471. "adc r4, r5\n\t"
  472. "str r4, [%[r], #28]\n\t"
  473. "mov %[c], #0\n\t"
  474. "adc %[c], %[c]\n\t"
  475. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  476. :
  477. : "memory", "r4", "r5"
  478. );
  479. return c;
  480. }
  481. /* Sub b from a into r. (r = a - b)
  482. *
  483. * r A single precision integer.
  484. * a A single precision integer.
  485. * b A single precision integer.
  486. */
  487. SP_NOINLINE static sp_digit sp_2048_sub_in_place_16(sp_digit* a,
  488. const sp_digit* b)
  489. {
  490. sp_digit c = 0;
  491. __asm__ __volatile__ (
  492. "ldr r3, [%[a], #0]\n\t"
  493. "ldr r4, [%[a], #4]\n\t"
  494. "ldr r5, [%[b], #0]\n\t"
  495. "ldr r6, [%[b], #4]\n\t"
  496. "sub r3, r5\n\t"
  497. "sbc r4, r6\n\t"
  498. "str r3, [%[a], #0]\n\t"
  499. "str r4, [%[a], #4]\n\t"
  500. "ldr r3, [%[a], #8]\n\t"
  501. "ldr r4, [%[a], #12]\n\t"
  502. "ldr r5, [%[b], #8]\n\t"
  503. "ldr r6, [%[b], #12]\n\t"
  504. "sbc r3, r5\n\t"
  505. "sbc r4, r6\n\t"
  506. "str r3, [%[a], #8]\n\t"
  507. "str r4, [%[a], #12]\n\t"
  508. "ldr r3, [%[a], #16]\n\t"
  509. "ldr r4, [%[a], #20]\n\t"
  510. "ldr r5, [%[b], #16]\n\t"
  511. "ldr r6, [%[b], #20]\n\t"
  512. "sbc r3, r5\n\t"
  513. "sbc r4, r6\n\t"
  514. "str r3, [%[a], #16]\n\t"
  515. "str r4, [%[a], #20]\n\t"
  516. "ldr r3, [%[a], #24]\n\t"
  517. "ldr r4, [%[a], #28]\n\t"
  518. "ldr r5, [%[b], #24]\n\t"
  519. "ldr r6, [%[b], #28]\n\t"
  520. "sbc r3, r5\n\t"
  521. "sbc r4, r6\n\t"
  522. "str r3, [%[a], #24]\n\t"
  523. "str r4, [%[a], #28]\n\t"
  524. "ldr r3, [%[a], #32]\n\t"
  525. "ldr r4, [%[a], #36]\n\t"
  526. "ldr r5, [%[b], #32]\n\t"
  527. "ldr r6, [%[b], #36]\n\t"
  528. "sbc r3, r5\n\t"
  529. "sbc r4, r6\n\t"
  530. "str r3, [%[a], #32]\n\t"
  531. "str r4, [%[a], #36]\n\t"
  532. "ldr r3, [%[a], #40]\n\t"
  533. "ldr r4, [%[a], #44]\n\t"
  534. "ldr r5, [%[b], #40]\n\t"
  535. "ldr r6, [%[b], #44]\n\t"
  536. "sbc r3, r5\n\t"
  537. "sbc r4, r6\n\t"
  538. "str r3, [%[a], #40]\n\t"
  539. "str r4, [%[a], #44]\n\t"
  540. "ldr r3, [%[a], #48]\n\t"
  541. "ldr r4, [%[a], #52]\n\t"
  542. "ldr r5, [%[b], #48]\n\t"
  543. "ldr r6, [%[b], #52]\n\t"
  544. "sbc r3, r5\n\t"
  545. "sbc r4, r6\n\t"
  546. "str r3, [%[a], #48]\n\t"
  547. "str r4, [%[a], #52]\n\t"
  548. "ldr r3, [%[a], #56]\n\t"
  549. "ldr r4, [%[a], #60]\n\t"
  550. "ldr r5, [%[b], #56]\n\t"
  551. "ldr r6, [%[b], #60]\n\t"
  552. "sbc r3, r5\n\t"
  553. "sbc r4, r6\n\t"
  554. "str r3, [%[a], #56]\n\t"
  555. "str r4, [%[a], #60]\n\t"
  556. "sbc %[c], %[c]\n\t"
  557. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  558. :
  559. : "memory", "r3", "r4", "r5", "r6"
  560. );
  561. return c;
  562. }
  563. /* Add b to a into r. (r = a + b)
  564. *
  565. * r A single precision integer.
  566. * a A single precision integer.
  567. * b A single precision integer.
  568. */
  569. SP_NOINLINE static sp_digit sp_2048_add_16(sp_digit* r, const sp_digit* a,
  570. const sp_digit* b)
  571. {
  572. sp_digit c = 0;
  573. __asm__ __volatile__ (
  574. "ldr r4, [%[a], #0]\n\t"
  575. "ldr r5, [%[b], #0]\n\t"
  576. "add r4, r5\n\t"
  577. "str r4, [%[r], #0]\n\t"
  578. "ldr r4, [%[a], #4]\n\t"
  579. "ldr r5, [%[b], #4]\n\t"
  580. "adc r4, r5\n\t"
  581. "str r4, [%[r], #4]\n\t"
  582. "ldr r4, [%[a], #8]\n\t"
  583. "ldr r5, [%[b], #8]\n\t"
  584. "adc r4, r5\n\t"
  585. "str r4, [%[r], #8]\n\t"
  586. "ldr r4, [%[a], #12]\n\t"
  587. "ldr r5, [%[b], #12]\n\t"
  588. "adc r4, r5\n\t"
  589. "str r4, [%[r], #12]\n\t"
  590. "ldr r4, [%[a], #16]\n\t"
  591. "ldr r5, [%[b], #16]\n\t"
  592. "adc r4, r5\n\t"
  593. "str r4, [%[r], #16]\n\t"
  594. "ldr r4, [%[a], #20]\n\t"
  595. "ldr r5, [%[b], #20]\n\t"
  596. "adc r4, r5\n\t"
  597. "str r4, [%[r], #20]\n\t"
  598. "ldr r4, [%[a], #24]\n\t"
  599. "ldr r5, [%[b], #24]\n\t"
  600. "adc r4, r5\n\t"
  601. "str r4, [%[r], #24]\n\t"
  602. "ldr r4, [%[a], #28]\n\t"
  603. "ldr r5, [%[b], #28]\n\t"
  604. "adc r4, r5\n\t"
  605. "str r4, [%[r], #28]\n\t"
  606. "ldr r4, [%[a], #32]\n\t"
  607. "ldr r5, [%[b], #32]\n\t"
  608. "adc r4, r5\n\t"
  609. "str r4, [%[r], #32]\n\t"
  610. "ldr r4, [%[a], #36]\n\t"
  611. "ldr r5, [%[b], #36]\n\t"
  612. "adc r4, r5\n\t"
  613. "str r4, [%[r], #36]\n\t"
  614. "ldr r4, [%[a], #40]\n\t"
  615. "ldr r5, [%[b], #40]\n\t"
  616. "adc r4, r5\n\t"
  617. "str r4, [%[r], #40]\n\t"
  618. "ldr r4, [%[a], #44]\n\t"
  619. "ldr r5, [%[b], #44]\n\t"
  620. "adc r4, r5\n\t"
  621. "str r4, [%[r], #44]\n\t"
  622. "ldr r4, [%[a], #48]\n\t"
  623. "ldr r5, [%[b], #48]\n\t"
  624. "adc r4, r5\n\t"
  625. "str r4, [%[r], #48]\n\t"
  626. "ldr r4, [%[a], #52]\n\t"
  627. "ldr r5, [%[b], #52]\n\t"
  628. "adc r4, r5\n\t"
  629. "str r4, [%[r], #52]\n\t"
  630. "ldr r4, [%[a], #56]\n\t"
  631. "ldr r5, [%[b], #56]\n\t"
  632. "adc r4, r5\n\t"
  633. "str r4, [%[r], #56]\n\t"
  634. "ldr r4, [%[a], #60]\n\t"
  635. "ldr r5, [%[b], #60]\n\t"
  636. "adc r4, r5\n\t"
  637. "str r4, [%[r], #60]\n\t"
  638. "mov %[c], #0\n\t"
  639. "adc %[c], %[c]\n\t"
  640. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  641. :
  642. : "memory", "r4", "r5"
  643. );
  644. return c;
  645. }
  646. /* AND m into each word of a and store in r.
  647. *
  648. * r A single precision integer.
  649. * a A single precision integer.
  650. * m Mask to AND against each digit.
  651. */
  652. static void sp_2048_mask_8(sp_digit* r, const sp_digit* a, sp_digit m)
  653. {
  654. #ifdef WOLFSSL_SP_SMALL
  655. int i;
  656. for (i=0; i<8; i++) {
  657. r[i] = a[i] & m;
  658. }
  659. #else
  660. r[0] = a[0] & m;
  661. r[1] = a[1] & m;
  662. r[2] = a[2] & m;
  663. r[3] = a[3] & m;
  664. r[4] = a[4] & m;
  665. r[5] = a[5] & m;
  666. r[6] = a[6] & m;
  667. r[7] = a[7] & m;
  668. #endif
  669. }
  670. /* Multiply a and b into r. (r = a * b)
  671. *
  672. * r A single precision integer.
  673. * a A single precision integer.
  674. * b A single precision integer.
  675. */
  676. SP_NOINLINE static void sp_2048_mul_16(sp_digit* r, const sp_digit* a,
  677. const sp_digit* b)
  678. {
  679. sp_digit* z0 = r;
  680. sp_digit z1[16];
  681. sp_digit a1[8];
  682. sp_digit b1[8];
  683. sp_digit z2[16];
  684. sp_digit u, ca, cb;
  685. ca = sp_2048_add_8(a1, a, &a[8]);
  686. cb = sp_2048_add_8(b1, b, &b[8]);
  687. u = ca & cb;
  688. sp_2048_mul_8(z1, a1, b1);
  689. sp_2048_mul_8(z2, &a[8], &b[8]);
  690. sp_2048_mul_8(z0, a, b);
  691. sp_2048_mask_8(r + 16, a1, 0 - cb);
  692. sp_2048_mask_8(b1, b1, 0 - ca);
  693. u += sp_2048_add_8(r + 16, r + 16, b1);
  694. u += sp_2048_sub_in_place_16(z1, z2);
  695. u += sp_2048_sub_in_place_16(z1, z0);
  696. u += sp_2048_add_16(r + 8, r + 8, z1);
  697. r[24] = u;
  698. XMEMSET(r + 24 + 1, 0, sizeof(sp_digit) * (8 - 1));
  699. (void)sp_2048_add_16(r + 16, r + 16, z2);
  700. }
  701. /* Square a and put result in r. (r = a * a)
  702. *
  703. * r A single precision integer.
  704. * a A single precision integer.
  705. */
  706. SP_NOINLINE static void sp_2048_sqr_16(sp_digit* r, const sp_digit* a)
  707. {
  708. sp_digit* z0 = r;
  709. sp_digit z2[16];
  710. sp_digit z1[16];
  711. sp_digit a1[8];
  712. sp_digit u;
  713. u = sp_2048_add_8(a1, a, &a[8]);
  714. sp_2048_sqr_8(z1, a1);
  715. sp_2048_sqr_8(z2, &a[8]);
  716. sp_2048_sqr_8(z0, a);
  717. sp_2048_mask_8(r + 16, a1, 0 - u);
  718. u += sp_2048_add_8(r + 16, r + 16, r + 16);
  719. u += sp_2048_sub_in_place_16(z1, z2);
  720. u += sp_2048_sub_in_place_16(z1, z0);
  721. u += sp_2048_add_16(r + 8, r + 8, z1);
  722. r[24] = u;
  723. XMEMSET(r + 24 + 1, 0, sizeof(sp_digit) * (8 - 1));
  724. (void)sp_2048_add_16(r + 16, r + 16, z2);
  725. }
  726. /* Sub b from a into r. (r = a - b)
  727. *
  728. * r A single precision integer.
  729. * a A single precision integer.
  730. * b A single precision integer.
  731. */
  732. SP_NOINLINE static sp_digit sp_2048_sub_in_place_32(sp_digit* a,
  733. const sp_digit* b)
  734. {
  735. sp_digit c = 0;
  736. __asm__ __volatile__ (
  737. "ldr r3, [%[a], #0]\n\t"
  738. "ldr r4, [%[a], #4]\n\t"
  739. "ldr r5, [%[b], #0]\n\t"
  740. "ldr r6, [%[b], #4]\n\t"
  741. "sub r3, r5\n\t"
  742. "sbc r4, r6\n\t"
  743. "str r3, [%[a], #0]\n\t"
  744. "str r4, [%[a], #4]\n\t"
  745. "ldr r3, [%[a], #8]\n\t"
  746. "ldr r4, [%[a], #12]\n\t"
  747. "ldr r5, [%[b], #8]\n\t"
  748. "ldr r6, [%[b], #12]\n\t"
  749. "sbc r3, r5\n\t"
  750. "sbc r4, r6\n\t"
  751. "str r3, [%[a], #8]\n\t"
  752. "str r4, [%[a], #12]\n\t"
  753. "ldr r3, [%[a], #16]\n\t"
  754. "ldr r4, [%[a], #20]\n\t"
  755. "ldr r5, [%[b], #16]\n\t"
  756. "ldr r6, [%[b], #20]\n\t"
  757. "sbc r3, r5\n\t"
  758. "sbc r4, r6\n\t"
  759. "str r3, [%[a], #16]\n\t"
  760. "str r4, [%[a], #20]\n\t"
  761. "ldr r3, [%[a], #24]\n\t"
  762. "ldr r4, [%[a], #28]\n\t"
  763. "ldr r5, [%[b], #24]\n\t"
  764. "ldr r6, [%[b], #28]\n\t"
  765. "sbc r3, r5\n\t"
  766. "sbc r4, r6\n\t"
  767. "str r3, [%[a], #24]\n\t"
  768. "str r4, [%[a], #28]\n\t"
  769. "ldr r3, [%[a], #32]\n\t"
  770. "ldr r4, [%[a], #36]\n\t"
  771. "ldr r5, [%[b], #32]\n\t"
  772. "ldr r6, [%[b], #36]\n\t"
  773. "sbc r3, r5\n\t"
  774. "sbc r4, r6\n\t"
  775. "str r3, [%[a], #32]\n\t"
  776. "str r4, [%[a], #36]\n\t"
  777. "ldr r3, [%[a], #40]\n\t"
  778. "ldr r4, [%[a], #44]\n\t"
  779. "ldr r5, [%[b], #40]\n\t"
  780. "ldr r6, [%[b], #44]\n\t"
  781. "sbc r3, r5\n\t"
  782. "sbc r4, r6\n\t"
  783. "str r3, [%[a], #40]\n\t"
  784. "str r4, [%[a], #44]\n\t"
  785. "ldr r3, [%[a], #48]\n\t"
  786. "ldr r4, [%[a], #52]\n\t"
  787. "ldr r5, [%[b], #48]\n\t"
  788. "ldr r6, [%[b], #52]\n\t"
  789. "sbc r3, r5\n\t"
  790. "sbc r4, r6\n\t"
  791. "str r3, [%[a], #48]\n\t"
  792. "str r4, [%[a], #52]\n\t"
  793. "ldr r3, [%[a], #56]\n\t"
  794. "ldr r4, [%[a], #60]\n\t"
  795. "ldr r5, [%[b], #56]\n\t"
  796. "ldr r6, [%[b], #60]\n\t"
  797. "sbc r3, r5\n\t"
  798. "sbc r4, r6\n\t"
  799. "str r3, [%[a], #56]\n\t"
  800. "str r4, [%[a], #60]\n\t"
  801. "ldr r3, [%[a], #64]\n\t"
  802. "ldr r4, [%[a], #68]\n\t"
  803. "ldr r5, [%[b], #64]\n\t"
  804. "ldr r6, [%[b], #68]\n\t"
  805. "sbc r3, r5\n\t"
  806. "sbc r4, r6\n\t"
  807. "str r3, [%[a], #64]\n\t"
  808. "str r4, [%[a], #68]\n\t"
  809. "ldr r3, [%[a], #72]\n\t"
  810. "ldr r4, [%[a], #76]\n\t"
  811. "ldr r5, [%[b], #72]\n\t"
  812. "ldr r6, [%[b], #76]\n\t"
  813. "sbc r3, r5\n\t"
  814. "sbc r4, r6\n\t"
  815. "str r3, [%[a], #72]\n\t"
  816. "str r4, [%[a], #76]\n\t"
  817. "ldr r3, [%[a], #80]\n\t"
  818. "ldr r4, [%[a], #84]\n\t"
  819. "ldr r5, [%[b], #80]\n\t"
  820. "ldr r6, [%[b], #84]\n\t"
  821. "sbc r3, r5\n\t"
  822. "sbc r4, r6\n\t"
  823. "str r3, [%[a], #80]\n\t"
  824. "str r4, [%[a], #84]\n\t"
  825. "ldr r3, [%[a], #88]\n\t"
  826. "ldr r4, [%[a], #92]\n\t"
  827. "ldr r5, [%[b], #88]\n\t"
  828. "ldr r6, [%[b], #92]\n\t"
  829. "sbc r3, r5\n\t"
  830. "sbc r4, r6\n\t"
  831. "str r3, [%[a], #88]\n\t"
  832. "str r4, [%[a], #92]\n\t"
  833. "ldr r3, [%[a], #96]\n\t"
  834. "ldr r4, [%[a], #100]\n\t"
  835. "ldr r5, [%[b], #96]\n\t"
  836. "ldr r6, [%[b], #100]\n\t"
  837. "sbc r3, r5\n\t"
  838. "sbc r4, r6\n\t"
  839. "str r3, [%[a], #96]\n\t"
  840. "str r4, [%[a], #100]\n\t"
  841. "ldr r3, [%[a], #104]\n\t"
  842. "ldr r4, [%[a], #108]\n\t"
  843. "ldr r5, [%[b], #104]\n\t"
  844. "ldr r6, [%[b], #108]\n\t"
  845. "sbc r3, r5\n\t"
  846. "sbc r4, r6\n\t"
  847. "str r3, [%[a], #104]\n\t"
  848. "str r4, [%[a], #108]\n\t"
  849. "ldr r3, [%[a], #112]\n\t"
  850. "ldr r4, [%[a], #116]\n\t"
  851. "ldr r5, [%[b], #112]\n\t"
  852. "ldr r6, [%[b], #116]\n\t"
  853. "sbc r3, r5\n\t"
  854. "sbc r4, r6\n\t"
  855. "str r3, [%[a], #112]\n\t"
  856. "str r4, [%[a], #116]\n\t"
  857. "ldr r3, [%[a], #120]\n\t"
  858. "ldr r4, [%[a], #124]\n\t"
  859. "ldr r5, [%[b], #120]\n\t"
  860. "ldr r6, [%[b], #124]\n\t"
  861. "sbc r3, r5\n\t"
  862. "sbc r4, r6\n\t"
  863. "str r3, [%[a], #120]\n\t"
  864. "str r4, [%[a], #124]\n\t"
  865. "sbc %[c], %[c]\n\t"
  866. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  867. :
  868. : "memory", "r3", "r4", "r5", "r6"
  869. );
  870. return c;
  871. }
  872. /* Add b to a into r. (r = a + b)
  873. *
  874. * r A single precision integer.
  875. * a A single precision integer.
  876. * b A single precision integer.
  877. */
  878. SP_NOINLINE static sp_digit sp_2048_add_32(sp_digit* r, const sp_digit* a,
  879. const sp_digit* b)
  880. {
  881. sp_digit c = 0;
  882. __asm__ __volatile__ (
  883. "ldr r4, [%[a], #0]\n\t"
  884. "ldr r5, [%[b], #0]\n\t"
  885. "add r4, r5\n\t"
  886. "str r4, [%[r], #0]\n\t"
  887. "ldr r4, [%[a], #4]\n\t"
  888. "ldr r5, [%[b], #4]\n\t"
  889. "adc r4, r5\n\t"
  890. "str r4, [%[r], #4]\n\t"
  891. "ldr r4, [%[a], #8]\n\t"
  892. "ldr r5, [%[b], #8]\n\t"
  893. "adc r4, r5\n\t"
  894. "str r4, [%[r], #8]\n\t"
  895. "ldr r4, [%[a], #12]\n\t"
  896. "ldr r5, [%[b], #12]\n\t"
  897. "adc r4, r5\n\t"
  898. "str r4, [%[r], #12]\n\t"
  899. "ldr r4, [%[a], #16]\n\t"
  900. "ldr r5, [%[b], #16]\n\t"
  901. "adc r4, r5\n\t"
  902. "str r4, [%[r], #16]\n\t"
  903. "ldr r4, [%[a], #20]\n\t"
  904. "ldr r5, [%[b], #20]\n\t"
  905. "adc r4, r5\n\t"
  906. "str r4, [%[r], #20]\n\t"
  907. "ldr r4, [%[a], #24]\n\t"
  908. "ldr r5, [%[b], #24]\n\t"
  909. "adc r4, r5\n\t"
  910. "str r4, [%[r], #24]\n\t"
  911. "ldr r4, [%[a], #28]\n\t"
  912. "ldr r5, [%[b], #28]\n\t"
  913. "adc r4, r5\n\t"
  914. "str r4, [%[r], #28]\n\t"
  915. "ldr r4, [%[a], #32]\n\t"
  916. "ldr r5, [%[b], #32]\n\t"
  917. "adc r4, r5\n\t"
  918. "str r4, [%[r], #32]\n\t"
  919. "ldr r4, [%[a], #36]\n\t"
  920. "ldr r5, [%[b], #36]\n\t"
  921. "adc r4, r5\n\t"
  922. "str r4, [%[r], #36]\n\t"
  923. "ldr r4, [%[a], #40]\n\t"
  924. "ldr r5, [%[b], #40]\n\t"
  925. "adc r4, r5\n\t"
  926. "str r4, [%[r], #40]\n\t"
  927. "ldr r4, [%[a], #44]\n\t"
  928. "ldr r5, [%[b], #44]\n\t"
  929. "adc r4, r5\n\t"
  930. "str r4, [%[r], #44]\n\t"
  931. "ldr r4, [%[a], #48]\n\t"
  932. "ldr r5, [%[b], #48]\n\t"
  933. "adc r4, r5\n\t"
  934. "str r4, [%[r], #48]\n\t"
  935. "ldr r4, [%[a], #52]\n\t"
  936. "ldr r5, [%[b], #52]\n\t"
  937. "adc r4, r5\n\t"
  938. "str r4, [%[r], #52]\n\t"
  939. "ldr r4, [%[a], #56]\n\t"
  940. "ldr r5, [%[b], #56]\n\t"
  941. "adc r4, r5\n\t"
  942. "str r4, [%[r], #56]\n\t"
  943. "ldr r4, [%[a], #60]\n\t"
  944. "ldr r5, [%[b], #60]\n\t"
  945. "adc r4, r5\n\t"
  946. "str r4, [%[r], #60]\n\t"
  947. "ldr r4, [%[a], #64]\n\t"
  948. "ldr r5, [%[b], #64]\n\t"
  949. "adc r4, r5\n\t"
  950. "str r4, [%[r], #64]\n\t"
  951. "ldr r4, [%[a], #68]\n\t"
  952. "ldr r5, [%[b], #68]\n\t"
  953. "adc r4, r5\n\t"
  954. "str r4, [%[r], #68]\n\t"
  955. "ldr r4, [%[a], #72]\n\t"
  956. "ldr r5, [%[b], #72]\n\t"
  957. "adc r4, r5\n\t"
  958. "str r4, [%[r], #72]\n\t"
  959. "ldr r4, [%[a], #76]\n\t"
  960. "ldr r5, [%[b], #76]\n\t"
  961. "adc r4, r5\n\t"
  962. "str r4, [%[r], #76]\n\t"
  963. "ldr r4, [%[a], #80]\n\t"
  964. "ldr r5, [%[b], #80]\n\t"
  965. "adc r4, r5\n\t"
  966. "str r4, [%[r], #80]\n\t"
  967. "ldr r4, [%[a], #84]\n\t"
  968. "ldr r5, [%[b], #84]\n\t"
  969. "adc r4, r5\n\t"
  970. "str r4, [%[r], #84]\n\t"
  971. "ldr r4, [%[a], #88]\n\t"
  972. "ldr r5, [%[b], #88]\n\t"
  973. "adc r4, r5\n\t"
  974. "str r4, [%[r], #88]\n\t"
  975. "ldr r4, [%[a], #92]\n\t"
  976. "ldr r5, [%[b], #92]\n\t"
  977. "adc r4, r5\n\t"
  978. "str r4, [%[r], #92]\n\t"
  979. "ldr r4, [%[a], #96]\n\t"
  980. "ldr r5, [%[b], #96]\n\t"
  981. "adc r4, r5\n\t"
  982. "str r4, [%[r], #96]\n\t"
  983. "ldr r4, [%[a], #100]\n\t"
  984. "ldr r5, [%[b], #100]\n\t"
  985. "adc r4, r5\n\t"
  986. "str r4, [%[r], #100]\n\t"
  987. "ldr r4, [%[a], #104]\n\t"
  988. "ldr r5, [%[b], #104]\n\t"
  989. "adc r4, r5\n\t"
  990. "str r4, [%[r], #104]\n\t"
  991. "ldr r4, [%[a], #108]\n\t"
  992. "ldr r5, [%[b], #108]\n\t"
  993. "adc r4, r5\n\t"
  994. "str r4, [%[r], #108]\n\t"
  995. "ldr r4, [%[a], #112]\n\t"
  996. "ldr r5, [%[b], #112]\n\t"
  997. "adc r4, r5\n\t"
  998. "str r4, [%[r], #112]\n\t"
  999. "ldr r4, [%[a], #116]\n\t"
  1000. "ldr r5, [%[b], #116]\n\t"
  1001. "adc r4, r5\n\t"
  1002. "str r4, [%[r], #116]\n\t"
  1003. "ldr r4, [%[a], #120]\n\t"
  1004. "ldr r5, [%[b], #120]\n\t"
  1005. "adc r4, r5\n\t"
  1006. "str r4, [%[r], #120]\n\t"
  1007. "ldr r4, [%[a], #124]\n\t"
  1008. "ldr r5, [%[b], #124]\n\t"
  1009. "adc r4, r5\n\t"
  1010. "str r4, [%[r], #124]\n\t"
  1011. "mov %[c], #0\n\t"
  1012. "adc %[c], %[c]\n\t"
  1013. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  1014. :
  1015. : "memory", "r4", "r5"
  1016. );
  1017. return c;
  1018. }
  1019. /* AND m into each word of a and store in r.
  1020. *
  1021. * r A single precision integer.
  1022. * a A single precision integer.
  1023. * m Mask to AND against each digit.
  1024. */
  1025. static void sp_2048_mask_16(sp_digit* r, const sp_digit* a, sp_digit m)
  1026. {
  1027. #ifdef WOLFSSL_SP_SMALL
  1028. int i;
  1029. for (i=0; i<16; i++) {
  1030. r[i] = a[i] & m;
  1031. }
  1032. #else
  1033. int i;
  1034. for (i = 0; i < 16; i += 8) {
  1035. r[i+0] = a[i+0] & m;
  1036. r[i+1] = a[i+1] & m;
  1037. r[i+2] = a[i+2] & m;
  1038. r[i+3] = a[i+3] & m;
  1039. r[i+4] = a[i+4] & m;
  1040. r[i+5] = a[i+5] & m;
  1041. r[i+6] = a[i+6] & m;
  1042. r[i+7] = a[i+7] & m;
  1043. }
  1044. #endif
  1045. }
  1046. /* Multiply a and b into r. (r = a * b)
  1047. *
  1048. * r A single precision integer.
  1049. * a A single precision integer.
  1050. * b A single precision integer.
  1051. */
  1052. SP_NOINLINE static void sp_2048_mul_32(sp_digit* r, const sp_digit* a,
  1053. const sp_digit* b)
  1054. {
  1055. sp_digit* z0 = r;
  1056. sp_digit z1[32];
  1057. sp_digit a1[16];
  1058. sp_digit b1[16];
  1059. sp_digit z2[32];
  1060. sp_digit u, ca, cb;
  1061. ca = sp_2048_add_16(a1, a, &a[16]);
  1062. cb = sp_2048_add_16(b1, b, &b[16]);
  1063. u = ca & cb;
  1064. sp_2048_mul_16(z1, a1, b1);
  1065. sp_2048_mul_16(z2, &a[16], &b[16]);
  1066. sp_2048_mul_16(z0, a, b);
  1067. sp_2048_mask_16(r + 32, a1, 0 - cb);
  1068. sp_2048_mask_16(b1, b1, 0 - ca);
  1069. u += sp_2048_add_16(r + 32, r + 32, b1);
  1070. u += sp_2048_sub_in_place_32(z1, z2);
  1071. u += sp_2048_sub_in_place_32(z1, z0);
  1072. u += sp_2048_add_32(r + 16, r + 16, z1);
  1073. r[48] = u;
  1074. XMEMSET(r + 48 + 1, 0, sizeof(sp_digit) * (16 - 1));
  1075. (void)sp_2048_add_32(r + 32, r + 32, z2);
  1076. }
  1077. /* Square a and put result in r. (r = a * a)
  1078. *
  1079. * r A single precision integer.
  1080. * a A single precision integer.
  1081. */
  1082. SP_NOINLINE static void sp_2048_sqr_32(sp_digit* r, const sp_digit* a)
  1083. {
  1084. sp_digit* z0 = r;
  1085. sp_digit z2[32];
  1086. sp_digit z1[32];
  1087. sp_digit a1[16];
  1088. sp_digit u;
  1089. u = sp_2048_add_16(a1, a, &a[16]);
  1090. sp_2048_sqr_16(z1, a1);
  1091. sp_2048_sqr_16(z2, &a[16]);
  1092. sp_2048_sqr_16(z0, a);
  1093. sp_2048_mask_16(r + 32, a1, 0 - u);
  1094. u += sp_2048_add_16(r + 32, r + 32, r + 32);
  1095. u += sp_2048_sub_in_place_32(z1, z2);
  1096. u += sp_2048_sub_in_place_32(z1, z0);
  1097. u += sp_2048_add_32(r + 16, r + 16, z1);
  1098. r[48] = u;
  1099. XMEMSET(r + 48 + 1, 0, sizeof(sp_digit) * (16 - 1));
  1100. (void)sp_2048_add_32(r + 32, r + 32, z2);
  1101. }
  1102. /* Sub b from a into r. (r = a - b)
  1103. *
  1104. * r A single precision integer.
  1105. * a A single precision integer.
  1106. * b A single precision integer.
  1107. */
  1108. SP_NOINLINE static sp_digit sp_2048_sub_in_place_64(sp_digit* a,
  1109. const sp_digit* b)
  1110. {
  1111. sp_digit c = 0;
  1112. __asm__ __volatile__ (
  1113. "ldr r3, [%[a], #0]\n\t"
  1114. "ldr r4, [%[a], #4]\n\t"
  1115. "ldr r5, [%[b], #0]\n\t"
  1116. "ldr r6, [%[b], #4]\n\t"
  1117. "sub r3, r5\n\t"
  1118. "sbc r4, r6\n\t"
  1119. "str r3, [%[a], #0]\n\t"
  1120. "str r4, [%[a], #4]\n\t"
  1121. "ldr r3, [%[a], #8]\n\t"
  1122. "ldr r4, [%[a], #12]\n\t"
  1123. "ldr r5, [%[b], #8]\n\t"
  1124. "ldr r6, [%[b], #12]\n\t"
  1125. "sbc r3, r5\n\t"
  1126. "sbc r4, r6\n\t"
  1127. "str r3, [%[a], #8]\n\t"
  1128. "str r4, [%[a], #12]\n\t"
  1129. "ldr r3, [%[a], #16]\n\t"
  1130. "ldr r4, [%[a], #20]\n\t"
  1131. "ldr r5, [%[b], #16]\n\t"
  1132. "ldr r6, [%[b], #20]\n\t"
  1133. "sbc r3, r5\n\t"
  1134. "sbc r4, r6\n\t"
  1135. "str r3, [%[a], #16]\n\t"
  1136. "str r4, [%[a], #20]\n\t"
  1137. "ldr r3, [%[a], #24]\n\t"
  1138. "ldr r4, [%[a], #28]\n\t"
  1139. "ldr r5, [%[b], #24]\n\t"
  1140. "ldr r6, [%[b], #28]\n\t"
  1141. "sbc r3, r5\n\t"
  1142. "sbc r4, r6\n\t"
  1143. "str r3, [%[a], #24]\n\t"
  1144. "str r4, [%[a], #28]\n\t"
  1145. "ldr r3, [%[a], #32]\n\t"
  1146. "ldr r4, [%[a], #36]\n\t"
  1147. "ldr r5, [%[b], #32]\n\t"
  1148. "ldr r6, [%[b], #36]\n\t"
  1149. "sbc r3, r5\n\t"
  1150. "sbc r4, r6\n\t"
  1151. "str r3, [%[a], #32]\n\t"
  1152. "str r4, [%[a], #36]\n\t"
  1153. "ldr r3, [%[a], #40]\n\t"
  1154. "ldr r4, [%[a], #44]\n\t"
  1155. "ldr r5, [%[b], #40]\n\t"
  1156. "ldr r6, [%[b], #44]\n\t"
  1157. "sbc r3, r5\n\t"
  1158. "sbc r4, r6\n\t"
  1159. "str r3, [%[a], #40]\n\t"
  1160. "str r4, [%[a], #44]\n\t"
  1161. "ldr r3, [%[a], #48]\n\t"
  1162. "ldr r4, [%[a], #52]\n\t"
  1163. "ldr r5, [%[b], #48]\n\t"
  1164. "ldr r6, [%[b], #52]\n\t"
  1165. "sbc r3, r5\n\t"
  1166. "sbc r4, r6\n\t"
  1167. "str r3, [%[a], #48]\n\t"
  1168. "str r4, [%[a], #52]\n\t"
  1169. "ldr r3, [%[a], #56]\n\t"
  1170. "ldr r4, [%[a], #60]\n\t"
  1171. "ldr r5, [%[b], #56]\n\t"
  1172. "ldr r6, [%[b], #60]\n\t"
  1173. "sbc r3, r5\n\t"
  1174. "sbc r4, r6\n\t"
  1175. "str r3, [%[a], #56]\n\t"
  1176. "str r4, [%[a], #60]\n\t"
  1177. "ldr r3, [%[a], #64]\n\t"
  1178. "ldr r4, [%[a], #68]\n\t"
  1179. "ldr r5, [%[b], #64]\n\t"
  1180. "ldr r6, [%[b], #68]\n\t"
  1181. "sbc r3, r5\n\t"
  1182. "sbc r4, r6\n\t"
  1183. "str r3, [%[a], #64]\n\t"
  1184. "str r4, [%[a], #68]\n\t"
  1185. "ldr r3, [%[a], #72]\n\t"
  1186. "ldr r4, [%[a], #76]\n\t"
  1187. "ldr r5, [%[b], #72]\n\t"
  1188. "ldr r6, [%[b], #76]\n\t"
  1189. "sbc r3, r5\n\t"
  1190. "sbc r4, r6\n\t"
  1191. "str r3, [%[a], #72]\n\t"
  1192. "str r4, [%[a], #76]\n\t"
  1193. "ldr r3, [%[a], #80]\n\t"
  1194. "ldr r4, [%[a], #84]\n\t"
  1195. "ldr r5, [%[b], #80]\n\t"
  1196. "ldr r6, [%[b], #84]\n\t"
  1197. "sbc r3, r5\n\t"
  1198. "sbc r4, r6\n\t"
  1199. "str r3, [%[a], #80]\n\t"
  1200. "str r4, [%[a], #84]\n\t"
  1201. "ldr r3, [%[a], #88]\n\t"
  1202. "ldr r4, [%[a], #92]\n\t"
  1203. "ldr r5, [%[b], #88]\n\t"
  1204. "ldr r6, [%[b], #92]\n\t"
  1205. "sbc r3, r5\n\t"
  1206. "sbc r4, r6\n\t"
  1207. "str r3, [%[a], #88]\n\t"
  1208. "str r4, [%[a], #92]\n\t"
  1209. "ldr r3, [%[a], #96]\n\t"
  1210. "ldr r4, [%[a], #100]\n\t"
  1211. "ldr r5, [%[b], #96]\n\t"
  1212. "ldr r6, [%[b], #100]\n\t"
  1213. "sbc r3, r5\n\t"
  1214. "sbc r4, r6\n\t"
  1215. "str r3, [%[a], #96]\n\t"
  1216. "str r4, [%[a], #100]\n\t"
  1217. "ldr r3, [%[a], #104]\n\t"
  1218. "ldr r4, [%[a], #108]\n\t"
  1219. "ldr r5, [%[b], #104]\n\t"
  1220. "ldr r6, [%[b], #108]\n\t"
  1221. "sbc r3, r5\n\t"
  1222. "sbc r4, r6\n\t"
  1223. "str r3, [%[a], #104]\n\t"
  1224. "str r4, [%[a], #108]\n\t"
  1225. "ldr r3, [%[a], #112]\n\t"
  1226. "ldr r4, [%[a], #116]\n\t"
  1227. "ldr r5, [%[b], #112]\n\t"
  1228. "ldr r6, [%[b], #116]\n\t"
  1229. "sbc r3, r5\n\t"
  1230. "sbc r4, r6\n\t"
  1231. "str r3, [%[a], #112]\n\t"
  1232. "str r4, [%[a], #116]\n\t"
  1233. "ldr r3, [%[a], #120]\n\t"
  1234. "ldr r4, [%[a], #124]\n\t"
  1235. "ldr r5, [%[b], #120]\n\t"
  1236. "ldr r6, [%[b], #124]\n\t"
  1237. "sbc r3, r5\n\t"
  1238. "sbc r4, r6\n\t"
  1239. "str r3, [%[a], #120]\n\t"
  1240. "str r4, [%[a], #124]\n\t"
  1241. "sbc %[c], %[c]\n\t"
  1242. "add %[a], #0x80\n\t"
  1243. "add %[b], #0x80\n\t"
  1244. "mov r5, #0\n\t"
  1245. "sub r5, %[c]\n\t"
  1246. "ldr r3, [%[a], #0]\n\t"
  1247. "ldr r4, [%[a], #4]\n\t"
  1248. "ldr r5, [%[b], #0]\n\t"
  1249. "ldr r6, [%[b], #4]\n\t"
  1250. "sbc r3, r5\n\t"
  1251. "sbc r4, r6\n\t"
  1252. "str r3, [%[a], #0]\n\t"
  1253. "str r4, [%[a], #4]\n\t"
  1254. "ldr r3, [%[a], #8]\n\t"
  1255. "ldr r4, [%[a], #12]\n\t"
  1256. "ldr r5, [%[b], #8]\n\t"
  1257. "ldr r6, [%[b], #12]\n\t"
  1258. "sbc r3, r5\n\t"
  1259. "sbc r4, r6\n\t"
  1260. "str r3, [%[a], #8]\n\t"
  1261. "str r4, [%[a], #12]\n\t"
  1262. "ldr r3, [%[a], #16]\n\t"
  1263. "ldr r4, [%[a], #20]\n\t"
  1264. "ldr r5, [%[b], #16]\n\t"
  1265. "ldr r6, [%[b], #20]\n\t"
  1266. "sbc r3, r5\n\t"
  1267. "sbc r4, r6\n\t"
  1268. "str r3, [%[a], #16]\n\t"
  1269. "str r4, [%[a], #20]\n\t"
  1270. "ldr r3, [%[a], #24]\n\t"
  1271. "ldr r4, [%[a], #28]\n\t"
  1272. "ldr r5, [%[b], #24]\n\t"
  1273. "ldr r6, [%[b], #28]\n\t"
  1274. "sbc r3, r5\n\t"
  1275. "sbc r4, r6\n\t"
  1276. "str r3, [%[a], #24]\n\t"
  1277. "str r4, [%[a], #28]\n\t"
  1278. "ldr r3, [%[a], #32]\n\t"
  1279. "ldr r4, [%[a], #36]\n\t"
  1280. "ldr r5, [%[b], #32]\n\t"
  1281. "ldr r6, [%[b], #36]\n\t"
  1282. "sbc r3, r5\n\t"
  1283. "sbc r4, r6\n\t"
  1284. "str r3, [%[a], #32]\n\t"
  1285. "str r4, [%[a], #36]\n\t"
  1286. "ldr r3, [%[a], #40]\n\t"
  1287. "ldr r4, [%[a], #44]\n\t"
  1288. "ldr r5, [%[b], #40]\n\t"
  1289. "ldr r6, [%[b], #44]\n\t"
  1290. "sbc r3, r5\n\t"
  1291. "sbc r4, r6\n\t"
  1292. "str r3, [%[a], #40]\n\t"
  1293. "str r4, [%[a], #44]\n\t"
  1294. "ldr r3, [%[a], #48]\n\t"
  1295. "ldr r4, [%[a], #52]\n\t"
  1296. "ldr r5, [%[b], #48]\n\t"
  1297. "ldr r6, [%[b], #52]\n\t"
  1298. "sbc r3, r5\n\t"
  1299. "sbc r4, r6\n\t"
  1300. "str r3, [%[a], #48]\n\t"
  1301. "str r4, [%[a], #52]\n\t"
  1302. "ldr r3, [%[a], #56]\n\t"
  1303. "ldr r4, [%[a], #60]\n\t"
  1304. "ldr r5, [%[b], #56]\n\t"
  1305. "ldr r6, [%[b], #60]\n\t"
  1306. "sbc r3, r5\n\t"
  1307. "sbc r4, r6\n\t"
  1308. "str r3, [%[a], #56]\n\t"
  1309. "str r4, [%[a], #60]\n\t"
  1310. "ldr r3, [%[a], #64]\n\t"
  1311. "ldr r4, [%[a], #68]\n\t"
  1312. "ldr r5, [%[b], #64]\n\t"
  1313. "ldr r6, [%[b], #68]\n\t"
  1314. "sbc r3, r5\n\t"
  1315. "sbc r4, r6\n\t"
  1316. "str r3, [%[a], #64]\n\t"
  1317. "str r4, [%[a], #68]\n\t"
  1318. "ldr r3, [%[a], #72]\n\t"
  1319. "ldr r4, [%[a], #76]\n\t"
  1320. "ldr r5, [%[b], #72]\n\t"
  1321. "ldr r6, [%[b], #76]\n\t"
  1322. "sbc r3, r5\n\t"
  1323. "sbc r4, r6\n\t"
  1324. "str r3, [%[a], #72]\n\t"
  1325. "str r4, [%[a], #76]\n\t"
  1326. "ldr r3, [%[a], #80]\n\t"
  1327. "ldr r4, [%[a], #84]\n\t"
  1328. "ldr r5, [%[b], #80]\n\t"
  1329. "ldr r6, [%[b], #84]\n\t"
  1330. "sbc r3, r5\n\t"
  1331. "sbc r4, r6\n\t"
  1332. "str r3, [%[a], #80]\n\t"
  1333. "str r4, [%[a], #84]\n\t"
  1334. "ldr r3, [%[a], #88]\n\t"
  1335. "ldr r4, [%[a], #92]\n\t"
  1336. "ldr r5, [%[b], #88]\n\t"
  1337. "ldr r6, [%[b], #92]\n\t"
  1338. "sbc r3, r5\n\t"
  1339. "sbc r4, r6\n\t"
  1340. "str r3, [%[a], #88]\n\t"
  1341. "str r4, [%[a], #92]\n\t"
  1342. "ldr r3, [%[a], #96]\n\t"
  1343. "ldr r4, [%[a], #100]\n\t"
  1344. "ldr r5, [%[b], #96]\n\t"
  1345. "ldr r6, [%[b], #100]\n\t"
  1346. "sbc r3, r5\n\t"
  1347. "sbc r4, r6\n\t"
  1348. "str r3, [%[a], #96]\n\t"
  1349. "str r4, [%[a], #100]\n\t"
  1350. "ldr r3, [%[a], #104]\n\t"
  1351. "ldr r4, [%[a], #108]\n\t"
  1352. "ldr r5, [%[b], #104]\n\t"
  1353. "ldr r6, [%[b], #108]\n\t"
  1354. "sbc r3, r5\n\t"
  1355. "sbc r4, r6\n\t"
  1356. "str r3, [%[a], #104]\n\t"
  1357. "str r4, [%[a], #108]\n\t"
  1358. "ldr r3, [%[a], #112]\n\t"
  1359. "ldr r4, [%[a], #116]\n\t"
  1360. "ldr r5, [%[b], #112]\n\t"
  1361. "ldr r6, [%[b], #116]\n\t"
  1362. "sbc r3, r5\n\t"
  1363. "sbc r4, r6\n\t"
  1364. "str r3, [%[a], #112]\n\t"
  1365. "str r4, [%[a], #116]\n\t"
  1366. "ldr r3, [%[a], #120]\n\t"
  1367. "ldr r4, [%[a], #124]\n\t"
  1368. "ldr r5, [%[b], #120]\n\t"
  1369. "ldr r6, [%[b], #124]\n\t"
  1370. "sbc r3, r5\n\t"
  1371. "sbc r4, r6\n\t"
  1372. "str r3, [%[a], #120]\n\t"
  1373. "str r4, [%[a], #124]\n\t"
  1374. "sbc %[c], %[c]\n\t"
  1375. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  1376. :
  1377. : "memory", "r3", "r4", "r5", "r6"
  1378. );
  1379. return c;
  1380. }
  1381. /* Add b to a into r. (r = a + b)
  1382. *
  1383. * r A single precision integer.
  1384. * a A single precision integer.
  1385. * b A single precision integer.
  1386. */
  1387. SP_NOINLINE static sp_digit sp_2048_add_64(sp_digit* r, const sp_digit* a,
  1388. const sp_digit* b)
  1389. {
  1390. sp_digit c = 0;
  1391. __asm__ __volatile__ (
  1392. "mov r7, #0\n\t"
  1393. "mvn r7, r7\n\t"
  1394. "ldr r4, [%[a], #0]\n\t"
  1395. "ldr r5, [%[b], #0]\n\t"
  1396. "add r4, r5\n\t"
  1397. "str r4, [%[r], #0]\n\t"
  1398. "ldr r4, [%[a], #4]\n\t"
  1399. "ldr r5, [%[b], #4]\n\t"
  1400. "adc r4, r5\n\t"
  1401. "str r4, [%[r], #4]\n\t"
  1402. "ldr r4, [%[a], #8]\n\t"
  1403. "ldr r5, [%[b], #8]\n\t"
  1404. "adc r4, r5\n\t"
  1405. "str r4, [%[r], #8]\n\t"
  1406. "ldr r4, [%[a], #12]\n\t"
  1407. "ldr r5, [%[b], #12]\n\t"
  1408. "adc r4, r5\n\t"
  1409. "str r4, [%[r], #12]\n\t"
  1410. "ldr r4, [%[a], #16]\n\t"
  1411. "ldr r5, [%[b], #16]\n\t"
  1412. "adc r4, r5\n\t"
  1413. "str r4, [%[r], #16]\n\t"
  1414. "ldr r4, [%[a], #20]\n\t"
  1415. "ldr r5, [%[b], #20]\n\t"
  1416. "adc r4, r5\n\t"
  1417. "str r4, [%[r], #20]\n\t"
  1418. "ldr r4, [%[a], #24]\n\t"
  1419. "ldr r5, [%[b], #24]\n\t"
  1420. "adc r4, r5\n\t"
  1421. "str r4, [%[r], #24]\n\t"
  1422. "ldr r4, [%[a], #28]\n\t"
  1423. "ldr r5, [%[b], #28]\n\t"
  1424. "adc r4, r5\n\t"
  1425. "str r4, [%[r], #28]\n\t"
  1426. "ldr r4, [%[a], #32]\n\t"
  1427. "ldr r5, [%[b], #32]\n\t"
  1428. "adc r4, r5\n\t"
  1429. "str r4, [%[r], #32]\n\t"
  1430. "ldr r4, [%[a], #36]\n\t"
  1431. "ldr r5, [%[b], #36]\n\t"
  1432. "adc r4, r5\n\t"
  1433. "str r4, [%[r], #36]\n\t"
  1434. "ldr r4, [%[a], #40]\n\t"
  1435. "ldr r5, [%[b], #40]\n\t"
  1436. "adc r4, r5\n\t"
  1437. "str r4, [%[r], #40]\n\t"
  1438. "ldr r4, [%[a], #44]\n\t"
  1439. "ldr r5, [%[b], #44]\n\t"
  1440. "adc r4, r5\n\t"
  1441. "str r4, [%[r], #44]\n\t"
  1442. "ldr r4, [%[a], #48]\n\t"
  1443. "ldr r5, [%[b], #48]\n\t"
  1444. "adc r4, r5\n\t"
  1445. "str r4, [%[r], #48]\n\t"
  1446. "ldr r4, [%[a], #52]\n\t"
  1447. "ldr r5, [%[b], #52]\n\t"
  1448. "adc r4, r5\n\t"
  1449. "str r4, [%[r], #52]\n\t"
  1450. "ldr r4, [%[a], #56]\n\t"
  1451. "ldr r5, [%[b], #56]\n\t"
  1452. "adc r4, r5\n\t"
  1453. "str r4, [%[r], #56]\n\t"
  1454. "ldr r4, [%[a], #60]\n\t"
  1455. "ldr r5, [%[b], #60]\n\t"
  1456. "adc r4, r5\n\t"
  1457. "str r4, [%[r], #60]\n\t"
  1458. "ldr r4, [%[a], #64]\n\t"
  1459. "ldr r5, [%[b], #64]\n\t"
  1460. "adc r4, r5\n\t"
  1461. "str r4, [%[r], #64]\n\t"
  1462. "ldr r4, [%[a], #68]\n\t"
  1463. "ldr r5, [%[b], #68]\n\t"
  1464. "adc r4, r5\n\t"
  1465. "str r4, [%[r], #68]\n\t"
  1466. "ldr r4, [%[a], #72]\n\t"
  1467. "ldr r5, [%[b], #72]\n\t"
  1468. "adc r4, r5\n\t"
  1469. "str r4, [%[r], #72]\n\t"
  1470. "ldr r4, [%[a], #76]\n\t"
  1471. "ldr r5, [%[b], #76]\n\t"
  1472. "adc r4, r5\n\t"
  1473. "str r4, [%[r], #76]\n\t"
  1474. "ldr r4, [%[a], #80]\n\t"
  1475. "ldr r5, [%[b], #80]\n\t"
  1476. "adc r4, r5\n\t"
  1477. "str r4, [%[r], #80]\n\t"
  1478. "ldr r4, [%[a], #84]\n\t"
  1479. "ldr r5, [%[b], #84]\n\t"
  1480. "adc r4, r5\n\t"
  1481. "str r4, [%[r], #84]\n\t"
  1482. "ldr r4, [%[a], #88]\n\t"
  1483. "ldr r5, [%[b], #88]\n\t"
  1484. "adc r4, r5\n\t"
  1485. "str r4, [%[r], #88]\n\t"
  1486. "ldr r4, [%[a], #92]\n\t"
  1487. "ldr r5, [%[b], #92]\n\t"
  1488. "adc r4, r5\n\t"
  1489. "str r4, [%[r], #92]\n\t"
  1490. "ldr r4, [%[a], #96]\n\t"
  1491. "ldr r5, [%[b], #96]\n\t"
  1492. "adc r4, r5\n\t"
  1493. "str r4, [%[r], #96]\n\t"
  1494. "ldr r4, [%[a], #100]\n\t"
  1495. "ldr r5, [%[b], #100]\n\t"
  1496. "adc r4, r5\n\t"
  1497. "str r4, [%[r], #100]\n\t"
  1498. "ldr r4, [%[a], #104]\n\t"
  1499. "ldr r5, [%[b], #104]\n\t"
  1500. "adc r4, r5\n\t"
  1501. "str r4, [%[r], #104]\n\t"
  1502. "ldr r4, [%[a], #108]\n\t"
  1503. "ldr r5, [%[b], #108]\n\t"
  1504. "adc r4, r5\n\t"
  1505. "str r4, [%[r], #108]\n\t"
  1506. "ldr r4, [%[a], #112]\n\t"
  1507. "ldr r5, [%[b], #112]\n\t"
  1508. "adc r4, r5\n\t"
  1509. "str r4, [%[r], #112]\n\t"
  1510. "ldr r4, [%[a], #116]\n\t"
  1511. "ldr r5, [%[b], #116]\n\t"
  1512. "adc r4, r5\n\t"
  1513. "str r4, [%[r], #116]\n\t"
  1514. "ldr r4, [%[a], #120]\n\t"
  1515. "ldr r5, [%[b], #120]\n\t"
  1516. "adc r4, r5\n\t"
  1517. "str r4, [%[r], #120]\n\t"
  1518. "ldr r4, [%[a], #124]\n\t"
  1519. "ldr r5, [%[b], #124]\n\t"
  1520. "adc r4, r5\n\t"
  1521. "str r4, [%[r], #124]\n\t"
  1522. "mov %[c], #0\n\t"
  1523. "adc %[c], %[c]\n\t"
  1524. "add %[a], #0x80\n\t"
  1525. "add %[b], #0x80\n\t"
  1526. "add %[r], #0x80\n\t"
  1527. "add %[c], r7\n\t"
  1528. "ldr r4, [%[a], #0]\n\t"
  1529. "ldr r5, [%[b], #0]\n\t"
  1530. "adc r4, r5\n\t"
  1531. "str r4, [%[r], #0]\n\t"
  1532. "ldr r4, [%[a], #4]\n\t"
  1533. "ldr r5, [%[b], #4]\n\t"
  1534. "adc r4, r5\n\t"
  1535. "str r4, [%[r], #4]\n\t"
  1536. "ldr r4, [%[a], #8]\n\t"
  1537. "ldr r5, [%[b], #8]\n\t"
  1538. "adc r4, r5\n\t"
  1539. "str r4, [%[r], #8]\n\t"
  1540. "ldr r4, [%[a], #12]\n\t"
  1541. "ldr r5, [%[b], #12]\n\t"
  1542. "adc r4, r5\n\t"
  1543. "str r4, [%[r], #12]\n\t"
  1544. "ldr r4, [%[a], #16]\n\t"
  1545. "ldr r5, [%[b], #16]\n\t"
  1546. "adc r4, r5\n\t"
  1547. "str r4, [%[r], #16]\n\t"
  1548. "ldr r4, [%[a], #20]\n\t"
  1549. "ldr r5, [%[b], #20]\n\t"
  1550. "adc r4, r5\n\t"
  1551. "str r4, [%[r], #20]\n\t"
  1552. "ldr r4, [%[a], #24]\n\t"
  1553. "ldr r5, [%[b], #24]\n\t"
  1554. "adc r4, r5\n\t"
  1555. "str r4, [%[r], #24]\n\t"
  1556. "ldr r4, [%[a], #28]\n\t"
  1557. "ldr r5, [%[b], #28]\n\t"
  1558. "adc r4, r5\n\t"
  1559. "str r4, [%[r], #28]\n\t"
  1560. "ldr r4, [%[a], #32]\n\t"
  1561. "ldr r5, [%[b], #32]\n\t"
  1562. "adc r4, r5\n\t"
  1563. "str r4, [%[r], #32]\n\t"
  1564. "ldr r4, [%[a], #36]\n\t"
  1565. "ldr r5, [%[b], #36]\n\t"
  1566. "adc r4, r5\n\t"
  1567. "str r4, [%[r], #36]\n\t"
  1568. "ldr r4, [%[a], #40]\n\t"
  1569. "ldr r5, [%[b], #40]\n\t"
  1570. "adc r4, r5\n\t"
  1571. "str r4, [%[r], #40]\n\t"
  1572. "ldr r4, [%[a], #44]\n\t"
  1573. "ldr r5, [%[b], #44]\n\t"
  1574. "adc r4, r5\n\t"
  1575. "str r4, [%[r], #44]\n\t"
  1576. "ldr r4, [%[a], #48]\n\t"
  1577. "ldr r5, [%[b], #48]\n\t"
  1578. "adc r4, r5\n\t"
  1579. "str r4, [%[r], #48]\n\t"
  1580. "ldr r4, [%[a], #52]\n\t"
  1581. "ldr r5, [%[b], #52]\n\t"
  1582. "adc r4, r5\n\t"
  1583. "str r4, [%[r], #52]\n\t"
  1584. "ldr r4, [%[a], #56]\n\t"
  1585. "ldr r5, [%[b], #56]\n\t"
  1586. "adc r4, r5\n\t"
  1587. "str r4, [%[r], #56]\n\t"
  1588. "ldr r4, [%[a], #60]\n\t"
  1589. "ldr r5, [%[b], #60]\n\t"
  1590. "adc r4, r5\n\t"
  1591. "str r4, [%[r], #60]\n\t"
  1592. "ldr r4, [%[a], #64]\n\t"
  1593. "ldr r5, [%[b], #64]\n\t"
  1594. "adc r4, r5\n\t"
  1595. "str r4, [%[r], #64]\n\t"
  1596. "ldr r4, [%[a], #68]\n\t"
  1597. "ldr r5, [%[b], #68]\n\t"
  1598. "adc r4, r5\n\t"
  1599. "str r4, [%[r], #68]\n\t"
  1600. "ldr r4, [%[a], #72]\n\t"
  1601. "ldr r5, [%[b], #72]\n\t"
  1602. "adc r4, r5\n\t"
  1603. "str r4, [%[r], #72]\n\t"
  1604. "ldr r4, [%[a], #76]\n\t"
  1605. "ldr r5, [%[b], #76]\n\t"
  1606. "adc r4, r5\n\t"
  1607. "str r4, [%[r], #76]\n\t"
  1608. "ldr r4, [%[a], #80]\n\t"
  1609. "ldr r5, [%[b], #80]\n\t"
  1610. "adc r4, r5\n\t"
  1611. "str r4, [%[r], #80]\n\t"
  1612. "ldr r4, [%[a], #84]\n\t"
  1613. "ldr r5, [%[b], #84]\n\t"
  1614. "adc r4, r5\n\t"
  1615. "str r4, [%[r], #84]\n\t"
  1616. "ldr r4, [%[a], #88]\n\t"
  1617. "ldr r5, [%[b], #88]\n\t"
  1618. "adc r4, r5\n\t"
  1619. "str r4, [%[r], #88]\n\t"
  1620. "ldr r4, [%[a], #92]\n\t"
  1621. "ldr r5, [%[b], #92]\n\t"
  1622. "adc r4, r5\n\t"
  1623. "str r4, [%[r], #92]\n\t"
  1624. "ldr r4, [%[a], #96]\n\t"
  1625. "ldr r5, [%[b], #96]\n\t"
  1626. "adc r4, r5\n\t"
  1627. "str r4, [%[r], #96]\n\t"
  1628. "ldr r4, [%[a], #100]\n\t"
  1629. "ldr r5, [%[b], #100]\n\t"
  1630. "adc r4, r5\n\t"
  1631. "str r4, [%[r], #100]\n\t"
  1632. "ldr r4, [%[a], #104]\n\t"
  1633. "ldr r5, [%[b], #104]\n\t"
  1634. "adc r4, r5\n\t"
  1635. "str r4, [%[r], #104]\n\t"
  1636. "ldr r4, [%[a], #108]\n\t"
  1637. "ldr r5, [%[b], #108]\n\t"
  1638. "adc r4, r5\n\t"
  1639. "str r4, [%[r], #108]\n\t"
  1640. "ldr r4, [%[a], #112]\n\t"
  1641. "ldr r5, [%[b], #112]\n\t"
  1642. "adc r4, r5\n\t"
  1643. "str r4, [%[r], #112]\n\t"
  1644. "ldr r4, [%[a], #116]\n\t"
  1645. "ldr r5, [%[b], #116]\n\t"
  1646. "adc r4, r5\n\t"
  1647. "str r4, [%[r], #116]\n\t"
  1648. "ldr r4, [%[a], #120]\n\t"
  1649. "ldr r5, [%[b], #120]\n\t"
  1650. "adc r4, r5\n\t"
  1651. "str r4, [%[r], #120]\n\t"
  1652. "ldr r4, [%[a], #124]\n\t"
  1653. "ldr r5, [%[b], #124]\n\t"
  1654. "adc r4, r5\n\t"
  1655. "str r4, [%[r], #124]\n\t"
  1656. "mov %[c], #0\n\t"
  1657. "adc %[c], %[c]\n\t"
  1658. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  1659. :
  1660. : "memory", "r4", "r5", "r7"
  1661. );
  1662. return c;
  1663. }
  1664. /* AND m into each word of a and store in r.
  1665. *
  1666. * r A single precision integer.
  1667. * a A single precision integer.
  1668. * m Mask to AND against each digit.
  1669. */
  1670. static void sp_2048_mask_32(sp_digit* r, const sp_digit* a, sp_digit m)
  1671. {
  1672. #ifdef WOLFSSL_SP_SMALL
  1673. int i;
  1674. for (i=0; i<32; i++) {
  1675. r[i] = a[i] & m;
  1676. }
  1677. #else
  1678. int i;
  1679. for (i = 0; i < 32; i += 8) {
  1680. r[i+0] = a[i+0] & m;
  1681. r[i+1] = a[i+1] & m;
  1682. r[i+2] = a[i+2] & m;
  1683. r[i+3] = a[i+3] & m;
  1684. r[i+4] = a[i+4] & m;
  1685. r[i+5] = a[i+5] & m;
  1686. r[i+6] = a[i+6] & m;
  1687. r[i+7] = a[i+7] & m;
  1688. }
  1689. #endif
  1690. }
  1691. /* Multiply a and b into r. (r = a * b)
  1692. *
  1693. * r A single precision integer.
  1694. * a A single precision integer.
  1695. * b A single precision integer.
  1696. */
  1697. SP_NOINLINE static void sp_2048_mul_64(sp_digit* r, const sp_digit* a,
  1698. const sp_digit* b)
  1699. {
  1700. sp_digit* z0 = r;
  1701. sp_digit z1[64];
  1702. sp_digit a1[32];
  1703. sp_digit b1[32];
  1704. sp_digit z2[64];
  1705. sp_digit u, ca, cb;
  1706. ca = sp_2048_add_32(a1, a, &a[32]);
  1707. cb = sp_2048_add_32(b1, b, &b[32]);
  1708. u = ca & cb;
  1709. sp_2048_mul_32(z1, a1, b1);
  1710. sp_2048_mul_32(z2, &a[32], &b[32]);
  1711. sp_2048_mul_32(z0, a, b);
  1712. sp_2048_mask_32(r + 64, a1, 0 - cb);
  1713. sp_2048_mask_32(b1, b1, 0 - ca);
  1714. u += sp_2048_add_32(r + 64, r + 64, b1);
  1715. u += sp_2048_sub_in_place_64(z1, z2);
  1716. u += sp_2048_sub_in_place_64(z1, z0);
  1717. u += sp_2048_add_64(r + 32, r + 32, z1);
  1718. r[96] = u;
  1719. XMEMSET(r + 96 + 1, 0, sizeof(sp_digit) * (32 - 1));
  1720. (void)sp_2048_add_64(r + 64, r + 64, z2);
  1721. }
  1722. /* Square a and put result in r. (r = a * a)
  1723. *
  1724. * r A single precision integer.
  1725. * a A single precision integer.
  1726. */
  1727. SP_NOINLINE static void sp_2048_sqr_64(sp_digit* r, const sp_digit* a)
  1728. {
  1729. sp_digit* z0 = r;
  1730. sp_digit z2[64];
  1731. sp_digit z1[64];
  1732. sp_digit a1[32];
  1733. sp_digit u;
  1734. u = sp_2048_add_32(a1, a, &a[32]);
  1735. sp_2048_sqr_32(z1, a1);
  1736. sp_2048_sqr_32(z2, &a[32]);
  1737. sp_2048_sqr_32(z0, a);
  1738. sp_2048_mask_32(r + 64, a1, 0 - u);
  1739. u += sp_2048_add_32(r + 64, r + 64, r + 64);
  1740. u += sp_2048_sub_in_place_64(z1, z2);
  1741. u += sp_2048_sub_in_place_64(z1, z0);
  1742. u += sp_2048_add_64(r + 32, r + 32, z1);
  1743. r[96] = u;
  1744. XMEMSET(r + 96 + 1, 0, sizeof(sp_digit) * (32 - 1));
  1745. (void)sp_2048_add_64(r + 64, r + 64, z2);
  1746. }
  1747. #endif /* !WOLFSSL_SP_SMALL */
  1748. #ifdef WOLFSSL_SP_SMALL
  1749. /* Add b to a into r. (r = a + b)
  1750. *
  1751. * r A single precision integer.
  1752. * a A single precision integer.
  1753. * b A single precision integer.
  1754. */
  1755. SP_NOINLINE static sp_digit sp_2048_add_64(sp_digit* r, const sp_digit* a,
  1756. const sp_digit* b)
  1757. {
  1758. sp_digit c = 0;
  1759. __asm__ __volatile__ (
  1760. "mov r6, %[a]\n\t"
  1761. "mov r7, #0\n\t"
  1762. "mov r4, #1\n\t"
  1763. "lsl r4, #8\n\t"
  1764. "sub r7, #1\n\t"
  1765. "add r6, r4\n\t"
  1766. "\n1:\n\t"
  1767. "add %[c], r7\n\t"
  1768. "ldr r4, [%[a]]\n\t"
  1769. "ldr r5, [%[b]]\n\t"
  1770. "adc r4, r5\n\t"
  1771. "str r4, [%[r]]\n\t"
  1772. "mov %[c], #0\n\t"
  1773. "adc %[c], %[c]\n\t"
  1774. "add %[a], #4\n\t"
  1775. "add %[b], #4\n\t"
  1776. "add %[r], #4\n\t"
  1777. "cmp %[a], r6\n\t"
  1778. "bne 1b\n\t"
  1779. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  1780. :
  1781. : "memory", "r4", "r5", "r6", "r7"
  1782. );
  1783. return c;
  1784. }
  1785. #endif /* WOLFSSL_SP_SMALL */
  1786. #ifdef WOLFSSL_SP_SMALL
  1787. /* Sub b from a into a. (a -= b)
  1788. *
  1789. * a A single precision integer.
  1790. * b A single precision integer.
  1791. */
  1792. SP_NOINLINE static sp_digit sp_2048_sub_in_place_64(sp_digit* a,
  1793. const sp_digit* b)
  1794. {
  1795. sp_digit c = 0;
  1796. __asm__ __volatile__ (
  1797. "mov r7, %[a]\n\t"
  1798. "mov r5, #1\n\t"
  1799. "lsl r5, #8\n\t"
  1800. "add r7, r5\n\t"
  1801. "\n1:\n\t"
  1802. "mov r5, #0\n\t"
  1803. "sub r5, %[c]\n\t"
  1804. "ldr r3, [%[a]]\n\t"
  1805. "ldr r4, [%[a], #4]\n\t"
  1806. "ldr r5, [%[b]]\n\t"
  1807. "ldr r6, [%[b], #4]\n\t"
  1808. "sbc r3, r5\n\t"
  1809. "sbc r4, r6\n\t"
  1810. "str r3, [%[a]]\n\t"
  1811. "str r4, [%[a], #4]\n\t"
  1812. "sbc %[c], %[c]\n\t"
  1813. "add %[a], #8\n\t"
  1814. "add %[b], #8\n\t"
  1815. "cmp %[a], r7\n\t"
  1816. "bne 1b\n\t"
  1817. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  1818. :
  1819. : "memory", "r3", "r4", "r5", "r6", "r7"
  1820. );
  1821. return c;
  1822. }
  1823. #endif /* WOLFSSL_SP_SMALL */
  1824. #ifdef WOLFSSL_SP_SMALL
  1825. /* Multiply a and b into r. (r = a * b)
  1826. *
  1827. * r A single precision integer.
  1828. * a A single precision integer.
  1829. * b A single precision integer.
  1830. */
  1831. SP_NOINLINE static void sp_2048_mul_64(sp_digit* r, const sp_digit* a,
  1832. const sp_digit* b)
  1833. {
  1834. sp_digit tmp[64 * 2];
  1835. __asm__ __volatile__ (
  1836. "mov r3, #0\n\t"
  1837. "mov r4, #0\n\t"
  1838. "mov r8, r3\n\t"
  1839. "mov r11, %[r]\n\t"
  1840. "mov r9, %[a]\n\t"
  1841. "mov r10, %[b]\n\t"
  1842. "mov r6, #1\n\t"
  1843. "lsl r6, r6, #8\n\t"
  1844. "add r6, r9\n\t"
  1845. "mov r12, r6\n\t"
  1846. "\n1:\n\t"
  1847. "mov %[r], #0\n\t"
  1848. "mov r5, #0\n\t"
  1849. "mov r6, #252\n\t"
  1850. "mov %[a], r8\n\t"
  1851. "sub %[a], r6\n\t"
  1852. "sbc r6, r6\n\t"
  1853. "mvn r6, r6\n\t"
  1854. "and %[a], r6\n\t"
  1855. "mov %[b], r8\n\t"
  1856. "sub %[b], %[a]\n\t"
  1857. "add %[a], r9\n\t"
  1858. "add %[b], r10\n\t"
  1859. "\n2:\n\t"
  1860. "# Multiply Start\n\t"
  1861. "ldr r6, [%[a]]\n\t"
  1862. "ldr r7, [%[b]]\n\t"
  1863. "lsl r6, r6, #16\n\t"
  1864. "lsl r7, r7, #16\n\t"
  1865. "lsr r6, r6, #16\n\t"
  1866. "lsr r7, r7, #16\n\t"
  1867. "mul r7, r6\n\t"
  1868. "add r3, r7\n\t"
  1869. "adc r4, %[r]\n\t"
  1870. "adc r5, %[r]\n\t"
  1871. "ldr r7, [%[b]]\n\t"
  1872. "lsr r7, r7, #16\n\t"
  1873. "mul r6, r7\n\t"
  1874. "lsr r7, r6, #16\n\t"
  1875. "lsl r6, r6, #16\n\t"
  1876. "add r3, r6\n\t"
  1877. "adc r4, r7\n\t"
  1878. "adc r5, %[r]\n\t"
  1879. "ldr r6, [%[a]]\n\t"
  1880. "ldr r7, [%[b]]\n\t"
  1881. "lsr r6, r6, #16\n\t"
  1882. "lsr r7, r7, #16\n\t"
  1883. "mul r7, r6\n\t"
  1884. "add r4, r7\n\t"
  1885. "adc r5, %[r]\n\t"
  1886. "ldr r7, [%[b]]\n\t"
  1887. "lsl r7, r7, #16\n\t"
  1888. "lsr r7, r7, #16\n\t"
  1889. "mul r6, r7\n\t"
  1890. "lsr r7, r6, #16\n\t"
  1891. "lsl r6, r6, #16\n\t"
  1892. "add r3, r6\n\t"
  1893. "adc r4, r7\n\t"
  1894. "adc r5, %[r]\n\t"
  1895. "# Multiply Done\n\t"
  1896. "add %[a], #4\n\t"
  1897. "sub %[b], #4\n\t"
  1898. "cmp %[a], r12\n\t"
  1899. "beq 3f\n\t"
  1900. "mov r6, r8\n\t"
  1901. "add r6, r9\n\t"
  1902. "cmp %[a], r6\n\t"
  1903. "ble 2b\n\t"
  1904. "\n3:\n\t"
  1905. "mov %[r], r11\n\t"
  1906. "mov r7, r8\n\t"
  1907. "str r3, [%[r], r7]\n\t"
  1908. "mov r3, r4\n\t"
  1909. "mov r4, r5\n\t"
  1910. "add r7, #4\n\t"
  1911. "mov r8, r7\n\t"
  1912. "mov r6, #1\n\t"
  1913. "lsl r6, r6, #8\n\t"
  1914. "add r6, #248\n\t"
  1915. "cmp r7, r6\n\t"
  1916. "ble 1b\n\t"
  1917. "str r3, [%[r], r7]\n\t"
  1918. "mov %[a], r9\n\t"
  1919. "mov %[b], r10\n\t"
  1920. :
  1921. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  1922. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  1923. );
  1924. XMEMCPY(r, tmp, sizeof(tmp));
  1925. }
  1926. /* Square a and put result in r. (r = a * a)
  1927. *
  1928. * r A single precision integer.
  1929. * a A single precision integer.
  1930. */
  1931. SP_NOINLINE static void sp_2048_sqr_64(sp_digit* r, const sp_digit* a)
  1932. {
  1933. __asm__ __volatile__ (
  1934. "mov r3, #0\n\t"
  1935. "mov r4, #0\n\t"
  1936. "mov r5, #0\n\t"
  1937. "mov r8, r3\n\t"
  1938. "mov r11, %[r]\n\t"
  1939. "mov r6, #2\n\t"
  1940. "lsl r6, r6, #8\n\t"
  1941. "neg r6, r6\n\t"
  1942. "add sp, r6\n\t"
  1943. "mov r10, sp\n\t"
  1944. "mov r9, %[a]\n\t"
  1945. "\n1:\n\t"
  1946. "mov %[r], #0\n\t"
  1947. "mov r6, #252\n\t"
  1948. "mov %[a], r8\n\t"
  1949. "sub %[a], r6\n\t"
  1950. "sbc r6, r6\n\t"
  1951. "mvn r6, r6\n\t"
  1952. "and %[a], r6\n\t"
  1953. "mov r2, r8\n\t"
  1954. "sub r2, %[a]\n\t"
  1955. "add %[a], r9\n\t"
  1956. "add r2, r9\n\t"
  1957. "\n2:\n\t"
  1958. "cmp r2, %[a]\n\t"
  1959. "beq 4f\n\t"
  1960. "# Multiply * 2: Start\n\t"
  1961. "ldr r6, [%[a]]\n\t"
  1962. "ldr r7, [r2]\n\t"
  1963. "lsl r6, r6, #16\n\t"
  1964. "lsl r7, r7, #16\n\t"
  1965. "lsr r6, r6, #16\n\t"
  1966. "lsr r7, r7, #16\n\t"
  1967. "mul r7, r6\n\t"
  1968. "add r3, r7\n\t"
  1969. "adc r4, %[r]\n\t"
  1970. "adc r5, %[r]\n\t"
  1971. "add r3, r7\n\t"
  1972. "adc r4, %[r]\n\t"
  1973. "adc r5, %[r]\n\t"
  1974. "ldr r7, [r2]\n\t"
  1975. "lsr r7, r7, #16\n\t"
  1976. "mul r6, r7\n\t"
  1977. "lsr r7, r6, #16\n\t"
  1978. "lsl r6, r6, #16\n\t"
  1979. "add r3, r6\n\t"
  1980. "adc r4, r7\n\t"
  1981. "adc r5, %[r]\n\t"
  1982. "add r3, r6\n\t"
  1983. "adc r4, r7\n\t"
  1984. "adc r5, %[r]\n\t"
  1985. "ldr r6, [%[a]]\n\t"
  1986. "ldr r7, [r2]\n\t"
  1987. "lsr r6, r6, #16\n\t"
  1988. "lsr r7, r7, #16\n\t"
  1989. "mul r7, r6\n\t"
  1990. "add r4, r7\n\t"
  1991. "adc r5, %[r]\n\t"
  1992. "add r4, r7\n\t"
  1993. "adc r5, %[r]\n\t"
  1994. "ldr r7, [r2]\n\t"
  1995. "lsl r7, r7, #16\n\t"
  1996. "lsr r7, r7, #16\n\t"
  1997. "mul r6, r7\n\t"
  1998. "lsr r7, r6, #16\n\t"
  1999. "lsl r6, r6, #16\n\t"
  2000. "add r3, r6\n\t"
  2001. "adc r4, r7\n\t"
  2002. "adc r5, %[r]\n\t"
  2003. "add r3, r6\n\t"
  2004. "adc r4, r7\n\t"
  2005. "adc r5, %[r]\n\t"
  2006. "# Multiply * 2: Done\n\t"
  2007. "bal 5f\n\t"
  2008. "\n4:\n\t"
  2009. "# Square: Start\n\t"
  2010. "ldr r6, [%[a]]\n\t"
  2011. "lsr r7, r6, #16\n\t"
  2012. "lsl r6, r6, #16\n\t"
  2013. "lsr r6, r6, #16\n\t"
  2014. "mul r6, r6\n\t"
  2015. "add r3, r6\n\t"
  2016. "adc r4, %[r]\n\t"
  2017. "adc r5, %[r]\n\t"
  2018. "mul r7, r7\n\t"
  2019. "add r4, r7\n\t"
  2020. "adc r5, %[r]\n\t"
  2021. "ldr r6, [%[a]]\n\t"
  2022. "lsr r7, r6, #16\n\t"
  2023. "lsl r6, r6, #16\n\t"
  2024. "lsr r6, r6, #16\n\t"
  2025. "mul r6, r7\n\t"
  2026. "lsr r7, r6, #15\n\t"
  2027. "lsl r6, r6, #17\n\t"
  2028. "add r3, r6\n\t"
  2029. "adc r4, r7\n\t"
  2030. "adc r5, %[r]\n\t"
  2031. "# Square: Done\n\t"
  2032. "\n5:\n\t"
  2033. "add %[a], #4\n\t"
  2034. "sub r2, #4\n\t"
  2035. "mov r6, #1\n\t"
  2036. "lsl r6, r6, #8\n\t"
  2037. "add r6, r9\n\t"
  2038. "cmp %[a], r6\n\t"
  2039. "beq 3f\n\t"
  2040. "cmp %[a], r2\n\t"
  2041. "bgt 3f\n\t"
  2042. "mov r7, r8\n\t"
  2043. "add r7, r9\n\t"
  2044. "cmp %[a], r7\n\t"
  2045. "ble 2b\n\t"
  2046. "\n3:\n\t"
  2047. "mov %[r], r10\n\t"
  2048. "mov r7, r8\n\t"
  2049. "str r3, [%[r], r7]\n\t"
  2050. "mov r3, r4\n\t"
  2051. "mov r4, r5\n\t"
  2052. "mov r5, #0\n\t"
  2053. "add r7, #4\n\t"
  2054. "mov r8, r7\n\t"
  2055. "mov r6, #1\n\t"
  2056. "lsl r6, r6, #8\n\t"
  2057. "add r6, #248\n\t"
  2058. "cmp r7, r6\n\t"
  2059. "ble 1b\n\t"
  2060. "mov %[a], r9\n\t"
  2061. "str r3, [%[r], r7]\n\t"
  2062. "mov %[r], r11\n\t"
  2063. "mov %[a], r10\n\t"
  2064. "mov r3, #1\n\t"
  2065. "lsl r3, r3, #8\n\t"
  2066. "add r3, #252\n\t"
  2067. "\n4:\n\t"
  2068. "ldr r6, [%[a], r3]\n\t"
  2069. "str r6, [%[r], r3]\n\t"
  2070. "sub r3, #4\n\t"
  2071. "bge 4b\n\t"
  2072. "mov r6, #2\n\t"
  2073. "lsl r6, r6, #8\n\t"
  2074. "add sp, r6\n\t"
  2075. :
  2076. : [r] "r" (r), [a] "r" (a)
  2077. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  2078. );
  2079. }
  2080. #endif /* WOLFSSL_SP_SMALL */
  2081. #if (defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)
  2082. #ifdef WOLFSSL_SP_SMALL
  2083. /* AND m into each word of a and store in r.
  2084. *
  2085. * r A single precision integer.
  2086. * a A single precision integer.
  2087. * m Mask to AND against each digit.
  2088. */
  2089. static void sp_2048_mask_32(sp_digit* r, const sp_digit* a, sp_digit m)
  2090. {
  2091. int i;
  2092. for (i=0; i<32; i++) {
  2093. r[i] = a[i] & m;
  2094. }
  2095. }
  2096. #endif /* WOLFSSL_SP_SMALL */
  2097. #ifdef WOLFSSL_SP_SMALL
  2098. /* Add b to a into r. (r = a + b)
  2099. *
  2100. * r A single precision integer.
  2101. * a A single precision integer.
  2102. * b A single precision integer.
  2103. */
  2104. SP_NOINLINE static sp_digit sp_2048_add_32(sp_digit* r, const sp_digit* a,
  2105. const sp_digit* b)
  2106. {
  2107. sp_digit c = 0;
  2108. __asm__ __volatile__ (
  2109. "mov r6, %[a]\n\t"
  2110. "mov r7, #0\n\t"
  2111. "add r6, #128\n\t"
  2112. "sub r7, #1\n\t"
  2113. "\n1:\n\t"
  2114. "add %[c], r7\n\t"
  2115. "ldr r4, [%[a]]\n\t"
  2116. "ldr r5, [%[b]]\n\t"
  2117. "adc r4, r5\n\t"
  2118. "str r4, [%[r]]\n\t"
  2119. "mov %[c], #0\n\t"
  2120. "adc %[c], %[c]\n\t"
  2121. "add %[a], #4\n\t"
  2122. "add %[b], #4\n\t"
  2123. "add %[r], #4\n\t"
  2124. "cmp %[a], r6\n\t"
  2125. "bne 1b\n\t"
  2126. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  2127. :
  2128. : "memory", "r4", "r5", "r6", "r7"
  2129. );
  2130. return c;
  2131. }
  2132. #endif /* WOLFSSL_SP_SMALL */
  2133. #ifdef WOLFSSL_SP_SMALL
  2134. /* Sub b from a into a. (a -= b)
  2135. *
  2136. * a A single precision integer.
  2137. * b A single precision integer.
  2138. */
  2139. SP_NOINLINE static sp_digit sp_2048_sub_in_place_32(sp_digit* a,
  2140. const sp_digit* b)
  2141. {
  2142. sp_digit c = 0;
  2143. __asm__ __volatile__ (
  2144. "mov r7, %[a]\n\t"
  2145. "add r7, #128\n\t"
  2146. "\n1:\n\t"
  2147. "mov r5, #0\n\t"
  2148. "sub r5, %[c]\n\t"
  2149. "ldr r3, [%[a]]\n\t"
  2150. "ldr r4, [%[a], #4]\n\t"
  2151. "ldr r5, [%[b]]\n\t"
  2152. "ldr r6, [%[b], #4]\n\t"
  2153. "sbc r3, r5\n\t"
  2154. "sbc r4, r6\n\t"
  2155. "str r3, [%[a]]\n\t"
  2156. "str r4, [%[a], #4]\n\t"
  2157. "sbc %[c], %[c]\n\t"
  2158. "add %[a], #8\n\t"
  2159. "add %[b], #8\n\t"
  2160. "cmp %[a], r7\n\t"
  2161. "bne 1b\n\t"
  2162. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  2163. :
  2164. : "memory", "r3", "r4", "r5", "r6", "r7"
  2165. );
  2166. return c;
  2167. }
  2168. #endif /* WOLFSSL_SP_SMALL */
  2169. #ifdef WOLFSSL_SP_SMALL
  2170. /* Multiply a and b into r. (r = a * b)
  2171. *
  2172. * r A single precision integer.
  2173. * a A single precision integer.
  2174. * b A single precision integer.
  2175. */
  2176. SP_NOINLINE static void sp_2048_mul_32(sp_digit* r, const sp_digit* a,
  2177. const sp_digit* b)
  2178. {
  2179. sp_digit tmp[32 * 2];
  2180. __asm__ __volatile__ (
  2181. "mov r3, #0\n\t"
  2182. "mov r4, #0\n\t"
  2183. "mov r8, r3\n\t"
  2184. "mov r11, %[r]\n\t"
  2185. "mov r9, %[a]\n\t"
  2186. "mov r10, %[b]\n\t"
  2187. "mov r6, #128\n\t"
  2188. "add r6, r9\n\t"
  2189. "mov r12, r6\n\t"
  2190. "\n1:\n\t"
  2191. "mov %[r], #0\n\t"
  2192. "mov r5, #0\n\t"
  2193. "mov r6, #124\n\t"
  2194. "mov %[a], r8\n\t"
  2195. "sub %[a], r6\n\t"
  2196. "sbc r6, r6\n\t"
  2197. "mvn r6, r6\n\t"
  2198. "and %[a], r6\n\t"
  2199. "mov %[b], r8\n\t"
  2200. "sub %[b], %[a]\n\t"
  2201. "add %[a], r9\n\t"
  2202. "add %[b], r10\n\t"
  2203. "\n2:\n\t"
  2204. "# Multiply Start\n\t"
  2205. "ldr r6, [%[a]]\n\t"
  2206. "ldr r7, [%[b]]\n\t"
  2207. "lsl r6, r6, #16\n\t"
  2208. "lsl r7, r7, #16\n\t"
  2209. "lsr r6, r6, #16\n\t"
  2210. "lsr r7, r7, #16\n\t"
  2211. "mul r7, r6\n\t"
  2212. "add r3, r7\n\t"
  2213. "adc r4, %[r]\n\t"
  2214. "adc r5, %[r]\n\t"
  2215. "ldr r7, [%[b]]\n\t"
  2216. "lsr r7, r7, #16\n\t"
  2217. "mul r6, r7\n\t"
  2218. "lsr r7, r6, #16\n\t"
  2219. "lsl r6, r6, #16\n\t"
  2220. "add r3, r6\n\t"
  2221. "adc r4, r7\n\t"
  2222. "adc r5, %[r]\n\t"
  2223. "ldr r6, [%[a]]\n\t"
  2224. "ldr r7, [%[b]]\n\t"
  2225. "lsr r6, r6, #16\n\t"
  2226. "lsr r7, r7, #16\n\t"
  2227. "mul r7, r6\n\t"
  2228. "add r4, r7\n\t"
  2229. "adc r5, %[r]\n\t"
  2230. "ldr r7, [%[b]]\n\t"
  2231. "lsl r7, r7, #16\n\t"
  2232. "lsr r7, r7, #16\n\t"
  2233. "mul r6, r7\n\t"
  2234. "lsr r7, r6, #16\n\t"
  2235. "lsl r6, r6, #16\n\t"
  2236. "add r3, r6\n\t"
  2237. "adc r4, r7\n\t"
  2238. "adc r5, %[r]\n\t"
  2239. "# Multiply Done\n\t"
  2240. "add %[a], #4\n\t"
  2241. "sub %[b], #4\n\t"
  2242. "cmp %[a], r12\n\t"
  2243. "beq 3f\n\t"
  2244. "mov r6, r8\n\t"
  2245. "add r6, r9\n\t"
  2246. "cmp %[a], r6\n\t"
  2247. "ble 2b\n\t"
  2248. "\n3:\n\t"
  2249. "mov %[r], r11\n\t"
  2250. "mov r7, r8\n\t"
  2251. "str r3, [%[r], r7]\n\t"
  2252. "mov r3, r4\n\t"
  2253. "mov r4, r5\n\t"
  2254. "add r7, #4\n\t"
  2255. "mov r8, r7\n\t"
  2256. "mov r6, #248\n\t"
  2257. "cmp r7, r6\n\t"
  2258. "ble 1b\n\t"
  2259. "str r3, [%[r], r7]\n\t"
  2260. "mov %[a], r9\n\t"
  2261. "mov %[b], r10\n\t"
  2262. :
  2263. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  2264. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  2265. );
  2266. XMEMCPY(r, tmp, sizeof(tmp));
  2267. }
  2268. /* Square a and put result in r. (r = a * a)
  2269. *
  2270. * r A single precision integer.
  2271. * a A single precision integer.
  2272. */
  2273. SP_NOINLINE static void sp_2048_sqr_32(sp_digit* r, const sp_digit* a)
  2274. {
  2275. __asm__ __volatile__ (
  2276. "mov r3, #0\n\t"
  2277. "mov r4, #0\n\t"
  2278. "mov r5, #0\n\t"
  2279. "mov r8, r3\n\t"
  2280. "mov r11, %[r]\n\t"
  2281. "mov r6, #1\n\t"
  2282. "lsl r6, r6, #8\n\t"
  2283. "neg r6, r6\n\t"
  2284. "add sp, r6\n\t"
  2285. "mov r10, sp\n\t"
  2286. "mov r9, %[a]\n\t"
  2287. "\n1:\n\t"
  2288. "mov %[r], #0\n\t"
  2289. "mov r6, #124\n\t"
  2290. "mov %[a], r8\n\t"
  2291. "sub %[a], r6\n\t"
  2292. "sbc r6, r6\n\t"
  2293. "mvn r6, r6\n\t"
  2294. "and %[a], r6\n\t"
  2295. "mov r2, r8\n\t"
  2296. "sub r2, %[a]\n\t"
  2297. "add %[a], r9\n\t"
  2298. "add r2, r9\n\t"
  2299. "\n2:\n\t"
  2300. "cmp r2, %[a]\n\t"
  2301. "beq 4f\n\t"
  2302. "# Multiply * 2: Start\n\t"
  2303. "ldr r6, [%[a]]\n\t"
  2304. "ldr r7, [r2]\n\t"
  2305. "lsl r6, r6, #16\n\t"
  2306. "lsl r7, r7, #16\n\t"
  2307. "lsr r6, r6, #16\n\t"
  2308. "lsr r7, r7, #16\n\t"
  2309. "mul r7, r6\n\t"
  2310. "add r3, r7\n\t"
  2311. "adc r4, %[r]\n\t"
  2312. "adc r5, %[r]\n\t"
  2313. "add r3, r7\n\t"
  2314. "adc r4, %[r]\n\t"
  2315. "adc r5, %[r]\n\t"
  2316. "ldr r7, [r2]\n\t"
  2317. "lsr r7, r7, #16\n\t"
  2318. "mul r6, r7\n\t"
  2319. "lsr r7, r6, #16\n\t"
  2320. "lsl r6, r6, #16\n\t"
  2321. "add r3, r6\n\t"
  2322. "adc r4, r7\n\t"
  2323. "adc r5, %[r]\n\t"
  2324. "add r3, r6\n\t"
  2325. "adc r4, r7\n\t"
  2326. "adc r5, %[r]\n\t"
  2327. "ldr r6, [%[a]]\n\t"
  2328. "ldr r7, [r2]\n\t"
  2329. "lsr r6, r6, #16\n\t"
  2330. "lsr r7, r7, #16\n\t"
  2331. "mul r7, r6\n\t"
  2332. "add r4, r7\n\t"
  2333. "adc r5, %[r]\n\t"
  2334. "add r4, r7\n\t"
  2335. "adc r5, %[r]\n\t"
  2336. "ldr r7, [r2]\n\t"
  2337. "lsl r7, r7, #16\n\t"
  2338. "lsr r7, r7, #16\n\t"
  2339. "mul r6, r7\n\t"
  2340. "lsr r7, r6, #16\n\t"
  2341. "lsl r6, r6, #16\n\t"
  2342. "add r3, r6\n\t"
  2343. "adc r4, r7\n\t"
  2344. "adc r5, %[r]\n\t"
  2345. "add r3, r6\n\t"
  2346. "adc r4, r7\n\t"
  2347. "adc r5, %[r]\n\t"
  2348. "# Multiply * 2: Done\n\t"
  2349. "bal 5f\n\t"
  2350. "\n4:\n\t"
  2351. "# Square: Start\n\t"
  2352. "ldr r6, [%[a]]\n\t"
  2353. "lsr r7, r6, #16\n\t"
  2354. "lsl r6, r6, #16\n\t"
  2355. "lsr r6, r6, #16\n\t"
  2356. "mul r6, r6\n\t"
  2357. "add r3, r6\n\t"
  2358. "adc r4, %[r]\n\t"
  2359. "adc r5, %[r]\n\t"
  2360. "mul r7, r7\n\t"
  2361. "add r4, r7\n\t"
  2362. "adc r5, %[r]\n\t"
  2363. "ldr r6, [%[a]]\n\t"
  2364. "lsr r7, r6, #16\n\t"
  2365. "lsl r6, r6, #16\n\t"
  2366. "lsr r6, r6, #16\n\t"
  2367. "mul r6, r7\n\t"
  2368. "lsr r7, r6, #15\n\t"
  2369. "lsl r6, r6, #17\n\t"
  2370. "add r3, r6\n\t"
  2371. "adc r4, r7\n\t"
  2372. "adc r5, %[r]\n\t"
  2373. "# Square: Done\n\t"
  2374. "\n5:\n\t"
  2375. "add %[a], #4\n\t"
  2376. "sub r2, #4\n\t"
  2377. "mov r6, #128\n\t"
  2378. "add r6, r9\n\t"
  2379. "cmp %[a], r6\n\t"
  2380. "beq 3f\n\t"
  2381. "cmp %[a], r2\n\t"
  2382. "bgt 3f\n\t"
  2383. "mov r7, r8\n\t"
  2384. "add r7, r9\n\t"
  2385. "cmp %[a], r7\n\t"
  2386. "ble 2b\n\t"
  2387. "\n3:\n\t"
  2388. "mov %[r], r10\n\t"
  2389. "mov r7, r8\n\t"
  2390. "str r3, [%[r], r7]\n\t"
  2391. "mov r3, r4\n\t"
  2392. "mov r4, r5\n\t"
  2393. "mov r5, #0\n\t"
  2394. "add r7, #4\n\t"
  2395. "mov r8, r7\n\t"
  2396. "mov r6, #248\n\t"
  2397. "cmp r7, r6\n\t"
  2398. "ble 1b\n\t"
  2399. "mov %[a], r9\n\t"
  2400. "str r3, [%[r], r7]\n\t"
  2401. "mov %[r], r11\n\t"
  2402. "mov %[a], r10\n\t"
  2403. "mov r3, #252\n\t"
  2404. "\n4:\n\t"
  2405. "ldr r6, [%[a], r3]\n\t"
  2406. "str r6, [%[r], r3]\n\t"
  2407. "sub r3, #4\n\t"
  2408. "bge 4b\n\t"
  2409. "mov r6, #1\n\t"
  2410. "lsl r6, r6, #8\n\t"
  2411. "add sp, r6\n\t"
  2412. :
  2413. : [r] "r" (r), [a] "r" (a)
  2414. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  2415. );
  2416. }
  2417. #endif /* WOLFSSL_SP_SMALL */
  2418. #endif /* (WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH) && !WOLFSSL_RSA_PUBLIC_ONLY */
  2419. /* Caclulate the bottom digit of -1/a mod 2^n.
  2420. *
  2421. * a A single precision number.
  2422. * rho Bottom word of inverse.
  2423. */
  2424. static void sp_2048_mont_setup(const sp_digit* a, sp_digit* rho)
  2425. {
  2426. sp_digit x, b;
  2427. b = a[0];
  2428. x = (((b + 2) & 4) << 1) + b; /* here x*a==1 mod 2**4 */
  2429. x *= 2 - b * x; /* here x*a==1 mod 2**8 */
  2430. x *= 2 - b * x; /* here x*a==1 mod 2**16 */
  2431. x *= 2 - b * x; /* here x*a==1 mod 2**32 */
  2432. /* rho = -1/m mod b */
  2433. *rho = -x;
  2434. }
  2435. /* Mul a by digit b into r. (r = a * b)
  2436. *
  2437. * r A single precision integer.
  2438. * a A single precision integer.
  2439. * b A single precision digit.
  2440. */
  2441. SP_NOINLINE static void sp_2048_mul_d_64(sp_digit* r, const sp_digit* a,
  2442. sp_digit b)
  2443. {
  2444. __asm__ __volatile__ (
  2445. "mov r6, #1\n\t"
  2446. "lsl r6, r6, #8\n\t"
  2447. "add r6, %[a]\n\t"
  2448. "mov r8, %[r]\n\t"
  2449. "mov r9, r6\n\t"
  2450. "mov r3, #0\n\t"
  2451. "mov r4, #0\n\t"
  2452. "1:\n\t"
  2453. "mov %[r], #0\n\t"
  2454. "mov r5, #0\n\t"
  2455. "# A[] * B\n\t"
  2456. "ldr r6, [%[a]]\n\t"
  2457. "lsl r6, r6, #16\n\t"
  2458. "lsl r7, %[b], #16\n\t"
  2459. "lsr r6, r6, #16\n\t"
  2460. "lsr r7, r7, #16\n\t"
  2461. "mul r7, r6\n\t"
  2462. "add r3, r7\n\t"
  2463. "adc r4, %[r]\n\t"
  2464. "adc r5, %[r]\n\t"
  2465. "lsr r7, %[b], #16\n\t"
  2466. "mul r6, r7\n\t"
  2467. "lsr r7, r6, #16\n\t"
  2468. "lsl r6, r6, #16\n\t"
  2469. "add r3, r6\n\t"
  2470. "adc r4, r7\n\t"
  2471. "adc r5, %[r]\n\t"
  2472. "ldr r6, [%[a]]\n\t"
  2473. "lsr r6, r6, #16\n\t"
  2474. "lsr r7, %[b], #16\n\t"
  2475. "mul r7, r6\n\t"
  2476. "add r4, r7\n\t"
  2477. "adc r5, %[r]\n\t"
  2478. "lsl r7, %[b], #16\n\t"
  2479. "lsr r7, r7, #16\n\t"
  2480. "mul r6, r7\n\t"
  2481. "lsr r7, r6, #16\n\t"
  2482. "lsl r6, r6, #16\n\t"
  2483. "add r3, r6\n\t"
  2484. "adc r4, r7\n\t"
  2485. "adc r5, %[r]\n\t"
  2486. "# A[] * B - Done\n\t"
  2487. "mov %[r], r8\n\t"
  2488. "str r3, [%[r]]\n\t"
  2489. "mov r3, r4\n\t"
  2490. "mov r4, r5\n\t"
  2491. "add %[r], #4\n\t"
  2492. "add %[a], #4\n\t"
  2493. "mov r8, %[r]\n\t"
  2494. "cmp %[a], r9\n\t"
  2495. "blt 1b\n\t"
  2496. "str r3, [%[r]]\n\t"
  2497. : [r] "+r" (r), [a] "+r" (a)
  2498. : [b] "r" (b)
  2499. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  2500. );
  2501. }
  2502. #if (defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)
  2503. /* r = 2^n mod m where n is the number of bits to reduce by.
  2504. * Given m must be 2048 bits, just need to subtract.
  2505. *
  2506. * r A single precision number.
  2507. * m A single precision number.
  2508. */
  2509. static void sp_2048_mont_norm_32(sp_digit* r, const sp_digit* m)
  2510. {
  2511. XMEMSET(r, 0, sizeof(sp_digit) * 32);
  2512. /* r = 2^n mod m */
  2513. sp_2048_sub_in_place_32(r, m);
  2514. }
  2515. /* Conditionally subtract b from a using the mask m.
  2516. * m is -1 to subtract and 0 when not copying.
  2517. *
  2518. * r A single precision number representing condition subtract result.
  2519. * a A single precision number to subtract from.
  2520. * b A single precision number to subtract.
  2521. * m Mask value to apply.
  2522. */
  2523. SP_NOINLINE static sp_digit sp_2048_cond_sub_32(sp_digit* r, const sp_digit* a,
  2524. const sp_digit* b, sp_digit m)
  2525. {
  2526. sp_digit c = 0;
  2527. __asm__ __volatile__ (
  2528. "mov r5, #128\n\t"
  2529. "mov r8, r5\n\t"
  2530. "mov r7, #0\n\t"
  2531. "1:\n\t"
  2532. "ldr r6, [%[b], r7]\n\t"
  2533. "and r6, %[m]\n\t"
  2534. "mov r5, #0\n\t"
  2535. "sub r5, %[c]\n\t"
  2536. "ldr r5, [%[a], r7]\n\t"
  2537. "sbc r5, r6\n\t"
  2538. "sbc %[c], %[c]\n\t"
  2539. "str r5, [%[r], r7]\n\t"
  2540. "add r7, #4\n\t"
  2541. "cmp r7, r8\n\t"
  2542. "blt 1b\n\t"
  2543. : [c] "+r" (c)
  2544. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  2545. : "memory", "r5", "r6", "r7", "r8"
  2546. );
  2547. return c;
  2548. }
  2549. /* Reduce the number back to 2048 bits using Montgomery reduction.
  2550. *
  2551. * a A single precision number to reduce in place.
  2552. * m The single precision number representing the modulus.
  2553. * mp The digit representing the negative inverse of m mod 2^n.
  2554. */
  2555. SP_NOINLINE static void sp_2048_mont_reduce_32(sp_digit* a, const sp_digit* m,
  2556. sp_digit mp)
  2557. {
  2558. sp_digit ca = 0;
  2559. __asm__ __volatile__ (
  2560. "mov r8, %[mp]\n\t"
  2561. "mov r12, %[ca]\n\t"
  2562. "mov r14, %[m]\n\t"
  2563. "mov r9, %[a]\n\t"
  2564. "mov r4, #0\n\t"
  2565. "# i = 0\n\t"
  2566. "mov r11, r4\n\t"
  2567. "\n1:\n\t"
  2568. "mov r5, #0\n\t"
  2569. "mov %[ca], #0\n\t"
  2570. "# mu = a[i] * mp\n\t"
  2571. "mov %[mp], r8\n\t"
  2572. "ldr %[a], [%[a]]\n\t"
  2573. "mul %[mp], %[a]\n\t"
  2574. "mov %[m], r14\n\t"
  2575. "mov r10, r9\n\t"
  2576. "\n2:\n\t"
  2577. "# a[i+j] += m[j] * mu\n\t"
  2578. "mov %[a], r10\n\t"
  2579. "ldr %[a], [%[a]]\n\t"
  2580. "mov %[ca], #0\n\t"
  2581. "mov r4, r5\n\t"
  2582. "mov r5, #0\n\t"
  2583. "# Multiply m[j] and mu - Start\n\t"
  2584. "ldr r7, [%[m]]\n\t"
  2585. "lsl r6, %[mp], #16\n\t"
  2586. "lsl r7, r7, #16\n\t"
  2587. "lsr r6, r6, #16\n\t"
  2588. "lsr r7, r7, #16\n\t"
  2589. "mul r7, r6\n\t"
  2590. "add %[a], r7\n\t"
  2591. "adc r5, %[ca]\n\t"
  2592. "ldr r7, [%[m]]\n\t"
  2593. "lsr r7, r7, #16\n\t"
  2594. "mul r6, r7\n\t"
  2595. "lsr r7, r6, #16\n\t"
  2596. "lsl r6, r6, #16\n\t"
  2597. "add %[a], r6\n\t"
  2598. "adc r5, r7\n\t"
  2599. "ldr r7, [%[m]]\n\t"
  2600. "lsr r6, %[mp], #16\n\t"
  2601. "lsr r7, r7, #16\n\t"
  2602. "mul r7, r6\n\t"
  2603. "add r5, r7\n\t"
  2604. "ldr r7, [%[m]]\n\t"
  2605. "lsl r7, r7, #16\n\t"
  2606. "lsr r7, r7, #16\n\t"
  2607. "mul r6, r7\n\t"
  2608. "lsr r7, r6, #16\n\t"
  2609. "lsl r6, r6, #16\n\t"
  2610. "add %[a], r6\n\t"
  2611. "adc r5, r7\n\t"
  2612. "# Multiply m[j] and mu - Done\n\t"
  2613. "add r4, %[a]\n\t"
  2614. "adc r5, %[ca]\n\t"
  2615. "mov %[a], r10\n\t"
  2616. "str r4, [%[a]]\n\t"
  2617. "mov r6, #4\n\t"
  2618. "add %[m], #4\n\t"
  2619. "add r10, r6\n\t"
  2620. "mov r4, #124\n\t"
  2621. "add r4, r9\n\t"
  2622. "cmp r10, r4\n\t"
  2623. "blt 2b\n\t"
  2624. "# a[i+31] += m[31] * mu\n\t"
  2625. "mov %[ca], #0\n\t"
  2626. "mov r4, r12\n\t"
  2627. "mov %[a], #0\n\t"
  2628. "# Multiply m[31] and mu - Start\n\t"
  2629. "ldr r7, [%[m]]\n\t"
  2630. "lsl r6, %[mp], #16\n\t"
  2631. "lsl r7, r7, #16\n\t"
  2632. "lsr r6, r6, #16\n\t"
  2633. "lsr r7, r7, #16\n\t"
  2634. "mul r7, r6\n\t"
  2635. "add r5, r7\n\t"
  2636. "adc r4, %[ca]\n\t"
  2637. "adc %[a], %[ca]\n\t"
  2638. "ldr r7, [%[m]]\n\t"
  2639. "lsr r7, r7, #16\n\t"
  2640. "mul r6, r7\n\t"
  2641. "lsr r7, r6, #16\n\t"
  2642. "lsl r6, r6, #16\n\t"
  2643. "add r5, r6\n\t"
  2644. "adc r4, r7\n\t"
  2645. "adc %[a], %[ca]\n\t"
  2646. "ldr r7, [%[m]]\n\t"
  2647. "lsr r6, %[mp], #16\n\t"
  2648. "lsr r7, r7, #16\n\t"
  2649. "mul r7, r6\n\t"
  2650. "add r4, r7\n\t"
  2651. "adc %[a], %[ca]\n\t"
  2652. "ldr r7, [%[m]]\n\t"
  2653. "lsl r7, r7, #16\n\t"
  2654. "lsr r7, r7, #16\n\t"
  2655. "mul r6, r7\n\t"
  2656. "lsr r7, r6, #16\n\t"
  2657. "lsl r6, r6, #16\n\t"
  2658. "add r5, r6\n\t"
  2659. "adc r4, r7\n\t"
  2660. "adc %[a], %[ca]\n\t"
  2661. "# Multiply m[31] and mu - Done\n\t"
  2662. "mov %[ca], %[a]\n\t"
  2663. "mov %[a], r10\n\t"
  2664. "ldr r7, [%[a], #4]\n\t"
  2665. "ldr %[a], [%[a]]\n\t"
  2666. "mov r6, #0\n\t"
  2667. "add r5, %[a]\n\t"
  2668. "adc r7, r4\n\t"
  2669. "adc %[ca], r6\n\t"
  2670. "mov %[a], r10\n\t"
  2671. "str r5, [%[a]]\n\t"
  2672. "str r7, [%[a], #4]\n\t"
  2673. "# i += 1\n\t"
  2674. "mov r6, #4\n\t"
  2675. "add r9, r6\n\t"
  2676. "add r11, r6\n\t"
  2677. "mov r12, %[ca]\n\t"
  2678. "mov %[a], r9\n\t"
  2679. "mov r4, #128\n\t"
  2680. "cmp r11, r4\n\t"
  2681. "blt 1b\n\t"
  2682. "mov %[m], r14\n\t"
  2683. : [ca] "+r" (ca), [a] "+r" (a)
  2684. : [m] "r" (m), [mp] "r" (mp)
  2685. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  2686. );
  2687. sp_2048_cond_sub_32(a - 32, a, m, (sp_digit)0 - ca);
  2688. }
  2689. /* Multiply two Montogmery form numbers mod the modulus (prime).
  2690. * (r = a * b mod m)
  2691. *
  2692. * r Result of multiplication.
  2693. * a First number to multiply in Montogmery form.
  2694. * b Second number to multiply in Montogmery form.
  2695. * m Modulus (prime).
  2696. * mp Montogmery mulitplier.
  2697. */
  2698. static void sp_2048_mont_mul_32(sp_digit* r, const sp_digit* a, const sp_digit* b,
  2699. const sp_digit* m, sp_digit mp)
  2700. {
  2701. sp_2048_mul_32(r, a, b);
  2702. sp_2048_mont_reduce_32(r, m, mp);
  2703. }
  2704. /* Square the Montgomery form number. (r = a * a mod m)
  2705. *
  2706. * r Result of squaring.
  2707. * a Number to square in Montogmery form.
  2708. * m Modulus (prime).
  2709. * mp Montogmery mulitplier.
  2710. */
  2711. static void sp_2048_mont_sqr_32(sp_digit* r, const sp_digit* a, const sp_digit* m,
  2712. sp_digit mp)
  2713. {
  2714. sp_2048_sqr_32(r, a);
  2715. sp_2048_mont_reduce_32(r, m, mp);
  2716. }
  2717. /* Mul a by digit b into r. (r = a * b)
  2718. *
  2719. * r A single precision integer.
  2720. * a A single precision integer.
  2721. * b A single precision digit.
  2722. */
  2723. SP_NOINLINE static void sp_2048_mul_d_32(sp_digit* r, const sp_digit* a,
  2724. sp_digit b)
  2725. {
  2726. __asm__ __volatile__ (
  2727. "mov r6, #128\n\t"
  2728. "add r6, %[a]\n\t"
  2729. "mov r8, %[r]\n\t"
  2730. "mov r9, r6\n\t"
  2731. "mov r3, #0\n\t"
  2732. "mov r4, #0\n\t"
  2733. "1:\n\t"
  2734. "mov %[r], #0\n\t"
  2735. "mov r5, #0\n\t"
  2736. "# A[] * B\n\t"
  2737. "ldr r6, [%[a]]\n\t"
  2738. "lsl r6, r6, #16\n\t"
  2739. "lsl r7, %[b], #16\n\t"
  2740. "lsr r6, r6, #16\n\t"
  2741. "lsr r7, r7, #16\n\t"
  2742. "mul r7, r6\n\t"
  2743. "add r3, r7\n\t"
  2744. "adc r4, %[r]\n\t"
  2745. "adc r5, %[r]\n\t"
  2746. "lsr r7, %[b], #16\n\t"
  2747. "mul r6, r7\n\t"
  2748. "lsr r7, r6, #16\n\t"
  2749. "lsl r6, r6, #16\n\t"
  2750. "add r3, r6\n\t"
  2751. "adc r4, r7\n\t"
  2752. "adc r5, %[r]\n\t"
  2753. "ldr r6, [%[a]]\n\t"
  2754. "lsr r6, r6, #16\n\t"
  2755. "lsr r7, %[b], #16\n\t"
  2756. "mul r7, r6\n\t"
  2757. "add r4, r7\n\t"
  2758. "adc r5, %[r]\n\t"
  2759. "lsl r7, %[b], #16\n\t"
  2760. "lsr r7, r7, #16\n\t"
  2761. "mul r6, r7\n\t"
  2762. "lsr r7, r6, #16\n\t"
  2763. "lsl r6, r6, #16\n\t"
  2764. "add r3, r6\n\t"
  2765. "adc r4, r7\n\t"
  2766. "adc r5, %[r]\n\t"
  2767. "# A[] * B - Done\n\t"
  2768. "mov %[r], r8\n\t"
  2769. "str r3, [%[r]]\n\t"
  2770. "mov r3, r4\n\t"
  2771. "mov r4, r5\n\t"
  2772. "add %[r], #4\n\t"
  2773. "add %[a], #4\n\t"
  2774. "mov r8, %[r]\n\t"
  2775. "cmp %[a], r9\n\t"
  2776. "blt 1b\n\t"
  2777. "str r3, [%[r]]\n\t"
  2778. : [r] "+r" (r), [a] "+r" (a)
  2779. : [b] "r" (b)
  2780. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  2781. );
  2782. }
  2783. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  2784. *
  2785. * d1 The high order half of the number to divide.
  2786. * d0 The low order half of the number to divide.
  2787. * div The dividend.
  2788. * returns the result of the division.
  2789. *
  2790. * Note that this is an approximate div. It may give an answer 1 larger.
  2791. */
  2792. SP_NOINLINE static sp_digit div_2048_word_32(sp_digit d1, sp_digit d0,
  2793. sp_digit div)
  2794. {
  2795. sp_digit r = 0;
  2796. __asm__ __volatile__ (
  2797. "lsr r5, %[div], #1\n\t"
  2798. "add r5, #1\n\t"
  2799. "mov r8, %[d0]\n\t"
  2800. "mov r9, %[d1]\n\t"
  2801. "# Do top 32\n\t"
  2802. "mov r6, r5\n\t"
  2803. "sub r6, %[d1]\n\t"
  2804. "sbc r6, r6\n\t"
  2805. "add %[r], %[r]\n\t"
  2806. "sub %[r], r6\n\t"
  2807. "and r6, r5\n\t"
  2808. "sub %[d1], r6\n\t"
  2809. "# Next 30 bits\n\t"
  2810. "mov r4, #29\n\t"
  2811. "1:\n\t"
  2812. "lsl %[d0], %[d0], #1\n\t"
  2813. "adc %[d1], %[d1]\n\t"
  2814. "mov r6, r5\n\t"
  2815. "sub r6, %[d1]\n\t"
  2816. "sbc r6, r6\n\t"
  2817. "add %[r], %[r]\n\t"
  2818. "sub %[r], r6\n\t"
  2819. "and r6, r5\n\t"
  2820. "sub %[d1], r6\n\t"
  2821. "sub r4, #1\n\t"
  2822. "bpl 1b\n\t"
  2823. "mov r7, #0\n\t"
  2824. "add %[r], %[r]\n\t"
  2825. "add %[r], #1\n\t"
  2826. "# r * div - Start\n\t"
  2827. "lsl %[d1], %[r], #16\n\t"
  2828. "lsl r4, %[div], #16\n\t"
  2829. "lsr %[d1], %[d1], #16\n\t"
  2830. "lsr r4, r4, #16\n\t"
  2831. "mul r4, %[d1]\n\t"
  2832. "lsr r6, %[div], #16\n\t"
  2833. "mul %[d1], r6\n\t"
  2834. "lsr r5, %[d1], #16\n\t"
  2835. "lsl %[d1], %[d1], #16\n\t"
  2836. "add r4, %[d1]\n\t"
  2837. "adc r5, r7\n\t"
  2838. "lsr %[d1], %[r], #16\n\t"
  2839. "mul r6, %[d1]\n\t"
  2840. "add r5, r6\n\t"
  2841. "lsl r6, %[div], #16\n\t"
  2842. "lsr r6, r6, #16\n\t"
  2843. "mul %[d1], r6\n\t"
  2844. "lsr r6, %[d1], #16\n\t"
  2845. "lsl %[d1], %[d1], #16\n\t"
  2846. "add r4, %[d1]\n\t"
  2847. "adc r5, r6\n\t"
  2848. "# r * div - Done\n\t"
  2849. "mov %[d1], r8\n\t"
  2850. "sub %[d1], r4\n\t"
  2851. "mov r4, %[d1]\n\t"
  2852. "mov %[d1], r9\n\t"
  2853. "sbc %[d1], r5\n\t"
  2854. "mov r5, %[d1]\n\t"
  2855. "add %[r], r5\n\t"
  2856. "# r * div - Start\n\t"
  2857. "lsl %[d1], %[r], #16\n\t"
  2858. "lsl r4, %[div], #16\n\t"
  2859. "lsr %[d1], %[d1], #16\n\t"
  2860. "lsr r4, r4, #16\n\t"
  2861. "mul r4, %[d1]\n\t"
  2862. "lsr r6, %[div], #16\n\t"
  2863. "mul %[d1], r6\n\t"
  2864. "lsr r5, %[d1], #16\n\t"
  2865. "lsl %[d1], %[d1], #16\n\t"
  2866. "add r4, %[d1]\n\t"
  2867. "adc r5, r7\n\t"
  2868. "lsr %[d1], %[r], #16\n\t"
  2869. "mul r6, %[d1]\n\t"
  2870. "add r5, r6\n\t"
  2871. "lsl r6, %[div], #16\n\t"
  2872. "lsr r6, r6, #16\n\t"
  2873. "mul %[d1], r6\n\t"
  2874. "lsr r6, %[d1], #16\n\t"
  2875. "lsl %[d1], %[d1], #16\n\t"
  2876. "add r4, %[d1]\n\t"
  2877. "adc r5, r6\n\t"
  2878. "# r * div - Done\n\t"
  2879. "mov %[d1], r8\n\t"
  2880. "mov r6, r9\n\t"
  2881. "sub r4, %[d1], r4\n\t"
  2882. "sbc r6, r5\n\t"
  2883. "mov r5, r6\n\t"
  2884. "add %[r], r5\n\t"
  2885. "# r * div - Start\n\t"
  2886. "lsl %[d1], %[r], #16\n\t"
  2887. "lsl r4, %[div], #16\n\t"
  2888. "lsr %[d1], %[d1], #16\n\t"
  2889. "lsr r4, r4, #16\n\t"
  2890. "mul r4, %[d1]\n\t"
  2891. "lsr r6, %[div], #16\n\t"
  2892. "mul %[d1], r6\n\t"
  2893. "lsr r5, %[d1], #16\n\t"
  2894. "lsl %[d1], %[d1], #16\n\t"
  2895. "add r4, %[d1]\n\t"
  2896. "adc r5, r7\n\t"
  2897. "lsr %[d1], %[r], #16\n\t"
  2898. "mul r6, %[d1]\n\t"
  2899. "add r5, r6\n\t"
  2900. "lsl r6, %[div], #16\n\t"
  2901. "lsr r6, r6, #16\n\t"
  2902. "mul %[d1], r6\n\t"
  2903. "lsr r6, %[d1], #16\n\t"
  2904. "lsl %[d1], %[d1], #16\n\t"
  2905. "add r4, %[d1]\n\t"
  2906. "adc r5, r6\n\t"
  2907. "# r * div - Done\n\t"
  2908. "mov %[d1], r8\n\t"
  2909. "mov r6, r9\n\t"
  2910. "sub r4, %[d1], r4\n\t"
  2911. "sbc r6, r5\n\t"
  2912. "mov r5, r6\n\t"
  2913. "add %[r], r5\n\t"
  2914. "mov r6, %[div]\n\t"
  2915. "sub r6, r4\n\t"
  2916. "sbc r6, r6\n\t"
  2917. "sub %[r], r6\n\t"
  2918. : [r] "+r" (r)
  2919. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  2920. : "r4", "r5", "r7", "r6", "r8", "r9"
  2921. );
  2922. return r;
  2923. }
  2924. /* Compare a with b in constant time.
  2925. *
  2926. * a A single precision integer.
  2927. * b A single precision integer.
  2928. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  2929. * respectively.
  2930. */
  2931. SP_NOINLINE static int32_t sp_2048_cmp_32(const sp_digit* a, const sp_digit* b)
  2932. {
  2933. sp_digit r = 0;
  2934. __asm__ __volatile__ (
  2935. "mov r3, #0\n\t"
  2936. "mvn r3, r3\n\t"
  2937. "mov r6, #124\n\t"
  2938. "1:\n\t"
  2939. "ldr r7, [%[a], r6]\n\t"
  2940. "ldr r5, [%[b], r6]\n\t"
  2941. "and r7, r3\n\t"
  2942. "and r5, r3\n\t"
  2943. "mov r4, r7\n\t"
  2944. "sub r7, r5\n\t"
  2945. "sbc r7, r7\n\t"
  2946. "add %[r], r7\n\t"
  2947. "mvn r7, r7\n\t"
  2948. "and r3, r7\n\t"
  2949. "sub r5, r4\n\t"
  2950. "sbc r7, r7\n\t"
  2951. "sub %[r], r7\n\t"
  2952. "mvn r7, r7\n\t"
  2953. "and r3, r7\n\t"
  2954. "sub r6, #4\n\t"
  2955. "cmp r6, #0\n\t"
  2956. "bge 1b\n\t"
  2957. : [r] "+r" (r)
  2958. : [a] "r" (a), [b] "r" (b)
  2959. : "r3", "r4", "r5", "r6", "r7"
  2960. );
  2961. return r;
  2962. }
  2963. /* Divide d in a and put remainder into r (m*d + r = a)
  2964. * m is not calculated as it is not needed at this time.
  2965. *
  2966. * a Nmber to be divided.
  2967. * d Number to divide with.
  2968. * m Multiplier result.
  2969. * r Remainder from the division.
  2970. * returns MP_OKAY indicating success.
  2971. */
  2972. static WC_INLINE int sp_2048_div_32(const sp_digit* a, const sp_digit* d, sp_digit* m,
  2973. sp_digit* r)
  2974. {
  2975. sp_digit t1[64], t2[33];
  2976. sp_digit div, r1;
  2977. int i;
  2978. (void)m;
  2979. div = d[31];
  2980. XMEMCPY(t1, a, sizeof(*t1) * 2 * 32);
  2981. for (i=31; i>=0; i--) {
  2982. r1 = div_2048_word_32(t1[32 + i], t1[32 + i - 1], div);
  2983. sp_2048_mul_d_32(t2, d, r1);
  2984. t1[32 + i] += sp_2048_sub_in_place_32(&t1[i], t2);
  2985. t1[32 + i] -= t2[32];
  2986. sp_2048_mask_32(t2, d, t1[32 + i]);
  2987. t1[32 + i] += sp_2048_add_32(&t1[i], &t1[i], t2);
  2988. sp_2048_mask_32(t2, d, t1[32 + i]);
  2989. t1[32 + i] += sp_2048_add_32(&t1[i], &t1[i], t2);
  2990. }
  2991. r1 = sp_2048_cmp_32(t1, d) >= 0;
  2992. sp_2048_cond_sub_32(r, t1, d, (sp_digit)0 - r1);
  2993. return MP_OKAY;
  2994. }
  2995. /* Reduce a modulo m into r. (r = a mod m)
  2996. *
  2997. * r A single precision number that is the reduced result.
  2998. * a A single precision number that is to be reduced.
  2999. * m A single precision number that is the modulus to reduce with.
  3000. * returns MP_OKAY indicating success.
  3001. */
  3002. static WC_INLINE int sp_2048_mod_32(sp_digit* r, const sp_digit* a, const sp_digit* m)
  3003. {
  3004. return sp_2048_div_32(a, m, NULL, r);
  3005. }
  3006. #ifdef WOLFSSL_SP_SMALL
  3007. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  3008. *
  3009. * r A single precision number that is the result of the operation.
  3010. * a A single precision number being exponentiated.
  3011. * e A single precision number that is the exponent.
  3012. * bits The number of bits in the exponent.
  3013. * m A single precision number that is the modulus.
  3014. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  3015. */
  3016. static int sp_2048_mod_exp_32(sp_digit* r, const sp_digit* a, const sp_digit* e,
  3017. int bits, const sp_digit* m, int reduceA)
  3018. {
  3019. #ifndef WOLFSSL_SMALL_STACK
  3020. sp_digit t[16][64];
  3021. #else
  3022. sp_digit* t[16];
  3023. sp_digit* td;
  3024. #endif
  3025. sp_digit* norm;
  3026. sp_digit mp = 1;
  3027. sp_digit n;
  3028. sp_digit mask;
  3029. int i;
  3030. int c, y;
  3031. int err = MP_OKAY;
  3032. #ifdef WOLFSSL_SMALL_STACK
  3033. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 64, NULL,
  3034. DYNAMIC_TYPE_TMP_BUFFER);
  3035. if (td == NULL) {
  3036. err = MEMORY_E;
  3037. }
  3038. #endif
  3039. if (err == MP_OKAY) {
  3040. #ifdef WOLFSSL_SMALL_STACK
  3041. for (i=0; i<16; i++) {
  3042. t[i] = td + i * 64;
  3043. }
  3044. #endif
  3045. norm = t[0];
  3046. sp_2048_mont_setup(m, &mp);
  3047. sp_2048_mont_norm_32(norm, m);
  3048. XMEMSET(t[1], 0, sizeof(sp_digit) * 32U);
  3049. if (reduceA != 0) {
  3050. err = sp_2048_mod_32(t[1] + 32, a, m);
  3051. if (err == MP_OKAY) {
  3052. err = sp_2048_mod_32(t[1], t[1], m);
  3053. }
  3054. }
  3055. else {
  3056. XMEMCPY(t[1] + 32, a, sizeof(sp_digit) * 32);
  3057. err = sp_2048_mod_32(t[1], t[1], m);
  3058. }
  3059. }
  3060. if (err == MP_OKAY) {
  3061. sp_2048_mont_sqr_32(t[ 2], t[ 1], m, mp);
  3062. sp_2048_mont_mul_32(t[ 3], t[ 2], t[ 1], m, mp);
  3063. sp_2048_mont_sqr_32(t[ 4], t[ 2], m, mp);
  3064. sp_2048_mont_mul_32(t[ 5], t[ 3], t[ 2], m, mp);
  3065. sp_2048_mont_sqr_32(t[ 6], t[ 3], m, mp);
  3066. sp_2048_mont_mul_32(t[ 7], t[ 4], t[ 3], m, mp);
  3067. sp_2048_mont_sqr_32(t[ 8], t[ 4], m, mp);
  3068. sp_2048_mont_mul_32(t[ 9], t[ 5], t[ 4], m, mp);
  3069. sp_2048_mont_sqr_32(t[10], t[ 5], m, mp);
  3070. sp_2048_mont_mul_32(t[11], t[ 6], t[ 5], m, mp);
  3071. sp_2048_mont_sqr_32(t[12], t[ 6], m, mp);
  3072. sp_2048_mont_mul_32(t[13], t[ 7], t[ 6], m, mp);
  3073. sp_2048_mont_sqr_32(t[14], t[ 7], m, mp);
  3074. sp_2048_mont_mul_32(t[15], t[ 8], t[ 7], m, mp);
  3075. i = (bits - 1) / 32;
  3076. n = e[i--];
  3077. c = bits & 31;
  3078. if (c == 0) {
  3079. c = 32;
  3080. }
  3081. c -= bits % 4;
  3082. if (c == 32) {
  3083. c = 28;
  3084. }
  3085. y = (int)(n >> c);
  3086. n <<= 32 - c;
  3087. XMEMCPY(r, t[y], sizeof(sp_digit) * 32);
  3088. for (; i>=0 || c>=4; ) {
  3089. if (c == 0) {
  3090. n = e[i--];
  3091. y = n >> 28;
  3092. n <<= 4;
  3093. c = 28;
  3094. }
  3095. else if (c < 4) {
  3096. y = n >> 28;
  3097. n = e[i--];
  3098. c = 4 - c;
  3099. y |= n >> (32 - c);
  3100. n <<= c;
  3101. c = 32 - c;
  3102. }
  3103. else {
  3104. y = (n >> 28) & 0xf;
  3105. n <<= 4;
  3106. c -= 4;
  3107. }
  3108. sp_2048_mont_sqr_32(r, r, m, mp);
  3109. sp_2048_mont_sqr_32(r, r, m, mp);
  3110. sp_2048_mont_sqr_32(r, r, m, mp);
  3111. sp_2048_mont_sqr_32(r, r, m, mp);
  3112. sp_2048_mont_mul_32(r, r, t[y], m, mp);
  3113. }
  3114. XMEMSET(&r[32], 0, sizeof(sp_digit) * 32U);
  3115. sp_2048_mont_reduce_32(r, m, mp);
  3116. mask = 0 - (sp_2048_cmp_32(r, m) >= 0);
  3117. sp_2048_cond_sub_32(r, r, m, mask);
  3118. }
  3119. #ifdef WOLFSSL_SMALL_STACK
  3120. if (td != NULL) {
  3121. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  3122. }
  3123. #endif
  3124. return err;
  3125. }
  3126. #else
  3127. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  3128. *
  3129. * r A single precision number that is the result of the operation.
  3130. * a A single precision number being exponentiated.
  3131. * e A single precision number that is the exponent.
  3132. * bits The number of bits in the exponent.
  3133. * m A single precision number that is the modulus.
  3134. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  3135. */
  3136. static int sp_2048_mod_exp_32(sp_digit* r, const sp_digit* a, const sp_digit* e,
  3137. int bits, const sp_digit* m, int reduceA)
  3138. {
  3139. #ifndef WOLFSSL_SMALL_STACK
  3140. sp_digit t[32][64];
  3141. #else
  3142. sp_digit* t[32];
  3143. sp_digit* td;
  3144. #endif
  3145. sp_digit* norm;
  3146. sp_digit mp = 1;
  3147. sp_digit n;
  3148. sp_digit mask;
  3149. int i;
  3150. int c, y;
  3151. int err = MP_OKAY;
  3152. #ifdef WOLFSSL_SMALL_STACK
  3153. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 64, NULL,
  3154. DYNAMIC_TYPE_TMP_BUFFER);
  3155. if (td == NULL) {
  3156. err = MEMORY_E;
  3157. }
  3158. #endif
  3159. if (err == MP_OKAY) {
  3160. #ifdef WOLFSSL_SMALL_STACK
  3161. for (i=0; i<32; i++) {
  3162. t[i] = td + i * 64;
  3163. }
  3164. #endif
  3165. norm = t[0];
  3166. sp_2048_mont_setup(m, &mp);
  3167. sp_2048_mont_norm_32(norm, m);
  3168. XMEMSET(t[1], 0, sizeof(sp_digit) * 32U);
  3169. if (reduceA != 0) {
  3170. err = sp_2048_mod_32(t[1] + 32, a, m);
  3171. if (err == MP_OKAY) {
  3172. err = sp_2048_mod_32(t[1], t[1], m);
  3173. }
  3174. }
  3175. else {
  3176. XMEMCPY(t[1] + 32, a, sizeof(sp_digit) * 32);
  3177. err = sp_2048_mod_32(t[1], t[1], m);
  3178. }
  3179. }
  3180. if (err == MP_OKAY) {
  3181. sp_2048_mont_sqr_32(t[ 2], t[ 1], m, mp);
  3182. sp_2048_mont_mul_32(t[ 3], t[ 2], t[ 1], m, mp);
  3183. sp_2048_mont_sqr_32(t[ 4], t[ 2], m, mp);
  3184. sp_2048_mont_mul_32(t[ 5], t[ 3], t[ 2], m, mp);
  3185. sp_2048_mont_sqr_32(t[ 6], t[ 3], m, mp);
  3186. sp_2048_mont_mul_32(t[ 7], t[ 4], t[ 3], m, mp);
  3187. sp_2048_mont_sqr_32(t[ 8], t[ 4], m, mp);
  3188. sp_2048_mont_mul_32(t[ 9], t[ 5], t[ 4], m, mp);
  3189. sp_2048_mont_sqr_32(t[10], t[ 5], m, mp);
  3190. sp_2048_mont_mul_32(t[11], t[ 6], t[ 5], m, mp);
  3191. sp_2048_mont_sqr_32(t[12], t[ 6], m, mp);
  3192. sp_2048_mont_mul_32(t[13], t[ 7], t[ 6], m, mp);
  3193. sp_2048_mont_sqr_32(t[14], t[ 7], m, mp);
  3194. sp_2048_mont_mul_32(t[15], t[ 8], t[ 7], m, mp);
  3195. sp_2048_mont_sqr_32(t[16], t[ 8], m, mp);
  3196. sp_2048_mont_mul_32(t[17], t[ 9], t[ 8], m, mp);
  3197. sp_2048_mont_sqr_32(t[18], t[ 9], m, mp);
  3198. sp_2048_mont_mul_32(t[19], t[10], t[ 9], m, mp);
  3199. sp_2048_mont_sqr_32(t[20], t[10], m, mp);
  3200. sp_2048_mont_mul_32(t[21], t[11], t[10], m, mp);
  3201. sp_2048_mont_sqr_32(t[22], t[11], m, mp);
  3202. sp_2048_mont_mul_32(t[23], t[12], t[11], m, mp);
  3203. sp_2048_mont_sqr_32(t[24], t[12], m, mp);
  3204. sp_2048_mont_mul_32(t[25], t[13], t[12], m, mp);
  3205. sp_2048_mont_sqr_32(t[26], t[13], m, mp);
  3206. sp_2048_mont_mul_32(t[27], t[14], t[13], m, mp);
  3207. sp_2048_mont_sqr_32(t[28], t[14], m, mp);
  3208. sp_2048_mont_mul_32(t[29], t[15], t[14], m, mp);
  3209. sp_2048_mont_sqr_32(t[30], t[15], m, mp);
  3210. sp_2048_mont_mul_32(t[31], t[16], t[15], m, mp);
  3211. i = (bits - 1) / 32;
  3212. n = e[i--];
  3213. c = bits & 31;
  3214. if (c == 0) {
  3215. c = 32;
  3216. }
  3217. c -= bits % 5;
  3218. if (c == 32) {
  3219. c = 27;
  3220. }
  3221. y = (int)(n >> c);
  3222. n <<= 32 - c;
  3223. XMEMCPY(r, t[y], sizeof(sp_digit) * 32);
  3224. for (; i>=0 || c>=5; ) {
  3225. if (c == 0) {
  3226. n = e[i--];
  3227. y = n >> 27;
  3228. n <<= 5;
  3229. c = 27;
  3230. }
  3231. else if (c < 5) {
  3232. y = n >> 27;
  3233. n = e[i--];
  3234. c = 5 - c;
  3235. y |= n >> (32 - c);
  3236. n <<= c;
  3237. c = 32 - c;
  3238. }
  3239. else {
  3240. y = (n >> 27) & 0x1f;
  3241. n <<= 5;
  3242. c -= 5;
  3243. }
  3244. sp_2048_mont_sqr_32(r, r, m, mp);
  3245. sp_2048_mont_sqr_32(r, r, m, mp);
  3246. sp_2048_mont_sqr_32(r, r, m, mp);
  3247. sp_2048_mont_sqr_32(r, r, m, mp);
  3248. sp_2048_mont_sqr_32(r, r, m, mp);
  3249. sp_2048_mont_mul_32(r, r, t[y], m, mp);
  3250. }
  3251. XMEMSET(&r[32], 0, sizeof(sp_digit) * 32U);
  3252. sp_2048_mont_reduce_32(r, m, mp);
  3253. mask = 0 - (sp_2048_cmp_32(r, m) >= 0);
  3254. sp_2048_cond_sub_32(r, r, m, mask);
  3255. }
  3256. #ifdef WOLFSSL_SMALL_STACK
  3257. if (td != NULL) {
  3258. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  3259. }
  3260. #endif
  3261. return err;
  3262. }
  3263. #endif /* WOLFSSL_SP_SMALL */
  3264. #endif /* (WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH) && !WOLFSSL_RSA_PUBLIC_ONLY */
  3265. #if defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)
  3266. /* r = 2^n mod m where n is the number of bits to reduce by.
  3267. * Given m must be 2048 bits, just need to subtract.
  3268. *
  3269. * r A single precision number.
  3270. * m A single precision number.
  3271. */
  3272. static void sp_2048_mont_norm_64(sp_digit* r, const sp_digit* m)
  3273. {
  3274. XMEMSET(r, 0, sizeof(sp_digit) * 64);
  3275. /* r = 2^n mod m */
  3276. sp_2048_sub_in_place_64(r, m);
  3277. }
  3278. #endif /* WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH */
  3279. /* Conditionally subtract b from a using the mask m.
  3280. * m is -1 to subtract and 0 when not copying.
  3281. *
  3282. * r A single precision number representing condition subtract result.
  3283. * a A single precision number to subtract from.
  3284. * b A single precision number to subtract.
  3285. * m Mask value to apply.
  3286. */
  3287. SP_NOINLINE static sp_digit sp_2048_cond_sub_64(sp_digit* r, const sp_digit* a,
  3288. const sp_digit* b, sp_digit m)
  3289. {
  3290. sp_digit c = 0;
  3291. __asm__ __volatile__ (
  3292. "mov r5, #1\n\t"
  3293. "lsl r5, r5, #8\n\t"
  3294. "mov r8, r5\n\t"
  3295. "mov r7, #0\n\t"
  3296. "1:\n\t"
  3297. "ldr r6, [%[b], r7]\n\t"
  3298. "and r6, %[m]\n\t"
  3299. "mov r5, #0\n\t"
  3300. "sub r5, %[c]\n\t"
  3301. "ldr r5, [%[a], r7]\n\t"
  3302. "sbc r5, r6\n\t"
  3303. "sbc %[c], %[c]\n\t"
  3304. "str r5, [%[r], r7]\n\t"
  3305. "add r7, #4\n\t"
  3306. "cmp r7, r8\n\t"
  3307. "blt 1b\n\t"
  3308. : [c] "+r" (c)
  3309. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  3310. : "memory", "r5", "r6", "r7", "r8"
  3311. );
  3312. return c;
  3313. }
  3314. /* Reduce the number back to 2048 bits using Montgomery reduction.
  3315. *
  3316. * a A single precision number to reduce in place.
  3317. * m The single precision number representing the modulus.
  3318. * mp The digit representing the negative inverse of m mod 2^n.
  3319. */
  3320. SP_NOINLINE static void sp_2048_mont_reduce_64(sp_digit* a, const sp_digit* m,
  3321. sp_digit mp)
  3322. {
  3323. sp_digit ca = 0;
  3324. __asm__ __volatile__ (
  3325. "mov r8, %[mp]\n\t"
  3326. "mov r12, %[ca]\n\t"
  3327. "mov r14, %[m]\n\t"
  3328. "mov r9, %[a]\n\t"
  3329. "mov r4, #0\n\t"
  3330. "# i = 0\n\t"
  3331. "mov r11, r4\n\t"
  3332. "\n1:\n\t"
  3333. "mov r5, #0\n\t"
  3334. "mov %[ca], #0\n\t"
  3335. "# mu = a[i] * mp\n\t"
  3336. "mov %[mp], r8\n\t"
  3337. "ldr %[a], [%[a]]\n\t"
  3338. "mul %[mp], %[a]\n\t"
  3339. "mov %[m], r14\n\t"
  3340. "mov r10, r9\n\t"
  3341. "\n2:\n\t"
  3342. "# a[i+j] += m[j] * mu\n\t"
  3343. "mov %[a], r10\n\t"
  3344. "ldr %[a], [%[a]]\n\t"
  3345. "mov %[ca], #0\n\t"
  3346. "mov r4, r5\n\t"
  3347. "mov r5, #0\n\t"
  3348. "# Multiply m[j] and mu - Start\n\t"
  3349. "ldr r7, [%[m]]\n\t"
  3350. "lsl r6, %[mp], #16\n\t"
  3351. "lsl r7, r7, #16\n\t"
  3352. "lsr r6, r6, #16\n\t"
  3353. "lsr r7, r7, #16\n\t"
  3354. "mul r7, r6\n\t"
  3355. "add %[a], r7\n\t"
  3356. "adc r5, %[ca]\n\t"
  3357. "ldr r7, [%[m]]\n\t"
  3358. "lsr r7, r7, #16\n\t"
  3359. "mul r6, r7\n\t"
  3360. "lsr r7, r6, #16\n\t"
  3361. "lsl r6, r6, #16\n\t"
  3362. "add %[a], r6\n\t"
  3363. "adc r5, r7\n\t"
  3364. "ldr r7, [%[m]]\n\t"
  3365. "lsr r6, %[mp], #16\n\t"
  3366. "lsr r7, r7, #16\n\t"
  3367. "mul r7, r6\n\t"
  3368. "add r5, r7\n\t"
  3369. "ldr r7, [%[m]]\n\t"
  3370. "lsl r7, r7, #16\n\t"
  3371. "lsr r7, r7, #16\n\t"
  3372. "mul r6, r7\n\t"
  3373. "lsr r7, r6, #16\n\t"
  3374. "lsl r6, r6, #16\n\t"
  3375. "add %[a], r6\n\t"
  3376. "adc r5, r7\n\t"
  3377. "# Multiply m[j] and mu - Done\n\t"
  3378. "add r4, %[a]\n\t"
  3379. "adc r5, %[ca]\n\t"
  3380. "mov %[a], r10\n\t"
  3381. "str r4, [%[a]]\n\t"
  3382. "mov r6, #4\n\t"
  3383. "add %[m], #4\n\t"
  3384. "add r10, r6\n\t"
  3385. "mov r4, #252\n\t"
  3386. "add r4, r9\n\t"
  3387. "cmp r10, r4\n\t"
  3388. "blt 2b\n\t"
  3389. "# a[i+63] += m[63] * mu\n\t"
  3390. "mov %[ca], #0\n\t"
  3391. "mov r4, r12\n\t"
  3392. "mov %[a], #0\n\t"
  3393. "# Multiply m[63] and mu - Start\n\t"
  3394. "ldr r7, [%[m]]\n\t"
  3395. "lsl r6, %[mp], #16\n\t"
  3396. "lsl r7, r7, #16\n\t"
  3397. "lsr r6, r6, #16\n\t"
  3398. "lsr r7, r7, #16\n\t"
  3399. "mul r7, r6\n\t"
  3400. "add r5, r7\n\t"
  3401. "adc r4, %[ca]\n\t"
  3402. "adc %[a], %[ca]\n\t"
  3403. "ldr r7, [%[m]]\n\t"
  3404. "lsr r7, r7, #16\n\t"
  3405. "mul r6, r7\n\t"
  3406. "lsr r7, r6, #16\n\t"
  3407. "lsl r6, r6, #16\n\t"
  3408. "add r5, r6\n\t"
  3409. "adc r4, r7\n\t"
  3410. "adc %[a], %[ca]\n\t"
  3411. "ldr r7, [%[m]]\n\t"
  3412. "lsr r6, %[mp], #16\n\t"
  3413. "lsr r7, r7, #16\n\t"
  3414. "mul r7, r6\n\t"
  3415. "add r4, r7\n\t"
  3416. "adc %[a], %[ca]\n\t"
  3417. "ldr r7, [%[m]]\n\t"
  3418. "lsl r7, r7, #16\n\t"
  3419. "lsr r7, r7, #16\n\t"
  3420. "mul r6, r7\n\t"
  3421. "lsr r7, r6, #16\n\t"
  3422. "lsl r6, r6, #16\n\t"
  3423. "add r5, r6\n\t"
  3424. "adc r4, r7\n\t"
  3425. "adc %[a], %[ca]\n\t"
  3426. "# Multiply m[63] and mu - Done\n\t"
  3427. "mov %[ca], %[a]\n\t"
  3428. "mov %[a], r10\n\t"
  3429. "ldr r7, [%[a], #4]\n\t"
  3430. "ldr %[a], [%[a]]\n\t"
  3431. "mov r6, #0\n\t"
  3432. "add r5, %[a]\n\t"
  3433. "adc r7, r4\n\t"
  3434. "adc %[ca], r6\n\t"
  3435. "mov %[a], r10\n\t"
  3436. "str r5, [%[a]]\n\t"
  3437. "str r7, [%[a], #4]\n\t"
  3438. "# i += 1\n\t"
  3439. "mov r6, #4\n\t"
  3440. "add r9, r6\n\t"
  3441. "add r11, r6\n\t"
  3442. "mov r12, %[ca]\n\t"
  3443. "mov %[a], r9\n\t"
  3444. "mov r4, #1\n\t"
  3445. "lsl r4, r4, #8\n\t"
  3446. "cmp r11, r4\n\t"
  3447. "blt 1b\n\t"
  3448. "mov %[m], r14\n\t"
  3449. : [ca] "+r" (ca), [a] "+r" (a)
  3450. : [m] "r" (m), [mp] "r" (mp)
  3451. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  3452. );
  3453. sp_2048_cond_sub_64(a - 64, a, m, (sp_digit)0 - ca);
  3454. }
  3455. /* Multiply two Montogmery form numbers mod the modulus (prime).
  3456. * (r = a * b mod m)
  3457. *
  3458. * r Result of multiplication.
  3459. * a First number to multiply in Montogmery form.
  3460. * b Second number to multiply in Montogmery form.
  3461. * m Modulus (prime).
  3462. * mp Montogmery mulitplier.
  3463. */
  3464. static void sp_2048_mont_mul_64(sp_digit* r, const sp_digit* a, const sp_digit* b,
  3465. const sp_digit* m, sp_digit mp)
  3466. {
  3467. sp_2048_mul_64(r, a, b);
  3468. sp_2048_mont_reduce_64(r, m, mp);
  3469. }
  3470. /* Square the Montgomery form number. (r = a * a mod m)
  3471. *
  3472. * r Result of squaring.
  3473. * a Number to square in Montogmery form.
  3474. * m Modulus (prime).
  3475. * mp Montogmery mulitplier.
  3476. */
  3477. static void sp_2048_mont_sqr_64(sp_digit* r, const sp_digit* a, const sp_digit* m,
  3478. sp_digit mp)
  3479. {
  3480. sp_2048_sqr_64(r, a);
  3481. sp_2048_mont_reduce_64(r, m, mp);
  3482. }
  3483. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  3484. *
  3485. * d1 The high order half of the number to divide.
  3486. * d0 The low order half of the number to divide.
  3487. * div The dividend.
  3488. * returns the result of the division.
  3489. *
  3490. * Note that this is an approximate div. It may give an answer 1 larger.
  3491. */
  3492. SP_NOINLINE static sp_digit div_2048_word_64(sp_digit d1, sp_digit d0,
  3493. sp_digit div)
  3494. {
  3495. sp_digit r = 0;
  3496. __asm__ __volatile__ (
  3497. "lsr r5, %[div], #1\n\t"
  3498. "add r5, #1\n\t"
  3499. "mov r8, %[d0]\n\t"
  3500. "mov r9, %[d1]\n\t"
  3501. "# Do top 32\n\t"
  3502. "mov r6, r5\n\t"
  3503. "sub r6, %[d1]\n\t"
  3504. "sbc r6, r6\n\t"
  3505. "add %[r], %[r]\n\t"
  3506. "sub %[r], r6\n\t"
  3507. "and r6, r5\n\t"
  3508. "sub %[d1], r6\n\t"
  3509. "# Next 30 bits\n\t"
  3510. "mov r4, #29\n\t"
  3511. "1:\n\t"
  3512. "lsl %[d0], %[d0], #1\n\t"
  3513. "adc %[d1], %[d1]\n\t"
  3514. "mov r6, r5\n\t"
  3515. "sub r6, %[d1]\n\t"
  3516. "sbc r6, r6\n\t"
  3517. "add %[r], %[r]\n\t"
  3518. "sub %[r], r6\n\t"
  3519. "and r6, r5\n\t"
  3520. "sub %[d1], r6\n\t"
  3521. "sub r4, #1\n\t"
  3522. "bpl 1b\n\t"
  3523. "mov r7, #0\n\t"
  3524. "add %[r], %[r]\n\t"
  3525. "add %[r], #1\n\t"
  3526. "# r * div - Start\n\t"
  3527. "lsl %[d1], %[r], #16\n\t"
  3528. "lsl r4, %[div], #16\n\t"
  3529. "lsr %[d1], %[d1], #16\n\t"
  3530. "lsr r4, r4, #16\n\t"
  3531. "mul r4, %[d1]\n\t"
  3532. "lsr r6, %[div], #16\n\t"
  3533. "mul %[d1], r6\n\t"
  3534. "lsr r5, %[d1], #16\n\t"
  3535. "lsl %[d1], %[d1], #16\n\t"
  3536. "add r4, %[d1]\n\t"
  3537. "adc r5, r7\n\t"
  3538. "lsr %[d1], %[r], #16\n\t"
  3539. "mul r6, %[d1]\n\t"
  3540. "add r5, r6\n\t"
  3541. "lsl r6, %[div], #16\n\t"
  3542. "lsr r6, r6, #16\n\t"
  3543. "mul %[d1], r6\n\t"
  3544. "lsr r6, %[d1], #16\n\t"
  3545. "lsl %[d1], %[d1], #16\n\t"
  3546. "add r4, %[d1]\n\t"
  3547. "adc r5, r6\n\t"
  3548. "# r * div - Done\n\t"
  3549. "mov %[d1], r8\n\t"
  3550. "sub %[d1], r4\n\t"
  3551. "mov r4, %[d1]\n\t"
  3552. "mov %[d1], r9\n\t"
  3553. "sbc %[d1], r5\n\t"
  3554. "mov r5, %[d1]\n\t"
  3555. "add %[r], r5\n\t"
  3556. "# r * div - Start\n\t"
  3557. "lsl %[d1], %[r], #16\n\t"
  3558. "lsl r4, %[div], #16\n\t"
  3559. "lsr %[d1], %[d1], #16\n\t"
  3560. "lsr r4, r4, #16\n\t"
  3561. "mul r4, %[d1]\n\t"
  3562. "lsr r6, %[div], #16\n\t"
  3563. "mul %[d1], r6\n\t"
  3564. "lsr r5, %[d1], #16\n\t"
  3565. "lsl %[d1], %[d1], #16\n\t"
  3566. "add r4, %[d1]\n\t"
  3567. "adc r5, r7\n\t"
  3568. "lsr %[d1], %[r], #16\n\t"
  3569. "mul r6, %[d1]\n\t"
  3570. "add r5, r6\n\t"
  3571. "lsl r6, %[div], #16\n\t"
  3572. "lsr r6, r6, #16\n\t"
  3573. "mul %[d1], r6\n\t"
  3574. "lsr r6, %[d1], #16\n\t"
  3575. "lsl %[d1], %[d1], #16\n\t"
  3576. "add r4, %[d1]\n\t"
  3577. "adc r5, r6\n\t"
  3578. "# r * div - Done\n\t"
  3579. "mov %[d1], r8\n\t"
  3580. "mov r6, r9\n\t"
  3581. "sub r4, %[d1], r4\n\t"
  3582. "sbc r6, r5\n\t"
  3583. "mov r5, r6\n\t"
  3584. "add %[r], r5\n\t"
  3585. "# r * div - Start\n\t"
  3586. "lsl %[d1], %[r], #16\n\t"
  3587. "lsl r4, %[div], #16\n\t"
  3588. "lsr %[d1], %[d1], #16\n\t"
  3589. "lsr r4, r4, #16\n\t"
  3590. "mul r4, %[d1]\n\t"
  3591. "lsr r6, %[div], #16\n\t"
  3592. "mul %[d1], r6\n\t"
  3593. "lsr r5, %[d1], #16\n\t"
  3594. "lsl %[d1], %[d1], #16\n\t"
  3595. "add r4, %[d1]\n\t"
  3596. "adc r5, r7\n\t"
  3597. "lsr %[d1], %[r], #16\n\t"
  3598. "mul r6, %[d1]\n\t"
  3599. "add r5, r6\n\t"
  3600. "lsl r6, %[div], #16\n\t"
  3601. "lsr r6, r6, #16\n\t"
  3602. "mul %[d1], r6\n\t"
  3603. "lsr r6, %[d1], #16\n\t"
  3604. "lsl %[d1], %[d1], #16\n\t"
  3605. "add r4, %[d1]\n\t"
  3606. "adc r5, r6\n\t"
  3607. "# r * div - Done\n\t"
  3608. "mov %[d1], r8\n\t"
  3609. "mov r6, r9\n\t"
  3610. "sub r4, %[d1], r4\n\t"
  3611. "sbc r6, r5\n\t"
  3612. "mov r5, r6\n\t"
  3613. "add %[r], r5\n\t"
  3614. "mov r6, %[div]\n\t"
  3615. "sub r6, r4\n\t"
  3616. "sbc r6, r6\n\t"
  3617. "sub %[r], r6\n\t"
  3618. : [r] "+r" (r)
  3619. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  3620. : "r4", "r5", "r7", "r6", "r8", "r9"
  3621. );
  3622. return r;
  3623. }
  3624. /* AND m into each word of a and store in r.
  3625. *
  3626. * r A single precision integer.
  3627. * a A single precision integer.
  3628. * m Mask to AND against each digit.
  3629. */
  3630. static void sp_2048_mask_64(sp_digit* r, const sp_digit* a, sp_digit m)
  3631. {
  3632. #ifdef WOLFSSL_SP_SMALL
  3633. int i;
  3634. for (i=0; i<64; i++) {
  3635. r[i] = a[i] & m;
  3636. }
  3637. #else
  3638. int i;
  3639. for (i = 0; i < 64; i += 8) {
  3640. r[i+0] = a[i+0] & m;
  3641. r[i+1] = a[i+1] & m;
  3642. r[i+2] = a[i+2] & m;
  3643. r[i+3] = a[i+3] & m;
  3644. r[i+4] = a[i+4] & m;
  3645. r[i+5] = a[i+5] & m;
  3646. r[i+6] = a[i+6] & m;
  3647. r[i+7] = a[i+7] & m;
  3648. }
  3649. #endif
  3650. }
  3651. /* Compare a with b in constant time.
  3652. *
  3653. * a A single precision integer.
  3654. * b A single precision integer.
  3655. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  3656. * respectively.
  3657. */
  3658. SP_NOINLINE static int32_t sp_2048_cmp_64(const sp_digit* a, const sp_digit* b)
  3659. {
  3660. sp_digit r = 0;
  3661. __asm__ __volatile__ (
  3662. "mov r3, #0\n\t"
  3663. "mvn r3, r3\n\t"
  3664. "mov r6, #252\n\t"
  3665. "1:\n\t"
  3666. "ldr r7, [%[a], r6]\n\t"
  3667. "ldr r5, [%[b], r6]\n\t"
  3668. "and r7, r3\n\t"
  3669. "and r5, r3\n\t"
  3670. "mov r4, r7\n\t"
  3671. "sub r7, r5\n\t"
  3672. "sbc r7, r7\n\t"
  3673. "add %[r], r7\n\t"
  3674. "mvn r7, r7\n\t"
  3675. "and r3, r7\n\t"
  3676. "sub r5, r4\n\t"
  3677. "sbc r7, r7\n\t"
  3678. "sub %[r], r7\n\t"
  3679. "mvn r7, r7\n\t"
  3680. "and r3, r7\n\t"
  3681. "sub r6, #4\n\t"
  3682. "cmp r6, #0\n\t"
  3683. "bge 1b\n\t"
  3684. : [r] "+r" (r)
  3685. : [a] "r" (a), [b] "r" (b)
  3686. : "r3", "r4", "r5", "r6", "r7"
  3687. );
  3688. return r;
  3689. }
  3690. /* Divide d in a and put remainder into r (m*d + r = a)
  3691. * m is not calculated as it is not needed at this time.
  3692. *
  3693. * a Nmber to be divided.
  3694. * d Number to divide with.
  3695. * m Multiplier result.
  3696. * r Remainder from the division.
  3697. * returns MP_OKAY indicating success.
  3698. */
  3699. static WC_INLINE int sp_2048_div_64(const sp_digit* a, const sp_digit* d, sp_digit* m,
  3700. sp_digit* r)
  3701. {
  3702. sp_digit t1[128], t2[65];
  3703. sp_digit div, r1;
  3704. int i;
  3705. (void)m;
  3706. div = d[63];
  3707. XMEMCPY(t1, a, sizeof(*t1) * 2 * 64);
  3708. for (i=63; i>=0; i--) {
  3709. r1 = div_2048_word_64(t1[64 + i], t1[64 + i - 1], div);
  3710. sp_2048_mul_d_64(t2, d, r1);
  3711. t1[64 + i] += sp_2048_sub_in_place_64(&t1[i], t2);
  3712. t1[64 + i] -= t2[64];
  3713. sp_2048_mask_64(t2, d, t1[64 + i]);
  3714. t1[64 + i] += sp_2048_add_64(&t1[i], &t1[i], t2);
  3715. sp_2048_mask_64(t2, d, t1[64 + i]);
  3716. t1[64 + i] += sp_2048_add_64(&t1[i], &t1[i], t2);
  3717. }
  3718. r1 = sp_2048_cmp_64(t1, d) >= 0;
  3719. sp_2048_cond_sub_64(r, t1, d, (sp_digit)0 - r1);
  3720. return MP_OKAY;
  3721. }
  3722. /* Reduce a modulo m into r. (r = a mod m)
  3723. *
  3724. * r A single precision number that is the reduced result.
  3725. * a A single precision number that is to be reduced.
  3726. * m A single precision number that is the modulus to reduce with.
  3727. * returns MP_OKAY indicating success.
  3728. */
  3729. static WC_INLINE int sp_2048_mod_64(sp_digit* r, const sp_digit* a, const sp_digit* m)
  3730. {
  3731. return sp_2048_div_64(a, m, NULL, r);
  3732. }
  3733. /* Divide d in a and put remainder into r (m*d + r = a)
  3734. * m is not calculated as it is not needed at this time.
  3735. *
  3736. * a Nmber to be divided.
  3737. * d Number to divide with.
  3738. * m Multiplier result.
  3739. * r Remainder from the division.
  3740. * returns MP_OKAY indicating success.
  3741. */
  3742. static WC_INLINE int sp_2048_div_64_cond(const sp_digit* a, const sp_digit* d, sp_digit* m,
  3743. sp_digit* r)
  3744. {
  3745. sp_digit t1[128], t2[65];
  3746. sp_digit div, r1;
  3747. int i;
  3748. (void)m;
  3749. div = d[63];
  3750. XMEMCPY(t1, a, sizeof(*t1) * 2 * 64);
  3751. for (i=63; i>=0; i--) {
  3752. r1 = div_2048_word_64(t1[64 + i], t1[64 + i - 1], div);
  3753. sp_2048_mul_d_64(t2, d, r1);
  3754. t1[64 + i] += sp_2048_sub_in_place_64(&t1[i], t2);
  3755. t1[64 + i] -= t2[64];
  3756. if (t1[64 + i] != 0) {
  3757. t1[64 + i] += sp_2048_add_64(&t1[i], &t1[i], d);
  3758. if (t1[64 + i] != 0)
  3759. t1[64 + i] += sp_2048_add_64(&t1[i], &t1[i], d);
  3760. }
  3761. }
  3762. r1 = sp_2048_cmp_64(t1, d) >= 0;
  3763. sp_2048_cond_sub_64(r, t1, d, (sp_digit)0 - r1);
  3764. return MP_OKAY;
  3765. }
  3766. /* Reduce a modulo m into r. (r = a mod m)
  3767. *
  3768. * r A single precision number that is the reduced result.
  3769. * a A single precision number that is to be reduced.
  3770. * m A single precision number that is the modulus to reduce with.
  3771. * returns MP_OKAY indicating success.
  3772. */
  3773. static WC_INLINE int sp_2048_mod_64_cond(sp_digit* r, const sp_digit* a, const sp_digit* m)
  3774. {
  3775. return sp_2048_div_64_cond(a, m, NULL, r);
  3776. }
  3777. #if (defined(WOLFSSL_HAVE_SP_RSA) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)) || \
  3778. defined(WOLFSSL_HAVE_SP_DH)
  3779. #ifdef WOLFSSL_SP_SMALL
  3780. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  3781. *
  3782. * r A single precision number that is the result of the operation.
  3783. * a A single precision number being exponentiated.
  3784. * e A single precision number that is the exponent.
  3785. * bits The number of bits in the exponent.
  3786. * m A single precision number that is the modulus.
  3787. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  3788. */
  3789. static int sp_2048_mod_exp_64(sp_digit* r, const sp_digit* a, const sp_digit* e,
  3790. int bits, const sp_digit* m, int reduceA)
  3791. {
  3792. #ifndef WOLFSSL_SMALL_STACK
  3793. sp_digit t[16][128];
  3794. #else
  3795. sp_digit* t[16];
  3796. sp_digit* td;
  3797. #endif
  3798. sp_digit* norm;
  3799. sp_digit mp = 1;
  3800. sp_digit n;
  3801. sp_digit mask;
  3802. int i;
  3803. int c, y;
  3804. int err = MP_OKAY;
  3805. #ifdef WOLFSSL_SMALL_STACK
  3806. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 128, NULL,
  3807. DYNAMIC_TYPE_TMP_BUFFER);
  3808. if (td == NULL) {
  3809. err = MEMORY_E;
  3810. }
  3811. #endif
  3812. if (err == MP_OKAY) {
  3813. #ifdef WOLFSSL_SMALL_STACK
  3814. for (i=0; i<16; i++) {
  3815. t[i] = td + i * 128;
  3816. }
  3817. #endif
  3818. norm = t[0];
  3819. sp_2048_mont_setup(m, &mp);
  3820. sp_2048_mont_norm_64(norm, m);
  3821. XMEMSET(t[1], 0, sizeof(sp_digit) * 64U);
  3822. if (reduceA != 0) {
  3823. err = sp_2048_mod_64(t[1] + 64, a, m);
  3824. if (err == MP_OKAY) {
  3825. err = sp_2048_mod_64(t[1], t[1], m);
  3826. }
  3827. }
  3828. else {
  3829. XMEMCPY(t[1] + 64, a, sizeof(sp_digit) * 64);
  3830. err = sp_2048_mod_64(t[1], t[1], m);
  3831. }
  3832. }
  3833. if (err == MP_OKAY) {
  3834. sp_2048_mont_sqr_64(t[ 2], t[ 1], m, mp);
  3835. sp_2048_mont_mul_64(t[ 3], t[ 2], t[ 1], m, mp);
  3836. sp_2048_mont_sqr_64(t[ 4], t[ 2], m, mp);
  3837. sp_2048_mont_mul_64(t[ 5], t[ 3], t[ 2], m, mp);
  3838. sp_2048_mont_sqr_64(t[ 6], t[ 3], m, mp);
  3839. sp_2048_mont_mul_64(t[ 7], t[ 4], t[ 3], m, mp);
  3840. sp_2048_mont_sqr_64(t[ 8], t[ 4], m, mp);
  3841. sp_2048_mont_mul_64(t[ 9], t[ 5], t[ 4], m, mp);
  3842. sp_2048_mont_sqr_64(t[10], t[ 5], m, mp);
  3843. sp_2048_mont_mul_64(t[11], t[ 6], t[ 5], m, mp);
  3844. sp_2048_mont_sqr_64(t[12], t[ 6], m, mp);
  3845. sp_2048_mont_mul_64(t[13], t[ 7], t[ 6], m, mp);
  3846. sp_2048_mont_sqr_64(t[14], t[ 7], m, mp);
  3847. sp_2048_mont_mul_64(t[15], t[ 8], t[ 7], m, mp);
  3848. i = (bits - 1) / 32;
  3849. n = e[i--];
  3850. c = bits & 31;
  3851. if (c == 0) {
  3852. c = 32;
  3853. }
  3854. c -= bits % 4;
  3855. if (c == 32) {
  3856. c = 28;
  3857. }
  3858. y = (int)(n >> c);
  3859. n <<= 32 - c;
  3860. XMEMCPY(r, t[y], sizeof(sp_digit) * 64);
  3861. for (; i>=0 || c>=4; ) {
  3862. if (c == 0) {
  3863. n = e[i--];
  3864. y = n >> 28;
  3865. n <<= 4;
  3866. c = 28;
  3867. }
  3868. else if (c < 4) {
  3869. y = n >> 28;
  3870. n = e[i--];
  3871. c = 4 - c;
  3872. y |= n >> (32 - c);
  3873. n <<= c;
  3874. c = 32 - c;
  3875. }
  3876. else {
  3877. y = (n >> 28) & 0xf;
  3878. n <<= 4;
  3879. c -= 4;
  3880. }
  3881. sp_2048_mont_sqr_64(r, r, m, mp);
  3882. sp_2048_mont_sqr_64(r, r, m, mp);
  3883. sp_2048_mont_sqr_64(r, r, m, mp);
  3884. sp_2048_mont_sqr_64(r, r, m, mp);
  3885. sp_2048_mont_mul_64(r, r, t[y], m, mp);
  3886. }
  3887. XMEMSET(&r[64], 0, sizeof(sp_digit) * 64U);
  3888. sp_2048_mont_reduce_64(r, m, mp);
  3889. mask = 0 - (sp_2048_cmp_64(r, m) >= 0);
  3890. sp_2048_cond_sub_64(r, r, m, mask);
  3891. }
  3892. #ifdef WOLFSSL_SMALL_STACK
  3893. if (td != NULL) {
  3894. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  3895. }
  3896. #endif
  3897. return err;
  3898. }
  3899. #else
  3900. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  3901. *
  3902. * r A single precision number that is the result of the operation.
  3903. * a A single precision number being exponentiated.
  3904. * e A single precision number that is the exponent.
  3905. * bits The number of bits in the exponent.
  3906. * m A single precision number that is the modulus.
  3907. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  3908. */
  3909. static int sp_2048_mod_exp_64(sp_digit* r, const sp_digit* a, const sp_digit* e,
  3910. int bits, const sp_digit* m, int reduceA)
  3911. {
  3912. #ifndef WOLFSSL_SMALL_STACK
  3913. sp_digit t[32][128];
  3914. #else
  3915. sp_digit* t[32];
  3916. sp_digit* td;
  3917. #endif
  3918. sp_digit* norm;
  3919. sp_digit mp = 1;
  3920. sp_digit n;
  3921. sp_digit mask;
  3922. int i;
  3923. int c, y;
  3924. int err = MP_OKAY;
  3925. #ifdef WOLFSSL_SMALL_STACK
  3926. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 128, NULL,
  3927. DYNAMIC_TYPE_TMP_BUFFER);
  3928. if (td == NULL) {
  3929. err = MEMORY_E;
  3930. }
  3931. #endif
  3932. if (err == MP_OKAY) {
  3933. #ifdef WOLFSSL_SMALL_STACK
  3934. for (i=0; i<32; i++) {
  3935. t[i] = td + i * 128;
  3936. }
  3937. #endif
  3938. norm = t[0];
  3939. sp_2048_mont_setup(m, &mp);
  3940. sp_2048_mont_norm_64(norm, m);
  3941. XMEMSET(t[1], 0, sizeof(sp_digit) * 64U);
  3942. if (reduceA != 0) {
  3943. err = sp_2048_mod_64(t[1] + 64, a, m);
  3944. if (err == MP_OKAY) {
  3945. err = sp_2048_mod_64(t[1], t[1], m);
  3946. }
  3947. }
  3948. else {
  3949. XMEMCPY(t[1] + 64, a, sizeof(sp_digit) * 64);
  3950. err = sp_2048_mod_64(t[1], t[1], m);
  3951. }
  3952. }
  3953. if (err == MP_OKAY) {
  3954. sp_2048_mont_sqr_64(t[ 2], t[ 1], m, mp);
  3955. sp_2048_mont_mul_64(t[ 3], t[ 2], t[ 1], m, mp);
  3956. sp_2048_mont_sqr_64(t[ 4], t[ 2], m, mp);
  3957. sp_2048_mont_mul_64(t[ 5], t[ 3], t[ 2], m, mp);
  3958. sp_2048_mont_sqr_64(t[ 6], t[ 3], m, mp);
  3959. sp_2048_mont_mul_64(t[ 7], t[ 4], t[ 3], m, mp);
  3960. sp_2048_mont_sqr_64(t[ 8], t[ 4], m, mp);
  3961. sp_2048_mont_mul_64(t[ 9], t[ 5], t[ 4], m, mp);
  3962. sp_2048_mont_sqr_64(t[10], t[ 5], m, mp);
  3963. sp_2048_mont_mul_64(t[11], t[ 6], t[ 5], m, mp);
  3964. sp_2048_mont_sqr_64(t[12], t[ 6], m, mp);
  3965. sp_2048_mont_mul_64(t[13], t[ 7], t[ 6], m, mp);
  3966. sp_2048_mont_sqr_64(t[14], t[ 7], m, mp);
  3967. sp_2048_mont_mul_64(t[15], t[ 8], t[ 7], m, mp);
  3968. sp_2048_mont_sqr_64(t[16], t[ 8], m, mp);
  3969. sp_2048_mont_mul_64(t[17], t[ 9], t[ 8], m, mp);
  3970. sp_2048_mont_sqr_64(t[18], t[ 9], m, mp);
  3971. sp_2048_mont_mul_64(t[19], t[10], t[ 9], m, mp);
  3972. sp_2048_mont_sqr_64(t[20], t[10], m, mp);
  3973. sp_2048_mont_mul_64(t[21], t[11], t[10], m, mp);
  3974. sp_2048_mont_sqr_64(t[22], t[11], m, mp);
  3975. sp_2048_mont_mul_64(t[23], t[12], t[11], m, mp);
  3976. sp_2048_mont_sqr_64(t[24], t[12], m, mp);
  3977. sp_2048_mont_mul_64(t[25], t[13], t[12], m, mp);
  3978. sp_2048_mont_sqr_64(t[26], t[13], m, mp);
  3979. sp_2048_mont_mul_64(t[27], t[14], t[13], m, mp);
  3980. sp_2048_mont_sqr_64(t[28], t[14], m, mp);
  3981. sp_2048_mont_mul_64(t[29], t[15], t[14], m, mp);
  3982. sp_2048_mont_sqr_64(t[30], t[15], m, mp);
  3983. sp_2048_mont_mul_64(t[31], t[16], t[15], m, mp);
  3984. i = (bits - 1) / 32;
  3985. n = e[i--];
  3986. c = bits & 31;
  3987. if (c == 0) {
  3988. c = 32;
  3989. }
  3990. c -= bits % 5;
  3991. if (c == 32) {
  3992. c = 27;
  3993. }
  3994. y = (int)(n >> c);
  3995. n <<= 32 - c;
  3996. XMEMCPY(r, t[y], sizeof(sp_digit) * 64);
  3997. for (; i>=0 || c>=5; ) {
  3998. if (c == 0) {
  3999. n = e[i--];
  4000. y = n >> 27;
  4001. n <<= 5;
  4002. c = 27;
  4003. }
  4004. else if (c < 5) {
  4005. y = n >> 27;
  4006. n = e[i--];
  4007. c = 5 - c;
  4008. y |= n >> (32 - c);
  4009. n <<= c;
  4010. c = 32 - c;
  4011. }
  4012. else {
  4013. y = (n >> 27) & 0x1f;
  4014. n <<= 5;
  4015. c -= 5;
  4016. }
  4017. sp_2048_mont_sqr_64(r, r, m, mp);
  4018. sp_2048_mont_sqr_64(r, r, m, mp);
  4019. sp_2048_mont_sqr_64(r, r, m, mp);
  4020. sp_2048_mont_sqr_64(r, r, m, mp);
  4021. sp_2048_mont_sqr_64(r, r, m, mp);
  4022. sp_2048_mont_mul_64(r, r, t[y], m, mp);
  4023. }
  4024. XMEMSET(&r[64], 0, sizeof(sp_digit) * 64U);
  4025. sp_2048_mont_reduce_64(r, m, mp);
  4026. mask = 0 - (sp_2048_cmp_64(r, m) >= 0);
  4027. sp_2048_cond_sub_64(r, r, m, mask);
  4028. }
  4029. #ifdef WOLFSSL_SMALL_STACK
  4030. if (td != NULL) {
  4031. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  4032. }
  4033. #endif
  4034. return err;
  4035. }
  4036. #endif /* WOLFSSL_SP_SMALL */
  4037. #endif /* (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) || WOLFSSL_HAVE_SP_DH */
  4038. #ifdef WOLFSSL_HAVE_SP_RSA
  4039. /* RSA public key operation.
  4040. *
  4041. * in Array of bytes representing the number to exponentiate, base.
  4042. * inLen Number of bytes in base.
  4043. * em Public exponent.
  4044. * mm Modulus.
  4045. * out Buffer to hold big-endian bytes of exponentiation result.
  4046. * Must be at least 256 bytes long.
  4047. * outLen Number of bytes in result.
  4048. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  4049. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  4050. */
  4051. int sp_RsaPublic_2048(const byte* in, word32 inLen, mp_int* em, mp_int* mm,
  4052. byte* out, word32* outLen)
  4053. {
  4054. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  4055. sp_digit ad[128], md[64], rd[128];
  4056. #else
  4057. sp_digit* d = NULL;
  4058. #endif
  4059. sp_digit* a;
  4060. sp_digit *ah;
  4061. sp_digit* m;
  4062. sp_digit* r;
  4063. sp_digit e[1];
  4064. int err = MP_OKAY;
  4065. if (*outLen < 256)
  4066. err = MP_TO_E;
  4067. if (err == MP_OKAY && (mp_count_bits(em) > 32 || inLen > 256 ||
  4068. mp_count_bits(mm) != 2048))
  4069. err = MP_READ_E;
  4070. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  4071. if (err == MP_OKAY) {
  4072. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 64 * 5, NULL,
  4073. DYNAMIC_TYPE_RSA);
  4074. if (d == NULL)
  4075. err = MEMORY_E;
  4076. }
  4077. if (err == MP_OKAY) {
  4078. a = d;
  4079. r = a + 64 * 2;
  4080. m = r + 64 * 2;
  4081. ah = a + 64;
  4082. }
  4083. #else
  4084. a = ad;
  4085. m = md;
  4086. r = rd;
  4087. ah = a + 64;
  4088. #endif
  4089. if (err == MP_OKAY) {
  4090. sp_2048_from_bin(ah, 64, in, inLen);
  4091. #if DIGIT_BIT >= 32
  4092. e[0] = em->dp[0];
  4093. #else
  4094. e[0] = em->dp[0];
  4095. if (em->used > 1)
  4096. e[0] |= ((sp_digit)em->dp[1]) << DIGIT_BIT;
  4097. #endif
  4098. if (e[0] == 0)
  4099. err = MP_EXPTMOD_E;
  4100. }
  4101. if (err == MP_OKAY) {
  4102. sp_2048_from_mp(m, 64, mm);
  4103. if (e[0] == 0x3) {
  4104. if (err == MP_OKAY) {
  4105. sp_2048_sqr_64(r, ah);
  4106. err = sp_2048_mod_64_cond(r, r, m);
  4107. }
  4108. if (err == MP_OKAY) {
  4109. sp_2048_mul_64(r, ah, r);
  4110. err = sp_2048_mod_64_cond(r, r, m);
  4111. }
  4112. }
  4113. else {
  4114. int i;
  4115. sp_digit mp;
  4116. sp_2048_mont_setup(m, &mp);
  4117. /* Convert to Montgomery form. */
  4118. XMEMSET(a, 0, sizeof(sp_digit) * 64);
  4119. err = sp_2048_mod_64_cond(a, a, m);
  4120. if (err == MP_OKAY) {
  4121. for (i=31; i>=0; i--)
  4122. if (e[0] >> i)
  4123. break;
  4124. XMEMCPY(r, a, sizeof(sp_digit) * 64);
  4125. for (i--; i>=0; i--) {
  4126. sp_2048_mont_sqr_64(r, r, m, mp);
  4127. if (((e[0] >> i) & 1) == 1)
  4128. sp_2048_mont_mul_64(r, r, a, m, mp);
  4129. }
  4130. XMEMSET(&r[64], 0, sizeof(sp_digit) * 64);
  4131. sp_2048_mont_reduce_64(r, m, mp);
  4132. for (i = 63; i > 0; i--) {
  4133. if (r[i] != m[i])
  4134. break;
  4135. }
  4136. if (r[i] >= m[i])
  4137. sp_2048_sub_in_place_64(r, m);
  4138. }
  4139. }
  4140. }
  4141. if (err == MP_OKAY) {
  4142. sp_2048_to_bin(r, out);
  4143. *outLen = 256;
  4144. }
  4145. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  4146. if (d != NULL)
  4147. XFREE(d, NULL, DYNAMIC_TYPE_RSA);
  4148. #endif
  4149. return err;
  4150. }
  4151. /* RSA private key operation.
  4152. *
  4153. * in Array of bytes representing the number to exponentiate, base.
  4154. * inLen Number of bytes in base.
  4155. * dm Private exponent.
  4156. * pm First prime.
  4157. * qm Second prime.
  4158. * dpm First prime's CRT exponent.
  4159. * dqm Second prime's CRT exponent.
  4160. * qim Inverse of second prime mod p.
  4161. * mm Modulus.
  4162. * out Buffer to hold big-endian bytes of exponentiation result.
  4163. * Must be at least 256 bytes long.
  4164. * outLen Number of bytes in result.
  4165. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  4166. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  4167. */
  4168. int sp_RsaPrivate_2048(const byte* in, word32 inLen, mp_int* dm,
  4169. mp_int* pm, mp_int* qm, mp_int* dpm, mp_int* dqm, mp_int* qim, mp_int* mm,
  4170. byte* out, word32* outLen)
  4171. {
  4172. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  4173. sp_digit ad[64 * 2];
  4174. sp_digit pd[32], qd[32], dpd[32];
  4175. sp_digit tmpad[64], tmpbd[64];
  4176. #else
  4177. sp_digit* t = NULL;
  4178. #endif
  4179. sp_digit* a;
  4180. sp_digit* p;
  4181. sp_digit* q;
  4182. sp_digit* dp;
  4183. sp_digit* dq;
  4184. sp_digit* qi;
  4185. sp_digit* tmp;
  4186. sp_digit* tmpa;
  4187. sp_digit* tmpb;
  4188. sp_digit* r;
  4189. sp_digit c;
  4190. int err = MP_OKAY;
  4191. (void)dm;
  4192. (void)mm;
  4193. if (*outLen < 256)
  4194. err = MP_TO_E;
  4195. if (err == MP_OKAY && (inLen > 256 || mp_count_bits(mm) != 2048))
  4196. err = MP_READ_E;
  4197. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  4198. if (err == MP_OKAY) {
  4199. t = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 11, NULL,
  4200. DYNAMIC_TYPE_RSA);
  4201. if (t == NULL)
  4202. err = MEMORY_E;
  4203. }
  4204. if (err == MP_OKAY) {
  4205. a = t;
  4206. p = a + 64 * 2;
  4207. q = p + 32;
  4208. qi = dq = dp = q + 32;
  4209. tmpa = qi + 32;
  4210. tmpb = tmpa + 64;
  4211. tmp = t;
  4212. r = tmp + 64;
  4213. }
  4214. #else
  4215. r = a = ad;
  4216. p = pd;
  4217. q = qd;
  4218. qi = dq = dp = dpd;
  4219. tmpa = tmpad;
  4220. tmpb = tmpbd;
  4221. tmp = a + 64;
  4222. #endif
  4223. if (err == MP_OKAY) {
  4224. sp_2048_from_bin(a, 64, in, inLen);
  4225. sp_2048_from_mp(p, 32, pm);
  4226. sp_2048_from_mp(q, 32, qm);
  4227. sp_2048_from_mp(dp, 32, dpm);
  4228. err = sp_2048_mod_exp_32(tmpa, a, dp, 1024, p, 1);
  4229. }
  4230. if (err == MP_OKAY) {
  4231. sp_2048_from_mp(dq, 32, dqm);
  4232. err = sp_2048_mod_exp_32(tmpb, a, dq, 1024, q, 1);
  4233. }
  4234. if (err == MP_OKAY) {
  4235. c = sp_2048_sub_in_place_32(tmpa, tmpb);
  4236. sp_2048_mask_32(tmp, p, c);
  4237. sp_2048_add_32(tmpa, tmpa, tmp);
  4238. sp_2048_from_mp(qi, 32, qim);
  4239. sp_2048_mul_32(tmpa, tmpa, qi);
  4240. err = sp_2048_mod_32(tmpa, tmpa, p);
  4241. }
  4242. if (err == MP_OKAY) {
  4243. sp_2048_mul_32(tmpa, q, tmpa);
  4244. XMEMSET(&tmpb[32], 0, sizeof(sp_digit) * 32);
  4245. sp_2048_add_64(r, tmpb, tmpa);
  4246. sp_2048_to_bin(r, out);
  4247. *outLen = 256;
  4248. }
  4249. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  4250. if (t != NULL) {
  4251. XMEMSET(t, 0, sizeof(sp_digit) * 32 * 11);
  4252. XFREE(t, NULL, DYNAMIC_TYPE_RSA);
  4253. }
  4254. #else
  4255. XMEMSET(tmpad, 0, sizeof(tmpad));
  4256. XMEMSET(tmpbd, 0, sizeof(tmpbd));
  4257. XMEMSET(pd, 0, sizeof(pd));
  4258. XMEMSET(qd, 0, sizeof(qd));
  4259. XMEMSET(dpd, 0, sizeof(dpd));
  4260. #endif
  4261. return err;
  4262. }
  4263. #endif /* WOLFSSL_HAVE_SP_RSA */
  4264. #if defined(WOLFSSL_HAVE_SP_DH) || (defined(WOLFSSL_HAVE_SP_RSA) && \
  4265. !defined(WOLFSSL_RSA_PUBLIC_ONLY))
  4266. /* Convert an array of sp_digit to an mp_int.
  4267. *
  4268. * a A single precision integer.
  4269. * r A multi-precision integer.
  4270. */
  4271. static int sp_2048_to_mp(const sp_digit* a, mp_int* r)
  4272. {
  4273. int err;
  4274. err = mp_grow(r, (2048 + DIGIT_BIT - 1) / DIGIT_BIT);
  4275. if (err == MP_OKAY) { /*lint !e774 case where err is always MP_OKAY*/
  4276. #if DIGIT_BIT == 32
  4277. XMEMCPY(r->dp, a, sizeof(sp_digit) * 64);
  4278. r->used = 64;
  4279. mp_clamp(r);
  4280. #elif DIGIT_BIT < 32
  4281. int i, j = 0, s = 0;
  4282. r->dp[0] = 0;
  4283. for (i = 0; i < 64; i++) {
  4284. r->dp[j] |= a[i] << s;
  4285. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  4286. s = DIGIT_BIT - s;
  4287. r->dp[++j] = a[i] >> s;
  4288. while (s + DIGIT_BIT <= 32) {
  4289. s += DIGIT_BIT;
  4290. r->dp[j++] &= (1L << DIGIT_BIT) - 1;
  4291. if (s == SP_WORD_SIZE) {
  4292. r->dp[j] = 0;
  4293. }
  4294. else {
  4295. r->dp[j] = a[i] >> s;
  4296. }
  4297. }
  4298. s = 32 - s;
  4299. }
  4300. r->used = (2048 + DIGIT_BIT - 1) / DIGIT_BIT;
  4301. mp_clamp(r);
  4302. #else
  4303. int i, j = 0, s = 0;
  4304. r->dp[0] = 0;
  4305. for (i = 0; i < 64; i++) {
  4306. r->dp[j] |= ((mp_digit)a[i]) << s;
  4307. if (s + 32 >= DIGIT_BIT) {
  4308. #if DIGIT_BIT != 32 && DIGIT_BIT != 64
  4309. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  4310. #endif
  4311. s = DIGIT_BIT - s;
  4312. r->dp[++j] = a[i] >> s;
  4313. s = 32 - s;
  4314. }
  4315. else {
  4316. s += 32;
  4317. }
  4318. }
  4319. r->used = (2048 + DIGIT_BIT - 1) / DIGIT_BIT;
  4320. mp_clamp(r);
  4321. #endif
  4322. }
  4323. return err;
  4324. }
  4325. /* Perform the modular exponentiation for Diffie-Hellman.
  4326. *
  4327. * base Base. MP integer.
  4328. * exp Exponent. MP integer.
  4329. * mod Modulus. MP integer.
  4330. * res Result. MP integer.
  4331. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  4332. * and MEMORY_E if memory allocation fails.
  4333. */
  4334. int sp_ModExp_2048(mp_int* base, mp_int* exp, mp_int* mod, mp_int* res)
  4335. {
  4336. int err = MP_OKAY;
  4337. sp_digit b[128], e[64], m[64];
  4338. sp_digit* r = b;
  4339. int expBits = mp_count_bits(exp);
  4340. if (mp_count_bits(base) > 2048) {
  4341. err = MP_READ_E;
  4342. }
  4343. if (err == MP_OKAY) {
  4344. if (expBits > 2048) {
  4345. err = MP_READ_E;
  4346. }
  4347. }
  4348. if (err == MP_OKAY) {
  4349. if (mp_count_bits(mod) != 2048) {
  4350. err = MP_READ_E;
  4351. }
  4352. }
  4353. if (err == MP_OKAY) {
  4354. sp_2048_from_mp(b, 64, base);
  4355. sp_2048_from_mp(e, 64, exp);
  4356. sp_2048_from_mp(m, 64, mod);
  4357. err = sp_2048_mod_exp_64(r, b, e, expBits, m, 0);
  4358. }
  4359. if (err == MP_OKAY) {
  4360. err = sp_2048_to_mp(r, res);
  4361. }
  4362. XMEMSET(e, 0, sizeof(e));
  4363. return err;
  4364. }
  4365. #ifdef WOLFSSL_HAVE_SP_DH
  4366. #ifdef HAVE_FFDHE_2048
  4367. static void sp_2048_lshift_64(sp_digit* r, sp_digit* a, byte n)
  4368. {
  4369. __asm__ __volatile__ (
  4370. "mov r6, #31\n\t"
  4371. "sub r6, r6, %[n]\n\t"
  4372. "add %[a], %[a], #192\n\t"
  4373. "add %[r], %[r], #192\n\t"
  4374. "ldr r3, [%[a], #60]\n\t"
  4375. "lsr r4, r3, #1\n\t"
  4376. "lsl r3, r3, %[n]\n\t"
  4377. "lsr r4, r4, r6\n\t"
  4378. "ldr r2, [%[a], #56]\n\t"
  4379. "str r4, [%[r], #64]\n\t"
  4380. "lsr r5, r2, #1\n\t"
  4381. "lsl r2, r2, %[n]\n\t"
  4382. "lsr r5, r5, r6\n\t"
  4383. "orr r3, r3, r5\n\t"
  4384. "ldr r4, [%[a], #52]\n\t"
  4385. "str r3, [%[r], #60]\n\t"
  4386. "lsr r5, r4, #1\n\t"
  4387. "lsl r4, r4, %[n]\n\t"
  4388. "lsr r5, r5, r6\n\t"
  4389. "orr r2, r2, r5\n\t"
  4390. "ldr r3, [%[a], #48]\n\t"
  4391. "str r2, [%[r], #56]\n\t"
  4392. "lsr r5, r3, #1\n\t"
  4393. "lsl r3, r3, %[n]\n\t"
  4394. "lsr r5, r5, r6\n\t"
  4395. "orr r4, r4, r5\n\t"
  4396. "ldr r2, [%[a], #44]\n\t"
  4397. "str r4, [%[r], #52]\n\t"
  4398. "lsr r5, r2, #1\n\t"
  4399. "lsl r2, r2, %[n]\n\t"
  4400. "lsr r5, r5, r6\n\t"
  4401. "orr r3, r3, r5\n\t"
  4402. "ldr r4, [%[a], #40]\n\t"
  4403. "str r3, [%[r], #48]\n\t"
  4404. "lsr r5, r4, #1\n\t"
  4405. "lsl r4, r4, %[n]\n\t"
  4406. "lsr r5, r5, r6\n\t"
  4407. "orr r2, r2, r5\n\t"
  4408. "ldr r3, [%[a], #36]\n\t"
  4409. "str r2, [%[r], #44]\n\t"
  4410. "lsr r5, r3, #1\n\t"
  4411. "lsl r3, r3, %[n]\n\t"
  4412. "lsr r5, r5, r6\n\t"
  4413. "orr r4, r4, r5\n\t"
  4414. "ldr r2, [%[a], #32]\n\t"
  4415. "str r4, [%[r], #40]\n\t"
  4416. "lsr r5, r2, #1\n\t"
  4417. "lsl r2, r2, %[n]\n\t"
  4418. "lsr r5, r5, r6\n\t"
  4419. "orr r3, r3, r5\n\t"
  4420. "ldr r4, [%[a], #28]\n\t"
  4421. "str r3, [%[r], #36]\n\t"
  4422. "lsr r5, r4, #1\n\t"
  4423. "lsl r4, r4, %[n]\n\t"
  4424. "lsr r5, r5, r6\n\t"
  4425. "orr r2, r2, r5\n\t"
  4426. "ldr r3, [%[a], #24]\n\t"
  4427. "str r2, [%[r], #32]\n\t"
  4428. "lsr r5, r3, #1\n\t"
  4429. "lsl r3, r3, %[n]\n\t"
  4430. "lsr r5, r5, r6\n\t"
  4431. "orr r4, r4, r5\n\t"
  4432. "ldr r2, [%[a], #20]\n\t"
  4433. "str r4, [%[r], #28]\n\t"
  4434. "lsr r5, r2, #1\n\t"
  4435. "lsl r2, r2, %[n]\n\t"
  4436. "lsr r5, r5, r6\n\t"
  4437. "orr r3, r3, r5\n\t"
  4438. "ldr r4, [%[a], #16]\n\t"
  4439. "str r3, [%[r], #24]\n\t"
  4440. "lsr r5, r4, #1\n\t"
  4441. "lsl r4, r4, %[n]\n\t"
  4442. "lsr r5, r5, r6\n\t"
  4443. "orr r2, r2, r5\n\t"
  4444. "ldr r3, [%[a], #12]\n\t"
  4445. "str r2, [%[r], #20]\n\t"
  4446. "lsr r5, r3, #1\n\t"
  4447. "lsl r3, r3, %[n]\n\t"
  4448. "lsr r5, r5, r6\n\t"
  4449. "orr r4, r4, r5\n\t"
  4450. "ldr r2, [%[a], #8]\n\t"
  4451. "str r4, [%[r], #16]\n\t"
  4452. "lsr r5, r2, #1\n\t"
  4453. "lsl r2, r2, %[n]\n\t"
  4454. "lsr r5, r5, r6\n\t"
  4455. "orr r3, r3, r5\n\t"
  4456. "ldr r4, [%[a], #4]\n\t"
  4457. "str r3, [%[r], #12]\n\t"
  4458. "lsr r5, r4, #1\n\t"
  4459. "lsl r4, r4, %[n]\n\t"
  4460. "lsr r5, r5, r6\n\t"
  4461. "orr r2, r2, r5\n\t"
  4462. "ldr r3, [%[a], #0]\n\t"
  4463. "str r2, [%[r], #8]\n\t"
  4464. "lsr r5, r3, #1\n\t"
  4465. "lsl r3, r3, %[n]\n\t"
  4466. "lsr r5, r5, r6\n\t"
  4467. "orr r4, r4, r5\n\t"
  4468. "sub %[a], %[a], #64\n\t"
  4469. "sub %[r], %[r], #64\n\t"
  4470. "ldr r2, [%[a], #60]\n\t"
  4471. "str r4, [%[r], #68]\n\t"
  4472. "lsr r5, r2, #1\n\t"
  4473. "lsl r2, r2, %[n]\n\t"
  4474. "lsr r5, r5, r6\n\t"
  4475. "orr r3, r3, r5\n\t"
  4476. "ldr r4, [%[a], #56]\n\t"
  4477. "str r3, [%[r], #64]\n\t"
  4478. "lsr r5, r4, #1\n\t"
  4479. "lsl r4, r4, %[n]\n\t"
  4480. "lsr r5, r5, r6\n\t"
  4481. "orr r2, r2, r5\n\t"
  4482. "ldr r3, [%[a], #52]\n\t"
  4483. "str r2, [%[r], #60]\n\t"
  4484. "lsr r5, r3, #1\n\t"
  4485. "lsl r3, r3, %[n]\n\t"
  4486. "lsr r5, r5, r6\n\t"
  4487. "orr r4, r4, r5\n\t"
  4488. "ldr r2, [%[a], #48]\n\t"
  4489. "str r4, [%[r], #56]\n\t"
  4490. "lsr r5, r2, #1\n\t"
  4491. "lsl r2, r2, %[n]\n\t"
  4492. "lsr r5, r5, r6\n\t"
  4493. "orr r3, r3, r5\n\t"
  4494. "ldr r4, [%[a], #44]\n\t"
  4495. "str r3, [%[r], #52]\n\t"
  4496. "lsr r5, r4, #1\n\t"
  4497. "lsl r4, r4, %[n]\n\t"
  4498. "lsr r5, r5, r6\n\t"
  4499. "orr r2, r2, r5\n\t"
  4500. "ldr r3, [%[a], #40]\n\t"
  4501. "str r2, [%[r], #48]\n\t"
  4502. "lsr r5, r3, #1\n\t"
  4503. "lsl r3, r3, %[n]\n\t"
  4504. "lsr r5, r5, r6\n\t"
  4505. "orr r4, r4, r5\n\t"
  4506. "ldr r2, [%[a], #36]\n\t"
  4507. "str r4, [%[r], #44]\n\t"
  4508. "lsr r5, r2, #1\n\t"
  4509. "lsl r2, r2, %[n]\n\t"
  4510. "lsr r5, r5, r6\n\t"
  4511. "orr r3, r3, r5\n\t"
  4512. "ldr r4, [%[a], #32]\n\t"
  4513. "str r3, [%[r], #40]\n\t"
  4514. "lsr r5, r4, #1\n\t"
  4515. "lsl r4, r4, %[n]\n\t"
  4516. "lsr r5, r5, r6\n\t"
  4517. "orr r2, r2, r5\n\t"
  4518. "ldr r3, [%[a], #28]\n\t"
  4519. "str r2, [%[r], #36]\n\t"
  4520. "lsr r5, r3, #1\n\t"
  4521. "lsl r3, r3, %[n]\n\t"
  4522. "lsr r5, r5, r6\n\t"
  4523. "orr r4, r4, r5\n\t"
  4524. "ldr r2, [%[a], #24]\n\t"
  4525. "str r4, [%[r], #32]\n\t"
  4526. "lsr r5, r2, #1\n\t"
  4527. "lsl r2, r2, %[n]\n\t"
  4528. "lsr r5, r5, r6\n\t"
  4529. "orr r3, r3, r5\n\t"
  4530. "ldr r4, [%[a], #20]\n\t"
  4531. "str r3, [%[r], #28]\n\t"
  4532. "lsr r5, r4, #1\n\t"
  4533. "lsl r4, r4, %[n]\n\t"
  4534. "lsr r5, r5, r6\n\t"
  4535. "orr r2, r2, r5\n\t"
  4536. "ldr r3, [%[a], #16]\n\t"
  4537. "str r2, [%[r], #24]\n\t"
  4538. "lsr r5, r3, #1\n\t"
  4539. "lsl r3, r3, %[n]\n\t"
  4540. "lsr r5, r5, r6\n\t"
  4541. "orr r4, r4, r5\n\t"
  4542. "ldr r2, [%[a], #12]\n\t"
  4543. "str r4, [%[r], #20]\n\t"
  4544. "lsr r5, r2, #1\n\t"
  4545. "lsl r2, r2, %[n]\n\t"
  4546. "lsr r5, r5, r6\n\t"
  4547. "orr r3, r3, r5\n\t"
  4548. "ldr r4, [%[a], #8]\n\t"
  4549. "str r3, [%[r], #16]\n\t"
  4550. "lsr r5, r4, #1\n\t"
  4551. "lsl r4, r4, %[n]\n\t"
  4552. "lsr r5, r5, r6\n\t"
  4553. "orr r2, r2, r5\n\t"
  4554. "ldr r3, [%[a], #4]\n\t"
  4555. "str r2, [%[r], #12]\n\t"
  4556. "lsr r5, r3, #1\n\t"
  4557. "lsl r3, r3, %[n]\n\t"
  4558. "lsr r5, r5, r6\n\t"
  4559. "orr r4, r4, r5\n\t"
  4560. "ldr r2, [%[a], #0]\n\t"
  4561. "str r4, [%[r], #8]\n\t"
  4562. "lsr r5, r2, #1\n\t"
  4563. "lsl r2, r2, %[n]\n\t"
  4564. "lsr r5, r5, r6\n\t"
  4565. "orr r3, r3, r5\n\t"
  4566. "sub %[a], %[a], #64\n\t"
  4567. "sub %[r], %[r], #64\n\t"
  4568. "ldr r4, [%[a], #60]\n\t"
  4569. "str r3, [%[r], #68]\n\t"
  4570. "lsr r5, r4, #1\n\t"
  4571. "lsl r4, r4, %[n]\n\t"
  4572. "lsr r5, r5, r6\n\t"
  4573. "orr r2, r2, r5\n\t"
  4574. "ldr r3, [%[a], #56]\n\t"
  4575. "str r2, [%[r], #64]\n\t"
  4576. "lsr r5, r3, #1\n\t"
  4577. "lsl r3, r3, %[n]\n\t"
  4578. "lsr r5, r5, r6\n\t"
  4579. "orr r4, r4, r5\n\t"
  4580. "ldr r2, [%[a], #52]\n\t"
  4581. "str r4, [%[r], #60]\n\t"
  4582. "lsr r5, r2, #1\n\t"
  4583. "lsl r2, r2, %[n]\n\t"
  4584. "lsr r5, r5, r6\n\t"
  4585. "orr r3, r3, r5\n\t"
  4586. "ldr r4, [%[a], #48]\n\t"
  4587. "str r3, [%[r], #56]\n\t"
  4588. "lsr r5, r4, #1\n\t"
  4589. "lsl r4, r4, %[n]\n\t"
  4590. "lsr r5, r5, r6\n\t"
  4591. "orr r2, r2, r5\n\t"
  4592. "ldr r3, [%[a], #44]\n\t"
  4593. "str r2, [%[r], #52]\n\t"
  4594. "lsr r5, r3, #1\n\t"
  4595. "lsl r3, r3, %[n]\n\t"
  4596. "lsr r5, r5, r6\n\t"
  4597. "orr r4, r4, r5\n\t"
  4598. "ldr r2, [%[a], #40]\n\t"
  4599. "str r4, [%[r], #48]\n\t"
  4600. "lsr r5, r2, #1\n\t"
  4601. "lsl r2, r2, %[n]\n\t"
  4602. "lsr r5, r5, r6\n\t"
  4603. "orr r3, r3, r5\n\t"
  4604. "ldr r4, [%[a], #36]\n\t"
  4605. "str r3, [%[r], #44]\n\t"
  4606. "lsr r5, r4, #1\n\t"
  4607. "lsl r4, r4, %[n]\n\t"
  4608. "lsr r5, r5, r6\n\t"
  4609. "orr r2, r2, r5\n\t"
  4610. "ldr r3, [%[a], #32]\n\t"
  4611. "str r2, [%[r], #40]\n\t"
  4612. "lsr r5, r3, #1\n\t"
  4613. "lsl r3, r3, %[n]\n\t"
  4614. "lsr r5, r5, r6\n\t"
  4615. "orr r4, r4, r5\n\t"
  4616. "ldr r2, [%[a], #28]\n\t"
  4617. "str r4, [%[r], #36]\n\t"
  4618. "lsr r5, r2, #1\n\t"
  4619. "lsl r2, r2, %[n]\n\t"
  4620. "lsr r5, r5, r6\n\t"
  4621. "orr r3, r3, r5\n\t"
  4622. "ldr r4, [%[a], #24]\n\t"
  4623. "str r3, [%[r], #32]\n\t"
  4624. "lsr r5, r4, #1\n\t"
  4625. "lsl r4, r4, %[n]\n\t"
  4626. "lsr r5, r5, r6\n\t"
  4627. "orr r2, r2, r5\n\t"
  4628. "ldr r3, [%[a], #20]\n\t"
  4629. "str r2, [%[r], #28]\n\t"
  4630. "lsr r5, r3, #1\n\t"
  4631. "lsl r3, r3, %[n]\n\t"
  4632. "lsr r5, r5, r6\n\t"
  4633. "orr r4, r4, r5\n\t"
  4634. "ldr r2, [%[a], #16]\n\t"
  4635. "str r4, [%[r], #24]\n\t"
  4636. "lsr r5, r2, #1\n\t"
  4637. "lsl r2, r2, %[n]\n\t"
  4638. "lsr r5, r5, r6\n\t"
  4639. "orr r3, r3, r5\n\t"
  4640. "ldr r4, [%[a], #12]\n\t"
  4641. "str r3, [%[r], #20]\n\t"
  4642. "lsr r5, r4, #1\n\t"
  4643. "lsl r4, r4, %[n]\n\t"
  4644. "lsr r5, r5, r6\n\t"
  4645. "orr r2, r2, r5\n\t"
  4646. "ldr r3, [%[a], #8]\n\t"
  4647. "str r2, [%[r], #16]\n\t"
  4648. "lsr r5, r3, #1\n\t"
  4649. "lsl r3, r3, %[n]\n\t"
  4650. "lsr r5, r5, r6\n\t"
  4651. "orr r4, r4, r5\n\t"
  4652. "ldr r2, [%[a], #4]\n\t"
  4653. "str r4, [%[r], #12]\n\t"
  4654. "lsr r5, r2, #1\n\t"
  4655. "lsl r2, r2, %[n]\n\t"
  4656. "lsr r5, r5, r6\n\t"
  4657. "orr r3, r3, r5\n\t"
  4658. "ldr r4, [%[a], #0]\n\t"
  4659. "str r3, [%[r], #8]\n\t"
  4660. "lsr r5, r4, #1\n\t"
  4661. "lsl r4, r4, %[n]\n\t"
  4662. "lsr r5, r5, r6\n\t"
  4663. "orr r2, r2, r5\n\t"
  4664. "sub %[a], %[a], #64\n\t"
  4665. "sub %[r], %[r], #64\n\t"
  4666. "ldr r3, [%[a], #60]\n\t"
  4667. "str r2, [%[r], #68]\n\t"
  4668. "lsr r5, r3, #1\n\t"
  4669. "lsl r3, r3, %[n]\n\t"
  4670. "lsr r5, r5, r6\n\t"
  4671. "orr r4, r4, r5\n\t"
  4672. "ldr r2, [%[a], #56]\n\t"
  4673. "str r4, [%[r], #64]\n\t"
  4674. "lsr r5, r2, #1\n\t"
  4675. "lsl r2, r2, %[n]\n\t"
  4676. "lsr r5, r5, r6\n\t"
  4677. "orr r3, r3, r5\n\t"
  4678. "ldr r4, [%[a], #52]\n\t"
  4679. "str r3, [%[r], #60]\n\t"
  4680. "lsr r5, r4, #1\n\t"
  4681. "lsl r4, r4, %[n]\n\t"
  4682. "lsr r5, r5, r6\n\t"
  4683. "orr r2, r2, r5\n\t"
  4684. "ldr r3, [%[a], #48]\n\t"
  4685. "str r2, [%[r], #56]\n\t"
  4686. "lsr r5, r3, #1\n\t"
  4687. "lsl r3, r3, %[n]\n\t"
  4688. "lsr r5, r5, r6\n\t"
  4689. "orr r4, r4, r5\n\t"
  4690. "ldr r2, [%[a], #44]\n\t"
  4691. "str r4, [%[r], #52]\n\t"
  4692. "lsr r5, r2, #1\n\t"
  4693. "lsl r2, r2, %[n]\n\t"
  4694. "lsr r5, r5, r6\n\t"
  4695. "orr r3, r3, r5\n\t"
  4696. "ldr r4, [%[a], #40]\n\t"
  4697. "str r3, [%[r], #48]\n\t"
  4698. "lsr r5, r4, #1\n\t"
  4699. "lsl r4, r4, %[n]\n\t"
  4700. "lsr r5, r5, r6\n\t"
  4701. "orr r2, r2, r5\n\t"
  4702. "ldr r3, [%[a], #36]\n\t"
  4703. "str r2, [%[r], #44]\n\t"
  4704. "lsr r5, r3, #1\n\t"
  4705. "lsl r3, r3, %[n]\n\t"
  4706. "lsr r5, r5, r6\n\t"
  4707. "orr r4, r4, r5\n\t"
  4708. "ldr r2, [%[a], #32]\n\t"
  4709. "str r4, [%[r], #40]\n\t"
  4710. "lsr r5, r2, #1\n\t"
  4711. "lsl r2, r2, %[n]\n\t"
  4712. "lsr r5, r5, r6\n\t"
  4713. "orr r3, r3, r5\n\t"
  4714. "ldr r4, [%[a], #28]\n\t"
  4715. "str r3, [%[r], #36]\n\t"
  4716. "lsr r5, r4, #1\n\t"
  4717. "lsl r4, r4, %[n]\n\t"
  4718. "lsr r5, r5, r6\n\t"
  4719. "orr r2, r2, r5\n\t"
  4720. "ldr r3, [%[a], #24]\n\t"
  4721. "str r2, [%[r], #32]\n\t"
  4722. "lsr r5, r3, #1\n\t"
  4723. "lsl r3, r3, %[n]\n\t"
  4724. "lsr r5, r5, r6\n\t"
  4725. "orr r4, r4, r5\n\t"
  4726. "ldr r2, [%[a], #20]\n\t"
  4727. "str r4, [%[r], #28]\n\t"
  4728. "lsr r5, r2, #1\n\t"
  4729. "lsl r2, r2, %[n]\n\t"
  4730. "lsr r5, r5, r6\n\t"
  4731. "orr r3, r3, r5\n\t"
  4732. "ldr r4, [%[a], #16]\n\t"
  4733. "str r3, [%[r], #24]\n\t"
  4734. "lsr r5, r4, #1\n\t"
  4735. "lsl r4, r4, %[n]\n\t"
  4736. "lsr r5, r5, r6\n\t"
  4737. "orr r2, r2, r5\n\t"
  4738. "ldr r3, [%[a], #12]\n\t"
  4739. "str r2, [%[r], #20]\n\t"
  4740. "lsr r5, r3, #1\n\t"
  4741. "lsl r3, r3, %[n]\n\t"
  4742. "lsr r5, r5, r6\n\t"
  4743. "orr r4, r4, r5\n\t"
  4744. "ldr r2, [%[a], #8]\n\t"
  4745. "str r4, [%[r], #16]\n\t"
  4746. "lsr r5, r2, #1\n\t"
  4747. "lsl r2, r2, %[n]\n\t"
  4748. "lsr r5, r5, r6\n\t"
  4749. "orr r3, r3, r5\n\t"
  4750. "ldr r4, [%[a], #4]\n\t"
  4751. "str r3, [%[r], #12]\n\t"
  4752. "lsr r5, r4, #1\n\t"
  4753. "lsl r4, r4, %[n]\n\t"
  4754. "lsr r5, r5, r6\n\t"
  4755. "orr r2, r2, r5\n\t"
  4756. "ldr r3, [%[a], #0]\n\t"
  4757. "str r2, [%[r], #8]\n\t"
  4758. "lsr r5, r3, #1\n\t"
  4759. "lsl r3, r3, %[n]\n\t"
  4760. "lsr r5, r5, r6\n\t"
  4761. "orr r4, r4, r5\n\t"
  4762. "str r3, [%[r]]\n\t"
  4763. "str r4, [%[r], #4]\n\t"
  4764. :
  4765. : [r] "r" (r), [a] "r" (a), [n] "r" (n)
  4766. : "memory", "r2", "r3", "r4", "r5", "r6"
  4767. );
  4768. }
  4769. /* Modular exponentiate 2 to the e mod m. (r = 2^e mod m)
  4770. *
  4771. * r A single precision number that is the result of the operation.
  4772. * e A single precision number that is the exponent.
  4773. * bits The number of bits in the exponent.
  4774. * m A single precision number that is the modulus.
  4775. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  4776. */
  4777. static int sp_2048_mod_exp_2_64(sp_digit* r, const sp_digit* e, int bits,
  4778. const sp_digit* m)
  4779. {
  4780. #ifndef WOLFSSL_SMALL_STACK
  4781. sp_digit nd[128];
  4782. sp_digit td[65];
  4783. #else
  4784. sp_digit* td;
  4785. #endif
  4786. sp_digit* norm;
  4787. sp_digit* tmp;
  4788. sp_digit mp = 1;
  4789. sp_digit n, o;
  4790. sp_digit mask;
  4791. int i;
  4792. int c, y;
  4793. int err = MP_OKAY;
  4794. #ifdef WOLFSSL_SMALL_STACK
  4795. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 193, NULL,
  4796. DYNAMIC_TYPE_TMP_BUFFER);
  4797. if (td == NULL) {
  4798. err = MEMORY_E;
  4799. }
  4800. #endif
  4801. if (err == MP_OKAY) {
  4802. #ifdef WOLFSSL_SMALL_STACK
  4803. norm = td;
  4804. tmp = td + 128;
  4805. #else
  4806. norm = nd;
  4807. tmp = td;
  4808. #endif
  4809. sp_2048_mont_setup(m, &mp);
  4810. sp_2048_mont_norm_64(norm, m);
  4811. i = (bits - 1) / 32;
  4812. n = e[i--];
  4813. c = bits & 31;
  4814. if (c == 0) {
  4815. c = 32;
  4816. }
  4817. c -= bits % 5;
  4818. if (c == 32) {
  4819. c = 27;
  4820. }
  4821. y = (int)(n >> c);
  4822. n <<= 32 - c;
  4823. sp_2048_lshift_64(r, norm, y);
  4824. for (; i>=0 || c>=5; ) {
  4825. if (c == 0) {
  4826. n = e[i--];
  4827. y = n >> 27;
  4828. n <<= 5;
  4829. c = 27;
  4830. }
  4831. else if (c < 5) {
  4832. y = n >> 27;
  4833. n = e[i--];
  4834. c = 5 - c;
  4835. y |= n >> (32 - c);
  4836. n <<= c;
  4837. c = 32 - c;
  4838. }
  4839. else {
  4840. y = (n >> 27) & 0x1f;
  4841. n <<= 5;
  4842. c -= 5;
  4843. }
  4844. sp_2048_mont_sqr_64(r, r, m, mp);
  4845. sp_2048_mont_sqr_64(r, r, m, mp);
  4846. sp_2048_mont_sqr_64(r, r, m, mp);
  4847. sp_2048_mont_sqr_64(r, r, m, mp);
  4848. sp_2048_mont_sqr_64(r, r, m, mp);
  4849. sp_2048_lshift_64(r, r, y);
  4850. sp_2048_mul_d_64(tmp, norm, r[64]);
  4851. r[64] = 0;
  4852. o = sp_2048_add_64(r, r, tmp);
  4853. sp_2048_cond_sub_64(r, r, m, (sp_digit)0 - o);
  4854. }
  4855. XMEMSET(&r[64], 0, sizeof(sp_digit) * 64U);
  4856. sp_2048_mont_reduce_64(r, m, mp);
  4857. mask = 0 - (sp_2048_cmp_64(r, m) >= 0);
  4858. sp_2048_cond_sub_64(r, r, m, mask);
  4859. }
  4860. #ifdef WOLFSSL_SMALL_STACK
  4861. if (td != NULL) {
  4862. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  4863. }
  4864. #endif
  4865. return err;
  4866. }
  4867. #endif /* HAVE_FFDHE_2048 */
  4868. /* Perform the modular exponentiation for Diffie-Hellman.
  4869. *
  4870. * base Base.
  4871. * exp Array of bytes that is the exponent.
  4872. * expLen Length of data, in bytes, in exponent.
  4873. * mod Modulus.
  4874. * out Buffer to hold big-endian bytes of exponentiation result.
  4875. * Must be at least 256 bytes long.
  4876. * outLen Length, in bytes, of exponentiation result.
  4877. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  4878. * and MEMORY_E if memory allocation fails.
  4879. */
  4880. int sp_DhExp_2048(mp_int* base, const byte* exp, word32 expLen,
  4881. mp_int* mod, byte* out, word32* outLen)
  4882. {
  4883. int err = MP_OKAY;
  4884. sp_digit b[128], e[64], m[64];
  4885. sp_digit* r = b;
  4886. word32 i;
  4887. if (mp_count_bits(base) > 2048) {
  4888. err = MP_READ_E;
  4889. }
  4890. if (err == MP_OKAY) {
  4891. if (expLen > 256) {
  4892. err = MP_READ_E;
  4893. }
  4894. }
  4895. if (err == MP_OKAY) {
  4896. if (mp_count_bits(mod) != 2048) {
  4897. err = MP_READ_E;
  4898. }
  4899. }
  4900. if (err == MP_OKAY) {
  4901. sp_2048_from_mp(b, 64, base);
  4902. sp_2048_from_bin(e, 64, exp, expLen);
  4903. sp_2048_from_mp(m, 64, mod);
  4904. #ifdef HAVE_FFDHE_2048
  4905. if (base->used == 1 && base->dp[0] == 2 && m[63] == (sp_digit)-1)
  4906. err = sp_2048_mod_exp_2_64(r, e, expLen * 8, m);
  4907. else
  4908. #endif
  4909. err = sp_2048_mod_exp_64(r, b, e, expLen * 8, m, 0);
  4910. }
  4911. if (err == MP_OKAY) {
  4912. sp_2048_to_bin(r, out);
  4913. *outLen = 256;
  4914. for (i=0; i<256 && out[i] == 0; i++) {
  4915. }
  4916. *outLen -= i;
  4917. XMEMMOVE(out, out + i, *outLen);
  4918. }
  4919. XMEMSET(e, 0, sizeof(e));
  4920. return err;
  4921. }
  4922. #endif /* WOLFSSL_HAVE_SP_DH */
  4923. /* Perform the modular exponentiation for Diffie-Hellman.
  4924. *
  4925. * base Base. MP integer.
  4926. * exp Exponent. MP integer.
  4927. * mod Modulus. MP integer.
  4928. * res Result. MP integer.
  4929. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  4930. * and MEMORY_E if memory allocation fails.
  4931. */
  4932. int sp_ModExp_1024(mp_int* base, mp_int* exp, mp_int* mod, mp_int* res)
  4933. {
  4934. int err = MP_OKAY;
  4935. sp_digit b[64], e[32], m[32];
  4936. sp_digit* r = b;
  4937. int expBits = mp_count_bits(exp);
  4938. if (mp_count_bits(base) > 1024) {
  4939. err = MP_READ_E;
  4940. }
  4941. if (err == MP_OKAY) {
  4942. if (expBits > 1024) {
  4943. err = MP_READ_E;
  4944. }
  4945. }
  4946. if (err == MP_OKAY) {
  4947. if (mp_count_bits(mod) != 1024) {
  4948. err = MP_READ_E;
  4949. }
  4950. }
  4951. if (err == MP_OKAY) {
  4952. sp_2048_from_mp(b, 32, base);
  4953. sp_2048_from_mp(e, 32, exp);
  4954. sp_2048_from_mp(m, 32, mod);
  4955. err = sp_2048_mod_exp_32(r, b, e, expBits, m, 0);
  4956. }
  4957. if (err == MP_OKAY) {
  4958. XMEMSET(r + 32, 0, sizeof(*r) * 32U);
  4959. err = sp_2048_to_mp(r, res);
  4960. res->used = mod->used;
  4961. mp_clamp(res);
  4962. }
  4963. XMEMSET(e, 0, sizeof(e));
  4964. return err;
  4965. }
  4966. #endif /* WOLFSSL_HAVE_SP_DH || (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) */
  4967. #endif /* !WOLFSSL_SP_NO_2048 */
  4968. #ifndef WOLFSSL_SP_NO_3072
  4969. /* Read big endian unsigned byte array into r.
  4970. *
  4971. * r A single precision integer.
  4972. * size Maximum number of bytes to convert
  4973. * a Byte array.
  4974. * n Number of bytes in array to read.
  4975. */
  4976. static void sp_3072_from_bin(sp_digit* r, int size, const byte* a, int n)
  4977. {
  4978. int i, j = 0;
  4979. word32 s = 0;
  4980. r[0] = 0;
  4981. for (i = n-1; i >= 0; i--) {
  4982. r[j] |= (((sp_digit)a[i]) << s);
  4983. if (s >= 24U) {
  4984. r[j] &= 0xffffffff;
  4985. s = 32U - s;
  4986. if (j + 1 >= size) {
  4987. break;
  4988. }
  4989. r[++j] = (sp_digit)a[i] >> s;
  4990. s = 8U - s;
  4991. }
  4992. else {
  4993. s += 8U;
  4994. }
  4995. }
  4996. for (j++; j < size; j++) {
  4997. r[j] = 0;
  4998. }
  4999. }
  5000. /* Convert an mp_int to an array of sp_digit.
  5001. *
  5002. * r A single precision integer.
  5003. * size Maximum number of bytes to convert
  5004. * a A multi-precision integer.
  5005. */
  5006. static void sp_3072_from_mp(sp_digit* r, int size, const mp_int* a)
  5007. {
  5008. #if DIGIT_BIT == 32
  5009. int j;
  5010. XMEMCPY(r, a->dp, sizeof(sp_digit) * a->used);
  5011. for (j = a->used; j < size; j++) {
  5012. r[j] = 0;
  5013. }
  5014. #elif DIGIT_BIT > 32
  5015. int i, j = 0;
  5016. word32 s = 0;
  5017. r[0] = 0;
  5018. for (i = 0; i < a->used && j < size; i++) {
  5019. r[j] |= ((sp_digit)a->dp[i] << s);
  5020. r[j] &= 0xffffffff;
  5021. s = 32U - s;
  5022. if (j + 1 >= size) {
  5023. break;
  5024. }
  5025. /* lint allow cast of mismatch word32 and mp_digit */
  5026. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  5027. while ((s + 32U) <= (word32)DIGIT_BIT) {
  5028. s += 32U;
  5029. r[j] &= 0xffffffff;
  5030. if (j + 1 >= size) {
  5031. break;
  5032. }
  5033. if (s < (word32)DIGIT_BIT) {
  5034. /* lint allow cast of mismatch word32 and mp_digit */
  5035. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  5036. }
  5037. else {
  5038. r[++j] = 0L;
  5039. }
  5040. }
  5041. s = (word32)DIGIT_BIT - s;
  5042. }
  5043. for (j++; j < size; j++) {
  5044. r[j] = 0;
  5045. }
  5046. #else
  5047. int i, j = 0, s = 0;
  5048. r[0] = 0;
  5049. for (i = 0; i < a->used && j < size; i++) {
  5050. r[j] |= ((sp_digit)a->dp[i]) << s;
  5051. if (s + DIGIT_BIT >= 32) {
  5052. r[j] &= 0xffffffff;
  5053. if (j + 1 >= size) {
  5054. break;
  5055. }
  5056. s = 32 - s;
  5057. if (s == DIGIT_BIT) {
  5058. r[++j] = 0;
  5059. s = 0;
  5060. }
  5061. else {
  5062. r[++j] = a->dp[i] >> s;
  5063. s = DIGIT_BIT - s;
  5064. }
  5065. }
  5066. else {
  5067. s += DIGIT_BIT;
  5068. }
  5069. }
  5070. for (j++; j < size; j++) {
  5071. r[j] = 0;
  5072. }
  5073. #endif
  5074. }
  5075. /* Write r as big endian to byte array.
  5076. * Fixed length number of bytes written: 384
  5077. *
  5078. * r A single precision integer.
  5079. * a Byte array.
  5080. */
  5081. static void sp_3072_to_bin(sp_digit* r, byte* a)
  5082. {
  5083. int i, j, s = 0, b;
  5084. j = 3072 / 8 - 1;
  5085. a[j] = 0;
  5086. for (i=0; i<96 && j>=0; i++) {
  5087. b = 0;
  5088. /* lint allow cast of mismatch sp_digit and int */
  5089. a[j--] |= (byte)(r[i] << s); b += 8 - s; /*lint !e9033*/
  5090. if (j < 0) {
  5091. break;
  5092. }
  5093. while (b < 32) {
  5094. a[j--] = r[i] >> b; b += 8;
  5095. if (j < 0) {
  5096. break;
  5097. }
  5098. }
  5099. s = 8 - (b - 32);
  5100. if (j >= 0) {
  5101. a[j] = 0;
  5102. }
  5103. if (s != 0) {
  5104. j++;
  5105. }
  5106. }
  5107. }
  5108. #ifndef WOLFSSL_SP_SMALL
  5109. /* Multiply a and b into r. (r = a * b)
  5110. *
  5111. * r A single precision integer.
  5112. * a A single precision integer.
  5113. * b A single precision integer.
  5114. */
  5115. SP_NOINLINE static void sp_3072_mul_12(sp_digit* r, const sp_digit* a,
  5116. const sp_digit* b)
  5117. {
  5118. sp_digit tmp[12 * 2];
  5119. __asm__ __volatile__ (
  5120. "mov r3, #0\n\t"
  5121. "mov r4, #0\n\t"
  5122. "mov r8, r3\n\t"
  5123. "mov r11, %[r]\n\t"
  5124. "mov r9, %[a]\n\t"
  5125. "mov r10, %[b]\n\t"
  5126. "mov r6, #48\n\t"
  5127. "add r6, r9\n\t"
  5128. "mov r12, r6\n\t"
  5129. "\n1:\n\t"
  5130. "mov %[r], #0\n\t"
  5131. "mov r5, #0\n\t"
  5132. "mov r6, #44\n\t"
  5133. "mov %[a], r8\n\t"
  5134. "sub %[a], r6\n\t"
  5135. "sbc r6, r6\n\t"
  5136. "mvn r6, r6\n\t"
  5137. "and %[a], r6\n\t"
  5138. "mov %[b], r8\n\t"
  5139. "sub %[b], %[a]\n\t"
  5140. "add %[a], r9\n\t"
  5141. "add %[b], r10\n\t"
  5142. "\n2:\n\t"
  5143. "# Multiply Start\n\t"
  5144. "ldr r6, [%[a]]\n\t"
  5145. "ldr r7, [%[b]]\n\t"
  5146. "lsl r6, r6, #16\n\t"
  5147. "lsl r7, r7, #16\n\t"
  5148. "lsr r6, r6, #16\n\t"
  5149. "lsr r7, r7, #16\n\t"
  5150. "mul r7, r6\n\t"
  5151. "add r3, r7\n\t"
  5152. "adc r4, %[r]\n\t"
  5153. "adc r5, %[r]\n\t"
  5154. "ldr r7, [%[b]]\n\t"
  5155. "lsr r7, r7, #16\n\t"
  5156. "mul r6, r7\n\t"
  5157. "lsr r7, r6, #16\n\t"
  5158. "lsl r6, r6, #16\n\t"
  5159. "add r3, r6\n\t"
  5160. "adc r4, r7\n\t"
  5161. "adc r5, %[r]\n\t"
  5162. "ldr r6, [%[a]]\n\t"
  5163. "ldr r7, [%[b]]\n\t"
  5164. "lsr r6, r6, #16\n\t"
  5165. "lsr r7, r7, #16\n\t"
  5166. "mul r7, r6\n\t"
  5167. "add r4, r7\n\t"
  5168. "adc r5, %[r]\n\t"
  5169. "ldr r7, [%[b]]\n\t"
  5170. "lsl r7, r7, #16\n\t"
  5171. "lsr r7, r7, #16\n\t"
  5172. "mul r6, r7\n\t"
  5173. "lsr r7, r6, #16\n\t"
  5174. "lsl r6, r6, #16\n\t"
  5175. "add r3, r6\n\t"
  5176. "adc r4, r7\n\t"
  5177. "adc r5, %[r]\n\t"
  5178. "# Multiply Done\n\t"
  5179. "add %[a], #4\n\t"
  5180. "sub %[b], #4\n\t"
  5181. "cmp %[a], r12\n\t"
  5182. "beq 3f\n\t"
  5183. "mov r6, r8\n\t"
  5184. "add r6, r9\n\t"
  5185. "cmp %[a], r6\n\t"
  5186. "ble 2b\n\t"
  5187. "\n3:\n\t"
  5188. "mov %[r], r11\n\t"
  5189. "mov r7, r8\n\t"
  5190. "str r3, [%[r], r7]\n\t"
  5191. "mov r3, r4\n\t"
  5192. "mov r4, r5\n\t"
  5193. "add r7, #4\n\t"
  5194. "mov r8, r7\n\t"
  5195. "mov r6, #88\n\t"
  5196. "cmp r7, r6\n\t"
  5197. "ble 1b\n\t"
  5198. "str r3, [%[r], r7]\n\t"
  5199. "mov %[a], r9\n\t"
  5200. "mov %[b], r10\n\t"
  5201. :
  5202. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  5203. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  5204. );
  5205. XMEMCPY(r, tmp, sizeof(tmp));
  5206. }
  5207. /* Square a and put result in r. (r = a * a)
  5208. *
  5209. * r A single precision integer.
  5210. * a A single precision integer.
  5211. */
  5212. SP_NOINLINE static void sp_3072_sqr_12(sp_digit* r, const sp_digit* a)
  5213. {
  5214. __asm__ __volatile__ (
  5215. "mov r3, #0\n\t"
  5216. "mov r4, #0\n\t"
  5217. "mov r5, #0\n\t"
  5218. "mov r8, r3\n\t"
  5219. "mov r11, %[r]\n\t"
  5220. "mov r6, #96\n\t"
  5221. "neg r6, r6\n\t"
  5222. "add sp, r6\n\t"
  5223. "mov r10, sp\n\t"
  5224. "mov r9, %[a]\n\t"
  5225. "\n1:\n\t"
  5226. "mov %[r], #0\n\t"
  5227. "mov r6, #44\n\t"
  5228. "mov %[a], r8\n\t"
  5229. "sub %[a], r6\n\t"
  5230. "sbc r6, r6\n\t"
  5231. "mvn r6, r6\n\t"
  5232. "and %[a], r6\n\t"
  5233. "mov r2, r8\n\t"
  5234. "sub r2, %[a]\n\t"
  5235. "add %[a], r9\n\t"
  5236. "add r2, r9\n\t"
  5237. "\n2:\n\t"
  5238. "cmp r2, %[a]\n\t"
  5239. "beq 4f\n\t"
  5240. "# Multiply * 2: Start\n\t"
  5241. "ldr r6, [%[a]]\n\t"
  5242. "ldr r7, [r2]\n\t"
  5243. "lsl r6, r6, #16\n\t"
  5244. "lsl r7, r7, #16\n\t"
  5245. "lsr r6, r6, #16\n\t"
  5246. "lsr r7, r7, #16\n\t"
  5247. "mul r7, r6\n\t"
  5248. "add r3, r7\n\t"
  5249. "adc r4, %[r]\n\t"
  5250. "adc r5, %[r]\n\t"
  5251. "add r3, r7\n\t"
  5252. "adc r4, %[r]\n\t"
  5253. "adc r5, %[r]\n\t"
  5254. "ldr r7, [r2]\n\t"
  5255. "lsr r7, r7, #16\n\t"
  5256. "mul r6, r7\n\t"
  5257. "lsr r7, r6, #16\n\t"
  5258. "lsl r6, r6, #16\n\t"
  5259. "add r3, r6\n\t"
  5260. "adc r4, r7\n\t"
  5261. "adc r5, %[r]\n\t"
  5262. "add r3, r6\n\t"
  5263. "adc r4, r7\n\t"
  5264. "adc r5, %[r]\n\t"
  5265. "ldr r6, [%[a]]\n\t"
  5266. "ldr r7, [r2]\n\t"
  5267. "lsr r6, r6, #16\n\t"
  5268. "lsr r7, r7, #16\n\t"
  5269. "mul r7, r6\n\t"
  5270. "add r4, r7\n\t"
  5271. "adc r5, %[r]\n\t"
  5272. "add r4, r7\n\t"
  5273. "adc r5, %[r]\n\t"
  5274. "ldr r7, [r2]\n\t"
  5275. "lsl r7, r7, #16\n\t"
  5276. "lsr r7, r7, #16\n\t"
  5277. "mul r6, r7\n\t"
  5278. "lsr r7, r6, #16\n\t"
  5279. "lsl r6, r6, #16\n\t"
  5280. "add r3, r6\n\t"
  5281. "adc r4, r7\n\t"
  5282. "adc r5, %[r]\n\t"
  5283. "add r3, r6\n\t"
  5284. "adc r4, r7\n\t"
  5285. "adc r5, %[r]\n\t"
  5286. "# Multiply * 2: Done\n\t"
  5287. "bal 5f\n\t"
  5288. "\n4:\n\t"
  5289. "# Square: Start\n\t"
  5290. "ldr r6, [%[a]]\n\t"
  5291. "lsr r7, r6, #16\n\t"
  5292. "lsl r6, r6, #16\n\t"
  5293. "lsr r6, r6, #16\n\t"
  5294. "mul r6, r6\n\t"
  5295. "add r3, r6\n\t"
  5296. "adc r4, %[r]\n\t"
  5297. "adc r5, %[r]\n\t"
  5298. "mul r7, r7\n\t"
  5299. "add r4, r7\n\t"
  5300. "adc r5, %[r]\n\t"
  5301. "ldr r6, [%[a]]\n\t"
  5302. "lsr r7, r6, #16\n\t"
  5303. "lsl r6, r6, #16\n\t"
  5304. "lsr r6, r6, #16\n\t"
  5305. "mul r6, r7\n\t"
  5306. "lsr r7, r6, #15\n\t"
  5307. "lsl r6, r6, #17\n\t"
  5308. "add r3, r6\n\t"
  5309. "adc r4, r7\n\t"
  5310. "adc r5, %[r]\n\t"
  5311. "# Square: Done\n\t"
  5312. "\n5:\n\t"
  5313. "add %[a], #4\n\t"
  5314. "sub r2, #4\n\t"
  5315. "mov r6, #48\n\t"
  5316. "add r6, r9\n\t"
  5317. "cmp %[a], r6\n\t"
  5318. "beq 3f\n\t"
  5319. "cmp %[a], r2\n\t"
  5320. "bgt 3f\n\t"
  5321. "mov r7, r8\n\t"
  5322. "add r7, r9\n\t"
  5323. "cmp %[a], r7\n\t"
  5324. "ble 2b\n\t"
  5325. "\n3:\n\t"
  5326. "mov %[r], r10\n\t"
  5327. "mov r7, r8\n\t"
  5328. "str r3, [%[r], r7]\n\t"
  5329. "mov r3, r4\n\t"
  5330. "mov r4, r5\n\t"
  5331. "mov r5, #0\n\t"
  5332. "add r7, #4\n\t"
  5333. "mov r8, r7\n\t"
  5334. "mov r6, #88\n\t"
  5335. "cmp r7, r6\n\t"
  5336. "ble 1b\n\t"
  5337. "mov %[a], r9\n\t"
  5338. "str r3, [%[r], r7]\n\t"
  5339. "mov %[r], r11\n\t"
  5340. "mov %[a], r10\n\t"
  5341. "mov r3, #92\n\t"
  5342. "\n4:\n\t"
  5343. "ldr r6, [%[a], r3]\n\t"
  5344. "str r6, [%[r], r3]\n\t"
  5345. "sub r3, #4\n\t"
  5346. "bge 4b\n\t"
  5347. "mov r6, #96\n\t"
  5348. "add sp, r6\n\t"
  5349. :
  5350. : [r] "r" (r), [a] "r" (a)
  5351. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  5352. );
  5353. }
  5354. /* Add b to a into r. (r = a + b)
  5355. *
  5356. * r A single precision integer.
  5357. * a A single precision integer.
  5358. * b A single precision integer.
  5359. */
  5360. SP_NOINLINE static sp_digit sp_3072_add_12(sp_digit* r, const sp_digit* a,
  5361. const sp_digit* b)
  5362. {
  5363. sp_digit c = 0;
  5364. __asm__ __volatile__ (
  5365. "ldr r4, [%[a], #0]\n\t"
  5366. "ldr r5, [%[b], #0]\n\t"
  5367. "add r4, r5\n\t"
  5368. "str r4, [%[r], #0]\n\t"
  5369. "ldr r4, [%[a], #4]\n\t"
  5370. "ldr r5, [%[b], #4]\n\t"
  5371. "adc r4, r5\n\t"
  5372. "str r4, [%[r], #4]\n\t"
  5373. "ldr r4, [%[a], #8]\n\t"
  5374. "ldr r5, [%[b], #8]\n\t"
  5375. "adc r4, r5\n\t"
  5376. "str r4, [%[r], #8]\n\t"
  5377. "ldr r4, [%[a], #12]\n\t"
  5378. "ldr r5, [%[b], #12]\n\t"
  5379. "adc r4, r5\n\t"
  5380. "str r4, [%[r], #12]\n\t"
  5381. "ldr r4, [%[a], #16]\n\t"
  5382. "ldr r5, [%[b], #16]\n\t"
  5383. "adc r4, r5\n\t"
  5384. "str r4, [%[r], #16]\n\t"
  5385. "ldr r4, [%[a], #20]\n\t"
  5386. "ldr r5, [%[b], #20]\n\t"
  5387. "adc r4, r5\n\t"
  5388. "str r4, [%[r], #20]\n\t"
  5389. "ldr r4, [%[a], #24]\n\t"
  5390. "ldr r5, [%[b], #24]\n\t"
  5391. "adc r4, r5\n\t"
  5392. "str r4, [%[r], #24]\n\t"
  5393. "ldr r4, [%[a], #28]\n\t"
  5394. "ldr r5, [%[b], #28]\n\t"
  5395. "adc r4, r5\n\t"
  5396. "str r4, [%[r], #28]\n\t"
  5397. "ldr r4, [%[a], #32]\n\t"
  5398. "ldr r5, [%[b], #32]\n\t"
  5399. "adc r4, r5\n\t"
  5400. "str r4, [%[r], #32]\n\t"
  5401. "ldr r4, [%[a], #36]\n\t"
  5402. "ldr r5, [%[b], #36]\n\t"
  5403. "adc r4, r5\n\t"
  5404. "str r4, [%[r], #36]\n\t"
  5405. "ldr r4, [%[a], #40]\n\t"
  5406. "ldr r5, [%[b], #40]\n\t"
  5407. "adc r4, r5\n\t"
  5408. "str r4, [%[r], #40]\n\t"
  5409. "ldr r4, [%[a], #44]\n\t"
  5410. "ldr r5, [%[b], #44]\n\t"
  5411. "adc r4, r5\n\t"
  5412. "str r4, [%[r], #44]\n\t"
  5413. "mov %[c], #0\n\t"
  5414. "adc %[c], %[c]\n\t"
  5415. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  5416. :
  5417. : "memory", "r4", "r5"
  5418. );
  5419. return c;
  5420. }
  5421. /* Sub b from a into r. (r = a - b)
  5422. *
  5423. * r A single precision integer.
  5424. * a A single precision integer.
  5425. * b A single precision integer.
  5426. */
  5427. SP_NOINLINE static sp_digit sp_3072_sub_in_place_24(sp_digit* a,
  5428. const sp_digit* b)
  5429. {
  5430. sp_digit c = 0;
  5431. __asm__ __volatile__ (
  5432. "ldr r3, [%[a], #0]\n\t"
  5433. "ldr r4, [%[a], #4]\n\t"
  5434. "ldr r5, [%[b], #0]\n\t"
  5435. "ldr r6, [%[b], #4]\n\t"
  5436. "sub r3, r5\n\t"
  5437. "sbc r4, r6\n\t"
  5438. "str r3, [%[a], #0]\n\t"
  5439. "str r4, [%[a], #4]\n\t"
  5440. "ldr r3, [%[a], #8]\n\t"
  5441. "ldr r4, [%[a], #12]\n\t"
  5442. "ldr r5, [%[b], #8]\n\t"
  5443. "ldr r6, [%[b], #12]\n\t"
  5444. "sbc r3, r5\n\t"
  5445. "sbc r4, r6\n\t"
  5446. "str r3, [%[a], #8]\n\t"
  5447. "str r4, [%[a], #12]\n\t"
  5448. "ldr r3, [%[a], #16]\n\t"
  5449. "ldr r4, [%[a], #20]\n\t"
  5450. "ldr r5, [%[b], #16]\n\t"
  5451. "ldr r6, [%[b], #20]\n\t"
  5452. "sbc r3, r5\n\t"
  5453. "sbc r4, r6\n\t"
  5454. "str r3, [%[a], #16]\n\t"
  5455. "str r4, [%[a], #20]\n\t"
  5456. "ldr r3, [%[a], #24]\n\t"
  5457. "ldr r4, [%[a], #28]\n\t"
  5458. "ldr r5, [%[b], #24]\n\t"
  5459. "ldr r6, [%[b], #28]\n\t"
  5460. "sbc r3, r5\n\t"
  5461. "sbc r4, r6\n\t"
  5462. "str r3, [%[a], #24]\n\t"
  5463. "str r4, [%[a], #28]\n\t"
  5464. "ldr r3, [%[a], #32]\n\t"
  5465. "ldr r4, [%[a], #36]\n\t"
  5466. "ldr r5, [%[b], #32]\n\t"
  5467. "ldr r6, [%[b], #36]\n\t"
  5468. "sbc r3, r5\n\t"
  5469. "sbc r4, r6\n\t"
  5470. "str r3, [%[a], #32]\n\t"
  5471. "str r4, [%[a], #36]\n\t"
  5472. "ldr r3, [%[a], #40]\n\t"
  5473. "ldr r4, [%[a], #44]\n\t"
  5474. "ldr r5, [%[b], #40]\n\t"
  5475. "ldr r6, [%[b], #44]\n\t"
  5476. "sbc r3, r5\n\t"
  5477. "sbc r4, r6\n\t"
  5478. "str r3, [%[a], #40]\n\t"
  5479. "str r4, [%[a], #44]\n\t"
  5480. "ldr r3, [%[a], #48]\n\t"
  5481. "ldr r4, [%[a], #52]\n\t"
  5482. "ldr r5, [%[b], #48]\n\t"
  5483. "ldr r6, [%[b], #52]\n\t"
  5484. "sbc r3, r5\n\t"
  5485. "sbc r4, r6\n\t"
  5486. "str r3, [%[a], #48]\n\t"
  5487. "str r4, [%[a], #52]\n\t"
  5488. "ldr r3, [%[a], #56]\n\t"
  5489. "ldr r4, [%[a], #60]\n\t"
  5490. "ldr r5, [%[b], #56]\n\t"
  5491. "ldr r6, [%[b], #60]\n\t"
  5492. "sbc r3, r5\n\t"
  5493. "sbc r4, r6\n\t"
  5494. "str r3, [%[a], #56]\n\t"
  5495. "str r4, [%[a], #60]\n\t"
  5496. "ldr r3, [%[a], #64]\n\t"
  5497. "ldr r4, [%[a], #68]\n\t"
  5498. "ldr r5, [%[b], #64]\n\t"
  5499. "ldr r6, [%[b], #68]\n\t"
  5500. "sbc r3, r5\n\t"
  5501. "sbc r4, r6\n\t"
  5502. "str r3, [%[a], #64]\n\t"
  5503. "str r4, [%[a], #68]\n\t"
  5504. "ldr r3, [%[a], #72]\n\t"
  5505. "ldr r4, [%[a], #76]\n\t"
  5506. "ldr r5, [%[b], #72]\n\t"
  5507. "ldr r6, [%[b], #76]\n\t"
  5508. "sbc r3, r5\n\t"
  5509. "sbc r4, r6\n\t"
  5510. "str r3, [%[a], #72]\n\t"
  5511. "str r4, [%[a], #76]\n\t"
  5512. "ldr r3, [%[a], #80]\n\t"
  5513. "ldr r4, [%[a], #84]\n\t"
  5514. "ldr r5, [%[b], #80]\n\t"
  5515. "ldr r6, [%[b], #84]\n\t"
  5516. "sbc r3, r5\n\t"
  5517. "sbc r4, r6\n\t"
  5518. "str r3, [%[a], #80]\n\t"
  5519. "str r4, [%[a], #84]\n\t"
  5520. "ldr r3, [%[a], #88]\n\t"
  5521. "ldr r4, [%[a], #92]\n\t"
  5522. "ldr r5, [%[b], #88]\n\t"
  5523. "ldr r6, [%[b], #92]\n\t"
  5524. "sbc r3, r5\n\t"
  5525. "sbc r4, r6\n\t"
  5526. "str r3, [%[a], #88]\n\t"
  5527. "str r4, [%[a], #92]\n\t"
  5528. "sbc %[c], %[c]\n\t"
  5529. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  5530. :
  5531. : "memory", "r3", "r4", "r5", "r6"
  5532. );
  5533. return c;
  5534. }
  5535. /* Add b to a into r. (r = a + b)
  5536. *
  5537. * r A single precision integer.
  5538. * a A single precision integer.
  5539. * b A single precision integer.
  5540. */
  5541. SP_NOINLINE static sp_digit sp_3072_add_24(sp_digit* r, const sp_digit* a,
  5542. const sp_digit* b)
  5543. {
  5544. sp_digit c = 0;
  5545. __asm__ __volatile__ (
  5546. "ldr r4, [%[a], #0]\n\t"
  5547. "ldr r5, [%[b], #0]\n\t"
  5548. "add r4, r5\n\t"
  5549. "str r4, [%[r], #0]\n\t"
  5550. "ldr r4, [%[a], #4]\n\t"
  5551. "ldr r5, [%[b], #4]\n\t"
  5552. "adc r4, r5\n\t"
  5553. "str r4, [%[r], #4]\n\t"
  5554. "ldr r4, [%[a], #8]\n\t"
  5555. "ldr r5, [%[b], #8]\n\t"
  5556. "adc r4, r5\n\t"
  5557. "str r4, [%[r], #8]\n\t"
  5558. "ldr r4, [%[a], #12]\n\t"
  5559. "ldr r5, [%[b], #12]\n\t"
  5560. "adc r4, r5\n\t"
  5561. "str r4, [%[r], #12]\n\t"
  5562. "ldr r4, [%[a], #16]\n\t"
  5563. "ldr r5, [%[b], #16]\n\t"
  5564. "adc r4, r5\n\t"
  5565. "str r4, [%[r], #16]\n\t"
  5566. "ldr r4, [%[a], #20]\n\t"
  5567. "ldr r5, [%[b], #20]\n\t"
  5568. "adc r4, r5\n\t"
  5569. "str r4, [%[r], #20]\n\t"
  5570. "ldr r4, [%[a], #24]\n\t"
  5571. "ldr r5, [%[b], #24]\n\t"
  5572. "adc r4, r5\n\t"
  5573. "str r4, [%[r], #24]\n\t"
  5574. "ldr r4, [%[a], #28]\n\t"
  5575. "ldr r5, [%[b], #28]\n\t"
  5576. "adc r4, r5\n\t"
  5577. "str r4, [%[r], #28]\n\t"
  5578. "ldr r4, [%[a], #32]\n\t"
  5579. "ldr r5, [%[b], #32]\n\t"
  5580. "adc r4, r5\n\t"
  5581. "str r4, [%[r], #32]\n\t"
  5582. "ldr r4, [%[a], #36]\n\t"
  5583. "ldr r5, [%[b], #36]\n\t"
  5584. "adc r4, r5\n\t"
  5585. "str r4, [%[r], #36]\n\t"
  5586. "ldr r4, [%[a], #40]\n\t"
  5587. "ldr r5, [%[b], #40]\n\t"
  5588. "adc r4, r5\n\t"
  5589. "str r4, [%[r], #40]\n\t"
  5590. "ldr r4, [%[a], #44]\n\t"
  5591. "ldr r5, [%[b], #44]\n\t"
  5592. "adc r4, r5\n\t"
  5593. "str r4, [%[r], #44]\n\t"
  5594. "ldr r4, [%[a], #48]\n\t"
  5595. "ldr r5, [%[b], #48]\n\t"
  5596. "adc r4, r5\n\t"
  5597. "str r4, [%[r], #48]\n\t"
  5598. "ldr r4, [%[a], #52]\n\t"
  5599. "ldr r5, [%[b], #52]\n\t"
  5600. "adc r4, r5\n\t"
  5601. "str r4, [%[r], #52]\n\t"
  5602. "ldr r4, [%[a], #56]\n\t"
  5603. "ldr r5, [%[b], #56]\n\t"
  5604. "adc r4, r5\n\t"
  5605. "str r4, [%[r], #56]\n\t"
  5606. "ldr r4, [%[a], #60]\n\t"
  5607. "ldr r5, [%[b], #60]\n\t"
  5608. "adc r4, r5\n\t"
  5609. "str r4, [%[r], #60]\n\t"
  5610. "ldr r4, [%[a], #64]\n\t"
  5611. "ldr r5, [%[b], #64]\n\t"
  5612. "adc r4, r5\n\t"
  5613. "str r4, [%[r], #64]\n\t"
  5614. "ldr r4, [%[a], #68]\n\t"
  5615. "ldr r5, [%[b], #68]\n\t"
  5616. "adc r4, r5\n\t"
  5617. "str r4, [%[r], #68]\n\t"
  5618. "ldr r4, [%[a], #72]\n\t"
  5619. "ldr r5, [%[b], #72]\n\t"
  5620. "adc r4, r5\n\t"
  5621. "str r4, [%[r], #72]\n\t"
  5622. "ldr r4, [%[a], #76]\n\t"
  5623. "ldr r5, [%[b], #76]\n\t"
  5624. "adc r4, r5\n\t"
  5625. "str r4, [%[r], #76]\n\t"
  5626. "ldr r4, [%[a], #80]\n\t"
  5627. "ldr r5, [%[b], #80]\n\t"
  5628. "adc r4, r5\n\t"
  5629. "str r4, [%[r], #80]\n\t"
  5630. "ldr r4, [%[a], #84]\n\t"
  5631. "ldr r5, [%[b], #84]\n\t"
  5632. "adc r4, r5\n\t"
  5633. "str r4, [%[r], #84]\n\t"
  5634. "ldr r4, [%[a], #88]\n\t"
  5635. "ldr r5, [%[b], #88]\n\t"
  5636. "adc r4, r5\n\t"
  5637. "str r4, [%[r], #88]\n\t"
  5638. "ldr r4, [%[a], #92]\n\t"
  5639. "ldr r5, [%[b], #92]\n\t"
  5640. "adc r4, r5\n\t"
  5641. "str r4, [%[r], #92]\n\t"
  5642. "mov %[c], #0\n\t"
  5643. "adc %[c], %[c]\n\t"
  5644. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  5645. :
  5646. : "memory", "r4", "r5"
  5647. );
  5648. return c;
  5649. }
  5650. /* AND m into each word of a and store in r.
  5651. *
  5652. * r A single precision integer.
  5653. * a A single precision integer.
  5654. * m Mask to AND against each digit.
  5655. */
  5656. static void sp_3072_mask_12(sp_digit* r, const sp_digit* a, sp_digit m)
  5657. {
  5658. #ifdef WOLFSSL_SP_SMALL
  5659. int i;
  5660. for (i=0; i<12; i++) {
  5661. r[i] = a[i] & m;
  5662. }
  5663. #else
  5664. r[0] = a[0] & m;
  5665. r[1] = a[1] & m;
  5666. r[2] = a[2] & m;
  5667. r[3] = a[3] & m;
  5668. r[4] = a[4] & m;
  5669. r[5] = a[5] & m;
  5670. r[6] = a[6] & m;
  5671. r[7] = a[7] & m;
  5672. r[8] = a[8] & m;
  5673. r[9] = a[9] & m;
  5674. r[10] = a[10] & m;
  5675. r[11] = a[11] & m;
  5676. #endif
  5677. }
  5678. /* Multiply a and b into r. (r = a * b)
  5679. *
  5680. * r A single precision integer.
  5681. * a A single precision integer.
  5682. * b A single precision integer.
  5683. */
  5684. SP_NOINLINE static void sp_3072_mul_24(sp_digit* r, const sp_digit* a,
  5685. const sp_digit* b)
  5686. {
  5687. sp_digit* z0 = r;
  5688. sp_digit z1[24];
  5689. sp_digit a1[12];
  5690. sp_digit b1[12];
  5691. sp_digit z2[24];
  5692. sp_digit u, ca, cb;
  5693. ca = sp_3072_add_12(a1, a, &a[12]);
  5694. cb = sp_3072_add_12(b1, b, &b[12]);
  5695. u = ca & cb;
  5696. sp_3072_mul_12(z1, a1, b1);
  5697. sp_3072_mul_12(z2, &a[12], &b[12]);
  5698. sp_3072_mul_12(z0, a, b);
  5699. sp_3072_mask_12(r + 24, a1, 0 - cb);
  5700. sp_3072_mask_12(b1, b1, 0 - ca);
  5701. u += sp_3072_add_12(r + 24, r + 24, b1);
  5702. u += sp_3072_sub_in_place_24(z1, z2);
  5703. u += sp_3072_sub_in_place_24(z1, z0);
  5704. u += sp_3072_add_24(r + 12, r + 12, z1);
  5705. r[36] = u;
  5706. XMEMSET(r + 36 + 1, 0, sizeof(sp_digit) * (12 - 1));
  5707. (void)sp_3072_add_24(r + 24, r + 24, z2);
  5708. }
  5709. /* Square a and put result in r. (r = a * a)
  5710. *
  5711. * r A single precision integer.
  5712. * a A single precision integer.
  5713. */
  5714. SP_NOINLINE static void sp_3072_sqr_24(sp_digit* r, const sp_digit* a)
  5715. {
  5716. sp_digit* z0 = r;
  5717. sp_digit z2[24];
  5718. sp_digit z1[24];
  5719. sp_digit a1[12];
  5720. sp_digit u;
  5721. u = sp_3072_add_12(a1, a, &a[12]);
  5722. sp_3072_sqr_12(z1, a1);
  5723. sp_3072_sqr_12(z2, &a[12]);
  5724. sp_3072_sqr_12(z0, a);
  5725. sp_3072_mask_12(r + 24, a1, 0 - u);
  5726. u += sp_3072_add_12(r + 24, r + 24, r + 24);
  5727. u += sp_3072_sub_in_place_24(z1, z2);
  5728. u += sp_3072_sub_in_place_24(z1, z0);
  5729. u += sp_3072_add_24(r + 12, r + 12, z1);
  5730. r[36] = u;
  5731. XMEMSET(r + 36 + 1, 0, sizeof(sp_digit) * (12 - 1));
  5732. (void)sp_3072_add_24(r + 24, r + 24, z2);
  5733. }
  5734. /* Sub b from a into r. (r = a - b)
  5735. *
  5736. * r A single precision integer.
  5737. * a A single precision integer.
  5738. * b A single precision integer.
  5739. */
  5740. SP_NOINLINE static sp_digit sp_3072_sub_in_place_48(sp_digit* a,
  5741. const sp_digit* b)
  5742. {
  5743. sp_digit c = 0;
  5744. __asm__ __volatile__ (
  5745. "ldr r3, [%[a], #0]\n\t"
  5746. "ldr r4, [%[a], #4]\n\t"
  5747. "ldr r5, [%[b], #0]\n\t"
  5748. "ldr r6, [%[b], #4]\n\t"
  5749. "sub r3, r5\n\t"
  5750. "sbc r4, r6\n\t"
  5751. "str r3, [%[a], #0]\n\t"
  5752. "str r4, [%[a], #4]\n\t"
  5753. "ldr r3, [%[a], #8]\n\t"
  5754. "ldr r4, [%[a], #12]\n\t"
  5755. "ldr r5, [%[b], #8]\n\t"
  5756. "ldr r6, [%[b], #12]\n\t"
  5757. "sbc r3, r5\n\t"
  5758. "sbc r4, r6\n\t"
  5759. "str r3, [%[a], #8]\n\t"
  5760. "str r4, [%[a], #12]\n\t"
  5761. "ldr r3, [%[a], #16]\n\t"
  5762. "ldr r4, [%[a], #20]\n\t"
  5763. "ldr r5, [%[b], #16]\n\t"
  5764. "ldr r6, [%[b], #20]\n\t"
  5765. "sbc r3, r5\n\t"
  5766. "sbc r4, r6\n\t"
  5767. "str r3, [%[a], #16]\n\t"
  5768. "str r4, [%[a], #20]\n\t"
  5769. "ldr r3, [%[a], #24]\n\t"
  5770. "ldr r4, [%[a], #28]\n\t"
  5771. "ldr r5, [%[b], #24]\n\t"
  5772. "ldr r6, [%[b], #28]\n\t"
  5773. "sbc r3, r5\n\t"
  5774. "sbc r4, r6\n\t"
  5775. "str r3, [%[a], #24]\n\t"
  5776. "str r4, [%[a], #28]\n\t"
  5777. "ldr r3, [%[a], #32]\n\t"
  5778. "ldr r4, [%[a], #36]\n\t"
  5779. "ldr r5, [%[b], #32]\n\t"
  5780. "ldr r6, [%[b], #36]\n\t"
  5781. "sbc r3, r5\n\t"
  5782. "sbc r4, r6\n\t"
  5783. "str r3, [%[a], #32]\n\t"
  5784. "str r4, [%[a], #36]\n\t"
  5785. "ldr r3, [%[a], #40]\n\t"
  5786. "ldr r4, [%[a], #44]\n\t"
  5787. "ldr r5, [%[b], #40]\n\t"
  5788. "ldr r6, [%[b], #44]\n\t"
  5789. "sbc r3, r5\n\t"
  5790. "sbc r4, r6\n\t"
  5791. "str r3, [%[a], #40]\n\t"
  5792. "str r4, [%[a], #44]\n\t"
  5793. "ldr r3, [%[a], #48]\n\t"
  5794. "ldr r4, [%[a], #52]\n\t"
  5795. "ldr r5, [%[b], #48]\n\t"
  5796. "ldr r6, [%[b], #52]\n\t"
  5797. "sbc r3, r5\n\t"
  5798. "sbc r4, r6\n\t"
  5799. "str r3, [%[a], #48]\n\t"
  5800. "str r4, [%[a], #52]\n\t"
  5801. "ldr r3, [%[a], #56]\n\t"
  5802. "ldr r4, [%[a], #60]\n\t"
  5803. "ldr r5, [%[b], #56]\n\t"
  5804. "ldr r6, [%[b], #60]\n\t"
  5805. "sbc r3, r5\n\t"
  5806. "sbc r4, r6\n\t"
  5807. "str r3, [%[a], #56]\n\t"
  5808. "str r4, [%[a], #60]\n\t"
  5809. "ldr r3, [%[a], #64]\n\t"
  5810. "ldr r4, [%[a], #68]\n\t"
  5811. "ldr r5, [%[b], #64]\n\t"
  5812. "ldr r6, [%[b], #68]\n\t"
  5813. "sbc r3, r5\n\t"
  5814. "sbc r4, r6\n\t"
  5815. "str r3, [%[a], #64]\n\t"
  5816. "str r4, [%[a], #68]\n\t"
  5817. "ldr r3, [%[a], #72]\n\t"
  5818. "ldr r4, [%[a], #76]\n\t"
  5819. "ldr r5, [%[b], #72]\n\t"
  5820. "ldr r6, [%[b], #76]\n\t"
  5821. "sbc r3, r5\n\t"
  5822. "sbc r4, r6\n\t"
  5823. "str r3, [%[a], #72]\n\t"
  5824. "str r4, [%[a], #76]\n\t"
  5825. "ldr r3, [%[a], #80]\n\t"
  5826. "ldr r4, [%[a], #84]\n\t"
  5827. "ldr r5, [%[b], #80]\n\t"
  5828. "ldr r6, [%[b], #84]\n\t"
  5829. "sbc r3, r5\n\t"
  5830. "sbc r4, r6\n\t"
  5831. "str r3, [%[a], #80]\n\t"
  5832. "str r4, [%[a], #84]\n\t"
  5833. "ldr r3, [%[a], #88]\n\t"
  5834. "ldr r4, [%[a], #92]\n\t"
  5835. "ldr r5, [%[b], #88]\n\t"
  5836. "ldr r6, [%[b], #92]\n\t"
  5837. "sbc r3, r5\n\t"
  5838. "sbc r4, r6\n\t"
  5839. "str r3, [%[a], #88]\n\t"
  5840. "str r4, [%[a], #92]\n\t"
  5841. "ldr r3, [%[a], #96]\n\t"
  5842. "ldr r4, [%[a], #100]\n\t"
  5843. "ldr r5, [%[b], #96]\n\t"
  5844. "ldr r6, [%[b], #100]\n\t"
  5845. "sbc r3, r5\n\t"
  5846. "sbc r4, r6\n\t"
  5847. "str r3, [%[a], #96]\n\t"
  5848. "str r4, [%[a], #100]\n\t"
  5849. "ldr r3, [%[a], #104]\n\t"
  5850. "ldr r4, [%[a], #108]\n\t"
  5851. "ldr r5, [%[b], #104]\n\t"
  5852. "ldr r6, [%[b], #108]\n\t"
  5853. "sbc r3, r5\n\t"
  5854. "sbc r4, r6\n\t"
  5855. "str r3, [%[a], #104]\n\t"
  5856. "str r4, [%[a], #108]\n\t"
  5857. "ldr r3, [%[a], #112]\n\t"
  5858. "ldr r4, [%[a], #116]\n\t"
  5859. "ldr r5, [%[b], #112]\n\t"
  5860. "ldr r6, [%[b], #116]\n\t"
  5861. "sbc r3, r5\n\t"
  5862. "sbc r4, r6\n\t"
  5863. "str r3, [%[a], #112]\n\t"
  5864. "str r4, [%[a], #116]\n\t"
  5865. "ldr r3, [%[a], #120]\n\t"
  5866. "ldr r4, [%[a], #124]\n\t"
  5867. "ldr r5, [%[b], #120]\n\t"
  5868. "ldr r6, [%[b], #124]\n\t"
  5869. "sbc r3, r5\n\t"
  5870. "sbc r4, r6\n\t"
  5871. "str r3, [%[a], #120]\n\t"
  5872. "str r4, [%[a], #124]\n\t"
  5873. "sbc %[c], %[c]\n\t"
  5874. "add %[a], #0x80\n\t"
  5875. "add %[b], #0x80\n\t"
  5876. "mov r5, #0\n\t"
  5877. "sub r5, %[c]\n\t"
  5878. "ldr r3, [%[a], #0]\n\t"
  5879. "ldr r4, [%[a], #4]\n\t"
  5880. "ldr r5, [%[b], #0]\n\t"
  5881. "ldr r6, [%[b], #4]\n\t"
  5882. "sbc r3, r5\n\t"
  5883. "sbc r4, r6\n\t"
  5884. "str r3, [%[a], #0]\n\t"
  5885. "str r4, [%[a], #4]\n\t"
  5886. "ldr r3, [%[a], #8]\n\t"
  5887. "ldr r4, [%[a], #12]\n\t"
  5888. "ldr r5, [%[b], #8]\n\t"
  5889. "ldr r6, [%[b], #12]\n\t"
  5890. "sbc r3, r5\n\t"
  5891. "sbc r4, r6\n\t"
  5892. "str r3, [%[a], #8]\n\t"
  5893. "str r4, [%[a], #12]\n\t"
  5894. "ldr r3, [%[a], #16]\n\t"
  5895. "ldr r4, [%[a], #20]\n\t"
  5896. "ldr r5, [%[b], #16]\n\t"
  5897. "ldr r6, [%[b], #20]\n\t"
  5898. "sbc r3, r5\n\t"
  5899. "sbc r4, r6\n\t"
  5900. "str r3, [%[a], #16]\n\t"
  5901. "str r4, [%[a], #20]\n\t"
  5902. "ldr r3, [%[a], #24]\n\t"
  5903. "ldr r4, [%[a], #28]\n\t"
  5904. "ldr r5, [%[b], #24]\n\t"
  5905. "ldr r6, [%[b], #28]\n\t"
  5906. "sbc r3, r5\n\t"
  5907. "sbc r4, r6\n\t"
  5908. "str r3, [%[a], #24]\n\t"
  5909. "str r4, [%[a], #28]\n\t"
  5910. "ldr r3, [%[a], #32]\n\t"
  5911. "ldr r4, [%[a], #36]\n\t"
  5912. "ldr r5, [%[b], #32]\n\t"
  5913. "ldr r6, [%[b], #36]\n\t"
  5914. "sbc r3, r5\n\t"
  5915. "sbc r4, r6\n\t"
  5916. "str r3, [%[a], #32]\n\t"
  5917. "str r4, [%[a], #36]\n\t"
  5918. "ldr r3, [%[a], #40]\n\t"
  5919. "ldr r4, [%[a], #44]\n\t"
  5920. "ldr r5, [%[b], #40]\n\t"
  5921. "ldr r6, [%[b], #44]\n\t"
  5922. "sbc r3, r5\n\t"
  5923. "sbc r4, r6\n\t"
  5924. "str r3, [%[a], #40]\n\t"
  5925. "str r4, [%[a], #44]\n\t"
  5926. "ldr r3, [%[a], #48]\n\t"
  5927. "ldr r4, [%[a], #52]\n\t"
  5928. "ldr r5, [%[b], #48]\n\t"
  5929. "ldr r6, [%[b], #52]\n\t"
  5930. "sbc r3, r5\n\t"
  5931. "sbc r4, r6\n\t"
  5932. "str r3, [%[a], #48]\n\t"
  5933. "str r4, [%[a], #52]\n\t"
  5934. "ldr r3, [%[a], #56]\n\t"
  5935. "ldr r4, [%[a], #60]\n\t"
  5936. "ldr r5, [%[b], #56]\n\t"
  5937. "ldr r6, [%[b], #60]\n\t"
  5938. "sbc r3, r5\n\t"
  5939. "sbc r4, r6\n\t"
  5940. "str r3, [%[a], #56]\n\t"
  5941. "str r4, [%[a], #60]\n\t"
  5942. "sbc %[c], %[c]\n\t"
  5943. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  5944. :
  5945. : "memory", "r3", "r4", "r5", "r6"
  5946. );
  5947. return c;
  5948. }
  5949. /* Add b to a into r. (r = a + b)
  5950. *
  5951. * r A single precision integer.
  5952. * a A single precision integer.
  5953. * b A single precision integer.
  5954. */
  5955. SP_NOINLINE static sp_digit sp_3072_add_48(sp_digit* r, const sp_digit* a,
  5956. const sp_digit* b)
  5957. {
  5958. sp_digit c = 0;
  5959. __asm__ __volatile__ (
  5960. "mov r7, #0\n\t"
  5961. "mvn r7, r7\n\t"
  5962. "ldr r4, [%[a], #0]\n\t"
  5963. "ldr r5, [%[b], #0]\n\t"
  5964. "add r4, r5\n\t"
  5965. "str r4, [%[r], #0]\n\t"
  5966. "ldr r4, [%[a], #4]\n\t"
  5967. "ldr r5, [%[b], #4]\n\t"
  5968. "adc r4, r5\n\t"
  5969. "str r4, [%[r], #4]\n\t"
  5970. "ldr r4, [%[a], #8]\n\t"
  5971. "ldr r5, [%[b], #8]\n\t"
  5972. "adc r4, r5\n\t"
  5973. "str r4, [%[r], #8]\n\t"
  5974. "ldr r4, [%[a], #12]\n\t"
  5975. "ldr r5, [%[b], #12]\n\t"
  5976. "adc r4, r5\n\t"
  5977. "str r4, [%[r], #12]\n\t"
  5978. "ldr r4, [%[a], #16]\n\t"
  5979. "ldr r5, [%[b], #16]\n\t"
  5980. "adc r4, r5\n\t"
  5981. "str r4, [%[r], #16]\n\t"
  5982. "ldr r4, [%[a], #20]\n\t"
  5983. "ldr r5, [%[b], #20]\n\t"
  5984. "adc r4, r5\n\t"
  5985. "str r4, [%[r], #20]\n\t"
  5986. "ldr r4, [%[a], #24]\n\t"
  5987. "ldr r5, [%[b], #24]\n\t"
  5988. "adc r4, r5\n\t"
  5989. "str r4, [%[r], #24]\n\t"
  5990. "ldr r4, [%[a], #28]\n\t"
  5991. "ldr r5, [%[b], #28]\n\t"
  5992. "adc r4, r5\n\t"
  5993. "str r4, [%[r], #28]\n\t"
  5994. "ldr r4, [%[a], #32]\n\t"
  5995. "ldr r5, [%[b], #32]\n\t"
  5996. "adc r4, r5\n\t"
  5997. "str r4, [%[r], #32]\n\t"
  5998. "ldr r4, [%[a], #36]\n\t"
  5999. "ldr r5, [%[b], #36]\n\t"
  6000. "adc r4, r5\n\t"
  6001. "str r4, [%[r], #36]\n\t"
  6002. "ldr r4, [%[a], #40]\n\t"
  6003. "ldr r5, [%[b], #40]\n\t"
  6004. "adc r4, r5\n\t"
  6005. "str r4, [%[r], #40]\n\t"
  6006. "ldr r4, [%[a], #44]\n\t"
  6007. "ldr r5, [%[b], #44]\n\t"
  6008. "adc r4, r5\n\t"
  6009. "str r4, [%[r], #44]\n\t"
  6010. "ldr r4, [%[a], #48]\n\t"
  6011. "ldr r5, [%[b], #48]\n\t"
  6012. "adc r4, r5\n\t"
  6013. "str r4, [%[r], #48]\n\t"
  6014. "ldr r4, [%[a], #52]\n\t"
  6015. "ldr r5, [%[b], #52]\n\t"
  6016. "adc r4, r5\n\t"
  6017. "str r4, [%[r], #52]\n\t"
  6018. "ldr r4, [%[a], #56]\n\t"
  6019. "ldr r5, [%[b], #56]\n\t"
  6020. "adc r4, r5\n\t"
  6021. "str r4, [%[r], #56]\n\t"
  6022. "ldr r4, [%[a], #60]\n\t"
  6023. "ldr r5, [%[b], #60]\n\t"
  6024. "adc r4, r5\n\t"
  6025. "str r4, [%[r], #60]\n\t"
  6026. "ldr r4, [%[a], #64]\n\t"
  6027. "ldr r5, [%[b], #64]\n\t"
  6028. "adc r4, r5\n\t"
  6029. "str r4, [%[r], #64]\n\t"
  6030. "ldr r4, [%[a], #68]\n\t"
  6031. "ldr r5, [%[b], #68]\n\t"
  6032. "adc r4, r5\n\t"
  6033. "str r4, [%[r], #68]\n\t"
  6034. "ldr r4, [%[a], #72]\n\t"
  6035. "ldr r5, [%[b], #72]\n\t"
  6036. "adc r4, r5\n\t"
  6037. "str r4, [%[r], #72]\n\t"
  6038. "ldr r4, [%[a], #76]\n\t"
  6039. "ldr r5, [%[b], #76]\n\t"
  6040. "adc r4, r5\n\t"
  6041. "str r4, [%[r], #76]\n\t"
  6042. "ldr r4, [%[a], #80]\n\t"
  6043. "ldr r5, [%[b], #80]\n\t"
  6044. "adc r4, r5\n\t"
  6045. "str r4, [%[r], #80]\n\t"
  6046. "ldr r4, [%[a], #84]\n\t"
  6047. "ldr r5, [%[b], #84]\n\t"
  6048. "adc r4, r5\n\t"
  6049. "str r4, [%[r], #84]\n\t"
  6050. "ldr r4, [%[a], #88]\n\t"
  6051. "ldr r5, [%[b], #88]\n\t"
  6052. "adc r4, r5\n\t"
  6053. "str r4, [%[r], #88]\n\t"
  6054. "ldr r4, [%[a], #92]\n\t"
  6055. "ldr r5, [%[b], #92]\n\t"
  6056. "adc r4, r5\n\t"
  6057. "str r4, [%[r], #92]\n\t"
  6058. "ldr r4, [%[a], #96]\n\t"
  6059. "ldr r5, [%[b], #96]\n\t"
  6060. "adc r4, r5\n\t"
  6061. "str r4, [%[r], #96]\n\t"
  6062. "ldr r4, [%[a], #100]\n\t"
  6063. "ldr r5, [%[b], #100]\n\t"
  6064. "adc r4, r5\n\t"
  6065. "str r4, [%[r], #100]\n\t"
  6066. "ldr r4, [%[a], #104]\n\t"
  6067. "ldr r5, [%[b], #104]\n\t"
  6068. "adc r4, r5\n\t"
  6069. "str r4, [%[r], #104]\n\t"
  6070. "ldr r4, [%[a], #108]\n\t"
  6071. "ldr r5, [%[b], #108]\n\t"
  6072. "adc r4, r5\n\t"
  6073. "str r4, [%[r], #108]\n\t"
  6074. "ldr r4, [%[a], #112]\n\t"
  6075. "ldr r5, [%[b], #112]\n\t"
  6076. "adc r4, r5\n\t"
  6077. "str r4, [%[r], #112]\n\t"
  6078. "ldr r4, [%[a], #116]\n\t"
  6079. "ldr r5, [%[b], #116]\n\t"
  6080. "adc r4, r5\n\t"
  6081. "str r4, [%[r], #116]\n\t"
  6082. "ldr r4, [%[a], #120]\n\t"
  6083. "ldr r5, [%[b], #120]\n\t"
  6084. "adc r4, r5\n\t"
  6085. "str r4, [%[r], #120]\n\t"
  6086. "ldr r4, [%[a], #124]\n\t"
  6087. "ldr r5, [%[b], #124]\n\t"
  6088. "adc r4, r5\n\t"
  6089. "str r4, [%[r], #124]\n\t"
  6090. "mov %[c], #0\n\t"
  6091. "adc %[c], %[c]\n\t"
  6092. "add %[a], #0x80\n\t"
  6093. "add %[b], #0x80\n\t"
  6094. "add %[r], #0x80\n\t"
  6095. "add %[c], r7\n\t"
  6096. "ldr r4, [%[a], #0]\n\t"
  6097. "ldr r5, [%[b], #0]\n\t"
  6098. "adc r4, r5\n\t"
  6099. "str r4, [%[r], #0]\n\t"
  6100. "ldr r4, [%[a], #4]\n\t"
  6101. "ldr r5, [%[b], #4]\n\t"
  6102. "adc r4, r5\n\t"
  6103. "str r4, [%[r], #4]\n\t"
  6104. "ldr r4, [%[a], #8]\n\t"
  6105. "ldr r5, [%[b], #8]\n\t"
  6106. "adc r4, r5\n\t"
  6107. "str r4, [%[r], #8]\n\t"
  6108. "ldr r4, [%[a], #12]\n\t"
  6109. "ldr r5, [%[b], #12]\n\t"
  6110. "adc r4, r5\n\t"
  6111. "str r4, [%[r], #12]\n\t"
  6112. "ldr r4, [%[a], #16]\n\t"
  6113. "ldr r5, [%[b], #16]\n\t"
  6114. "adc r4, r5\n\t"
  6115. "str r4, [%[r], #16]\n\t"
  6116. "ldr r4, [%[a], #20]\n\t"
  6117. "ldr r5, [%[b], #20]\n\t"
  6118. "adc r4, r5\n\t"
  6119. "str r4, [%[r], #20]\n\t"
  6120. "ldr r4, [%[a], #24]\n\t"
  6121. "ldr r5, [%[b], #24]\n\t"
  6122. "adc r4, r5\n\t"
  6123. "str r4, [%[r], #24]\n\t"
  6124. "ldr r4, [%[a], #28]\n\t"
  6125. "ldr r5, [%[b], #28]\n\t"
  6126. "adc r4, r5\n\t"
  6127. "str r4, [%[r], #28]\n\t"
  6128. "ldr r4, [%[a], #32]\n\t"
  6129. "ldr r5, [%[b], #32]\n\t"
  6130. "adc r4, r5\n\t"
  6131. "str r4, [%[r], #32]\n\t"
  6132. "ldr r4, [%[a], #36]\n\t"
  6133. "ldr r5, [%[b], #36]\n\t"
  6134. "adc r4, r5\n\t"
  6135. "str r4, [%[r], #36]\n\t"
  6136. "ldr r4, [%[a], #40]\n\t"
  6137. "ldr r5, [%[b], #40]\n\t"
  6138. "adc r4, r5\n\t"
  6139. "str r4, [%[r], #40]\n\t"
  6140. "ldr r4, [%[a], #44]\n\t"
  6141. "ldr r5, [%[b], #44]\n\t"
  6142. "adc r4, r5\n\t"
  6143. "str r4, [%[r], #44]\n\t"
  6144. "ldr r4, [%[a], #48]\n\t"
  6145. "ldr r5, [%[b], #48]\n\t"
  6146. "adc r4, r5\n\t"
  6147. "str r4, [%[r], #48]\n\t"
  6148. "ldr r4, [%[a], #52]\n\t"
  6149. "ldr r5, [%[b], #52]\n\t"
  6150. "adc r4, r5\n\t"
  6151. "str r4, [%[r], #52]\n\t"
  6152. "ldr r4, [%[a], #56]\n\t"
  6153. "ldr r5, [%[b], #56]\n\t"
  6154. "adc r4, r5\n\t"
  6155. "str r4, [%[r], #56]\n\t"
  6156. "ldr r4, [%[a], #60]\n\t"
  6157. "ldr r5, [%[b], #60]\n\t"
  6158. "adc r4, r5\n\t"
  6159. "str r4, [%[r], #60]\n\t"
  6160. "mov %[c], #0\n\t"
  6161. "adc %[c], %[c]\n\t"
  6162. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  6163. :
  6164. : "memory", "r4", "r5", "r7"
  6165. );
  6166. return c;
  6167. }
  6168. /* AND m into each word of a and store in r.
  6169. *
  6170. * r A single precision integer.
  6171. * a A single precision integer.
  6172. * m Mask to AND against each digit.
  6173. */
  6174. static void sp_3072_mask_24(sp_digit* r, const sp_digit* a, sp_digit m)
  6175. {
  6176. #ifdef WOLFSSL_SP_SMALL
  6177. int i;
  6178. for (i=0; i<24; i++) {
  6179. r[i] = a[i] & m;
  6180. }
  6181. #else
  6182. int i;
  6183. for (i = 0; i < 24; i += 8) {
  6184. r[i+0] = a[i+0] & m;
  6185. r[i+1] = a[i+1] & m;
  6186. r[i+2] = a[i+2] & m;
  6187. r[i+3] = a[i+3] & m;
  6188. r[i+4] = a[i+4] & m;
  6189. r[i+5] = a[i+5] & m;
  6190. r[i+6] = a[i+6] & m;
  6191. r[i+7] = a[i+7] & m;
  6192. }
  6193. #endif
  6194. }
  6195. /* Multiply a and b into r. (r = a * b)
  6196. *
  6197. * r A single precision integer.
  6198. * a A single precision integer.
  6199. * b A single precision integer.
  6200. */
  6201. SP_NOINLINE static void sp_3072_mul_48(sp_digit* r, const sp_digit* a,
  6202. const sp_digit* b)
  6203. {
  6204. sp_digit* z0 = r;
  6205. sp_digit z1[48];
  6206. sp_digit a1[24];
  6207. sp_digit b1[24];
  6208. sp_digit z2[48];
  6209. sp_digit u, ca, cb;
  6210. ca = sp_3072_add_24(a1, a, &a[24]);
  6211. cb = sp_3072_add_24(b1, b, &b[24]);
  6212. u = ca & cb;
  6213. sp_3072_mul_24(z1, a1, b1);
  6214. sp_3072_mul_24(z2, &a[24], &b[24]);
  6215. sp_3072_mul_24(z0, a, b);
  6216. sp_3072_mask_24(r + 48, a1, 0 - cb);
  6217. sp_3072_mask_24(b1, b1, 0 - ca);
  6218. u += sp_3072_add_24(r + 48, r + 48, b1);
  6219. u += sp_3072_sub_in_place_48(z1, z2);
  6220. u += sp_3072_sub_in_place_48(z1, z0);
  6221. u += sp_3072_add_48(r + 24, r + 24, z1);
  6222. r[72] = u;
  6223. XMEMSET(r + 72 + 1, 0, sizeof(sp_digit) * (24 - 1));
  6224. (void)sp_3072_add_48(r + 48, r + 48, z2);
  6225. }
  6226. /* Square a and put result in r. (r = a * a)
  6227. *
  6228. * r A single precision integer.
  6229. * a A single precision integer.
  6230. */
  6231. SP_NOINLINE static void sp_3072_sqr_48(sp_digit* r, const sp_digit* a)
  6232. {
  6233. sp_digit* z0 = r;
  6234. sp_digit z2[48];
  6235. sp_digit z1[48];
  6236. sp_digit a1[24];
  6237. sp_digit u;
  6238. u = sp_3072_add_24(a1, a, &a[24]);
  6239. sp_3072_sqr_24(z1, a1);
  6240. sp_3072_sqr_24(z2, &a[24]);
  6241. sp_3072_sqr_24(z0, a);
  6242. sp_3072_mask_24(r + 48, a1, 0 - u);
  6243. u += sp_3072_add_24(r + 48, r + 48, r + 48);
  6244. u += sp_3072_sub_in_place_48(z1, z2);
  6245. u += sp_3072_sub_in_place_48(z1, z0);
  6246. u += sp_3072_add_48(r + 24, r + 24, z1);
  6247. r[72] = u;
  6248. XMEMSET(r + 72 + 1, 0, sizeof(sp_digit) * (24 - 1));
  6249. (void)sp_3072_add_48(r + 48, r + 48, z2);
  6250. }
  6251. /* Sub b from a into r. (r = a - b)
  6252. *
  6253. * r A single precision integer.
  6254. * a A single precision integer.
  6255. * b A single precision integer.
  6256. */
  6257. SP_NOINLINE static sp_digit sp_3072_sub_in_place_96(sp_digit* a,
  6258. const sp_digit* b)
  6259. {
  6260. sp_digit c = 0;
  6261. __asm__ __volatile__ (
  6262. "ldr r3, [%[a], #0]\n\t"
  6263. "ldr r4, [%[a], #4]\n\t"
  6264. "ldr r5, [%[b], #0]\n\t"
  6265. "ldr r6, [%[b], #4]\n\t"
  6266. "sub r3, r5\n\t"
  6267. "sbc r4, r6\n\t"
  6268. "str r3, [%[a], #0]\n\t"
  6269. "str r4, [%[a], #4]\n\t"
  6270. "ldr r3, [%[a], #8]\n\t"
  6271. "ldr r4, [%[a], #12]\n\t"
  6272. "ldr r5, [%[b], #8]\n\t"
  6273. "ldr r6, [%[b], #12]\n\t"
  6274. "sbc r3, r5\n\t"
  6275. "sbc r4, r6\n\t"
  6276. "str r3, [%[a], #8]\n\t"
  6277. "str r4, [%[a], #12]\n\t"
  6278. "ldr r3, [%[a], #16]\n\t"
  6279. "ldr r4, [%[a], #20]\n\t"
  6280. "ldr r5, [%[b], #16]\n\t"
  6281. "ldr r6, [%[b], #20]\n\t"
  6282. "sbc r3, r5\n\t"
  6283. "sbc r4, r6\n\t"
  6284. "str r3, [%[a], #16]\n\t"
  6285. "str r4, [%[a], #20]\n\t"
  6286. "ldr r3, [%[a], #24]\n\t"
  6287. "ldr r4, [%[a], #28]\n\t"
  6288. "ldr r5, [%[b], #24]\n\t"
  6289. "ldr r6, [%[b], #28]\n\t"
  6290. "sbc r3, r5\n\t"
  6291. "sbc r4, r6\n\t"
  6292. "str r3, [%[a], #24]\n\t"
  6293. "str r4, [%[a], #28]\n\t"
  6294. "ldr r3, [%[a], #32]\n\t"
  6295. "ldr r4, [%[a], #36]\n\t"
  6296. "ldr r5, [%[b], #32]\n\t"
  6297. "ldr r6, [%[b], #36]\n\t"
  6298. "sbc r3, r5\n\t"
  6299. "sbc r4, r6\n\t"
  6300. "str r3, [%[a], #32]\n\t"
  6301. "str r4, [%[a], #36]\n\t"
  6302. "ldr r3, [%[a], #40]\n\t"
  6303. "ldr r4, [%[a], #44]\n\t"
  6304. "ldr r5, [%[b], #40]\n\t"
  6305. "ldr r6, [%[b], #44]\n\t"
  6306. "sbc r3, r5\n\t"
  6307. "sbc r4, r6\n\t"
  6308. "str r3, [%[a], #40]\n\t"
  6309. "str r4, [%[a], #44]\n\t"
  6310. "ldr r3, [%[a], #48]\n\t"
  6311. "ldr r4, [%[a], #52]\n\t"
  6312. "ldr r5, [%[b], #48]\n\t"
  6313. "ldr r6, [%[b], #52]\n\t"
  6314. "sbc r3, r5\n\t"
  6315. "sbc r4, r6\n\t"
  6316. "str r3, [%[a], #48]\n\t"
  6317. "str r4, [%[a], #52]\n\t"
  6318. "ldr r3, [%[a], #56]\n\t"
  6319. "ldr r4, [%[a], #60]\n\t"
  6320. "ldr r5, [%[b], #56]\n\t"
  6321. "ldr r6, [%[b], #60]\n\t"
  6322. "sbc r3, r5\n\t"
  6323. "sbc r4, r6\n\t"
  6324. "str r3, [%[a], #56]\n\t"
  6325. "str r4, [%[a], #60]\n\t"
  6326. "ldr r3, [%[a], #64]\n\t"
  6327. "ldr r4, [%[a], #68]\n\t"
  6328. "ldr r5, [%[b], #64]\n\t"
  6329. "ldr r6, [%[b], #68]\n\t"
  6330. "sbc r3, r5\n\t"
  6331. "sbc r4, r6\n\t"
  6332. "str r3, [%[a], #64]\n\t"
  6333. "str r4, [%[a], #68]\n\t"
  6334. "ldr r3, [%[a], #72]\n\t"
  6335. "ldr r4, [%[a], #76]\n\t"
  6336. "ldr r5, [%[b], #72]\n\t"
  6337. "ldr r6, [%[b], #76]\n\t"
  6338. "sbc r3, r5\n\t"
  6339. "sbc r4, r6\n\t"
  6340. "str r3, [%[a], #72]\n\t"
  6341. "str r4, [%[a], #76]\n\t"
  6342. "ldr r3, [%[a], #80]\n\t"
  6343. "ldr r4, [%[a], #84]\n\t"
  6344. "ldr r5, [%[b], #80]\n\t"
  6345. "ldr r6, [%[b], #84]\n\t"
  6346. "sbc r3, r5\n\t"
  6347. "sbc r4, r6\n\t"
  6348. "str r3, [%[a], #80]\n\t"
  6349. "str r4, [%[a], #84]\n\t"
  6350. "ldr r3, [%[a], #88]\n\t"
  6351. "ldr r4, [%[a], #92]\n\t"
  6352. "ldr r5, [%[b], #88]\n\t"
  6353. "ldr r6, [%[b], #92]\n\t"
  6354. "sbc r3, r5\n\t"
  6355. "sbc r4, r6\n\t"
  6356. "str r3, [%[a], #88]\n\t"
  6357. "str r4, [%[a], #92]\n\t"
  6358. "ldr r3, [%[a], #96]\n\t"
  6359. "ldr r4, [%[a], #100]\n\t"
  6360. "ldr r5, [%[b], #96]\n\t"
  6361. "ldr r6, [%[b], #100]\n\t"
  6362. "sbc r3, r5\n\t"
  6363. "sbc r4, r6\n\t"
  6364. "str r3, [%[a], #96]\n\t"
  6365. "str r4, [%[a], #100]\n\t"
  6366. "ldr r3, [%[a], #104]\n\t"
  6367. "ldr r4, [%[a], #108]\n\t"
  6368. "ldr r5, [%[b], #104]\n\t"
  6369. "ldr r6, [%[b], #108]\n\t"
  6370. "sbc r3, r5\n\t"
  6371. "sbc r4, r6\n\t"
  6372. "str r3, [%[a], #104]\n\t"
  6373. "str r4, [%[a], #108]\n\t"
  6374. "ldr r3, [%[a], #112]\n\t"
  6375. "ldr r4, [%[a], #116]\n\t"
  6376. "ldr r5, [%[b], #112]\n\t"
  6377. "ldr r6, [%[b], #116]\n\t"
  6378. "sbc r3, r5\n\t"
  6379. "sbc r4, r6\n\t"
  6380. "str r3, [%[a], #112]\n\t"
  6381. "str r4, [%[a], #116]\n\t"
  6382. "ldr r3, [%[a], #120]\n\t"
  6383. "ldr r4, [%[a], #124]\n\t"
  6384. "ldr r5, [%[b], #120]\n\t"
  6385. "ldr r6, [%[b], #124]\n\t"
  6386. "sbc r3, r5\n\t"
  6387. "sbc r4, r6\n\t"
  6388. "str r3, [%[a], #120]\n\t"
  6389. "str r4, [%[a], #124]\n\t"
  6390. "sbc %[c], %[c]\n\t"
  6391. "add %[a], #0x80\n\t"
  6392. "add %[b], #0x80\n\t"
  6393. "mov r5, #0\n\t"
  6394. "sub r5, %[c]\n\t"
  6395. "ldr r3, [%[a], #0]\n\t"
  6396. "ldr r4, [%[a], #4]\n\t"
  6397. "ldr r5, [%[b], #0]\n\t"
  6398. "ldr r6, [%[b], #4]\n\t"
  6399. "sbc r3, r5\n\t"
  6400. "sbc r4, r6\n\t"
  6401. "str r3, [%[a], #0]\n\t"
  6402. "str r4, [%[a], #4]\n\t"
  6403. "ldr r3, [%[a], #8]\n\t"
  6404. "ldr r4, [%[a], #12]\n\t"
  6405. "ldr r5, [%[b], #8]\n\t"
  6406. "ldr r6, [%[b], #12]\n\t"
  6407. "sbc r3, r5\n\t"
  6408. "sbc r4, r6\n\t"
  6409. "str r3, [%[a], #8]\n\t"
  6410. "str r4, [%[a], #12]\n\t"
  6411. "ldr r3, [%[a], #16]\n\t"
  6412. "ldr r4, [%[a], #20]\n\t"
  6413. "ldr r5, [%[b], #16]\n\t"
  6414. "ldr r6, [%[b], #20]\n\t"
  6415. "sbc r3, r5\n\t"
  6416. "sbc r4, r6\n\t"
  6417. "str r3, [%[a], #16]\n\t"
  6418. "str r4, [%[a], #20]\n\t"
  6419. "ldr r3, [%[a], #24]\n\t"
  6420. "ldr r4, [%[a], #28]\n\t"
  6421. "ldr r5, [%[b], #24]\n\t"
  6422. "ldr r6, [%[b], #28]\n\t"
  6423. "sbc r3, r5\n\t"
  6424. "sbc r4, r6\n\t"
  6425. "str r3, [%[a], #24]\n\t"
  6426. "str r4, [%[a], #28]\n\t"
  6427. "ldr r3, [%[a], #32]\n\t"
  6428. "ldr r4, [%[a], #36]\n\t"
  6429. "ldr r5, [%[b], #32]\n\t"
  6430. "ldr r6, [%[b], #36]\n\t"
  6431. "sbc r3, r5\n\t"
  6432. "sbc r4, r6\n\t"
  6433. "str r3, [%[a], #32]\n\t"
  6434. "str r4, [%[a], #36]\n\t"
  6435. "ldr r3, [%[a], #40]\n\t"
  6436. "ldr r4, [%[a], #44]\n\t"
  6437. "ldr r5, [%[b], #40]\n\t"
  6438. "ldr r6, [%[b], #44]\n\t"
  6439. "sbc r3, r5\n\t"
  6440. "sbc r4, r6\n\t"
  6441. "str r3, [%[a], #40]\n\t"
  6442. "str r4, [%[a], #44]\n\t"
  6443. "ldr r3, [%[a], #48]\n\t"
  6444. "ldr r4, [%[a], #52]\n\t"
  6445. "ldr r5, [%[b], #48]\n\t"
  6446. "ldr r6, [%[b], #52]\n\t"
  6447. "sbc r3, r5\n\t"
  6448. "sbc r4, r6\n\t"
  6449. "str r3, [%[a], #48]\n\t"
  6450. "str r4, [%[a], #52]\n\t"
  6451. "ldr r3, [%[a], #56]\n\t"
  6452. "ldr r4, [%[a], #60]\n\t"
  6453. "ldr r5, [%[b], #56]\n\t"
  6454. "ldr r6, [%[b], #60]\n\t"
  6455. "sbc r3, r5\n\t"
  6456. "sbc r4, r6\n\t"
  6457. "str r3, [%[a], #56]\n\t"
  6458. "str r4, [%[a], #60]\n\t"
  6459. "ldr r3, [%[a], #64]\n\t"
  6460. "ldr r4, [%[a], #68]\n\t"
  6461. "ldr r5, [%[b], #64]\n\t"
  6462. "ldr r6, [%[b], #68]\n\t"
  6463. "sbc r3, r5\n\t"
  6464. "sbc r4, r6\n\t"
  6465. "str r3, [%[a], #64]\n\t"
  6466. "str r4, [%[a], #68]\n\t"
  6467. "ldr r3, [%[a], #72]\n\t"
  6468. "ldr r4, [%[a], #76]\n\t"
  6469. "ldr r5, [%[b], #72]\n\t"
  6470. "ldr r6, [%[b], #76]\n\t"
  6471. "sbc r3, r5\n\t"
  6472. "sbc r4, r6\n\t"
  6473. "str r3, [%[a], #72]\n\t"
  6474. "str r4, [%[a], #76]\n\t"
  6475. "ldr r3, [%[a], #80]\n\t"
  6476. "ldr r4, [%[a], #84]\n\t"
  6477. "ldr r5, [%[b], #80]\n\t"
  6478. "ldr r6, [%[b], #84]\n\t"
  6479. "sbc r3, r5\n\t"
  6480. "sbc r4, r6\n\t"
  6481. "str r3, [%[a], #80]\n\t"
  6482. "str r4, [%[a], #84]\n\t"
  6483. "ldr r3, [%[a], #88]\n\t"
  6484. "ldr r4, [%[a], #92]\n\t"
  6485. "ldr r5, [%[b], #88]\n\t"
  6486. "ldr r6, [%[b], #92]\n\t"
  6487. "sbc r3, r5\n\t"
  6488. "sbc r4, r6\n\t"
  6489. "str r3, [%[a], #88]\n\t"
  6490. "str r4, [%[a], #92]\n\t"
  6491. "ldr r3, [%[a], #96]\n\t"
  6492. "ldr r4, [%[a], #100]\n\t"
  6493. "ldr r5, [%[b], #96]\n\t"
  6494. "ldr r6, [%[b], #100]\n\t"
  6495. "sbc r3, r5\n\t"
  6496. "sbc r4, r6\n\t"
  6497. "str r3, [%[a], #96]\n\t"
  6498. "str r4, [%[a], #100]\n\t"
  6499. "ldr r3, [%[a], #104]\n\t"
  6500. "ldr r4, [%[a], #108]\n\t"
  6501. "ldr r5, [%[b], #104]\n\t"
  6502. "ldr r6, [%[b], #108]\n\t"
  6503. "sbc r3, r5\n\t"
  6504. "sbc r4, r6\n\t"
  6505. "str r3, [%[a], #104]\n\t"
  6506. "str r4, [%[a], #108]\n\t"
  6507. "ldr r3, [%[a], #112]\n\t"
  6508. "ldr r4, [%[a], #116]\n\t"
  6509. "ldr r5, [%[b], #112]\n\t"
  6510. "ldr r6, [%[b], #116]\n\t"
  6511. "sbc r3, r5\n\t"
  6512. "sbc r4, r6\n\t"
  6513. "str r3, [%[a], #112]\n\t"
  6514. "str r4, [%[a], #116]\n\t"
  6515. "ldr r3, [%[a], #120]\n\t"
  6516. "ldr r4, [%[a], #124]\n\t"
  6517. "ldr r5, [%[b], #120]\n\t"
  6518. "ldr r6, [%[b], #124]\n\t"
  6519. "sbc r3, r5\n\t"
  6520. "sbc r4, r6\n\t"
  6521. "str r3, [%[a], #120]\n\t"
  6522. "str r4, [%[a], #124]\n\t"
  6523. "sbc %[c], %[c]\n\t"
  6524. "add %[a], #0x80\n\t"
  6525. "add %[b], #0x80\n\t"
  6526. "mov r5, #0\n\t"
  6527. "sub r5, %[c]\n\t"
  6528. "ldr r3, [%[a], #0]\n\t"
  6529. "ldr r4, [%[a], #4]\n\t"
  6530. "ldr r5, [%[b], #0]\n\t"
  6531. "ldr r6, [%[b], #4]\n\t"
  6532. "sbc r3, r5\n\t"
  6533. "sbc r4, r6\n\t"
  6534. "str r3, [%[a], #0]\n\t"
  6535. "str r4, [%[a], #4]\n\t"
  6536. "ldr r3, [%[a], #8]\n\t"
  6537. "ldr r4, [%[a], #12]\n\t"
  6538. "ldr r5, [%[b], #8]\n\t"
  6539. "ldr r6, [%[b], #12]\n\t"
  6540. "sbc r3, r5\n\t"
  6541. "sbc r4, r6\n\t"
  6542. "str r3, [%[a], #8]\n\t"
  6543. "str r4, [%[a], #12]\n\t"
  6544. "ldr r3, [%[a], #16]\n\t"
  6545. "ldr r4, [%[a], #20]\n\t"
  6546. "ldr r5, [%[b], #16]\n\t"
  6547. "ldr r6, [%[b], #20]\n\t"
  6548. "sbc r3, r5\n\t"
  6549. "sbc r4, r6\n\t"
  6550. "str r3, [%[a], #16]\n\t"
  6551. "str r4, [%[a], #20]\n\t"
  6552. "ldr r3, [%[a], #24]\n\t"
  6553. "ldr r4, [%[a], #28]\n\t"
  6554. "ldr r5, [%[b], #24]\n\t"
  6555. "ldr r6, [%[b], #28]\n\t"
  6556. "sbc r3, r5\n\t"
  6557. "sbc r4, r6\n\t"
  6558. "str r3, [%[a], #24]\n\t"
  6559. "str r4, [%[a], #28]\n\t"
  6560. "ldr r3, [%[a], #32]\n\t"
  6561. "ldr r4, [%[a], #36]\n\t"
  6562. "ldr r5, [%[b], #32]\n\t"
  6563. "ldr r6, [%[b], #36]\n\t"
  6564. "sbc r3, r5\n\t"
  6565. "sbc r4, r6\n\t"
  6566. "str r3, [%[a], #32]\n\t"
  6567. "str r4, [%[a], #36]\n\t"
  6568. "ldr r3, [%[a], #40]\n\t"
  6569. "ldr r4, [%[a], #44]\n\t"
  6570. "ldr r5, [%[b], #40]\n\t"
  6571. "ldr r6, [%[b], #44]\n\t"
  6572. "sbc r3, r5\n\t"
  6573. "sbc r4, r6\n\t"
  6574. "str r3, [%[a], #40]\n\t"
  6575. "str r4, [%[a], #44]\n\t"
  6576. "ldr r3, [%[a], #48]\n\t"
  6577. "ldr r4, [%[a], #52]\n\t"
  6578. "ldr r5, [%[b], #48]\n\t"
  6579. "ldr r6, [%[b], #52]\n\t"
  6580. "sbc r3, r5\n\t"
  6581. "sbc r4, r6\n\t"
  6582. "str r3, [%[a], #48]\n\t"
  6583. "str r4, [%[a], #52]\n\t"
  6584. "ldr r3, [%[a], #56]\n\t"
  6585. "ldr r4, [%[a], #60]\n\t"
  6586. "ldr r5, [%[b], #56]\n\t"
  6587. "ldr r6, [%[b], #60]\n\t"
  6588. "sbc r3, r5\n\t"
  6589. "sbc r4, r6\n\t"
  6590. "str r3, [%[a], #56]\n\t"
  6591. "str r4, [%[a], #60]\n\t"
  6592. "ldr r3, [%[a], #64]\n\t"
  6593. "ldr r4, [%[a], #68]\n\t"
  6594. "ldr r5, [%[b], #64]\n\t"
  6595. "ldr r6, [%[b], #68]\n\t"
  6596. "sbc r3, r5\n\t"
  6597. "sbc r4, r6\n\t"
  6598. "str r3, [%[a], #64]\n\t"
  6599. "str r4, [%[a], #68]\n\t"
  6600. "ldr r3, [%[a], #72]\n\t"
  6601. "ldr r4, [%[a], #76]\n\t"
  6602. "ldr r5, [%[b], #72]\n\t"
  6603. "ldr r6, [%[b], #76]\n\t"
  6604. "sbc r3, r5\n\t"
  6605. "sbc r4, r6\n\t"
  6606. "str r3, [%[a], #72]\n\t"
  6607. "str r4, [%[a], #76]\n\t"
  6608. "ldr r3, [%[a], #80]\n\t"
  6609. "ldr r4, [%[a], #84]\n\t"
  6610. "ldr r5, [%[b], #80]\n\t"
  6611. "ldr r6, [%[b], #84]\n\t"
  6612. "sbc r3, r5\n\t"
  6613. "sbc r4, r6\n\t"
  6614. "str r3, [%[a], #80]\n\t"
  6615. "str r4, [%[a], #84]\n\t"
  6616. "ldr r3, [%[a], #88]\n\t"
  6617. "ldr r4, [%[a], #92]\n\t"
  6618. "ldr r5, [%[b], #88]\n\t"
  6619. "ldr r6, [%[b], #92]\n\t"
  6620. "sbc r3, r5\n\t"
  6621. "sbc r4, r6\n\t"
  6622. "str r3, [%[a], #88]\n\t"
  6623. "str r4, [%[a], #92]\n\t"
  6624. "ldr r3, [%[a], #96]\n\t"
  6625. "ldr r4, [%[a], #100]\n\t"
  6626. "ldr r5, [%[b], #96]\n\t"
  6627. "ldr r6, [%[b], #100]\n\t"
  6628. "sbc r3, r5\n\t"
  6629. "sbc r4, r6\n\t"
  6630. "str r3, [%[a], #96]\n\t"
  6631. "str r4, [%[a], #100]\n\t"
  6632. "ldr r3, [%[a], #104]\n\t"
  6633. "ldr r4, [%[a], #108]\n\t"
  6634. "ldr r5, [%[b], #104]\n\t"
  6635. "ldr r6, [%[b], #108]\n\t"
  6636. "sbc r3, r5\n\t"
  6637. "sbc r4, r6\n\t"
  6638. "str r3, [%[a], #104]\n\t"
  6639. "str r4, [%[a], #108]\n\t"
  6640. "ldr r3, [%[a], #112]\n\t"
  6641. "ldr r4, [%[a], #116]\n\t"
  6642. "ldr r5, [%[b], #112]\n\t"
  6643. "ldr r6, [%[b], #116]\n\t"
  6644. "sbc r3, r5\n\t"
  6645. "sbc r4, r6\n\t"
  6646. "str r3, [%[a], #112]\n\t"
  6647. "str r4, [%[a], #116]\n\t"
  6648. "ldr r3, [%[a], #120]\n\t"
  6649. "ldr r4, [%[a], #124]\n\t"
  6650. "ldr r5, [%[b], #120]\n\t"
  6651. "ldr r6, [%[b], #124]\n\t"
  6652. "sbc r3, r5\n\t"
  6653. "sbc r4, r6\n\t"
  6654. "str r3, [%[a], #120]\n\t"
  6655. "str r4, [%[a], #124]\n\t"
  6656. "sbc %[c], %[c]\n\t"
  6657. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  6658. :
  6659. : "memory", "r3", "r4", "r5", "r6"
  6660. );
  6661. return c;
  6662. }
  6663. /* Add b to a into r. (r = a + b)
  6664. *
  6665. * r A single precision integer.
  6666. * a A single precision integer.
  6667. * b A single precision integer.
  6668. */
  6669. SP_NOINLINE static sp_digit sp_3072_add_96(sp_digit* r, const sp_digit* a,
  6670. const sp_digit* b)
  6671. {
  6672. sp_digit c = 0;
  6673. __asm__ __volatile__ (
  6674. "mov r7, #0\n\t"
  6675. "mvn r7, r7\n\t"
  6676. "ldr r4, [%[a], #0]\n\t"
  6677. "ldr r5, [%[b], #0]\n\t"
  6678. "add r4, r5\n\t"
  6679. "str r4, [%[r], #0]\n\t"
  6680. "ldr r4, [%[a], #4]\n\t"
  6681. "ldr r5, [%[b], #4]\n\t"
  6682. "adc r4, r5\n\t"
  6683. "str r4, [%[r], #4]\n\t"
  6684. "ldr r4, [%[a], #8]\n\t"
  6685. "ldr r5, [%[b], #8]\n\t"
  6686. "adc r4, r5\n\t"
  6687. "str r4, [%[r], #8]\n\t"
  6688. "ldr r4, [%[a], #12]\n\t"
  6689. "ldr r5, [%[b], #12]\n\t"
  6690. "adc r4, r5\n\t"
  6691. "str r4, [%[r], #12]\n\t"
  6692. "ldr r4, [%[a], #16]\n\t"
  6693. "ldr r5, [%[b], #16]\n\t"
  6694. "adc r4, r5\n\t"
  6695. "str r4, [%[r], #16]\n\t"
  6696. "ldr r4, [%[a], #20]\n\t"
  6697. "ldr r5, [%[b], #20]\n\t"
  6698. "adc r4, r5\n\t"
  6699. "str r4, [%[r], #20]\n\t"
  6700. "ldr r4, [%[a], #24]\n\t"
  6701. "ldr r5, [%[b], #24]\n\t"
  6702. "adc r4, r5\n\t"
  6703. "str r4, [%[r], #24]\n\t"
  6704. "ldr r4, [%[a], #28]\n\t"
  6705. "ldr r5, [%[b], #28]\n\t"
  6706. "adc r4, r5\n\t"
  6707. "str r4, [%[r], #28]\n\t"
  6708. "ldr r4, [%[a], #32]\n\t"
  6709. "ldr r5, [%[b], #32]\n\t"
  6710. "adc r4, r5\n\t"
  6711. "str r4, [%[r], #32]\n\t"
  6712. "ldr r4, [%[a], #36]\n\t"
  6713. "ldr r5, [%[b], #36]\n\t"
  6714. "adc r4, r5\n\t"
  6715. "str r4, [%[r], #36]\n\t"
  6716. "ldr r4, [%[a], #40]\n\t"
  6717. "ldr r5, [%[b], #40]\n\t"
  6718. "adc r4, r5\n\t"
  6719. "str r4, [%[r], #40]\n\t"
  6720. "ldr r4, [%[a], #44]\n\t"
  6721. "ldr r5, [%[b], #44]\n\t"
  6722. "adc r4, r5\n\t"
  6723. "str r4, [%[r], #44]\n\t"
  6724. "ldr r4, [%[a], #48]\n\t"
  6725. "ldr r5, [%[b], #48]\n\t"
  6726. "adc r4, r5\n\t"
  6727. "str r4, [%[r], #48]\n\t"
  6728. "ldr r4, [%[a], #52]\n\t"
  6729. "ldr r5, [%[b], #52]\n\t"
  6730. "adc r4, r5\n\t"
  6731. "str r4, [%[r], #52]\n\t"
  6732. "ldr r4, [%[a], #56]\n\t"
  6733. "ldr r5, [%[b], #56]\n\t"
  6734. "adc r4, r5\n\t"
  6735. "str r4, [%[r], #56]\n\t"
  6736. "ldr r4, [%[a], #60]\n\t"
  6737. "ldr r5, [%[b], #60]\n\t"
  6738. "adc r4, r5\n\t"
  6739. "str r4, [%[r], #60]\n\t"
  6740. "ldr r4, [%[a], #64]\n\t"
  6741. "ldr r5, [%[b], #64]\n\t"
  6742. "adc r4, r5\n\t"
  6743. "str r4, [%[r], #64]\n\t"
  6744. "ldr r4, [%[a], #68]\n\t"
  6745. "ldr r5, [%[b], #68]\n\t"
  6746. "adc r4, r5\n\t"
  6747. "str r4, [%[r], #68]\n\t"
  6748. "ldr r4, [%[a], #72]\n\t"
  6749. "ldr r5, [%[b], #72]\n\t"
  6750. "adc r4, r5\n\t"
  6751. "str r4, [%[r], #72]\n\t"
  6752. "ldr r4, [%[a], #76]\n\t"
  6753. "ldr r5, [%[b], #76]\n\t"
  6754. "adc r4, r5\n\t"
  6755. "str r4, [%[r], #76]\n\t"
  6756. "ldr r4, [%[a], #80]\n\t"
  6757. "ldr r5, [%[b], #80]\n\t"
  6758. "adc r4, r5\n\t"
  6759. "str r4, [%[r], #80]\n\t"
  6760. "ldr r4, [%[a], #84]\n\t"
  6761. "ldr r5, [%[b], #84]\n\t"
  6762. "adc r4, r5\n\t"
  6763. "str r4, [%[r], #84]\n\t"
  6764. "ldr r4, [%[a], #88]\n\t"
  6765. "ldr r5, [%[b], #88]\n\t"
  6766. "adc r4, r5\n\t"
  6767. "str r4, [%[r], #88]\n\t"
  6768. "ldr r4, [%[a], #92]\n\t"
  6769. "ldr r5, [%[b], #92]\n\t"
  6770. "adc r4, r5\n\t"
  6771. "str r4, [%[r], #92]\n\t"
  6772. "ldr r4, [%[a], #96]\n\t"
  6773. "ldr r5, [%[b], #96]\n\t"
  6774. "adc r4, r5\n\t"
  6775. "str r4, [%[r], #96]\n\t"
  6776. "ldr r4, [%[a], #100]\n\t"
  6777. "ldr r5, [%[b], #100]\n\t"
  6778. "adc r4, r5\n\t"
  6779. "str r4, [%[r], #100]\n\t"
  6780. "ldr r4, [%[a], #104]\n\t"
  6781. "ldr r5, [%[b], #104]\n\t"
  6782. "adc r4, r5\n\t"
  6783. "str r4, [%[r], #104]\n\t"
  6784. "ldr r4, [%[a], #108]\n\t"
  6785. "ldr r5, [%[b], #108]\n\t"
  6786. "adc r4, r5\n\t"
  6787. "str r4, [%[r], #108]\n\t"
  6788. "ldr r4, [%[a], #112]\n\t"
  6789. "ldr r5, [%[b], #112]\n\t"
  6790. "adc r4, r5\n\t"
  6791. "str r4, [%[r], #112]\n\t"
  6792. "ldr r4, [%[a], #116]\n\t"
  6793. "ldr r5, [%[b], #116]\n\t"
  6794. "adc r4, r5\n\t"
  6795. "str r4, [%[r], #116]\n\t"
  6796. "ldr r4, [%[a], #120]\n\t"
  6797. "ldr r5, [%[b], #120]\n\t"
  6798. "adc r4, r5\n\t"
  6799. "str r4, [%[r], #120]\n\t"
  6800. "ldr r4, [%[a], #124]\n\t"
  6801. "ldr r5, [%[b], #124]\n\t"
  6802. "adc r4, r5\n\t"
  6803. "str r4, [%[r], #124]\n\t"
  6804. "mov %[c], #0\n\t"
  6805. "adc %[c], %[c]\n\t"
  6806. "add %[a], #0x80\n\t"
  6807. "add %[b], #0x80\n\t"
  6808. "add %[r], #0x80\n\t"
  6809. "add %[c], r7\n\t"
  6810. "ldr r4, [%[a], #0]\n\t"
  6811. "ldr r5, [%[b], #0]\n\t"
  6812. "adc r4, r5\n\t"
  6813. "str r4, [%[r], #0]\n\t"
  6814. "ldr r4, [%[a], #4]\n\t"
  6815. "ldr r5, [%[b], #4]\n\t"
  6816. "adc r4, r5\n\t"
  6817. "str r4, [%[r], #4]\n\t"
  6818. "ldr r4, [%[a], #8]\n\t"
  6819. "ldr r5, [%[b], #8]\n\t"
  6820. "adc r4, r5\n\t"
  6821. "str r4, [%[r], #8]\n\t"
  6822. "ldr r4, [%[a], #12]\n\t"
  6823. "ldr r5, [%[b], #12]\n\t"
  6824. "adc r4, r5\n\t"
  6825. "str r4, [%[r], #12]\n\t"
  6826. "ldr r4, [%[a], #16]\n\t"
  6827. "ldr r5, [%[b], #16]\n\t"
  6828. "adc r4, r5\n\t"
  6829. "str r4, [%[r], #16]\n\t"
  6830. "ldr r4, [%[a], #20]\n\t"
  6831. "ldr r5, [%[b], #20]\n\t"
  6832. "adc r4, r5\n\t"
  6833. "str r4, [%[r], #20]\n\t"
  6834. "ldr r4, [%[a], #24]\n\t"
  6835. "ldr r5, [%[b], #24]\n\t"
  6836. "adc r4, r5\n\t"
  6837. "str r4, [%[r], #24]\n\t"
  6838. "ldr r4, [%[a], #28]\n\t"
  6839. "ldr r5, [%[b], #28]\n\t"
  6840. "adc r4, r5\n\t"
  6841. "str r4, [%[r], #28]\n\t"
  6842. "ldr r4, [%[a], #32]\n\t"
  6843. "ldr r5, [%[b], #32]\n\t"
  6844. "adc r4, r5\n\t"
  6845. "str r4, [%[r], #32]\n\t"
  6846. "ldr r4, [%[a], #36]\n\t"
  6847. "ldr r5, [%[b], #36]\n\t"
  6848. "adc r4, r5\n\t"
  6849. "str r4, [%[r], #36]\n\t"
  6850. "ldr r4, [%[a], #40]\n\t"
  6851. "ldr r5, [%[b], #40]\n\t"
  6852. "adc r4, r5\n\t"
  6853. "str r4, [%[r], #40]\n\t"
  6854. "ldr r4, [%[a], #44]\n\t"
  6855. "ldr r5, [%[b], #44]\n\t"
  6856. "adc r4, r5\n\t"
  6857. "str r4, [%[r], #44]\n\t"
  6858. "ldr r4, [%[a], #48]\n\t"
  6859. "ldr r5, [%[b], #48]\n\t"
  6860. "adc r4, r5\n\t"
  6861. "str r4, [%[r], #48]\n\t"
  6862. "ldr r4, [%[a], #52]\n\t"
  6863. "ldr r5, [%[b], #52]\n\t"
  6864. "adc r4, r5\n\t"
  6865. "str r4, [%[r], #52]\n\t"
  6866. "ldr r4, [%[a], #56]\n\t"
  6867. "ldr r5, [%[b], #56]\n\t"
  6868. "adc r4, r5\n\t"
  6869. "str r4, [%[r], #56]\n\t"
  6870. "ldr r4, [%[a], #60]\n\t"
  6871. "ldr r5, [%[b], #60]\n\t"
  6872. "adc r4, r5\n\t"
  6873. "str r4, [%[r], #60]\n\t"
  6874. "ldr r4, [%[a], #64]\n\t"
  6875. "ldr r5, [%[b], #64]\n\t"
  6876. "adc r4, r5\n\t"
  6877. "str r4, [%[r], #64]\n\t"
  6878. "ldr r4, [%[a], #68]\n\t"
  6879. "ldr r5, [%[b], #68]\n\t"
  6880. "adc r4, r5\n\t"
  6881. "str r4, [%[r], #68]\n\t"
  6882. "ldr r4, [%[a], #72]\n\t"
  6883. "ldr r5, [%[b], #72]\n\t"
  6884. "adc r4, r5\n\t"
  6885. "str r4, [%[r], #72]\n\t"
  6886. "ldr r4, [%[a], #76]\n\t"
  6887. "ldr r5, [%[b], #76]\n\t"
  6888. "adc r4, r5\n\t"
  6889. "str r4, [%[r], #76]\n\t"
  6890. "ldr r4, [%[a], #80]\n\t"
  6891. "ldr r5, [%[b], #80]\n\t"
  6892. "adc r4, r5\n\t"
  6893. "str r4, [%[r], #80]\n\t"
  6894. "ldr r4, [%[a], #84]\n\t"
  6895. "ldr r5, [%[b], #84]\n\t"
  6896. "adc r4, r5\n\t"
  6897. "str r4, [%[r], #84]\n\t"
  6898. "ldr r4, [%[a], #88]\n\t"
  6899. "ldr r5, [%[b], #88]\n\t"
  6900. "adc r4, r5\n\t"
  6901. "str r4, [%[r], #88]\n\t"
  6902. "ldr r4, [%[a], #92]\n\t"
  6903. "ldr r5, [%[b], #92]\n\t"
  6904. "adc r4, r5\n\t"
  6905. "str r4, [%[r], #92]\n\t"
  6906. "ldr r4, [%[a], #96]\n\t"
  6907. "ldr r5, [%[b], #96]\n\t"
  6908. "adc r4, r5\n\t"
  6909. "str r4, [%[r], #96]\n\t"
  6910. "ldr r4, [%[a], #100]\n\t"
  6911. "ldr r5, [%[b], #100]\n\t"
  6912. "adc r4, r5\n\t"
  6913. "str r4, [%[r], #100]\n\t"
  6914. "ldr r4, [%[a], #104]\n\t"
  6915. "ldr r5, [%[b], #104]\n\t"
  6916. "adc r4, r5\n\t"
  6917. "str r4, [%[r], #104]\n\t"
  6918. "ldr r4, [%[a], #108]\n\t"
  6919. "ldr r5, [%[b], #108]\n\t"
  6920. "adc r4, r5\n\t"
  6921. "str r4, [%[r], #108]\n\t"
  6922. "ldr r4, [%[a], #112]\n\t"
  6923. "ldr r5, [%[b], #112]\n\t"
  6924. "adc r4, r5\n\t"
  6925. "str r4, [%[r], #112]\n\t"
  6926. "ldr r4, [%[a], #116]\n\t"
  6927. "ldr r5, [%[b], #116]\n\t"
  6928. "adc r4, r5\n\t"
  6929. "str r4, [%[r], #116]\n\t"
  6930. "ldr r4, [%[a], #120]\n\t"
  6931. "ldr r5, [%[b], #120]\n\t"
  6932. "adc r4, r5\n\t"
  6933. "str r4, [%[r], #120]\n\t"
  6934. "ldr r4, [%[a], #124]\n\t"
  6935. "ldr r5, [%[b], #124]\n\t"
  6936. "adc r4, r5\n\t"
  6937. "str r4, [%[r], #124]\n\t"
  6938. "mov %[c], #0\n\t"
  6939. "adc %[c], %[c]\n\t"
  6940. "add %[a], #0x80\n\t"
  6941. "add %[b], #0x80\n\t"
  6942. "add %[r], #0x80\n\t"
  6943. "add %[c], r7\n\t"
  6944. "ldr r4, [%[a], #0]\n\t"
  6945. "ldr r5, [%[b], #0]\n\t"
  6946. "adc r4, r5\n\t"
  6947. "str r4, [%[r], #0]\n\t"
  6948. "ldr r4, [%[a], #4]\n\t"
  6949. "ldr r5, [%[b], #4]\n\t"
  6950. "adc r4, r5\n\t"
  6951. "str r4, [%[r], #4]\n\t"
  6952. "ldr r4, [%[a], #8]\n\t"
  6953. "ldr r5, [%[b], #8]\n\t"
  6954. "adc r4, r5\n\t"
  6955. "str r4, [%[r], #8]\n\t"
  6956. "ldr r4, [%[a], #12]\n\t"
  6957. "ldr r5, [%[b], #12]\n\t"
  6958. "adc r4, r5\n\t"
  6959. "str r4, [%[r], #12]\n\t"
  6960. "ldr r4, [%[a], #16]\n\t"
  6961. "ldr r5, [%[b], #16]\n\t"
  6962. "adc r4, r5\n\t"
  6963. "str r4, [%[r], #16]\n\t"
  6964. "ldr r4, [%[a], #20]\n\t"
  6965. "ldr r5, [%[b], #20]\n\t"
  6966. "adc r4, r5\n\t"
  6967. "str r4, [%[r], #20]\n\t"
  6968. "ldr r4, [%[a], #24]\n\t"
  6969. "ldr r5, [%[b], #24]\n\t"
  6970. "adc r4, r5\n\t"
  6971. "str r4, [%[r], #24]\n\t"
  6972. "ldr r4, [%[a], #28]\n\t"
  6973. "ldr r5, [%[b], #28]\n\t"
  6974. "adc r4, r5\n\t"
  6975. "str r4, [%[r], #28]\n\t"
  6976. "ldr r4, [%[a], #32]\n\t"
  6977. "ldr r5, [%[b], #32]\n\t"
  6978. "adc r4, r5\n\t"
  6979. "str r4, [%[r], #32]\n\t"
  6980. "ldr r4, [%[a], #36]\n\t"
  6981. "ldr r5, [%[b], #36]\n\t"
  6982. "adc r4, r5\n\t"
  6983. "str r4, [%[r], #36]\n\t"
  6984. "ldr r4, [%[a], #40]\n\t"
  6985. "ldr r5, [%[b], #40]\n\t"
  6986. "adc r4, r5\n\t"
  6987. "str r4, [%[r], #40]\n\t"
  6988. "ldr r4, [%[a], #44]\n\t"
  6989. "ldr r5, [%[b], #44]\n\t"
  6990. "adc r4, r5\n\t"
  6991. "str r4, [%[r], #44]\n\t"
  6992. "ldr r4, [%[a], #48]\n\t"
  6993. "ldr r5, [%[b], #48]\n\t"
  6994. "adc r4, r5\n\t"
  6995. "str r4, [%[r], #48]\n\t"
  6996. "ldr r4, [%[a], #52]\n\t"
  6997. "ldr r5, [%[b], #52]\n\t"
  6998. "adc r4, r5\n\t"
  6999. "str r4, [%[r], #52]\n\t"
  7000. "ldr r4, [%[a], #56]\n\t"
  7001. "ldr r5, [%[b], #56]\n\t"
  7002. "adc r4, r5\n\t"
  7003. "str r4, [%[r], #56]\n\t"
  7004. "ldr r4, [%[a], #60]\n\t"
  7005. "ldr r5, [%[b], #60]\n\t"
  7006. "adc r4, r5\n\t"
  7007. "str r4, [%[r], #60]\n\t"
  7008. "ldr r4, [%[a], #64]\n\t"
  7009. "ldr r5, [%[b], #64]\n\t"
  7010. "adc r4, r5\n\t"
  7011. "str r4, [%[r], #64]\n\t"
  7012. "ldr r4, [%[a], #68]\n\t"
  7013. "ldr r5, [%[b], #68]\n\t"
  7014. "adc r4, r5\n\t"
  7015. "str r4, [%[r], #68]\n\t"
  7016. "ldr r4, [%[a], #72]\n\t"
  7017. "ldr r5, [%[b], #72]\n\t"
  7018. "adc r4, r5\n\t"
  7019. "str r4, [%[r], #72]\n\t"
  7020. "ldr r4, [%[a], #76]\n\t"
  7021. "ldr r5, [%[b], #76]\n\t"
  7022. "adc r4, r5\n\t"
  7023. "str r4, [%[r], #76]\n\t"
  7024. "ldr r4, [%[a], #80]\n\t"
  7025. "ldr r5, [%[b], #80]\n\t"
  7026. "adc r4, r5\n\t"
  7027. "str r4, [%[r], #80]\n\t"
  7028. "ldr r4, [%[a], #84]\n\t"
  7029. "ldr r5, [%[b], #84]\n\t"
  7030. "adc r4, r5\n\t"
  7031. "str r4, [%[r], #84]\n\t"
  7032. "ldr r4, [%[a], #88]\n\t"
  7033. "ldr r5, [%[b], #88]\n\t"
  7034. "adc r4, r5\n\t"
  7035. "str r4, [%[r], #88]\n\t"
  7036. "ldr r4, [%[a], #92]\n\t"
  7037. "ldr r5, [%[b], #92]\n\t"
  7038. "adc r4, r5\n\t"
  7039. "str r4, [%[r], #92]\n\t"
  7040. "ldr r4, [%[a], #96]\n\t"
  7041. "ldr r5, [%[b], #96]\n\t"
  7042. "adc r4, r5\n\t"
  7043. "str r4, [%[r], #96]\n\t"
  7044. "ldr r4, [%[a], #100]\n\t"
  7045. "ldr r5, [%[b], #100]\n\t"
  7046. "adc r4, r5\n\t"
  7047. "str r4, [%[r], #100]\n\t"
  7048. "ldr r4, [%[a], #104]\n\t"
  7049. "ldr r5, [%[b], #104]\n\t"
  7050. "adc r4, r5\n\t"
  7051. "str r4, [%[r], #104]\n\t"
  7052. "ldr r4, [%[a], #108]\n\t"
  7053. "ldr r5, [%[b], #108]\n\t"
  7054. "adc r4, r5\n\t"
  7055. "str r4, [%[r], #108]\n\t"
  7056. "ldr r4, [%[a], #112]\n\t"
  7057. "ldr r5, [%[b], #112]\n\t"
  7058. "adc r4, r5\n\t"
  7059. "str r4, [%[r], #112]\n\t"
  7060. "ldr r4, [%[a], #116]\n\t"
  7061. "ldr r5, [%[b], #116]\n\t"
  7062. "adc r4, r5\n\t"
  7063. "str r4, [%[r], #116]\n\t"
  7064. "ldr r4, [%[a], #120]\n\t"
  7065. "ldr r5, [%[b], #120]\n\t"
  7066. "adc r4, r5\n\t"
  7067. "str r4, [%[r], #120]\n\t"
  7068. "ldr r4, [%[a], #124]\n\t"
  7069. "ldr r5, [%[b], #124]\n\t"
  7070. "adc r4, r5\n\t"
  7071. "str r4, [%[r], #124]\n\t"
  7072. "mov %[c], #0\n\t"
  7073. "adc %[c], %[c]\n\t"
  7074. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  7075. :
  7076. : "memory", "r4", "r5", "r7"
  7077. );
  7078. return c;
  7079. }
  7080. /* AND m into each word of a and store in r.
  7081. *
  7082. * r A single precision integer.
  7083. * a A single precision integer.
  7084. * m Mask to AND against each digit.
  7085. */
  7086. static void sp_3072_mask_48(sp_digit* r, const sp_digit* a, sp_digit m)
  7087. {
  7088. #ifdef WOLFSSL_SP_SMALL
  7089. int i;
  7090. for (i=0; i<48; i++) {
  7091. r[i] = a[i] & m;
  7092. }
  7093. #else
  7094. int i;
  7095. for (i = 0; i < 48; i += 8) {
  7096. r[i+0] = a[i+0] & m;
  7097. r[i+1] = a[i+1] & m;
  7098. r[i+2] = a[i+2] & m;
  7099. r[i+3] = a[i+3] & m;
  7100. r[i+4] = a[i+4] & m;
  7101. r[i+5] = a[i+5] & m;
  7102. r[i+6] = a[i+6] & m;
  7103. r[i+7] = a[i+7] & m;
  7104. }
  7105. #endif
  7106. }
  7107. /* Multiply a and b into r. (r = a * b)
  7108. *
  7109. * r A single precision integer.
  7110. * a A single precision integer.
  7111. * b A single precision integer.
  7112. */
  7113. SP_NOINLINE static void sp_3072_mul_96(sp_digit* r, const sp_digit* a,
  7114. const sp_digit* b)
  7115. {
  7116. sp_digit* z0 = r;
  7117. sp_digit z1[96];
  7118. sp_digit a1[48];
  7119. sp_digit b1[48];
  7120. sp_digit z2[96];
  7121. sp_digit u, ca, cb;
  7122. ca = sp_3072_add_48(a1, a, &a[48]);
  7123. cb = sp_3072_add_48(b1, b, &b[48]);
  7124. u = ca & cb;
  7125. sp_3072_mul_48(z1, a1, b1);
  7126. sp_3072_mul_48(z2, &a[48], &b[48]);
  7127. sp_3072_mul_48(z0, a, b);
  7128. sp_3072_mask_48(r + 96, a1, 0 - cb);
  7129. sp_3072_mask_48(b1, b1, 0 - ca);
  7130. u += sp_3072_add_48(r + 96, r + 96, b1);
  7131. u += sp_3072_sub_in_place_96(z1, z2);
  7132. u += sp_3072_sub_in_place_96(z1, z0);
  7133. u += sp_3072_add_96(r + 48, r + 48, z1);
  7134. r[144] = u;
  7135. XMEMSET(r + 144 + 1, 0, sizeof(sp_digit) * (48 - 1));
  7136. (void)sp_3072_add_96(r + 96, r + 96, z2);
  7137. }
  7138. /* Square a and put result in r. (r = a * a)
  7139. *
  7140. * r A single precision integer.
  7141. * a A single precision integer.
  7142. */
  7143. SP_NOINLINE static void sp_3072_sqr_96(sp_digit* r, const sp_digit* a)
  7144. {
  7145. sp_digit* z0 = r;
  7146. sp_digit z2[96];
  7147. sp_digit z1[96];
  7148. sp_digit a1[48];
  7149. sp_digit u;
  7150. u = sp_3072_add_48(a1, a, &a[48]);
  7151. sp_3072_sqr_48(z1, a1);
  7152. sp_3072_sqr_48(z2, &a[48]);
  7153. sp_3072_sqr_48(z0, a);
  7154. sp_3072_mask_48(r + 96, a1, 0 - u);
  7155. u += sp_3072_add_48(r + 96, r + 96, r + 96);
  7156. u += sp_3072_sub_in_place_96(z1, z2);
  7157. u += sp_3072_sub_in_place_96(z1, z0);
  7158. u += sp_3072_add_96(r + 48, r + 48, z1);
  7159. r[144] = u;
  7160. XMEMSET(r + 144 + 1, 0, sizeof(sp_digit) * (48 - 1));
  7161. (void)sp_3072_add_96(r + 96, r + 96, z2);
  7162. }
  7163. #endif /* !WOLFSSL_SP_SMALL */
  7164. #ifdef WOLFSSL_SP_SMALL
  7165. /* Add b to a into r. (r = a + b)
  7166. *
  7167. * r A single precision integer.
  7168. * a A single precision integer.
  7169. * b A single precision integer.
  7170. */
  7171. SP_NOINLINE static sp_digit sp_3072_add_96(sp_digit* r, const sp_digit* a,
  7172. const sp_digit* b)
  7173. {
  7174. sp_digit c = 0;
  7175. __asm__ __volatile__ (
  7176. "mov r6, %[a]\n\t"
  7177. "mov r7, #0\n\t"
  7178. "mov r4, #1\n\t"
  7179. "lsl r4, #8\n\t"
  7180. "add r4, #128\n\t"
  7181. "sub r7, #1\n\t"
  7182. "add r6, r4\n\t"
  7183. "\n1:\n\t"
  7184. "add %[c], r7\n\t"
  7185. "ldr r4, [%[a]]\n\t"
  7186. "ldr r5, [%[b]]\n\t"
  7187. "adc r4, r5\n\t"
  7188. "str r4, [%[r]]\n\t"
  7189. "mov %[c], #0\n\t"
  7190. "adc %[c], %[c]\n\t"
  7191. "add %[a], #4\n\t"
  7192. "add %[b], #4\n\t"
  7193. "add %[r], #4\n\t"
  7194. "cmp %[a], r6\n\t"
  7195. "bne 1b\n\t"
  7196. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  7197. :
  7198. : "memory", "r4", "r5", "r6", "r7"
  7199. );
  7200. return c;
  7201. }
  7202. #endif /* WOLFSSL_SP_SMALL */
  7203. #ifdef WOLFSSL_SP_SMALL
  7204. /* Sub b from a into a. (a -= b)
  7205. *
  7206. * a A single precision integer.
  7207. * b A single precision integer.
  7208. */
  7209. SP_NOINLINE static sp_digit sp_3072_sub_in_place_96(sp_digit* a,
  7210. const sp_digit* b)
  7211. {
  7212. sp_digit c = 0;
  7213. __asm__ __volatile__ (
  7214. "mov r7, %[a]\n\t"
  7215. "mov r5, #1\n\t"
  7216. "lsl r5, #8\n\t"
  7217. "add r5, #128\n\t"
  7218. "add r7, r5\n\t"
  7219. "\n1:\n\t"
  7220. "mov r5, #0\n\t"
  7221. "sub r5, %[c]\n\t"
  7222. "ldr r3, [%[a]]\n\t"
  7223. "ldr r4, [%[a], #4]\n\t"
  7224. "ldr r5, [%[b]]\n\t"
  7225. "ldr r6, [%[b], #4]\n\t"
  7226. "sbc r3, r5\n\t"
  7227. "sbc r4, r6\n\t"
  7228. "str r3, [%[a]]\n\t"
  7229. "str r4, [%[a], #4]\n\t"
  7230. "sbc %[c], %[c]\n\t"
  7231. "add %[a], #8\n\t"
  7232. "add %[b], #8\n\t"
  7233. "cmp %[a], r7\n\t"
  7234. "bne 1b\n\t"
  7235. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  7236. :
  7237. : "memory", "r3", "r4", "r5", "r6", "r7"
  7238. );
  7239. return c;
  7240. }
  7241. #endif /* WOLFSSL_SP_SMALL */
  7242. #ifdef WOLFSSL_SP_SMALL
  7243. /* Multiply a and b into r. (r = a * b)
  7244. *
  7245. * r A single precision integer.
  7246. * a A single precision integer.
  7247. * b A single precision integer.
  7248. */
  7249. SP_NOINLINE static void sp_3072_mul_96(sp_digit* r, const sp_digit* a,
  7250. const sp_digit* b)
  7251. {
  7252. sp_digit tmp[96 * 2];
  7253. __asm__ __volatile__ (
  7254. "mov r3, #0\n\t"
  7255. "mov r4, #0\n\t"
  7256. "mov r8, r3\n\t"
  7257. "mov r11, %[r]\n\t"
  7258. "mov r9, %[a]\n\t"
  7259. "mov r10, %[b]\n\t"
  7260. "mov r6, #1\n\t"
  7261. "lsl r6, r6, #8\n\t"
  7262. "add r6, #128\n\t"
  7263. "add r6, r9\n\t"
  7264. "mov r12, r6\n\t"
  7265. "\n1:\n\t"
  7266. "mov %[r], #0\n\t"
  7267. "mov r5, #0\n\t"
  7268. "mov r6, #1\n\t"
  7269. "lsl r6, r6, #8\n\t"
  7270. "add r6, #124\n\t"
  7271. "mov %[a], r8\n\t"
  7272. "sub %[a], r6\n\t"
  7273. "sbc r6, r6\n\t"
  7274. "mvn r6, r6\n\t"
  7275. "and %[a], r6\n\t"
  7276. "mov %[b], r8\n\t"
  7277. "sub %[b], %[a]\n\t"
  7278. "add %[a], r9\n\t"
  7279. "add %[b], r10\n\t"
  7280. "\n2:\n\t"
  7281. "# Multiply Start\n\t"
  7282. "ldr r6, [%[a]]\n\t"
  7283. "ldr r7, [%[b]]\n\t"
  7284. "lsl r6, r6, #16\n\t"
  7285. "lsl r7, r7, #16\n\t"
  7286. "lsr r6, r6, #16\n\t"
  7287. "lsr r7, r7, #16\n\t"
  7288. "mul r7, r6\n\t"
  7289. "add r3, r7\n\t"
  7290. "adc r4, %[r]\n\t"
  7291. "adc r5, %[r]\n\t"
  7292. "ldr r7, [%[b]]\n\t"
  7293. "lsr r7, r7, #16\n\t"
  7294. "mul r6, r7\n\t"
  7295. "lsr r7, r6, #16\n\t"
  7296. "lsl r6, r6, #16\n\t"
  7297. "add r3, r6\n\t"
  7298. "adc r4, r7\n\t"
  7299. "adc r5, %[r]\n\t"
  7300. "ldr r6, [%[a]]\n\t"
  7301. "ldr r7, [%[b]]\n\t"
  7302. "lsr r6, r6, #16\n\t"
  7303. "lsr r7, r7, #16\n\t"
  7304. "mul r7, r6\n\t"
  7305. "add r4, r7\n\t"
  7306. "adc r5, %[r]\n\t"
  7307. "ldr r7, [%[b]]\n\t"
  7308. "lsl r7, r7, #16\n\t"
  7309. "lsr r7, r7, #16\n\t"
  7310. "mul r6, r7\n\t"
  7311. "lsr r7, r6, #16\n\t"
  7312. "lsl r6, r6, #16\n\t"
  7313. "add r3, r6\n\t"
  7314. "adc r4, r7\n\t"
  7315. "adc r5, %[r]\n\t"
  7316. "# Multiply Done\n\t"
  7317. "add %[a], #4\n\t"
  7318. "sub %[b], #4\n\t"
  7319. "cmp %[a], r12\n\t"
  7320. "beq 3f\n\t"
  7321. "mov r6, r8\n\t"
  7322. "add r6, r9\n\t"
  7323. "cmp %[a], r6\n\t"
  7324. "ble 2b\n\t"
  7325. "\n3:\n\t"
  7326. "mov %[r], r11\n\t"
  7327. "mov r7, r8\n\t"
  7328. "str r3, [%[r], r7]\n\t"
  7329. "mov r3, r4\n\t"
  7330. "mov r4, r5\n\t"
  7331. "add r7, #4\n\t"
  7332. "mov r8, r7\n\t"
  7333. "mov r6, #2\n\t"
  7334. "lsl r6, r6, #8\n\t"
  7335. "add r6, #248\n\t"
  7336. "cmp r7, r6\n\t"
  7337. "ble 1b\n\t"
  7338. "str r3, [%[r], r7]\n\t"
  7339. "mov %[a], r9\n\t"
  7340. "mov %[b], r10\n\t"
  7341. :
  7342. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  7343. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  7344. );
  7345. XMEMCPY(r, tmp, sizeof(tmp));
  7346. }
  7347. /* Square a and put result in r. (r = a * a)
  7348. *
  7349. * r A single precision integer.
  7350. * a A single precision integer.
  7351. */
  7352. SP_NOINLINE static void sp_3072_sqr_96(sp_digit* r, const sp_digit* a)
  7353. {
  7354. __asm__ __volatile__ (
  7355. "mov r3, #0\n\t"
  7356. "mov r4, #0\n\t"
  7357. "mov r5, #0\n\t"
  7358. "mov r8, r3\n\t"
  7359. "mov r11, %[r]\n\t"
  7360. "mov r6, #3\n\t"
  7361. "lsl r6, r6, #8\n\t"
  7362. "neg r6, r6\n\t"
  7363. "add sp, r6\n\t"
  7364. "mov r10, sp\n\t"
  7365. "mov r9, %[a]\n\t"
  7366. "\n1:\n\t"
  7367. "mov %[r], #0\n\t"
  7368. "mov r6, #1\n\t"
  7369. "lsl r6, r6, #8\n\t"
  7370. "add r6, #124\n\t"
  7371. "mov %[a], r8\n\t"
  7372. "sub %[a], r6\n\t"
  7373. "sbc r6, r6\n\t"
  7374. "mvn r6, r6\n\t"
  7375. "and %[a], r6\n\t"
  7376. "mov r2, r8\n\t"
  7377. "sub r2, %[a]\n\t"
  7378. "add %[a], r9\n\t"
  7379. "add r2, r9\n\t"
  7380. "\n2:\n\t"
  7381. "cmp r2, %[a]\n\t"
  7382. "beq 4f\n\t"
  7383. "# Multiply * 2: Start\n\t"
  7384. "ldr r6, [%[a]]\n\t"
  7385. "ldr r7, [r2]\n\t"
  7386. "lsl r6, r6, #16\n\t"
  7387. "lsl r7, r7, #16\n\t"
  7388. "lsr r6, r6, #16\n\t"
  7389. "lsr r7, r7, #16\n\t"
  7390. "mul r7, r6\n\t"
  7391. "add r3, r7\n\t"
  7392. "adc r4, %[r]\n\t"
  7393. "adc r5, %[r]\n\t"
  7394. "add r3, r7\n\t"
  7395. "adc r4, %[r]\n\t"
  7396. "adc r5, %[r]\n\t"
  7397. "ldr r7, [r2]\n\t"
  7398. "lsr r7, r7, #16\n\t"
  7399. "mul r6, r7\n\t"
  7400. "lsr r7, r6, #16\n\t"
  7401. "lsl r6, r6, #16\n\t"
  7402. "add r3, r6\n\t"
  7403. "adc r4, r7\n\t"
  7404. "adc r5, %[r]\n\t"
  7405. "add r3, r6\n\t"
  7406. "adc r4, r7\n\t"
  7407. "adc r5, %[r]\n\t"
  7408. "ldr r6, [%[a]]\n\t"
  7409. "ldr r7, [r2]\n\t"
  7410. "lsr r6, r6, #16\n\t"
  7411. "lsr r7, r7, #16\n\t"
  7412. "mul r7, r6\n\t"
  7413. "add r4, r7\n\t"
  7414. "adc r5, %[r]\n\t"
  7415. "add r4, r7\n\t"
  7416. "adc r5, %[r]\n\t"
  7417. "ldr r7, [r2]\n\t"
  7418. "lsl r7, r7, #16\n\t"
  7419. "lsr r7, r7, #16\n\t"
  7420. "mul r6, r7\n\t"
  7421. "lsr r7, r6, #16\n\t"
  7422. "lsl r6, r6, #16\n\t"
  7423. "add r3, r6\n\t"
  7424. "adc r4, r7\n\t"
  7425. "adc r5, %[r]\n\t"
  7426. "add r3, r6\n\t"
  7427. "adc r4, r7\n\t"
  7428. "adc r5, %[r]\n\t"
  7429. "# Multiply * 2: Done\n\t"
  7430. "bal 5f\n\t"
  7431. "\n4:\n\t"
  7432. "# Square: Start\n\t"
  7433. "ldr r6, [%[a]]\n\t"
  7434. "lsr r7, r6, #16\n\t"
  7435. "lsl r6, r6, #16\n\t"
  7436. "lsr r6, r6, #16\n\t"
  7437. "mul r6, r6\n\t"
  7438. "add r3, r6\n\t"
  7439. "adc r4, %[r]\n\t"
  7440. "adc r5, %[r]\n\t"
  7441. "mul r7, r7\n\t"
  7442. "add r4, r7\n\t"
  7443. "adc r5, %[r]\n\t"
  7444. "ldr r6, [%[a]]\n\t"
  7445. "lsr r7, r6, #16\n\t"
  7446. "lsl r6, r6, #16\n\t"
  7447. "lsr r6, r6, #16\n\t"
  7448. "mul r6, r7\n\t"
  7449. "lsr r7, r6, #15\n\t"
  7450. "lsl r6, r6, #17\n\t"
  7451. "add r3, r6\n\t"
  7452. "adc r4, r7\n\t"
  7453. "adc r5, %[r]\n\t"
  7454. "# Square: Done\n\t"
  7455. "\n5:\n\t"
  7456. "add %[a], #4\n\t"
  7457. "sub r2, #4\n\t"
  7458. "mov r6, #1\n\t"
  7459. "lsl r6, r6, #8\n\t"
  7460. "add r6, #128\n\t"
  7461. "add r6, r9\n\t"
  7462. "cmp %[a], r6\n\t"
  7463. "beq 3f\n\t"
  7464. "cmp %[a], r2\n\t"
  7465. "bgt 3f\n\t"
  7466. "mov r7, r8\n\t"
  7467. "add r7, r9\n\t"
  7468. "cmp %[a], r7\n\t"
  7469. "ble 2b\n\t"
  7470. "\n3:\n\t"
  7471. "mov %[r], r10\n\t"
  7472. "mov r7, r8\n\t"
  7473. "str r3, [%[r], r7]\n\t"
  7474. "mov r3, r4\n\t"
  7475. "mov r4, r5\n\t"
  7476. "mov r5, #0\n\t"
  7477. "add r7, #4\n\t"
  7478. "mov r8, r7\n\t"
  7479. "mov r6, #2\n\t"
  7480. "lsl r6, r6, #8\n\t"
  7481. "add r6, #248\n\t"
  7482. "cmp r7, r6\n\t"
  7483. "ble 1b\n\t"
  7484. "mov %[a], r9\n\t"
  7485. "str r3, [%[r], r7]\n\t"
  7486. "mov %[r], r11\n\t"
  7487. "mov %[a], r10\n\t"
  7488. "mov r3, #2\n\t"
  7489. "lsl r3, r3, #8\n\t"
  7490. "add r3, #252\n\t"
  7491. "\n4:\n\t"
  7492. "ldr r6, [%[a], r3]\n\t"
  7493. "str r6, [%[r], r3]\n\t"
  7494. "sub r3, #4\n\t"
  7495. "bge 4b\n\t"
  7496. "mov r6, #3\n\t"
  7497. "lsl r6, r6, #8\n\t"
  7498. "add sp, r6\n\t"
  7499. :
  7500. : [r] "r" (r), [a] "r" (a)
  7501. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  7502. );
  7503. }
  7504. #endif /* WOLFSSL_SP_SMALL */
  7505. #if (defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)
  7506. #ifdef WOLFSSL_SP_SMALL
  7507. /* AND m into each word of a and store in r.
  7508. *
  7509. * r A single precision integer.
  7510. * a A single precision integer.
  7511. * m Mask to AND against each digit.
  7512. */
  7513. static void sp_3072_mask_48(sp_digit* r, const sp_digit* a, sp_digit m)
  7514. {
  7515. int i;
  7516. for (i=0; i<48; i++) {
  7517. r[i] = a[i] & m;
  7518. }
  7519. }
  7520. #endif /* WOLFSSL_SP_SMALL */
  7521. #ifdef WOLFSSL_SP_SMALL
  7522. /* Add b to a into r. (r = a + b)
  7523. *
  7524. * r A single precision integer.
  7525. * a A single precision integer.
  7526. * b A single precision integer.
  7527. */
  7528. SP_NOINLINE static sp_digit sp_3072_add_48(sp_digit* r, const sp_digit* a,
  7529. const sp_digit* b)
  7530. {
  7531. sp_digit c = 0;
  7532. __asm__ __volatile__ (
  7533. "mov r6, %[a]\n\t"
  7534. "mov r7, #0\n\t"
  7535. "add r6, #192\n\t"
  7536. "sub r7, #1\n\t"
  7537. "\n1:\n\t"
  7538. "add %[c], r7\n\t"
  7539. "ldr r4, [%[a]]\n\t"
  7540. "ldr r5, [%[b]]\n\t"
  7541. "adc r4, r5\n\t"
  7542. "str r4, [%[r]]\n\t"
  7543. "mov %[c], #0\n\t"
  7544. "adc %[c], %[c]\n\t"
  7545. "add %[a], #4\n\t"
  7546. "add %[b], #4\n\t"
  7547. "add %[r], #4\n\t"
  7548. "cmp %[a], r6\n\t"
  7549. "bne 1b\n\t"
  7550. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  7551. :
  7552. : "memory", "r4", "r5", "r6", "r7"
  7553. );
  7554. return c;
  7555. }
  7556. #endif /* WOLFSSL_SP_SMALL */
  7557. #ifdef WOLFSSL_SP_SMALL
  7558. /* Sub b from a into a. (a -= b)
  7559. *
  7560. * a A single precision integer.
  7561. * b A single precision integer.
  7562. */
  7563. SP_NOINLINE static sp_digit sp_3072_sub_in_place_48(sp_digit* a,
  7564. const sp_digit* b)
  7565. {
  7566. sp_digit c = 0;
  7567. __asm__ __volatile__ (
  7568. "mov r7, %[a]\n\t"
  7569. "add r7, #192\n\t"
  7570. "\n1:\n\t"
  7571. "mov r5, #0\n\t"
  7572. "sub r5, %[c]\n\t"
  7573. "ldr r3, [%[a]]\n\t"
  7574. "ldr r4, [%[a], #4]\n\t"
  7575. "ldr r5, [%[b]]\n\t"
  7576. "ldr r6, [%[b], #4]\n\t"
  7577. "sbc r3, r5\n\t"
  7578. "sbc r4, r6\n\t"
  7579. "str r3, [%[a]]\n\t"
  7580. "str r4, [%[a], #4]\n\t"
  7581. "sbc %[c], %[c]\n\t"
  7582. "add %[a], #8\n\t"
  7583. "add %[b], #8\n\t"
  7584. "cmp %[a], r7\n\t"
  7585. "bne 1b\n\t"
  7586. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  7587. :
  7588. : "memory", "r3", "r4", "r5", "r6", "r7"
  7589. );
  7590. return c;
  7591. }
  7592. #endif /* WOLFSSL_SP_SMALL */
  7593. #ifdef WOLFSSL_SP_SMALL
  7594. /* Multiply a and b into r. (r = a * b)
  7595. *
  7596. * r A single precision integer.
  7597. * a A single precision integer.
  7598. * b A single precision integer.
  7599. */
  7600. SP_NOINLINE static void sp_3072_mul_48(sp_digit* r, const sp_digit* a,
  7601. const sp_digit* b)
  7602. {
  7603. sp_digit tmp[48 * 2];
  7604. __asm__ __volatile__ (
  7605. "mov r3, #0\n\t"
  7606. "mov r4, #0\n\t"
  7607. "mov r8, r3\n\t"
  7608. "mov r11, %[r]\n\t"
  7609. "mov r9, %[a]\n\t"
  7610. "mov r10, %[b]\n\t"
  7611. "mov r6, #192\n\t"
  7612. "add r6, r9\n\t"
  7613. "mov r12, r6\n\t"
  7614. "\n1:\n\t"
  7615. "mov %[r], #0\n\t"
  7616. "mov r5, #0\n\t"
  7617. "mov r6, #188\n\t"
  7618. "mov %[a], r8\n\t"
  7619. "sub %[a], r6\n\t"
  7620. "sbc r6, r6\n\t"
  7621. "mvn r6, r6\n\t"
  7622. "and %[a], r6\n\t"
  7623. "mov %[b], r8\n\t"
  7624. "sub %[b], %[a]\n\t"
  7625. "add %[a], r9\n\t"
  7626. "add %[b], r10\n\t"
  7627. "\n2:\n\t"
  7628. "# Multiply Start\n\t"
  7629. "ldr r6, [%[a]]\n\t"
  7630. "ldr r7, [%[b]]\n\t"
  7631. "lsl r6, r6, #16\n\t"
  7632. "lsl r7, r7, #16\n\t"
  7633. "lsr r6, r6, #16\n\t"
  7634. "lsr r7, r7, #16\n\t"
  7635. "mul r7, r6\n\t"
  7636. "add r3, r7\n\t"
  7637. "adc r4, %[r]\n\t"
  7638. "adc r5, %[r]\n\t"
  7639. "ldr r7, [%[b]]\n\t"
  7640. "lsr r7, r7, #16\n\t"
  7641. "mul r6, r7\n\t"
  7642. "lsr r7, r6, #16\n\t"
  7643. "lsl r6, r6, #16\n\t"
  7644. "add r3, r6\n\t"
  7645. "adc r4, r7\n\t"
  7646. "adc r5, %[r]\n\t"
  7647. "ldr r6, [%[a]]\n\t"
  7648. "ldr r7, [%[b]]\n\t"
  7649. "lsr r6, r6, #16\n\t"
  7650. "lsr r7, r7, #16\n\t"
  7651. "mul r7, r6\n\t"
  7652. "add r4, r7\n\t"
  7653. "adc r5, %[r]\n\t"
  7654. "ldr r7, [%[b]]\n\t"
  7655. "lsl r7, r7, #16\n\t"
  7656. "lsr r7, r7, #16\n\t"
  7657. "mul r6, r7\n\t"
  7658. "lsr r7, r6, #16\n\t"
  7659. "lsl r6, r6, #16\n\t"
  7660. "add r3, r6\n\t"
  7661. "adc r4, r7\n\t"
  7662. "adc r5, %[r]\n\t"
  7663. "# Multiply Done\n\t"
  7664. "add %[a], #4\n\t"
  7665. "sub %[b], #4\n\t"
  7666. "cmp %[a], r12\n\t"
  7667. "beq 3f\n\t"
  7668. "mov r6, r8\n\t"
  7669. "add r6, r9\n\t"
  7670. "cmp %[a], r6\n\t"
  7671. "ble 2b\n\t"
  7672. "\n3:\n\t"
  7673. "mov %[r], r11\n\t"
  7674. "mov r7, r8\n\t"
  7675. "str r3, [%[r], r7]\n\t"
  7676. "mov r3, r4\n\t"
  7677. "mov r4, r5\n\t"
  7678. "add r7, #4\n\t"
  7679. "mov r8, r7\n\t"
  7680. "mov r6, #1\n\t"
  7681. "lsl r6, r6, #8\n\t"
  7682. "add r6, #120\n\t"
  7683. "cmp r7, r6\n\t"
  7684. "ble 1b\n\t"
  7685. "str r3, [%[r], r7]\n\t"
  7686. "mov %[a], r9\n\t"
  7687. "mov %[b], r10\n\t"
  7688. :
  7689. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  7690. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  7691. );
  7692. XMEMCPY(r, tmp, sizeof(tmp));
  7693. }
  7694. /* Square a and put result in r. (r = a * a)
  7695. *
  7696. * r A single precision integer.
  7697. * a A single precision integer.
  7698. */
  7699. SP_NOINLINE static void sp_3072_sqr_48(sp_digit* r, const sp_digit* a)
  7700. {
  7701. __asm__ __volatile__ (
  7702. "mov r3, #0\n\t"
  7703. "mov r4, #0\n\t"
  7704. "mov r5, #0\n\t"
  7705. "mov r8, r3\n\t"
  7706. "mov r11, %[r]\n\t"
  7707. "mov r6, #1\n\t"
  7708. "lsl r6, r6, #8\n\t"
  7709. "add r6, #128\n\t"
  7710. "neg r6, r6\n\t"
  7711. "add sp, r6\n\t"
  7712. "mov r10, sp\n\t"
  7713. "mov r9, %[a]\n\t"
  7714. "\n1:\n\t"
  7715. "mov %[r], #0\n\t"
  7716. "mov r6, #188\n\t"
  7717. "mov %[a], r8\n\t"
  7718. "sub %[a], r6\n\t"
  7719. "sbc r6, r6\n\t"
  7720. "mvn r6, r6\n\t"
  7721. "and %[a], r6\n\t"
  7722. "mov r2, r8\n\t"
  7723. "sub r2, %[a]\n\t"
  7724. "add %[a], r9\n\t"
  7725. "add r2, r9\n\t"
  7726. "\n2:\n\t"
  7727. "cmp r2, %[a]\n\t"
  7728. "beq 4f\n\t"
  7729. "# Multiply * 2: Start\n\t"
  7730. "ldr r6, [%[a]]\n\t"
  7731. "ldr r7, [r2]\n\t"
  7732. "lsl r6, r6, #16\n\t"
  7733. "lsl r7, r7, #16\n\t"
  7734. "lsr r6, r6, #16\n\t"
  7735. "lsr r7, r7, #16\n\t"
  7736. "mul r7, r6\n\t"
  7737. "add r3, r7\n\t"
  7738. "adc r4, %[r]\n\t"
  7739. "adc r5, %[r]\n\t"
  7740. "add r3, r7\n\t"
  7741. "adc r4, %[r]\n\t"
  7742. "adc r5, %[r]\n\t"
  7743. "ldr r7, [r2]\n\t"
  7744. "lsr r7, r7, #16\n\t"
  7745. "mul r6, r7\n\t"
  7746. "lsr r7, r6, #16\n\t"
  7747. "lsl r6, r6, #16\n\t"
  7748. "add r3, r6\n\t"
  7749. "adc r4, r7\n\t"
  7750. "adc r5, %[r]\n\t"
  7751. "add r3, r6\n\t"
  7752. "adc r4, r7\n\t"
  7753. "adc r5, %[r]\n\t"
  7754. "ldr r6, [%[a]]\n\t"
  7755. "ldr r7, [r2]\n\t"
  7756. "lsr r6, r6, #16\n\t"
  7757. "lsr r7, r7, #16\n\t"
  7758. "mul r7, r6\n\t"
  7759. "add r4, r7\n\t"
  7760. "adc r5, %[r]\n\t"
  7761. "add r4, r7\n\t"
  7762. "adc r5, %[r]\n\t"
  7763. "ldr r7, [r2]\n\t"
  7764. "lsl r7, r7, #16\n\t"
  7765. "lsr r7, r7, #16\n\t"
  7766. "mul r6, r7\n\t"
  7767. "lsr r7, r6, #16\n\t"
  7768. "lsl r6, r6, #16\n\t"
  7769. "add r3, r6\n\t"
  7770. "adc r4, r7\n\t"
  7771. "adc r5, %[r]\n\t"
  7772. "add r3, r6\n\t"
  7773. "adc r4, r7\n\t"
  7774. "adc r5, %[r]\n\t"
  7775. "# Multiply * 2: Done\n\t"
  7776. "bal 5f\n\t"
  7777. "\n4:\n\t"
  7778. "# Square: Start\n\t"
  7779. "ldr r6, [%[a]]\n\t"
  7780. "lsr r7, r6, #16\n\t"
  7781. "lsl r6, r6, #16\n\t"
  7782. "lsr r6, r6, #16\n\t"
  7783. "mul r6, r6\n\t"
  7784. "add r3, r6\n\t"
  7785. "adc r4, %[r]\n\t"
  7786. "adc r5, %[r]\n\t"
  7787. "mul r7, r7\n\t"
  7788. "add r4, r7\n\t"
  7789. "adc r5, %[r]\n\t"
  7790. "ldr r6, [%[a]]\n\t"
  7791. "lsr r7, r6, #16\n\t"
  7792. "lsl r6, r6, #16\n\t"
  7793. "lsr r6, r6, #16\n\t"
  7794. "mul r6, r7\n\t"
  7795. "lsr r7, r6, #15\n\t"
  7796. "lsl r6, r6, #17\n\t"
  7797. "add r3, r6\n\t"
  7798. "adc r4, r7\n\t"
  7799. "adc r5, %[r]\n\t"
  7800. "# Square: Done\n\t"
  7801. "\n5:\n\t"
  7802. "add %[a], #4\n\t"
  7803. "sub r2, #4\n\t"
  7804. "mov r6, #192\n\t"
  7805. "add r6, r9\n\t"
  7806. "cmp %[a], r6\n\t"
  7807. "beq 3f\n\t"
  7808. "cmp %[a], r2\n\t"
  7809. "bgt 3f\n\t"
  7810. "mov r7, r8\n\t"
  7811. "add r7, r9\n\t"
  7812. "cmp %[a], r7\n\t"
  7813. "ble 2b\n\t"
  7814. "\n3:\n\t"
  7815. "mov %[r], r10\n\t"
  7816. "mov r7, r8\n\t"
  7817. "str r3, [%[r], r7]\n\t"
  7818. "mov r3, r4\n\t"
  7819. "mov r4, r5\n\t"
  7820. "mov r5, #0\n\t"
  7821. "add r7, #4\n\t"
  7822. "mov r8, r7\n\t"
  7823. "mov r6, #1\n\t"
  7824. "lsl r6, r6, #8\n\t"
  7825. "add r6, #120\n\t"
  7826. "cmp r7, r6\n\t"
  7827. "ble 1b\n\t"
  7828. "mov %[a], r9\n\t"
  7829. "str r3, [%[r], r7]\n\t"
  7830. "mov %[r], r11\n\t"
  7831. "mov %[a], r10\n\t"
  7832. "mov r3, #1\n\t"
  7833. "lsl r3, r3, #8\n\t"
  7834. "add r3, #124\n\t"
  7835. "\n4:\n\t"
  7836. "ldr r6, [%[a], r3]\n\t"
  7837. "str r6, [%[r], r3]\n\t"
  7838. "sub r3, #4\n\t"
  7839. "bge 4b\n\t"
  7840. "mov r6, #1\n\t"
  7841. "lsl r6, r6, #8\n\t"
  7842. "add r6, #128\n\t"
  7843. "add sp, r6\n\t"
  7844. :
  7845. : [r] "r" (r), [a] "r" (a)
  7846. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  7847. );
  7848. }
  7849. #endif /* WOLFSSL_SP_SMALL */
  7850. #endif /* (WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH) && !WOLFSSL_RSA_PUBLIC_ONLY */
  7851. /* Caclulate the bottom digit of -1/a mod 2^n.
  7852. *
  7853. * a A single precision number.
  7854. * rho Bottom word of inverse.
  7855. */
  7856. static void sp_3072_mont_setup(const sp_digit* a, sp_digit* rho)
  7857. {
  7858. sp_digit x, b;
  7859. b = a[0];
  7860. x = (((b + 2) & 4) << 1) + b; /* here x*a==1 mod 2**4 */
  7861. x *= 2 - b * x; /* here x*a==1 mod 2**8 */
  7862. x *= 2 - b * x; /* here x*a==1 mod 2**16 */
  7863. x *= 2 - b * x; /* here x*a==1 mod 2**32 */
  7864. /* rho = -1/m mod b */
  7865. *rho = -x;
  7866. }
  7867. /* Mul a by digit b into r. (r = a * b)
  7868. *
  7869. * r A single precision integer.
  7870. * a A single precision integer.
  7871. * b A single precision digit.
  7872. */
  7873. SP_NOINLINE static void sp_3072_mul_d_96(sp_digit* r, const sp_digit* a,
  7874. sp_digit b)
  7875. {
  7876. __asm__ __volatile__ (
  7877. "mov r6, #1\n\t"
  7878. "lsl r6, r6, #8\n\t"
  7879. "add r6, #128\n\t"
  7880. "add r6, %[a]\n\t"
  7881. "mov r8, %[r]\n\t"
  7882. "mov r9, r6\n\t"
  7883. "mov r3, #0\n\t"
  7884. "mov r4, #0\n\t"
  7885. "1:\n\t"
  7886. "mov %[r], #0\n\t"
  7887. "mov r5, #0\n\t"
  7888. "# A[] * B\n\t"
  7889. "ldr r6, [%[a]]\n\t"
  7890. "lsl r6, r6, #16\n\t"
  7891. "lsl r7, %[b], #16\n\t"
  7892. "lsr r6, r6, #16\n\t"
  7893. "lsr r7, r7, #16\n\t"
  7894. "mul r7, r6\n\t"
  7895. "add r3, r7\n\t"
  7896. "adc r4, %[r]\n\t"
  7897. "adc r5, %[r]\n\t"
  7898. "lsr r7, %[b], #16\n\t"
  7899. "mul r6, r7\n\t"
  7900. "lsr r7, r6, #16\n\t"
  7901. "lsl r6, r6, #16\n\t"
  7902. "add r3, r6\n\t"
  7903. "adc r4, r7\n\t"
  7904. "adc r5, %[r]\n\t"
  7905. "ldr r6, [%[a]]\n\t"
  7906. "lsr r6, r6, #16\n\t"
  7907. "lsr r7, %[b], #16\n\t"
  7908. "mul r7, r6\n\t"
  7909. "add r4, r7\n\t"
  7910. "adc r5, %[r]\n\t"
  7911. "lsl r7, %[b], #16\n\t"
  7912. "lsr r7, r7, #16\n\t"
  7913. "mul r6, r7\n\t"
  7914. "lsr r7, r6, #16\n\t"
  7915. "lsl r6, r6, #16\n\t"
  7916. "add r3, r6\n\t"
  7917. "adc r4, r7\n\t"
  7918. "adc r5, %[r]\n\t"
  7919. "# A[] * B - Done\n\t"
  7920. "mov %[r], r8\n\t"
  7921. "str r3, [%[r]]\n\t"
  7922. "mov r3, r4\n\t"
  7923. "mov r4, r5\n\t"
  7924. "add %[r], #4\n\t"
  7925. "add %[a], #4\n\t"
  7926. "mov r8, %[r]\n\t"
  7927. "cmp %[a], r9\n\t"
  7928. "blt 1b\n\t"
  7929. "str r3, [%[r]]\n\t"
  7930. : [r] "+r" (r), [a] "+r" (a)
  7931. : [b] "r" (b)
  7932. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  7933. );
  7934. }
  7935. #if (defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)
  7936. /* r = 2^n mod m where n is the number of bits to reduce by.
  7937. * Given m must be 3072 bits, just need to subtract.
  7938. *
  7939. * r A single precision number.
  7940. * m A single precision number.
  7941. */
  7942. static void sp_3072_mont_norm_48(sp_digit* r, const sp_digit* m)
  7943. {
  7944. XMEMSET(r, 0, sizeof(sp_digit) * 48);
  7945. /* r = 2^n mod m */
  7946. sp_3072_sub_in_place_48(r, m);
  7947. }
  7948. /* Conditionally subtract b from a using the mask m.
  7949. * m is -1 to subtract and 0 when not copying.
  7950. *
  7951. * r A single precision number representing condition subtract result.
  7952. * a A single precision number to subtract from.
  7953. * b A single precision number to subtract.
  7954. * m Mask value to apply.
  7955. */
  7956. SP_NOINLINE static sp_digit sp_3072_cond_sub_48(sp_digit* r, const sp_digit* a,
  7957. const sp_digit* b, sp_digit m)
  7958. {
  7959. sp_digit c = 0;
  7960. __asm__ __volatile__ (
  7961. "mov r5, #192\n\t"
  7962. "mov r8, r5\n\t"
  7963. "mov r7, #0\n\t"
  7964. "1:\n\t"
  7965. "ldr r6, [%[b], r7]\n\t"
  7966. "and r6, %[m]\n\t"
  7967. "mov r5, #0\n\t"
  7968. "sub r5, %[c]\n\t"
  7969. "ldr r5, [%[a], r7]\n\t"
  7970. "sbc r5, r6\n\t"
  7971. "sbc %[c], %[c]\n\t"
  7972. "str r5, [%[r], r7]\n\t"
  7973. "add r7, #4\n\t"
  7974. "cmp r7, r8\n\t"
  7975. "blt 1b\n\t"
  7976. : [c] "+r" (c)
  7977. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  7978. : "memory", "r5", "r6", "r7", "r8"
  7979. );
  7980. return c;
  7981. }
  7982. /* Reduce the number back to 3072 bits using Montgomery reduction.
  7983. *
  7984. * a A single precision number to reduce in place.
  7985. * m The single precision number representing the modulus.
  7986. * mp The digit representing the negative inverse of m mod 2^n.
  7987. */
  7988. SP_NOINLINE static void sp_3072_mont_reduce_48(sp_digit* a, const sp_digit* m,
  7989. sp_digit mp)
  7990. {
  7991. sp_digit ca = 0;
  7992. __asm__ __volatile__ (
  7993. "mov r8, %[mp]\n\t"
  7994. "mov r12, %[ca]\n\t"
  7995. "mov r14, %[m]\n\t"
  7996. "mov r9, %[a]\n\t"
  7997. "mov r4, #0\n\t"
  7998. "# i = 0\n\t"
  7999. "mov r11, r4\n\t"
  8000. "\n1:\n\t"
  8001. "mov r5, #0\n\t"
  8002. "mov %[ca], #0\n\t"
  8003. "# mu = a[i] * mp\n\t"
  8004. "mov %[mp], r8\n\t"
  8005. "ldr %[a], [%[a]]\n\t"
  8006. "mul %[mp], %[a]\n\t"
  8007. "mov %[m], r14\n\t"
  8008. "mov r10, r9\n\t"
  8009. "\n2:\n\t"
  8010. "# a[i+j] += m[j] * mu\n\t"
  8011. "mov %[a], r10\n\t"
  8012. "ldr %[a], [%[a]]\n\t"
  8013. "mov %[ca], #0\n\t"
  8014. "mov r4, r5\n\t"
  8015. "mov r5, #0\n\t"
  8016. "# Multiply m[j] and mu - Start\n\t"
  8017. "ldr r7, [%[m]]\n\t"
  8018. "lsl r6, %[mp], #16\n\t"
  8019. "lsl r7, r7, #16\n\t"
  8020. "lsr r6, r6, #16\n\t"
  8021. "lsr r7, r7, #16\n\t"
  8022. "mul r7, r6\n\t"
  8023. "add %[a], r7\n\t"
  8024. "adc r5, %[ca]\n\t"
  8025. "ldr r7, [%[m]]\n\t"
  8026. "lsr r7, r7, #16\n\t"
  8027. "mul r6, r7\n\t"
  8028. "lsr r7, r6, #16\n\t"
  8029. "lsl r6, r6, #16\n\t"
  8030. "add %[a], r6\n\t"
  8031. "adc r5, r7\n\t"
  8032. "ldr r7, [%[m]]\n\t"
  8033. "lsr r6, %[mp], #16\n\t"
  8034. "lsr r7, r7, #16\n\t"
  8035. "mul r7, r6\n\t"
  8036. "add r5, r7\n\t"
  8037. "ldr r7, [%[m]]\n\t"
  8038. "lsl r7, r7, #16\n\t"
  8039. "lsr r7, r7, #16\n\t"
  8040. "mul r6, r7\n\t"
  8041. "lsr r7, r6, #16\n\t"
  8042. "lsl r6, r6, #16\n\t"
  8043. "add %[a], r6\n\t"
  8044. "adc r5, r7\n\t"
  8045. "# Multiply m[j] and mu - Done\n\t"
  8046. "add r4, %[a]\n\t"
  8047. "adc r5, %[ca]\n\t"
  8048. "mov %[a], r10\n\t"
  8049. "str r4, [%[a]]\n\t"
  8050. "mov r6, #4\n\t"
  8051. "add %[m], #4\n\t"
  8052. "add r10, r6\n\t"
  8053. "mov r4, #188\n\t"
  8054. "add r4, r9\n\t"
  8055. "cmp r10, r4\n\t"
  8056. "blt 2b\n\t"
  8057. "# a[i+47] += m[47] * mu\n\t"
  8058. "mov %[ca], #0\n\t"
  8059. "mov r4, r12\n\t"
  8060. "mov %[a], #0\n\t"
  8061. "# Multiply m[47] and mu - Start\n\t"
  8062. "ldr r7, [%[m]]\n\t"
  8063. "lsl r6, %[mp], #16\n\t"
  8064. "lsl r7, r7, #16\n\t"
  8065. "lsr r6, r6, #16\n\t"
  8066. "lsr r7, r7, #16\n\t"
  8067. "mul r7, r6\n\t"
  8068. "add r5, r7\n\t"
  8069. "adc r4, %[ca]\n\t"
  8070. "adc %[a], %[ca]\n\t"
  8071. "ldr r7, [%[m]]\n\t"
  8072. "lsr r7, r7, #16\n\t"
  8073. "mul r6, r7\n\t"
  8074. "lsr r7, r6, #16\n\t"
  8075. "lsl r6, r6, #16\n\t"
  8076. "add r5, r6\n\t"
  8077. "adc r4, r7\n\t"
  8078. "adc %[a], %[ca]\n\t"
  8079. "ldr r7, [%[m]]\n\t"
  8080. "lsr r6, %[mp], #16\n\t"
  8081. "lsr r7, r7, #16\n\t"
  8082. "mul r7, r6\n\t"
  8083. "add r4, r7\n\t"
  8084. "adc %[a], %[ca]\n\t"
  8085. "ldr r7, [%[m]]\n\t"
  8086. "lsl r7, r7, #16\n\t"
  8087. "lsr r7, r7, #16\n\t"
  8088. "mul r6, r7\n\t"
  8089. "lsr r7, r6, #16\n\t"
  8090. "lsl r6, r6, #16\n\t"
  8091. "add r5, r6\n\t"
  8092. "adc r4, r7\n\t"
  8093. "adc %[a], %[ca]\n\t"
  8094. "# Multiply m[47] and mu - Done\n\t"
  8095. "mov %[ca], %[a]\n\t"
  8096. "mov %[a], r10\n\t"
  8097. "ldr r7, [%[a], #4]\n\t"
  8098. "ldr %[a], [%[a]]\n\t"
  8099. "mov r6, #0\n\t"
  8100. "add r5, %[a]\n\t"
  8101. "adc r7, r4\n\t"
  8102. "adc %[ca], r6\n\t"
  8103. "mov %[a], r10\n\t"
  8104. "str r5, [%[a]]\n\t"
  8105. "str r7, [%[a], #4]\n\t"
  8106. "# i += 1\n\t"
  8107. "mov r6, #4\n\t"
  8108. "add r9, r6\n\t"
  8109. "add r11, r6\n\t"
  8110. "mov r12, %[ca]\n\t"
  8111. "mov %[a], r9\n\t"
  8112. "mov r4, #192\n\t"
  8113. "cmp r11, r4\n\t"
  8114. "blt 1b\n\t"
  8115. "mov %[m], r14\n\t"
  8116. : [ca] "+r" (ca), [a] "+r" (a)
  8117. : [m] "r" (m), [mp] "r" (mp)
  8118. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  8119. );
  8120. sp_3072_cond_sub_48(a - 48, a, m, (sp_digit)0 - ca);
  8121. }
  8122. /* Multiply two Montogmery form numbers mod the modulus (prime).
  8123. * (r = a * b mod m)
  8124. *
  8125. * r Result of multiplication.
  8126. * a First number to multiply in Montogmery form.
  8127. * b Second number to multiply in Montogmery form.
  8128. * m Modulus (prime).
  8129. * mp Montogmery mulitplier.
  8130. */
  8131. static void sp_3072_mont_mul_48(sp_digit* r, const sp_digit* a, const sp_digit* b,
  8132. const sp_digit* m, sp_digit mp)
  8133. {
  8134. sp_3072_mul_48(r, a, b);
  8135. sp_3072_mont_reduce_48(r, m, mp);
  8136. }
  8137. /* Square the Montgomery form number. (r = a * a mod m)
  8138. *
  8139. * r Result of squaring.
  8140. * a Number to square in Montogmery form.
  8141. * m Modulus (prime).
  8142. * mp Montogmery mulitplier.
  8143. */
  8144. static void sp_3072_mont_sqr_48(sp_digit* r, const sp_digit* a, const sp_digit* m,
  8145. sp_digit mp)
  8146. {
  8147. sp_3072_sqr_48(r, a);
  8148. sp_3072_mont_reduce_48(r, m, mp);
  8149. }
  8150. /* Mul a by digit b into r. (r = a * b)
  8151. *
  8152. * r A single precision integer.
  8153. * a A single precision integer.
  8154. * b A single precision digit.
  8155. */
  8156. SP_NOINLINE static void sp_3072_mul_d_48(sp_digit* r, const sp_digit* a,
  8157. sp_digit b)
  8158. {
  8159. __asm__ __volatile__ (
  8160. "mov r6, #192\n\t"
  8161. "add r6, %[a]\n\t"
  8162. "mov r8, %[r]\n\t"
  8163. "mov r9, r6\n\t"
  8164. "mov r3, #0\n\t"
  8165. "mov r4, #0\n\t"
  8166. "1:\n\t"
  8167. "mov %[r], #0\n\t"
  8168. "mov r5, #0\n\t"
  8169. "# A[] * B\n\t"
  8170. "ldr r6, [%[a]]\n\t"
  8171. "lsl r6, r6, #16\n\t"
  8172. "lsl r7, %[b], #16\n\t"
  8173. "lsr r6, r6, #16\n\t"
  8174. "lsr r7, r7, #16\n\t"
  8175. "mul r7, r6\n\t"
  8176. "add r3, r7\n\t"
  8177. "adc r4, %[r]\n\t"
  8178. "adc r5, %[r]\n\t"
  8179. "lsr r7, %[b], #16\n\t"
  8180. "mul r6, r7\n\t"
  8181. "lsr r7, r6, #16\n\t"
  8182. "lsl r6, r6, #16\n\t"
  8183. "add r3, r6\n\t"
  8184. "adc r4, r7\n\t"
  8185. "adc r5, %[r]\n\t"
  8186. "ldr r6, [%[a]]\n\t"
  8187. "lsr r6, r6, #16\n\t"
  8188. "lsr r7, %[b], #16\n\t"
  8189. "mul r7, r6\n\t"
  8190. "add r4, r7\n\t"
  8191. "adc r5, %[r]\n\t"
  8192. "lsl r7, %[b], #16\n\t"
  8193. "lsr r7, r7, #16\n\t"
  8194. "mul r6, r7\n\t"
  8195. "lsr r7, r6, #16\n\t"
  8196. "lsl r6, r6, #16\n\t"
  8197. "add r3, r6\n\t"
  8198. "adc r4, r7\n\t"
  8199. "adc r5, %[r]\n\t"
  8200. "# A[] * B - Done\n\t"
  8201. "mov %[r], r8\n\t"
  8202. "str r3, [%[r]]\n\t"
  8203. "mov r3, r4\n\t"
  8204. "mov r4, r5\n\t"
  8205. "add %[r], #4\n\t"
  8206. "add %[a], #4\n\t"
  8207. "mov r8, %[r]\n\t"
  8208. "cmp %[a], r9\n\t"
  8209. "blt 1b\n\t"
  8210. "str r3, [%[r]]\n\t"
  8211. : [r] "+r" (r), [a] "+r" (a)
  8212. : [b] "r" (b)
  8213. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  8214. );
  8215. }
  8216. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  8217. *
  8218. * d1 The high order half of the number to divide.
  8219. * d0 The low order half of the number to divide.
  8220. * div The dividend.
  8221. * returns the result of the division.
  8222. *
  8223. * Note that this is an approximate div. It may give an answer 1 larger.
  8224. */
  8225. SP_NOINLINE static sp_digit div_3072_word_48(sp_digit d1, sp_digit d0,
  8226. sp_digit div)
  8227. {
  8228. sp_digit r = 0;
  8229. __asm__ __volatile__ (
  8230. "lsr r5, %[div], #1\n\t"
  8231. "add r5, #1\n\t"
  8232. "mov r8, %[d0]\n\t"
  8233. "mov r9, %[d1]\n\t"
  8234. "# Do top 32\n\t"
  8235. "mov r6, r5\n\t"
  8236. "sub r6, %[d1]\n\t"
  8237. "sbc r6, r6\n\t"
  8238. "add %[r], %[r]\n\t"
  8239. "sub %[r], r6\n\t"
  8240. "and r6, r5\n\t"
  8241. "sub %[d1], r6\n\t"
  8242. "# Next 30 bits\n\t"
  8243. "mov r4, #29\n\t"
  8244. "1:\n\t"
  8245. "lsl %[d0], %[d0], #1\n\t"
  8246. "adc %[d1], %[d1]\n\t"
  8247. "mov r6, r5\n\t"
  8248. "sub r6, %[d1]\n\t"
  8249. "sbc r6, r6\n\t"
  8250. "add %[r], %[r]\n\t"
  8251. "sub %[r], r6\n\t"
  8252. "and r6, r5\n\t"
  8253. "sub %[d1], r6\n\t"
  8254. "sub r4, #1\n\t"
  8255. "bpl 1b\n\t"
  8256. "mov r7, #0\n\t"
  8257. "add %[r], %[r]\n\t"
  8258. "add %[r], #1\n\t"
  8259. "# r * div - Start\n\t"
  8260. "lsl %[d1], %[r], #16\n\t"
  8261. "lsl r4, %[div], #16\n\t"
  8262. "lsr %[d1], %[d1], #16\n\t"
  8263. "lsr r4, r4, #16\n\t"
  8264. "mul r4, %[d1]\n\t"
  8265. "lsr r6, %[div], #16\n\t"
  8266. "mul %[d1], r6\n\t"
  8267. "lsr r5, %[d1], #16\n\t"
  8268. "lsl %[d1], %[d1], #16\n\t"
  8269. "add r4, %[d1]\n\t"
  8270. "adc r5, r7\n\t"
  8271. "lsr %[d1], %[r], #16\n\t"
  8272. "mul r6, %[d1]\n\t"
  8273. "add r5, r6\n\t"
  8274. "lsl r6, %[div], #16\n\t"
  8275. "lsr r6, r6, #16\n\t"
  8276. "mul %[d1], r6\n\t"
  8277. "lsr r6, %[d1], #16\n\t"
  8278. "lsl %[d1], %[d1], #16\n\t"
  8279. "add r4, %[d1]\n\t"
  8280. "adc r5, r6\n\t"
  8281. "# r * div - Done\n\t"
  8282. "mov %[d1], r8\n\t"
  8283. "sub %[d1], r4\n\t"
  8284. "mov r4, %[d1]\n\t"
  8285. "mov %[d1], r9\n\t"
  8286. "sbc %[d1], r5\n\t"
  8287. "mov r5, %[d1]\n\t"
  8288. "add %[r], r5\n\t"
  8289. "# r * div - Start\n\t"
  8290. "lsl %[d1], %[r], #16\n\t"
  8291. "lsl r4, %[div], #16\n\t"
  8292. "lsr %[d1], %[d1], #16\n\t"
  8293. "lsr r4, r4, #16\n\t"
  8294. "mul r4, %[d1]\n\t"
  8295. "lsr r6, %[div], #16\n\t"
  8296. "mul %[d1], r6\n\t"
  8297. "lsr r5, %[d1], #16\n\t"
  8298. "lsl %[d1], %[d1], #16\n\t"
  8299. "add r4, %[d1]\n\t"
  8300. "adc r5, r7\n\t"
  8301. "lsr %[d1], %[r], #16\n\t"
  8302. "mul r6, %[d1]\n\t"
  8303. "add r5, r6\n\t"
  8304. "lsl r6, %[div], #16\n\t"
  8305. "lsr r6, r6, #16\n\t"
  8306. "mul %[d1], r6\n\t"
  8307. "lsr r6, %[d1], #16\n\t"
  8308. "lsl %[d1], %[d1], #16\n\t"
  8309. "add r4, %[d1]\n\t"
  8310. "adc r5, r6\n\t"
  8311. "# r * div - Done\n\t"
  8312. "mov %[d1], r8\n\t"
  8313. "mov r6, r9\n\t"
  8314. "sub r4, %[d1], r4\n\t"
  8315. "sbc r6, r5\n\t"
  8316. "mov r5, r6\n\t"
  8317. "add %[r], r5\n\t"
  8318. "# r * div - Start\n\t"
  8319. "lsl %[d1], %[r], #16\n\t"
  8320. "lsl r4, %[div], #16\n\t"
  8321. "lsr %[d1], %[d1], #16\n\t"
  8322. "lsr r4, r4, #16\n\t"
  8323. "mul r4, %[d1]\n\t"
  8324. "lsr r6, %[div], #16\n\t"
  8325. "mul %[d1], r6\n\t"
  8326. "lsr r5, %[d1], #16\n\t"
  8327. "lsl %[d1], %[d1], #16\n\t"
  8328. "add r4, %[d1]\n\t"
  8329. "adc r5, r7\n\t"
  8330. "lsr %[d1], %[r], #16\n\t"
  8331. "mul r6, %[d1]\n\t"
  8332. "add r5, r6\n\t"
  8333. "lsl r6, %[div], #16\n\t"
  8334. "lsr r6, r6, #16\n\t"
  8335. "mul %[d1], r6\n\t"
  8336. "lsr r6, %[d1], #16\n\t"
  8337. "lsl %[d1], %[d1], #16\n\t"
  8338. "add r4, %[d1]\n\t"
  8339. "adc r5, r6\n\t"
  8340. "# r * div - Done\n\t"
  8341. "mov %[d1], r8\n\t"
  8342. "mov r6, r9\n\t"
  8343. "sub r4, %[d1], r4\n\t"
  8344. "sbc r6, r5\n\t"
  8345. "mov r5, r6\n\t"
  8346. "add %[r], r5\n\t"
  8347. "mov r6, %[div]\n\t"
  8348. "sub r6, r4\n\t"
  8349. "sbc r6, r6\n\t"
  8350. "sub %[r], r6\n\t"
  8351. : [r] "+r" (r)
  8352. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  8353. : "r4", "r5", "r7", "r6", "r8", "r9"
  8354. );
  8355. return r;
  8356. }
  8357. /* Compare a with b in constant time.
  8358. *
  8359. * a A single precision integer.
  8360. * b A single precision integer.
  8361. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  8362. * respectively.
  8363. */
  8364. SP_NOINLINE static int32_t sp_3072_cmp_48(const sp_digit* a, const sp_digit* b)
  8365. {
  8366. sp_digit r = 0;
  8367. __asm__ __volatile__ (
  8368. "mov r3, #0\n\t"
  8369. "mvn r3, r3\n\t"
  8370. "mov r6, #188\n\t"
  8371. "1:\n\t"
  8372. "ldr r7, [%[a], r6]\n\t"
  8373. "ldr r5, [%[b], r6]\n\t"
  8374. "and r7, r3\n\t"
  8375. "and r5, r3\n\t"
  8376. "mov r4, r7\n\t"
  8377. "sub r7, r5\n\t"
  8378. "sbc r7, r7\n\t"
  8379. "add %[r], r7\n\t"
  8380. "mvn r7, r7\n\t"
  8381. "and r3, r7\n\t"
  8382. "sub r5, r4\n\t"
  8383. "sbc r7, r7\n\t"
  8384. "sub %[r], r7\n\t"
  8385. "mvn r7, r7\n\t"
  8386. "and r3, r7\n\t"
  8387. "sub r6, #4\n\t"
  8388. "cmp r6, #0\n\t"
  8389. "bge 1b\n\t"
  8390. : [r] "+r" (r)
  8391. : [a] "r" (a), [b] "r" (b)
  8392. : "r3", "r4", "r5", "r6", "r7"
  8393. );
  8394. return r;
  8395. }
  8396. /* Divide d in a and put remainder into r (m*d + r = a)
  8397. * m is not calculated as it is not needed at this time.
  8398. *
  8399. * a Nmber to be divided.
  8400. * d Number to divide with.
  8401. * m Multiplier result.
  8402. * r Remainder from the division.
  8403. * returns MP_OKAY indicating success.
  8404. */
  8405. static WC_INLINE int sp_3072_div_48(const sp_digit* a, const sp_digit* d, sp_digit* m,
  8406. sp_digit* r)
  8407. {
  8408. sp_digit t1[96], t2[49];
  8409. sp_digit div, r1;
  8410. int i;
  8411. (void)m;
  8412. div = d[47];
  8413. XMEMCPY(t1, a, sizeof(*t1) * 2 * 48);
  8414. for (i=47; i>=0; i--) {
  8415. r1 = div_3072_word_48(t1[48 + i], t1[48 + i - 1], div);
  8416. sp_3072_mul_d_48(t2, d, r1);
  8417. t1[48 + i] += sp_3072_sub_in_place_48(&t1[i], t2);
  8418. t1[48 + i] -= t2[48];
  8419. sp_3072_mask_48(t2, d, t1[48 + i]);
  8420. t1[48 + i] += sp_3072_add_48(&t1[i], &t1[i], t2);
  8421. sp_3072_mask_48(t2, d, t1[48 + i]);
  8422. t1[48 + i] += sp_3072_add_48(&t1[i], &t1[i], t2);
  8423. }
  8424. r1 = sp_3072_cmp_48(t1, d) >= 0;
  8425. sp_3072_cond_sub_48(r, t1, d, (sp_digit)0 - r1);
  8426. return MP_OKAY;
  8427. }
  8428. /* Reduce a modulo m into r. (r = a mod m)
  8429. *
  8430. * r A single precision number that is the reduced result.
  8431. * a A single precision number that is to be reduced.
  8432. * m A single precision number that is the modulus to reduce with.
  8433. * returns MP_OKAY indicating success.
  8434. */
  8435. static WC_INLINE int sp_3072_mod_48(sp_digit* r, const sp_digit* a, const sp_digit* m)
  8436. {
  8437. return sp_3072_div_48(a, m, NULL, r);
  8438. }
  8439. #ifdef WOLFSSL_SP_SMALL
  8440. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  8441. *
  8442. * r A single precision number that is the result of the operation.
  8443. * a A single precision number being exponentiated.
  8444. * e A single precision number that is the exponent.
  8445. * bits The number of bits in the exponent.
  8446. * m A single precision number that is the modulus.
  8447. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  8448. */
  8449. static int sp_3072_mod_exp_48(sp_digit* r, const sp_digit* a, const sp_digit* e,
  8450. int bits, const sp_digit* m, int reduceA)
  8451. {
  8452. #ifndef WOLFSSL_SMALL_STACK
  8453. sp_digit t[16][96];
  8454. #else
  8455. sp_digit* t[16];
  8456. sp_digit* td;
  8457. #endif
  8458. sp_digit* norm;
  8459. sp_digit mp = 1;
  8460. sp_digit n;
  8461. sp_digit mask;
  8462. int i;
  8463. int c, y;
  8464. int err = MP_OKAY;
  8465. #ifdef WOLFSSL_SMALL_STACK
  8466. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 96, NULL,
  8467. DYNAMIC_TYPE_TMP_BUFFER);
  8468. if (td == NULL) {
  8469. err = MEMORY_E;
  8470. }
  8471. #endif
  8472. if (err == MP_OKAY) {
  8473. #ifdef WOLFSSL_SMALL_STACK
  8474. for (i=0; i<16; i++) {
  8475. t[i] = td + i * 96;
  8476. }
  8477. #endif
  8478. norm = t[0];
  8479. sp_3072_mont_setup(m, &mp);
  8480. sp_3072_mont_norm_48(norm, m);
  8481. XMEMSET(t[1], 0, sizeof(sp_digit) * 48U);
  8482. if (reduceA != 0) {
  8483. err = sp_3072_mod_48(t[1] + 48, a, m);
  8484. if (err == MP_OKAY) {
  8485. err = sp_3072_mod_48(t[1], t[1], m);
  8486. }
  8487. }
  8488. else {
  8489. XMEMCPY(t[1] + 48, a, sizeof(sp_digit) * 48);
  8490. err = sp_3072_mod_48(t[1], t[1], m);
  8491. }
  8492. }
  8493. if (err == MP_OKAY) {
  8494. sp_3072_mont_sqr_48(t[ 2], t[ 1], m, mp);
  8495. sp_3072_mont_mul_48(t[ 3], t[ 2], t[ 1], m, mp);
  8496. sp_3072_mont_sqr_48(t[ 4], t[ 2], m, mp);
  8497. sp_3072_mont_mul_48(t[ 5], t[ 3], t[ 2], m, mp);
  8498. sp_3072_mont_sqr_48(t[ 6], t[ 3], m, mp);
  8499. sp_3072_mont_mul_48(t[ 7], t[ 4], t[ 3], m, mp);
  8500. sp_3072_mont_sqr_48(t[ 8], t[ 4], m, mp);
  8501. sp_3072_mont_mul_48(t[ 9], t[ 5], t[ 4], m, mp);
  8502. sp_3072_mont_sqr_48(t[10], t[ 5], m, mp);
  8503. sp_3072_mont_mul_48(t[11], t[ 6], t[ 5], m, mp);
  8504. sp_3072_mont_sqr_48(t[12], t[ 6], m, mp);
  8505. sp_3072_mont_mul_48(t[13], t[ 7], t[ 6], m, mp);
  8506. sp_3072_mont_sqr_48(t[14], t[ 7], m, mp);
  8507. sp_3072_mont_mul_48(t[15], t[ 8], t[ 7], m, mp);
  8508. i = (bits - 1) / 32;
  8509. n = e[i--];
  8510. c = bits & 31;
  8511. if (c == 0) {
  8512. c = 32;
  8513. }
  8514. c -= bits % 4;
  8515. if (c == 32) {
  8516. c = 28;
  8517. }
  8518. y = (int)(n >> c);
  8519. n <<= 32 - c;
  8520. XMEMCPY(r, t[y], sizeof(sp_digit) * 48);
  8521. for (; i>=0 || c>=4; ) {
  8522. if (c == 0) {
  8523. n = e[i--];
  8524. y = n >> 28;
  8525. n <<= 4;
  8526. c = 28;
  8527. }
  8528. else if (c < 4) {
  8529. y = n >> 28;
  8530. n = e[i--];
  8531. c = 4 - c;
  8532. y |= n >> (32 - c);
  8533. n <<= c;
  8534. c = 32 - c;
  8535. }
  8536. else {
  8537. y = (n >> 28) & 0xf;
  8538. n <<= 4;
  8539. c -= 4;
  8540. }
  8541. sp_3072_mont_sqr_48(r, r, m, mp);
  8542. sp_3072_mont_sqr_48(r, r, m, mp);
  8543. sp_3072_mont_sqr_48(r, r, m, mp);
  8544. sp_3072_mont_sqr_48(r, r, m, mp);
  8545. sp_3072_mont_mul_48(r, r, t[y], m, mp);
  8546. }
  8547. XMEMSET(&r[48], 0, sizeof(sp_digit) * 48U);
  8548. sp_3072_mont_reduce_48(r, m, mp);
  8549. mask = 0 - (sp_3072_cmp_48(r, m) >= 0);
  8550. sp_3072_cond_sub_48(r, r, m, mask);
  8551. }
  8552. #ifdef WOLFSSL_SMALL_STACK
  8553. if (td != NULL) {
  8554. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  8555. }
  8556. #endif
  8557. return err;
  8558. }
  8559. #else
  8560. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  8561. *
  8562. * r A single precision number that is the result of the operation.
  8563. * a A single precision number being exponentiated.
  8564. * e A single precision number that is the exponent.
  8565. * bits The number of bits in the exponent.
  8566. * m A single precision number that is the modulus.
  8567. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  8568. */
  8569. static int sp_3072_mod_exp_48(sp_digit* r, const sp_digit* a, const sp_digit* e,
  8570. int bits, const sp_digit* m, int reduceA)
  8571. {
  8572. #ifndef WOLFSSL_SMALL_STACK
  8573. sp_digit t[32][96];
  8574. #else
  8575. sp_digit* t[32];
  8576. sp_digit* td;
  8577. #endif
  8578. sp_digit* norm;
  8579. sp_digit mp = 1;
  8580. sp_digit n;
  8581. sp_digit mask;
  8582. int i;
  8583. int c, y;
  8584. int err = MP_OKAY;
  8585. #ifdef WOLFSSL_SMALL_STACK
  8586. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 96, NULL,
  8587. DYNAMIC_TYPE_TMP_BUFFER);
  8588. if (td == NULL) {
  8589. err = MEMORY_E;
  8590. }
  8591. #endif
  8592. if (err == MP_OKAY) {
  8593. #ifdef WOLFSSL_SMALL_STACK
  8594. for (i=0; i<32; i++) {
  8595. t[i] = td + i * 96;
  8596. }
  8597. #endif
  8598. norm = t[0];
  8599. sp_3072_mont_setup(m, &mp);
  8600. sp_3072_mont_norm_48(norm, m);
  8601. XMEMSET(t[1], 0, sizeof(sp_digit) * 48U);
  8602. if (reduceA != 0) {
  8603. err = sp_3072_mod_48(t[1] + 48, a, m);
  8604. if (err == MP_OKAY) {
  8605. err = sp_3072_mod_48(t[1], t[1], m);
  8606. }
  8607. }
  8608. else {
  8609. XMEMCPY(t[1] + 48, a, sizeof(sp_digit) * 48);
  8610. err = sp_3072_mod_48(t[1], t[1], m);
  8611. }
  8612. }
  8613. if (err == MP_OKAY) {
  8614. sp_3072_mont_sqr_48(t[ 2], t[ 1], m, mp);
  8615. sp_3072_mont_mul_48(t[ 3], t[ 2], t[ 1], m, mp);
  8616. sp_3072_mont_sqr_48(t[ 4], t[ 2], m, mp);
  8617. sp_3072_mont_mul_48(t[ 5], t[ 3], t[ 2], m, mp);
  8618. sp_3072_mont_sqr_48(t[ 6], t[ 3], m, mp);
  8619. sp_3072_mont_mul_48(t[ 7], t[ 4], t[ 3], m, mp);
  8620. sp_3072_mont_sqr_48(t[ 8], t[ 4], m, mp);
  8621. sp_3072_mont_mul_48(t[ 9], t[ 5], t[ 4], m, mp);
  8622. sp_3072_mont_sqr_48(t[10], t[ 5], m, mp);
  8623. sp_3072_mont_mul_48(t[11], t[ 6], t[ 5], m, mp);
  8624. sp_3072_mont_sqr_48(t[12], t[ 6], m, mp);
  8625. sp_3072_mont_mul_48(t[13], t[ 7], t[ 6], m, mp);
  8626. sp_3072_mont_sqr_48(t[14], t[ 7], m, mp);
  8627. sp_3072_mont_mul_48(t[15], t[ 8], t[ 7], m, mp);
  8628. sp_3072_mont_sqr_48(t[16], t[ 8], m, mp);
  8629. sp_3072_mont_mul_48(t[17], t[ 9], t[ 8], m, mp);
  8630. sp_3072_mont_sqr_48(t[18], t[ 9], m, mp);
  8631. sp_3072_mont_mul_48(t[19], t[10], t[ 9], m, mp);
  8632. sp_3072_mont_sqr_48(t[20], t[10], m, mp);
  8633. sp_3072_mont_mul_48(t[21], t[11], t[10], m, mp);
  8634. sp_3072_mont_sqr_48(t[22], t[11], m, mp);
  8635. sp_3072_mont_mul_48(t[23], t[12], t[11], m, mp);
  8636. sp_3072_mont_sqr_48(t[24], t[12], m, mp);
  8637. sp_3072_mont_mul_48(t[25], t[13], t[12], m, mp);
  8638. sp_3072_mont_sqr_48(t[26], t[13], m, mp);
  8639. sp_3072_mont_mul_48(t[27], t[14], t[13], m, mp);
  8640. sp_3072_mont_sqr_48(t[28], t[14], m, mp);
  8641. sp_3072_mont_mul_48(t[29], t[15], t[14], m, mp);
  8642. sp_3072_mont_sqr_48(t[30], t[15], m, mp);
  8643. sp_3072_mont_mul_48(t[31], t[16], t[15], m, mp);
  8644. i = (bits - 1) / 32;
  8645. n = e[i--];
  8646. c = bits & 31;
  8647. if (c == 0) {
  8648. c = 32;
  8649. }
  8650. c -= bits % 5;
  8651. if (c == 32) {
  8652. c = 27;
  8653. }
  8654. y = (int)(n >> c);
  8655. n <<= 32 - c;
  8656. XMEMCPY(r, t[y], sizeof(sp_digit) * 48);
  8657. for (; i>=0 || c>=5; ) {
  8658. if (c == 0) {
  8659. n = e[i--];
  8660. y = n >> 27;
  8661. n <<= 5;
  8662. c = 27;
  8663. }
  8664. else if (c < 5) {
  8665. y = n >> 27;
  8666. n = e[i--];
  8667. c = 5 - c;
  8668. y |= n >> (32 - c);
  8669. n <<= c;
  8670. c = 32 - c;
  8671. }
  8672. else {
  8673. y = (n >> 27) & 0x1f;
  8674. n <<= 5;
  8675. c -= 5;
  8676. }
  8677. sp_3072_mont_sqr_48(r, r, m, mp);
  8678. sp_3072_mont_sqr_48(r, r, m, mp);
  8679. sp_3072_mont_sqr_48(r, r, m, mp);
  8680. sp_3072_mont_sqr_48(r, r, m, mp);
  8681. sp_3072_mont_sqr_48(r, r, m, mp);
  8682. sp_3072_mont_mul_48(r, r, t[y], m, mp);
  8683. }
  8684. XMEMSET(&r[48], 0, sizeof(sp_digit) * 48U);
  8685. sp_3072_mont_reduce_48(r, m, mp);
  8686. mask = 0 - (sp_3072_cmp_48(r, m) >= 0);
  8687. sp_3072_cond_sub_48(r, r, m, mask);
  8688. }
  8689. #ifdef WOLFSSL_SMALL_STACK
  8690. if (td != NULL) {
  8691. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  8692. }
  8693. #endif
  8694. return err;
  8695. }
  8696. #endif /* WOLFSSL_SP_SMALL */
  8697. #endif /* (WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH) && !WOLFSSL_RSA_PUBLIC_ONLY */
  8698. #if defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)
  8699. /* r = 2^n mod m where n is the number of bits to reduce by.
  8700. * Given m must be 3072 bits, just need to subtract.
  8701. *
  8702. * r A single precision number.
  8703. * m A single precision number.
  8704. */
  8705. static void sp_3072_mont_norm_96(sp_digit* r, const sp_digit* m)
  8706. {
  8707. XMEMSET(r, 0, sizeof(sp_digit) * 96);
  8708. /* r = 2^n mod m */
  8709. sp_3072_sub_in_place_96(r, m);
  8710. }
  8711. #endif /* WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH */
  8712. /* Conditionally subtract b from a using the mask m.
  8713. * m is -1 to subtract and 0 when not copying.
  8714. *
  8715. * r A single precision number representing condition subtract result.
  8716. * a A single precision number to subtract from.
  8717. * b A single precision number to subtract.
  8718. * m Mask value to apply.
  8719. */
  8720. SP_NOINLINE static sp_digit sp_3072_cond_sub_96(sp_digit* r, const sp_digit* a,
  8721. const sp_digit* b, sp_digit m)
  8722. {
  8723. sp_digit c = 0;
  8724. __asm__ __volatile__ (
  8725. "mov r5, #1\n\t"
  8726. "lsl r5, r5, #8\n\t"
  8727. "add r5, #128\n\t"
  8728. "mov r8, r5\n\t"
  8729. "mov r7, #0\n\t"
  8730. "1:\n\t"
  8731. "ldr r6, [%[b], r7]\n\t"
  8732. "and r6, %[m]\n\t"
  8733. "mov r5, #0\n\t"
  8734. "sub r5, %[c]\n\t"
  8735. "ldr r5, [%[a], r7]\n\t"
  8736. "sbc r5, r6\n\t"
  8737. "sbc %[c], %[c]\n\t"
  8738. "str r5, [%[r], r7]\n\t"
  8739. "add r7, #4\n\t"
  8740. "cmp r7, r8\n\t"
  8741. "blt 1b\n\t"
  8742. : [c] "+r" (c)
  8743. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  8744. : "memory", "r5", "r6", "r7", "r8"
  8745. );
  8746. return c;
  8747. }
  8748. /* Reduce the number back to 3072 bits using Montgomery reduction.
  8749. *
  8750. * a A single precision number to reduce in place.
  8751. * m The single precision number representing the modulus.
  8752. * mp The digit representing the negative inverse of m mod 2^n.
  8753. */
  8754. SP_NOINLINE static void sp_3072_mont_reduce_96(sp_digit* a, const sp_digit* m,
  8755. sp_digit mp)
  8756. {
  8757. sp_digit ca = 0;
  8758. __asm__ __volatile__ (
  8759. "mov r8, %[mp]\n\t"
  8760. "mov r12, %[ca]\n\t"
  8761. "mov r14, %[m]\n\t"
  8762. "mov r9, %[a]\n\t"
  8763. "mov r4, #0\n\t"
  8764. "# i = 0\n\t"
  8765. "mov r11, r4\n\t"
  8766. "\n1:\n\t"
  8767. "mov r5, #0\n\t"
  8768. "mov %[ca], #0\n\t"
  8769. "# mu = a[i] * mp\n\t"
  8770. "mov %[mp], r8\n\t"
  8771. "ldr %[a], [%[a]]\n\t"
  8772. "mul %[mp], %[a]\n\t"
  8773. "mov %[m], r14\n\t"
  8774. "mov r10, r9\n\t"
  8775. "\n2:\n\t"
  8776. "# a[i+j] += m[j] * mu\n\t"
  8777. "mov %[a], r10\n\t"
  8778. "ldr %[a], [%[a]]\n\t"
  8779. "mov %[ca], #0\n\t"
  8780. "mov r4, r5\n\t"
  8781. "mov r5, #0\n\t"
  8782. "# Multiply m[j] and mu - Start\n\t"
  8783. "ldr r7, [%[m]]\n\t"
  8784. "lsl r6, %[mp], #16\n\t"
  8785. "lsl r7, r7, #16\n\t"
  8786. "lsr r6, r6, #16\n\t"
  8787. "lsr r7, r7, #16\n\t"
  8788. "mul r7, r6\n\t"
  8789. "add %[a], r7\n\t"
  8790. "adc r5, %[ca]\n\t"
  8791. "ldr r7, [%[m]]\n\t"
  8792. "lsr r7, r7, #16\n\t"
  8793. "mul r6, r7\n\t"
  8794. "lsr r7, r6, #16\n\t"
  8795. "lsl r6, r6, #16\n\t"
  8796. "add %[a], r6\n\t"
  8797. "adc r5, r7\n\t"
  8798. "ldr r7, [%[m]]\n\t"
  8799. "lsr r6, %[mp], #16\n\t"
  8800. "lsr r7, r7, #16\n\t"
  8801. "mul r7, r6\n\t"
  8802. "add r5, r7\n\t"
  8803. "ldr r7, [%[m]]\n\t"
  8804. "lsl r7, r7, #16\n\t"
  8805. "lsr r7, r7, #16\n\t"
  8806. "mul r6, r7\n\t"
  8807. "lsr r7, r6, #16\n\t"
  8808. "lsl r6, r6, #16\n\t"
  8809. "add %[a], r6\n\t"
  8810. "adc r5, r7\n\t"
  8811. "# Multiply m[j] and mu - Done\n\t"
  8812. "add r4, %[a]\n\t"
  8813. "adc r5, %[ca]\n\t"
  8814. "mov %[a], r10\n\t"
  8815. "str r4, [%[a]]\n\t"
  8816. "mov r6, #4\n\t"
  8817. "add %[m], #4\n\t"
  8818. "add r10, r6\n\t"
  8819. "mov r4, #1\n\t"
  8820. "lsl r4, r4, #8\n\t"
  8821. "add r4, #124\n\t"
  8822. "add r4, r9\n\t"
  8823. "cmp r10, r4\n\t"
  8824. "blt 2b\n\t"
  8825. "# a[i+95] += m[95] * mu\n\t"
  8826. "mov %[ca], #0\n\t"
  8827. "mov r4, r12\n\t"
  8828. "mov %[a], #0\n\t"
  8829. "# Multiply m[95] and mu - Start\n\t"
  8830. "ldr r7, [%[m]]\n\t"
  8831. "lsl r6, %[mp], #16\n\t"
  8832. "lsl r7, r7, #16\n\t"
  8833. "lsr r6, r6, #16\n\t"
  8834. "lsr r7, r7, #16\n\t"
  8835. "mul r7, r6\n\t"
  8836. "add r5, r7\n\t"
  8837. "adc r4, %[ca]\n\t"
  8838. "adc %[a], %[ca]\n\t"
  8839. "ldr r7, [%[m]]\n\t"
  8840. "lsr r7, r7, #16\n\t"
  8841. "mul r6, r7\n\t"
  8842. "lsr r7, r6, #16\n\t"
  8843. "lsl r6, r6, #16\n\t"
  8844. "add r5, r6\n\t"
  8845. "adc r4, r7\n\t"
  8846. "adc %[a], %[ca]\n\t"
  8847. "ldr r7, [%[m]]\n\t"
  8848. "lsr r6, %[mp], #16\n\t"
  8849. "lsr r7, r7, #16\n\t"
  8850. "mul r7, r6\n\t"
  8851. "add r4, r7\n\t"
  8852. "adc %[a], %[ca]\n\t"
  8853. "ldr r7, [%[m]]\n\t"
  8854. "lsl r7, r7, #16\n\t"
  8855. "lsr r7, r7, #16\n\t"
  8856. "mul r6, r7\n\t"
  8857. "lsr r7, r6, #16\n\t"
  8858. "lsl r6, r6, #16\n\t"
  8859. "add r5, r6\n\t"
  8860. "adc r4, r7\n\t"
  8861. "adc %[a], %[ca]\n\t"
  8862. "# Multiply m[95] and mu - Done\n\t"
  8863. "mov %[ca], %[a]\n\t"
  8864. "mov %[a], r10\n\t"
  8865. "ldr r7, [%[a], #4]\n\t"
  8866. "ldr %[a], [%[a]]\n\t"
  8867. "mov r6, #0\n\t"
  8868. "add r5, %[a]\n\t"
  8869. "adc r7, r4\n\t"
  8870. "adc %[ca], r6\n\t"
  8871. "mov %[a], r10\n\t"
  8872. "str r5, [%[a]]\n\t"
  8873. "str r7, [%[a], #4]\n\t"
  8874. "# i += 1\n\t"
  8875. "mov r6, #4\n\t"
  8876. "add r9, r6\n\t"
  8877. "add r11, r6\n\t"
  8878. "mov r12, %[ca]\n\t"
  8879. "mov %[a], r9\n\t"
  8880. "mov r4, #1\n\t"
  8881. "lsl r4, r4, #8\n\t"
  8882. "add r4, #128\n\t"
  8883. "cmp r11, r4\n\t"
  8884. "blt 1b\n\t"
  8885. "mov %[m], r14\n\t"
  8886. : [ca] "+r" (ca), [a] "+r" (a)
  8887. : [m] "r" (m), [mp] "r" (mp)
  8888. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  8889. );
  8890. sp_3072_cond_sub_96(a - 96, a, m, (sp_digit)0 - ca);
  8891. }
  8892. /* Multiply two Montogmery form numbers mod the modulus (prime).
  8893. * (r = a * b mod m)
  8894. *
  8895. * r Result of multiplication.
  8896. * a First number to multiply in Montogmery form.
  8897. * b Second number to multiply in Montogmery form.
  8898. * m Modulus (prime).
  8899. * mp Montogmery mulitplier.
  8900. */
  8901. static void sp_3072_mont_mul_96(sp_digit* r, const sp_digit* a, const sp_digit* b,
  8902. const sp_digit* m, sp_digit mp)
  8903. {
  8904. sp_3072_mul_96(r, a, b);
  8905. sp_3072_mont_reduce_96(r, m, mp);
  8906. }
  8907. /* Square the Montgomery form number. (r = a * a mod m)
  8908. *
  8909. * r Result of squaring.
  8910. * a Number to square in Montogmery form.
  8911. * m Modulus (prime).
  8912. * mp Montogmery mulitplier.
  8913. */
  8914. static void sp_3072_mont_sqr_96(sp_digit* r, const sp_digit* a, const sp_digit* m,
  8915. sp_digit mp)
  8916. {
  8917. sp_3072_sqr_96(r, a);
  8918. sp_3072_mont_reduce_96(r, m, mp);
  8919. }
  8920. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  8921. *
  8922. * d1 The high order half of the number to divide.
  8923. * d0 The low order half of the number to divide.
  8924. * div The dividend.
  8925. * returns the result of the division.
  8926. *
  8927. * Note that this is an approximate div. It may give an answer 1 larger.
  8928. */
  8929. SP_NOINLINE static sp_digit div_3072_word_96(sp_digit d1, sp_digit d0,
  8930. sp_digit div)
  8931. {
  8932. sp_digit r = 0;
  8933. __asm__ __volatile__ (
  8934. "lsr r5, %[div], #1\n\t"
  8935. "add r5, #1\n\t"
  8936. "mov r8, %[d0]\n\t"
  8937. "mov r9, %[d1]\n\t"
  8938. "# Do top 32\n\t"
  8939. "mov r6, r5\n\t"
  8940. "sub r6, %[d1]\n\t"
  8941. "sbc r6, r6\n\t"
  8942. "add %[r], %[r]\n\t"
  8943. "sub %[r], r6\n\t"
  8944. "and r6, r5\n\t"
  8945. "sub %[d1], r6\n\t"
  8946. "# Next 30 bits\n\t"
  8947. "mov r4, #29\n\t"
  8948. "1:\n\t"
  8949. "lsl %[d0], %[d0], #1\n\t"
  8950. "adc %[d1], %[d1]\n\t"
  8951. "mov r6, r5\n\t"
  8952. "sub r6, %[d1]\n\t"
  8953. "sbc r6, r6\n\t"
  8954. "add %[r], %[r]\n\t"
  8955. "sub %[r], r6\n\t"
  8956. "and r6, r5\n\t"
  8957. "sub %[d1], r6\n\t"
  8958. "sub r4, #1\n\t"
  8959. "bpl 1b\n\t"
  8960. "mov r7, #0\n\t"
  8961. "add %[r], %[r]\n\t"
  8962. "add %[r], #1\n\t"
  8963. "# r * div - Start\n\t"
  8964. "lsl %[d1], %[r], #16\n\t"
  8965. "lsl r4, %[div], #16\n\t"
  8966. "lsr %[d1], %[d1], #16\n\t"
  8967. "lsr r4, r4, #16\n\t"
  8968. "mul r4, %[d1]\n\t"
  8969. "lsr r6, %[div], #16\n\t"
  8970. "mul %[d1], r6\n\t"
  8971. "lsr r5, %[d1], #16\n\t"
  8972. "lsl %[d1], %[d1], #16\n\t"
  8973. "add r4, %[d1]\n\t"
  8974. "adc r5, r7\n\t"
  8975. "lsr %[d1], %[r], #16\n\t"
  8976. "mul r6, %[d1]\n\t"
  8977. "add r5, r6\n\t"
  8978. "lsl r6, %[div], #16\n\t"
  8979. "lsr r6, r6, #16\n\t"
  8980. "mul %[d1], r6\n\t"
  8981. "lsr r6, %[d1], #16\n\t"
  8982. "lsl %[d1], %[d1], #16\n\t"
  8983. "add r4, %[d1]\n\t"
  8984. "adc r5, r6\n\t"
  8985. "# r * div - Done\n\t"
  8986. "mov %[d1], r8\n\t"
  8987. "sub %[d1], r4\n\t"
  8988. "mov r4, %[d1]\n\t"
  8989. "mov %[d1], r9\n\t"
  8990. "sbc %[d1], r5\n\t"
  8991. "mov r5, %[d1]\n\t"
  8992. "add %[r], r5\n\t"
  8993. "# r * div - Start\n\t"
  8994. "lsl %[d1], %[r], #16\n\t"
  8995. "lsl r4, %[div], #16\n\t"
  8996. "lsr %[d1], %[d1], #16\n\t"
  8997. "lsr r4, r4, #16\n\t"
  8998. "mul r4, %[d1]\n\t"
  8999. "lsr r6, %[div], #16\n\t"
  9000. "mul %[d1], r6\n\t"
  9001. "lsr r5, %[d1], #16\n\t"
  9002. "lsl %[d1], %[d1], #16\n\t"
  9003. "add r4, %[d1]\n\t"
  9004. "adc r5, r7\n\t"
  9005. "lsr %[d1], %[r], #16\n\t"
  9006. "mul r6, %[d1]\n\t"
  9007. "add r5, r6\n\t"
  9008. "lsl r6, %[div], #16\n\t"
  9009. "lsr r6, r6, #16\n\t"
  9010. "mul %[d1], r6\n\t"
  9011. "lsr r6, %[d1], #16\n\t"
  9012. "lsl %[d1], %[d1], #16\n\t"
  9013. "add r4, %[d1]\n\t"
  9014. "adc r5, r6\n\t"
  9015. "# r * div - Done\n\t"
  9016. "mov %[d1], r8\n\t"
  9017. "mov r6, r9\n\t"
  9018. "sub r4, %[d1], r4\n\t"
  9019. "sbc r6, r5\n\t"
  9020. "mov r5, r6\n\t"
  9021. "add %[r], r5\n\t"
  9022. "# r * div - Start\n\t"
  9023. "lsl %[d1], %[r], #16\n\t"
  9024. "lsl r4, %[div], #16\n\t"
  9025. "lsr %[d1], %[d1], #16\n\t"
  9026. "lsr r4, r4, #16\n\t"
  9027. "mul r4, %[d1]\n\t"
  9028. "lsr r6, %[div], #16\n\t"
  9029. "mul %[d1], r6\n\t"
  9030. "lsr r5, %[d1], #16\n\t"
  9031. "lsl %[d1], %[d1], #16\n\t"
  9032. "add r4, %[d1]\n\t"
  9033. "adc r5, r7\n\t"
  9034. "lsr %[d1], %[r], #16\n\t"
  9035. "mul r6, %[d1]\n\t"
  9036. "add r5, r6\n\t"
  9037. "lsl r6, %[div], #16\n\t"
  9038. "lsr r6, r6, #16\n\t"
  9039. "mul %[d1], r6\n\t"
  9040. "lsr r6, %[d1], #16\n\t"
  9041. "lsl %[d1], %[d1], #16\n\t"
  9042. "add r4, %[d1]\n\t"
  9043. "adc r5, r6\n\t"
  9044. "# r * div - Done\n\t"
  9045. "mov %[d1], r8\n\t"
  9046. "mov r6, r9\n\t"
  9047. "sub r4, %[d1], r4\n\t"
  9048. "sbc r6, r5\n\t"
  9049. "mov r5, r6\n\t"
  9050. "add %[r], r5\n\t"
  9051. "mov r6, %[div]\n\t"
  9052. "sub r6, r4\n\t"
  9053. "sbc r6, r6\n\t"
  9054. "sub %[r], r6\n\t"
  9055. : [r] "+r" (r)
  9056. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  9057. : "r4", "r5", "r7", "r6", "r8", "r9"
  9058. );
  9059. return r;
  9060. }
  9061. /* AND m into each word of a and store in r.
  9062. *
  9063. * r A single precision integer.
  9064. * a A single precision integer.
  9065. * m Mask to AND against each digit.
  9066. */
  9067. static void sp_3072_mask_96(sp_digit* r, const sp_digit* a, sp_digit m)
  9068. {
  9069. #ifdef WOLFSSL_SP_SMALL
  9070. int i;
  9071. for (i=0; i<96; i++) {
  9072. r[i] = a[i] & m;
  9073. }
  9074. #else
  9075. int i;
  9076. for (i = 0; i < 96; i += 8) {
  9077. r[i+0] = a[i+0] & m;
  9078. r[i+1] = a[i+1] & m;
  9079. r[i+2] = a[i+2] & m;
  9080. r[i+3] = a[i+3] & m;
  9081. r[i+4] = a[i+4] & m;
  9082. r[i+5] = a[i+5] & m;
  9083. r[i+6] = a[i+6] & m;
  9084. r[i+7] = a[i+7] & m;
  9085. }
  9086. #endif
  9087. }
  9088. /* Compare a with b in constant time.
  9089. *
  9090. * a A single precision integer.
  9091. * b A single precision integer.
  9092. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  9093. * respectively.
  9094. */
  9095. SP_NOINLINE static int32_t sp_3072_cmp_96(const sp_digit* a, const sp_digit* b)
  9096. {
  9097. sp_digit r = 0;
  9098. __asm__ __volatile__ (
  9099. "mov r3, #0\n\t"
  9100. "mvn r3, r3\n\t"
  9101. "mov r6, #1\n\t"
  9102. "lsl r6, r6, #8\n\t"
  9103. "add r6, #124\n\t"
  9104. "1:\n\t"
  9105. "ldr r7, [%[a], r6]\n\t"
  9106. "ldr r5, [%[b], r6]\n\t"
  9107. "and r7, r3\n\t"
  9108. "and r5, r3\n\t"
  9109. "mov r4, r7\n\t"
  9110. "sub r7, r5\n\t"
  9111. "sbc r7, r7\n\t"
  9112. "add %[r], r7\n\t"
  9113. "mvn r7, r7\n\t"
  9114. "and r3, r7\n\t"
  9115. "sub r5, r4\n\t"
  9116. "sbc r7, r7\n\t"
  9117. "sub %[r], r7\n\t"
  9118. "mvn r7, r7\n\t"
  9119. "and r3, r7\n\t"
  9120. "sub r6, #4\n\t"
  9121. "cmp r6, #0\n\t"
  9122. "bge 1b\n\t"
  9123. : [r] "+r" (r)
  9124. : [a] "r" (a), [b] "r" (b)
  9125. : "r3", "r4", "r5", "r6", "r7"
  9126. );
  9127. return r;
  9128. }
  9129. /* Divide d in a and put remainder into r (m*d + r = a)
  9130. * m is not calculated as it is not needed at this time.
  9131. *
  9132. * a Nmber to be divided.
  9133. * d Number to divide with.
  9134. * m Multiplier result.
  9135. * r Remainder from the division.
  9136. * returns MP_OKAY indicating success.
  9137. */
  9138. static WC_INLINE int sp_3072_div_96(const sp_digit* a, const sp_digit* d, sp_digit* m,
  9139. sp_digit* r)
  9140. {
  9141. sp_digit t1[192], t2[97];
  9142. sp_digit div, r1;
  9143. int i;
  9144. (void)m;
  9145. div = d[95];
  9146. XMEMCPY(t1, a, sizeof(*t1) * 2 * 96);
  9147. for (i=95; i>=0; i--) {
  9148. r1 = div_3072_word_96(t1[96 + i], t1[96 + i - 1], div);
  9149. sp_3072_mul_d_96(t2, d, r1);
  9150. t1[96 + i] += sp_3072_sub_in_place_96(&t1[i], t2);
  9151. t1[96 + i] -= t2[96];
  9152. sp_3072_mask_96(t2, d, t1[96 + i]);
  9153. t1[96 + i] += sp_3072_add_96(&t1[i], &t1[i], t2);
  9154. sp_3072_mask_96(t2, d, t1[96 + i]);
  9155. t1[96 + i] += sp_3072_add_96(&t1[i], &t1[i], t2);
  9156. }
  9157. r1 = sp_3072_cmp_96(t1, d) >= 0;
  9158. sp_3072_cond_sub_96(r, t1, d, (sp_digit)0 - r1);
  9159. return MP_OKAY;
  9160. }
  9161. /* Reduce a modulo m into r. (r = a mod m)
  9162. *
  9163. * r A single precision number that is the reduced result.
  9164. * a A single precision number that is to be reduced.
  9165. * m A single precision number that is the modulus to reduce with.
  9166. * returns MP_OKAY indicating success.
  9167. */
  9168. static WC_INLINE int sp_3072_mod_96(sp_digit* r, const sp_digit* a, const sp_digit* m)
  9169. {
  9170. return sp_3072_div_96(a, m, NULL, r);
  9171. }
  9172. /* Divide d in a and put remainder into r (m*d + r = a)
  9173. * m is not calculated as it is not needed at this time.
  9174. *
  9175. * a Nmber to be divided.
  9176. * d Number to divide with.
  9177. * m Multiplier result.
  9178. * r Remainder from the division.
  9179. * returns MP_OKAY indicating success.
  9180. */
  9181. static WC_INLINE int sp_3072_div_96_cond(const sp_digit* a, const sp_digit* d, sp_digit* m,
  9182. sp_digit* r)
  9183. {
  9184. sp_digit t1[192], t2[97];
  9185. sp_digit div, r1;
  9186. int i;
  9187. (void)m;
  9188. div = d[95];
  9189. XMEMCPY(t1, a, sizeof(*t1) * 2 * 96);
  9190. for (i=95; i>=0; i--) {
  9191. r1 = div_3072_word_96(t1[96 + i], t1[96 + i - 1], div);
  9192. sp_3072_mul_d_96(t2, d, r1);
  9193. t1[96 + i] += sp_3072_sub_in_place_96(&t1[i], t2);
  9194. t1[96 + i] -= t2[96];
  9195. if (t1[96 + i] != 0) {
  9196. t1[96 + i] += sp_3072_add_96(&t1[i], &t1[i], d);
  9197. if (t1[96 + i] != 0)
  9198. t1[96 + i] += sp_3072_add_96(&t1[i], &t1[i], d);
  9199. }
  9200. }
  9201. r1 = sp_3072_cmp_96(t1, d) >= 0;
  9202. sp_3072_cond_sub_96(r, t1, d, (sp_digit)0 - r1);
  9203. return MP_OKAY;
  9204. }
  9205. /* Reduce a modulo m into r. (r = a mod m)
  9206. *
  9207. * r A single precision number that is the reduced result.
  9208. * a A single precision number that is to be reduced.
  9209. * m A single precision number that is the modulus to reduce with.
  9210. * returns MP_OKAY indicating success.
  9211. */
  9212. static WC_INLINE int sp_3072_mod_96_cond(sp_digit* r, const sp_digit* a, const sp_digit* m)
  9213. {
  9214. return sp_3072_div_96_cond(a, m, NULL, r);
  9215. }
  9216. #if (defined(WOLFSSL_HAVE_SP_RSA) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)) || \
  9217. defined(WOLFSSL_HAVE_SP_DH)
  9218. #ifdef WOLFSSL_SP_SMALL
  9219. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  9220. *
  9221. * r A single precision number that is the result of the operation.
  9222. * a A single precision number being exponentiated.
  9223. * e A single precision number that is the exponent.
  9224. * bits The number of bits in the exponent.
  9225. * m A single precision number that is the modulus.
  9226. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  9227. */
  9228. static int sp_3072_mod_exp_96(sp_digit* r, const sp_digit* a, const sp_digit* e,
  9229. int bits, const sp_digit* m, int reduceA)
  9230. {
  9231. #ifndef WOLFSSL_SMALL_STACK
  9232. sp_digit t[16][192];
  9233. #else
  9234. sp_digit* t[16];
  9235. sp_digit* td;
  9236. #endif
  9237. sp_digit* norm;
  9238. sp_digit mp = 1;
  9239. sp_digit n;
  9240. sp_digit mask;
  9241. int i;
  9242. int c, y;
  9243. int err = MP_OKAY;
  9244. #ifdef WOLFSSL_SMALL_STACK
  9245. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 192, NULL,
  9246. DYNAMIC_TYPE_TMP_BUFFER);
  9247. if (td == NULL) {
  9248. err = MEMORY_E;
  9249. }
  9250. #endif
  9251. if (err == MP_OKAY) {
  9252. #ifdef WOLFSSL_SMALL_STACK
  9253. for (i=0; i<16; i++) {
  9254. t[i] = td + i * 192;
  9255. }
  9256. #endif
  9257. norm = t[0];
  9258. sp_3072_mont_setup(m, &mp);
  9259. sp_3072_mont_norm_96(norm, m);
  9260. XMEMSET(t[1], 0, sizeof(sp_digit) * 96U);
  9261. if (reduceA != 0) {
  9262. err = sp_3072_mod_96(t[1] + 96, a, m);
  9263. if (err == MP_OKAY) {
  9264. err = sp_3072_mod_96(t[1], t[1], m);
  9265. }
  9266. }
  9267. else {
  9268. XMEMCPY(t[1] + 96, a, sizeof(sp_digit) * 96);
  9269. err = sp_3072_mod_96(t[1], t[1], m);
  9270. }
  9271. }
  9272. if (err == MP_OKAY) {
  9273. sp_3072_mont_sqr_96(t[ 2], t[ 1], m, mp);
  9274. sp_3072_mont_mul_96(t[ 3], t[ 2], t[ 1], m, mp);
  9275. sp_3072_mont_sqr_96(t[ 4], t[ 2], m, mp);
  9276. sp_3072_mont_mul_96(t[ 5], t[ 3], t[ 2], m, mp);
  9277. sp_3072_mont_sqr_96(t[ 6], t[ 3], m, mp);
  9278. sp_3072_mont_mul_96(t[ 7], t[ 4], t[ 3], m, mp);
  9279. sp_3072_mont_sqr_96(t[ 8], t[ 4], m, mp);
  9280. sp_3072_mont_mul_96(t[ 9], t[ 5], t[ 4], m, mp);
  9281. sp_3072_mont_sqr_96(t[10], t[ 5], m, mp);
  9282. sp_3072_mont_mul_96(t[11], t[ 6], t[ 5], m, mp);
  9283. sp_3072_mont_sqr_96(t[12], t[ 6], m, mp);
  9284. sp_3072_mont_mul_96(t[13], t[ 7], t[ 6], m, mp);
  9285. sp_3072_mont_sqr_96(t[14], t[ 7], m, mp);
  9286. sp_3072_mont_mul_96(t[15], t[ 8], t[ 7], m, mp);
  9287. i = (bits - 1) / 32;
  9288. n = e[i--];
  9289. c = bits & 31;
  9290. if (c == 0) {
  9291. c = 32;
  9292. }
  9293. c -= bits % 4;
  9294. if (c == 32) {
  9295. c = 28;
  9296. }
  9297. y = (int)(n >> c);
  9298. n <<= 32 - c;
  9299. XMEMCPY(r, t[y], sizeof(sp_digit) * 96);
  9300. for (; i>=0 || c>=4; ) {
  9301. if (c == 0) {
  9302. n = e[i--];
  9303. y = n >> 28;
  9304. n <<= 4;
  9305. c = 28;
  9306. }
  9307. else if (c < 4) {
  9308. y = n >> 28;
  9309. n = e[i--];
  9310. c = 4 - c;
  9311. y |= n >> (32 - c);
  9312. n <<= c;
  9313. c = 32 - c;
  9314. }
  9315. else {
  9316. y = (n >> 28) & 0xf;
  9317. n <<= 4;
  9318. c -= 4;
  9319. }
  9320. sp_3072_mont_sqr_96(r, r, m, mp);
  9321. sp_3072_mont_sqr_96(r, r, m, mp);
  9322. sp_3072_mont_sqr_96(r, r, m, mp);
  9323. sp_3072_mont_sqr_96(r, r, m, mp);
  9324. sp_3072_mont_mul_96(r, r, t[y], m, mp);
  9325. }
  9326. XMEMSET(&r[96], 0, sizeof(sp_digit) * 96U);
  9327. sp_3072_mont_reduce_96(r, m, mp);
  9328. mask = 0 - (sp_3072_cmp_96(r, m) >= 0);
  9329. sp_3072_cond_sub_96(r, r, m, mask);
  9330. }
  9331. #ifdef WOLFSSL_SMALL_STACK
  9332. if (td != NULL) {
  9333. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  9334. }
  9335. #endif
  9336. return err;
  9337. }
  9338. #else
  9339. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  9340. *
  9341. * r A single precision number that is the result of the operation.
  9342. * a A single precision number being exponentiated.
  9343. * e A single precision number that is the exponent.
  9344. * bits The number of bits in the exponent.
  9345. * m A single precision number that is the modulus.
  9346. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  9347. */
  9348. static int sp_3072_mod_exp_96(sp_digit* r, const sp_digit* a, const sp_digit* e,
  9349. int bits, const sp_digit* m, int reduceA)
  9350. {
  9351. #ifndef WOLFSSL_SMALL_STACK
  9352. sp_digit t[32][192];
  9353. #else
  9354. sp_digit* t[32];
  9355. sp_digit* td;
  9356. #endif
  9357. sp_digit* norm;
  9358. sp_digit mp = 1;
  9359. sp_digit n;
  9360. sp_digit mask;
  9361. int i;
  9362. int c, y;
  9363. int err = MP_OKAY;
  9364. #ifdef WOLFSSL_SMALL_STACK
  9365. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 192, NULL,
  9366. DYNAMIC_TYPE_TMP_BUFFER);
  9367. if (td == NULL) {
  9368. err = MEMORY_E;
  9369. }
  9370. #endif
  9371. if (err == MP_OKAY) {
  9372. #ifdef WOLFSSL_SMALL_STACK
  9373. for (i=0; i<32; i++) {
  9374. t[i] = td + i * 192;
  9375. }
  9376. #endif
  9377. norm = t[0];
  9378. sp_3072_mont_setup(m, &mp);
  9379. sp_3072_mont_norm_96(norm, m);
  9380. XMEMSET(t[1], 0, sizeof(sp_digit) * 96U);
  9381. if (reduceA != 0) {
  9382. err = sp_3072_mod_96(t[1] + 96, a, m);
  9383. if (err == MP_OKAY) {
  9384. err = sp_3072_mod_96(t[1], t[1], m);
  9385. }
  9386. }
  9387. else {
  9388. XMEMCPY(t[1] + 96, a, sizeof(sp_digit) * 96);
  9389. err = sp_3072_mod_96(t[1], t[1], m);
  9390. }
  9391. }
  9392. if (err == MP_OKAY) {
  9393. sp_3072_mont_sqr_96(t[ 2], t[ 1], m, mp);
  9394. sp_3072_mont_mul_96(t[ 3], t[ 2], t[ 1], m, mp);
  9395. sp_3072_mont_sqr_96(t[ 4], t[ 2], m, mp);
  9396. sp_3072_mont_mul_96(t[ 5], t[ 3], t[ 2], m, mp);
  9397. sp_3072_mont_sqr_96(t[ 6], t[ 3], m, mp);
  9398. sp_3072_mont_mul_96(t[ 7], t[ 4], t[ 3], m, mp);
  9399. sp_3072_mont_sqr_96(t[ 8], t[ 4], m, mp);
  9400. sp_3072_mont_mul_96(t[ 9], t[ 5], t[ 4], m, mp);
  9401. sp_3072_mont_sqr_96(t[10], t[ 5], m, mp);
  9402. sp_3072_mont_mul_96(t[11], t[ 6], t[ 5], m, mp);
  9403. sp_3072_mont_sqr_96(t[12], t[ 6], m, mp);
  9404. sp_3072_mont_mul_96(t[13], t[ 7], t[ 6], m, mp);
  9405. sp_3072_mont_sqr_96(t[14], t[ 7], m, mp);
  9406. sp_3072_mont_mul_96(t[15], t[ 8], t[ 7], m, mp);
  9407. sp_3072_mont_sqr_96(t[16], t[ 8], m, mp);
  9408. sp_3072_mont_mul_96(t[17], t[ 9], t[ 8], m, mp);
  9409. sp_3072_mont_sqr_96(t[18], t[ 9], m, mp);
  9410. sp_3072_mont_mul_96(t[19], t[10], t[ 9], m, mp);
  9411. sp_3072_mont_sqr_96(t[20], t[10], m, mp);
  9412. sp_3072_mont_mul_96(t[21], t[11], t[10], m, mp);
  9413. sp_3072_mont_sqr_96(t[22], t[11], m, mp);
  9414. sp_3072_mont_mul_96(t[23], t[12], t[11], m, mp);
  9415. sp_3072_mont_sqr_96(t[24], t[12], m, mp);
  9416. sp_3072_mont_mul_96(t[25], t[13], t[12], m, mp);
  9417. sp_3072_mont_sqr_96(t[26], t[13], m, mp);
  9418. sp_3072_mont_mul_96(t[27], t[14], t[13], m, mp);
  9419. sp_3072_mont_sqr_96(t[28], t[14], m, mp);
  9420. sp_3072_mont_mul_96(t[29], t[15], t[14], m, mp);
  9421. sp_3072_mont_sqr_96(t[30], t[15], m, mp);
  9422. sp_3072_mont_mul_96(t[31], t[16], t[15], m, mp);
  9423. i = (bits - 1) / 32;
  9424. n = e[i--];
  9425. c = bits & 31;
  9426. if (c == 0) {
  9427. c = 32;
  9428. }
  9429. c -= bits % 5;
  9430. if (c == 32) {
  9431. c = 27;
  9432. }
  9433. y = (int)(n >> c);
  9434. n <<= 32 - c;
  9435. XMEMCPY(r, t[y], sizeof(sp_digit) * 96);
  9436. for (; i>=0 || c>=5; ) {
  9437. if (c == 0) {
  9438. n = e[i--];
  9439. y = n >> 27;
  9440. n <<= 5;
  9441. c = 27;
  9442. }
  9443. else if (c < 5) {
  9444. y = n >> 27;
  9445. n = e[i--];
  9446. c = 5 - c;
  9447. y |= n >> (32 - c);
  9448. n <<= c;
  9449. c = 32 - c;
  9450. }
  9451. else {
  9452. y = (n >> 27) & 0x1f;
  9453. n <<= 5;
  9454. c -= 5;
  9455. }
  9456. sp_3072_mont_sqr_96(r, r, m, mp);
  9457. sp_3072_mont_sqr_96(r, r, m, mp);
  9458. sp_3072_mont_sqr_96(r, r, m, mp);
  9459. sp_3072_mont_sqr_96(r, r, m, mp);
  9460. sp_3072_mont_sqr_96(r, r, m, mp);
  9461. sp_3072_mont_mul_96(r, r, t[y], m, mp);
  9462. }
  9463. XMEMSET(&r[96], 0, sizeof(sp_digit) * 96U);
  9464. sp_3072_mont_reduce_96(r, m, mp);
  9465. mask = 0 - (sp_3072_cmp_96(r, m) >= 0);
  9466. sp_3072_cond_sub_96(r, r, m, mask);
  9467. }
  9468. #ifdef WOLFSSL_SMALL_STACK
  9469. if (td != NULL) {
  9470. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  9471. }
  9472. #endif
  9473. return err;
  9474. }
  9475. #endif /* WOLFSSL_SP_SMALL */
  9476. #endif /* (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) || WOLFSSL_HAVE_SP_DH */
  9477. #ifdef WOLFSSL_HAVE_SP_RSA
  9478. /* RSA public key operation.
  9479. *
  9480. * in Array of bytes representing the number to exponentiate, base.
  9481. * inLen Number of bytes in base.
  9482. * em Public exponent.
  9483. * mm Modulus.
  9484. * out Buffer to hold big-endian bytes of exponentiation result.
  9485. * Must be at least 384 bytes long.
  9486. * outLen Number of bytes in result.
  9487. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  9488. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  9489. */
  9490. int sp_RsaPublic_3072(const byte* in, word32 inLen, mp_int* em, mp_int* mm,
  9491. byte* out, word32* outLen)
  9492. {
  9493. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  9494. sp_digit ad[192], md[96], rd[192];
  9495. #else
  9496. sp_digit* d = NULL;
  9497. #endif
  9498. sp_digit* a;
  9499. sp_digit *ah;
  9500. sp_digit* m;
  9501. sp_digit* r;
  9502. sp_digit e[1];
  9503. int err = MP_OKAY;
  9504. if (*outLen < 384)
  9505. err = MP_TO_E;
  9506. if (err == MP_OKAY && (mp_count_bits(em) > 32 || inLen > 384 ||
  9507. mp_count_bits(mm) != 3072))
  9508. err = MP_READ_E;
  9509. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  9510. if (err == MP_OKAY) {
  9511. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 96 * 5, NULL,
  9512. DYNAMIC_TYPE_RSA);
  9513. if (d == NULL)
  9514. err = MEMORY_E;
  9515. }
  9516. if (err == MP_OKAY) {
  9517. a = d;
  9518. r = a + 96 * 2;
  9519. m = r + 96 * 2;
  9520. ah = a + 96;
  9521. }
  9522. #else
  9523. a = ad;
  9524. m = md;
  9525. r = rd;
  9526. ah = a + 96;
  9527. #endif
  9528. if (err == MP_OKAY) {
  9529. sp_3072_from_bin(ah, 96, in, inLen);
  9530. #if DIGIT_BIT >= 32
  9531. e[0] = em->dp[0];
  9532. #else
  9533. e[0] = em->dp[0];
  9534. if (em->used > 1)
  9535. e[0] |= ((sp_digit)em->dp[1]) << DIGIT_BIT;
  9536. #endif
  9537. if (e[0] == 0)
  9538. err = MP_EXPTMOD_E;
  9539. }
  9540. if (err == MP_OKAY) {
  9541. sp_3072_from_mp(m, 96, mm);
  9542. if (e[0] == 0x3) {
  9543. if (err == MP_OKAY) {
  9544. sp_3072_sqr_96(r, ah);
  9545. err = sp_3072_mod_96_cond(r, r, m);
  9546. }
  9547. if (err == MP_OKAY) {
  9548. sp_3072_mul_96(r, ah, r);
  9549. err = sp_3072_mod_96_cond(r, r, m);
  9550. }
  9551. }
  9552. else {
  9553. int i;
  9554. sp_digit mp;
  9555. sp_3072_mont_setup(m, &mp);
  9556. /* Convert to Montgomery form. */
  9557. XMEMSET(a, 0, sizeof(sp_digit) * 96);
  9558. err = sp_3072_mod_96_cond(a, a, m);
  9559. if (err == MP_OKAY) {
  9560. for (i=31; i>=0; i--)
  9561. if (e[0] >> i)
  9562. break;
  9563. XMEMCPY(r, a, sizeof(sp_digit) * 96);
  9564. for (i--; i>=0; i--) {
  9565. sp_3072_mont_sqr_96(r, r, m, mp);
  9566. if (((e[0] >> i) & 1) == 1)
  9567. sp_3072_mont_mul_96(r, r, a, m, mp);
  9568. }
  9569. XMEMSET(&r[96], 0, sizeof(sp_digit) * 96);
  9570. sp_3072_mont_reduce_96(r, m, mp);
  9571. for (i = 95; i > 0; i--) {
  9572. if (r[i] != m[i])
  9573. break;
  9574. }
  9575. if (r[i] >= m[i])
  9576. sp_3072_sub_in_place_96(r, m);
  9577. }
  9578. }
  9579. }
  9580. if (err == MP_OKAY) {
  9581. sp_3072_to_bin(r, out);
  9582. *outLen = 384;
  9583. }
  9584. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  9585. if (d != NULL)
  9586. XFREE(d, NULL, DYNAMIC_TYPE_RSA);
  9587. #endif
  9588. return err;
  9589. }
  9590. /* RSA private key operation.
  9591. *
  9592. * in Array of bytes representing the number to exponentiate, base.
  9593. * inLen Number of bytes in base.
  9594. * dm Private exponent.
  9595. * pm First prime.
  9596. * qm Second prime.
  9597. * dpm First prime's CRT exponent.
  9598. * dqm Second prime's CRT exponent.
  9599. * qim Inverse of second prime mod p.
  9600. * mm Modulus.
  9601. * out Buffer to hold big-endian bytes of exponentiation result.
  9602. * Must be at least 384 bytes long.
  9603. * outLen Number of bytes in result.
  9604. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  9605. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  9606. */
  9607. int sp_RsaPrivate_3072(const byte* in, word32 inLen, mp_int* dm,
  9608. mp_int* pm, mp_int* qm, mp_int* dpm, mp_int* dqm, mp_int* qim, mp_int* mm,
  9609. byte* out, word32* outLen)
  9610. {
  9611. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  9612. sp_digit ad[96 * 2];
  9613. sp_digit pd[48], qd[48], dpd[48];
  9614. sp_digit tmpad[96], tmpbd[96];
  9615. #else
  9616. sp_digit* t = NULL;
  9617. #endif
  9618. sp_digit* a;
  9619. sp_digit* p;
  9620. sp_digit* q;
  9621. sp_digit* dp;
  9622. sp_digit* dq;
  9623. sp_digit* qi;
  9624. sp_digit* tmp;
  9625. sp_digit* tmpa;
  9626. sp_digit* tmpb;
  9627. sp_digit* r;
  9628. sp_digit c;
  9629. int err = MP_OKAY;
  9630. (void)dm;
  9631. (void)mm;
  9632. if (*outLen < 384)
  9633. err = MP_TO_E;
  9634. if (err == MP_OKAY && (inLen > 384 || mp_count_bits(mm) != 3072))
  9635. err = MP_READ_E;
  9636. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  9637. if (err == MP_OKAY) {
  9638. t = (sp_digit*)XMALLOC(sizeof(sp_digit) * 48 * 11, NULL,
  9639. DYNAMIC_TYPE_RSA);
  9640. if (t == NULL)
  9641. err = MEMORY_E;
  9642. }
  9643. if (err == MP_OKAY) {
  9644. a = t;
  9645. p = a + 96 * 2;
  9646. q = p + 48;
  9647. qi = dq = dp = q + 48;
  9648. tmpa = qi + 48;
  9649. tmpb = tmpa + 96;
  9650. tmp = t;
  9651. r = tmp + 96;
  9652. }
  9653. #else
  9654. r = a = ad;
  9655. p = pd;
  9656. q = qd;
  9657. qi = dq = dp = dpd;
  9658. tmpa = tmpad;
  9659. tmpb = tmpbd;
  9660. tmp = a + 96;
  9661. #endif
  9662. if (err == MP_OKAY) {
  9663. sp_3072_from_bin(a, 96, in, inLen);
  9664. sp_3072_from_mp(p, 48, pm);
  9665. sp_3072_from_mp(q, 48, qm);
  9666. sp_3072_from_mp(dp, 48, dpm);
  9667. err = sp_3072_mod_exp_48(tmpa, a, dp, 1536, p, 1);
  9668. }
  9669. if (err == MP_OKAY) {
  9670. sp_3072_from_mp(dq, 48, dqm);
  9671. err = sp_3072_mod_exp_48(tmpb, a, dq, 1536, q, 1);
  9672. }
  9673. if (err == MP_OKAY) {
  9674. c = sp_3072_sub_in_place_48(tmpa, tmpb);
  9675. sp_3072_mask_48(tmp, p, c);
  9676. sp_3072_add_48(tmpa, tmpa, tmp);
  9677. sp_3072_from_mp(qi, 48, qim);
  9678. sp_3072_mul_48(tmpa, tmpa, qi);
  9679. err = sp_3072_mod_48(tmpa, tmpa, p);
  9680. }
  9681. if (err == MP_OKAY) {
  9682. sp_3072_mul_48(tmpa, q, tmpa);
  9683. XMEMSET(&tmpb[48], 0, sizeof(sp_digit) * 48);
  9684. sp_3072_add_96(r, tmpb, tmpa);
  9685. sp_3072_to_bin(r, out);
  9686. *outLen = 384;
  9687. }
  9688. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  9689. if (t != NULL) {
  9690. XMEMSET(t, 0, sizeof(sp_digit) * 48 * 11);
  9691. XFREE(t, NULL, DYNAMIC_TYPE_RSA);
  9692. }
  9693. #else
  9694. XMEMSET(tmpad, 0, sizeof(tmpad));
  9695. XMEMSET(tmpbd, 0, sizeof(tmpbd));
  9696. XMEMSET(pd, 0, sizeof(pd));
  9697. XMEMSET(qd, 0, sizeof(qd));
  9698. XMEMSET(dpd, 0, sizeof(dpd));
  9699. #endif
  9700. return err;
  9701. }
  9702. #endif /* WOLFSSL_HAVE_SP_RSA */
  9703. #if defined(WOLFSSL_HAVE_SP_DH) || (defined(WOLFSSL_HAVE_SP_RSA) && \
  9704. !defined(WOLFSSL_RSA_PUBLIC_ONLY))
  9705. /* Convert an array of sp_digit to an mp_int.
  9706. *
  9707. * a A single precision integer.
  9708. * r A multi-precision integer.
  9709. */
  9710. static int sp_3072_to_mp(const sp_digit* a, mp_int* r)
  9711. {
  9712. int err;
  9713. err = mp_grow(r, (3072 + DIGIT_BIT - 1) / DIGIT_BIT);
  9714. if (err == MP_OKAY) { /*lint !e774 case where err is always MP_OKAY*/
  9715. #if DIGIT_BIT == 32
  9716. XMEMCPY(r->dp, a, sizeof(sp_digit) * 96);
  9717. r->used = 96;
  9718. mp_clamp(r);
  9719. #elif DIGIT_BIT < 32
  9720. int i, j = 0, s = 0;
  9721. r->dp[0] = 0;
  9722. for (i = 0; i < 96; i++) {
  9723. r->dp[j] |= a[i] << s;
  9724. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  9725. s = DIGIT_BIT - s;
  9726. r->dp[++j] = a[i] >> s;
  9727. while (s + DIGIT_BIT <= 32) {
  9728. s += DIGIT_BIT;
  9729. r->dp[j++] &= (1L << DIGIT_BIT) - 1;
  9730. if (s == SP_WORD_SIZE) {
  9731. r->dp[j] = 0;
  9732. }
  9733. else {
  9734. r->dp[j] = a[i] >> s;
  9735. }
  9736. }
  9737. s = 32 - s;
  9738. }
  9739. r->used = (3072 + DIGIT_BIT - 1) / DIGIT_BIT;
  9740. mp_clamp(r);
  9741. #else
  9742. int i, j = 0, s = 0;
  9743. r->dp[0] = 0;
  9744. for (i = 0; i < 96; i++) {
  9745. r->dp[j] |= ((mp_digit)a[i]) << s;
  9746. if (s + 32 >= DIGIT_BIT) {
  9747. #if DIGIT_BIT != 32 && DIGIT_BIT != 64
  9748. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  9749. #endif
  9750. s = DIGIT_BIT - s;
  9751. r->dp[++j] = a[i] >> s;
  9752. s = 32 - s;
  9753. }
  9754. else {
  9755. s += 32;
  9756. }
  9757. }
  9758. r->used = (3072 + DIGIT_BIT - 1) / DIGIT_BIT;
  9759. mp_clamp(r);
  9760. #endif
  9761. }
  9762. return err;
  9763. }
  9764. /* Perform the modular exponentiation for Diffie-Hellman.
  9765. *
  9766. * base Base. MP integer.
  9767. * exp Exponent. MP integer.
  9768. * mod Modulus. MP integer.
  9769. * res Result. MP integer.
  9770. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  9771. * and MEMORY_E if memory allocation fails.
  9772. */
  9773. int sp_ModExp_3072(mp_int* base, mp_int* exp, mp_int* mod, mp_int* res)
  9774. {
  9775. int err = MP_OKAY;
  9776. sp_digit b[192], e[96], m[96];
  9777. sp_digit* r = b;
  9778. int expBits = mp_count_bits(exp);
  9779. if (mp_count_bits(base) > 3072) {
  9780. err = MP_READ_E;
  9781. }
  9782. if (err == MP_OKAY) {
  9783. if (expBits > 3072) {
  9784. err = MP_READ_E;
  9785. }
  9786. }
  9787. if (err == MP_OKAY) {
  9788. if (mp_count_bits(mod) != 3072) {
  9789. err = MP_READ_E;
  9790. }
  9791. }
  9792. if (err == MP_OKAY) {
  9793. sp_3072_from_mp(b, 96, base);
  9794. sp_3072_from_mp(e, 96, exp);
  9795. sp_3072_from_mp(m, 96, mod);
  9796. err = sp_3072_mod_exp_96(r, b, e, expBits, m, 0);
  9797. }
  9798. if (err == MP_OKAY) {
  9799. err = sp_3072_to_mp(r, res);
  9800. }
  9801. XMEMSET(e, 0, sizeof(e));
  9802. return err;
  9803. }
  9804. #ifdef WOLFSSL_HAVE_SP_DH
  9805. #ifdef HAVE_FFDHE_3072
  9806. static void sp_3072_lshift_96(sp_digit* r, sp_digit* a, byte n)
  9807. {
  9808. __asm__ __volatile__ (
  9809. "mov r6, #31\n\t"
  9810. "sub r6, r6, %[n]\n\t"
  9811. "add %[a], %[a], #255\n\t"
  9812. "add %[r], %[r], #255\n\t"
  9813. "add %[a], %[a], #65\n\t"
  9814. "add %[r], %[r], #65\n\t"
  9815. "ldr r3, [%[a], #60]\n\t"
  9816. "lsr r4, r3, #1\n\t"
  9817. "lsl r3, r3, %[n]\n\t"
  9818. "lsr r4, r4, r6\n\t"
  9819. "ldr r2, [%[a], #56]\n\t"
  9820. "str r4, [%[r], #64]\n\t"
  9821. "lsr r5, r2, #1\n\t"
  9822. "lsl r2, r2, %[n]\n\t"
  9823. "lsr r5, r5, r6\n\t"
  9824. "orr r3, r3, r5\n\t"
  9825. "ldr r4, [%[a], #52]\n\t"
  9826. "str r3, [%[r], #60]\n\t"
  9827. "lsr r5, r4, #1\n\t"
  9828. "lsl r4, r4, %[n]\n\t"
  9829. "lsr r5, r5, r6\n\t"
  9830. "orr r2, r2, r5\n\t"
  9831. "ldr r3, [%[a], #48]\n\t"
  9832. "str r2, [%[r], #56]\n\t"
  9833. "lsr r5, r3, #1\n\t"
  9834. "lsl r3, r3, %[n]\n\t"
  9835. "lsr r5, r5, r6\n\t"
  9836. "orr r4, r4, r5\n\t"
  9837. "ldr r2, [%[a], #44]\n\t"
  9838. "str r4, [%[r], #52]\n\t"
  9839. "lsr r5, r2, #1\n\t"
  9840. "lsl r2, r2, %[n]\n\t"
  9841. "lsr r5, r5, r6\n\t"
  9842. "orr r3, r3, r5\n\t"
  9843. "ldr r4, [%[a], #40]\n\t"
  9844. "str r3, [%[r], #48]\n\t"
  9845. "lsr r5, r4, #1\n\t"
  9846. "lsl r4, r4, %[n]\n\t"
  9847. "lsr r5, r5, r6\n\t"
  9848. "orr r2, r2, r5\n\t"
  9849. "ldr r3, [%[a], #36]\n\t"
  9850. "str r2, [%[r], #44]\n\t"
  9851. "lsr r5, r3, #1\n\t"
  9852. "lsl r3, r3, %[n]\n\t"
  9853. "lsr r5, r5, r6\n\t"
  9854. "orr r4, r4, r5\n\t"
  9855. "ldr r2, [%[a], #32]\n\t"
  9856. "str r4, [%[r], #40]\n\t"
  9857. "lsr r5, r2, #1\n\t"
  9858. "lsl r2, r2, %[n]\n\t"
  9859. "lsr r5, r5, r6\n\t"
  9860. "orr r3, r3, r5\n\t"
  9861. "ldr r4, [%[a], #28]\n\t"
  9862. "str r3, [%[r], #36]\n\t"
  9863. "lsr r5, r4, #1\n\t"
  9864. "lsl r4, r4, %[n]\n\t"
  9865. "lsr r5, r5, r6\n\t"
  9866. "orr r2, r2, r5\n\t"
  9867. "ldr r3, [%[a], #24]\n\t"
  9868. "str r2, [%[r], #32]\n\t"
  9869. "lsr r5, r3, #1\n\t"
  9870. "lsl r3, r3, %[n]\n\t"
  9871. "lsr r5, r5, r6\n\t"
  9872. "orr r4, r4, r5\n\t"
  9873. "ldr r2, [%[a], #20]\n\t"
  9874. "str r4, [%[r], #28]\n\t"
  9875. "lsr r5, r2, #1\n\t"
  9876. "lsl r2, r2, %[n]\n\t"
  9877. "lsr r5, r5, r6\n\t"
  9878. "orr r3, r3, r5\n\t"
  9879. "ldr r4, [%[a], #16]\n\t"
  9880. "str r3, [%[r], #24]\n\t"
  9881. "lsr r5, r4, #1\n\t"
  9882. "lsl r4, r4, %[n]\n\t"
  9883. "lsr r5, r5, r6\n\t"
  9884. "orr r2, r2, r5\n\t"
  9885. "ldr r3, [%[a], #12]\n\t"
  9886. "str r2, [%[r], #20]\n\t"
  9887. "lsr r5, r3, #1\n\t"
  9888. "lsl r3, r3, %[n]\n\t"
  9889. "lsr r5, r5, r6\n\t"
  9890. "orr r4, r4, r5\n\t"
  9891. "ldr r2, [%[a], #8]\n\t"
  9892. "str r4, [%[r], #16]\n\t"
  9893. "lsr r5, r2, #1\n\t"
  9894. "lsl r2, r2, %[n]\n\t"
  9895. "lsr r5, r5, r6\n\t"
  9896. "orr r3, r3, r5\n\t"
  9897. "ldr r4, [%[a], #4]\n\t"
  9898. "str r3, [%[r], #12]\n\t"
  9899. "lsr r5, r4, #1\n\t"
  9900. "lsl r4, r4, %[n]\n\t"
  9901. "lsr r5, r5, r6\n\t"
  9902. "orr r2, r2, r5\n\t"
  9903. "ldr r3, [%[a], #0]\n\t"
  9904. "str r2, [%[r], #8]\n\t"
  9905. "lsr r5, r3, #1\n\t"
  9906. "lsl r3, r3, %[n]\n\t"
  9907. "lsr r5, r5, r6\n\t"
  9908. "orr r4, r4, r5\n\t"
  9909. "sub %[a], %[a], #64\n\t"
  9910. "sub %[r], %[r], #64\n\t"
  9911. "ldr r2, [%[a], #60]\n\t"
  9912. "str r4, [%[r], #68]\n\t"
  9913. "lsr r5, r2, #1\n\t"
  9914. "lsl r2, r2, %[n]\n\t"
  9915. "lsr r5, r5, r6\n\t"
  9916. "orr r3, r3, r5\n\t"
  9917. "ldr r4, [%[a], #56]\n\t"
  9918. "str r3, [%[r], #64]\n\t"
  9919. "lsr r5, r4, #1\n\t"
  9920. "lsl r4, r4, %[n]\n\t"
  9921. "lsr r5, r5, r6\n\t"
  9922. "orr r2, r2, r5\n\t"
  9923. "ldr r3, [%[a], #52]\n\t"
  9924. "str r2, [%[r], #60]\n\t"
  9925. "lsr r5, r3, #1\n\t"
  9926. "lsl r3, r3, %[n]\n\t"
  9927. "lsr r5, r5, r6\n\t"
  9928. "orr r4, r4, r5\n\t"
  9929. "ldr r2, [%[a], #48]\n\t"
  9930. "str r4, [%[r], #56]\n\t"
  9931. "lsr r5, r2, #1\n\t"
  9932. "lsl r2, r2, %[n]\n\t"
  9933. "lsr r5, r5, r6\n\t"
  9934. "orr r3, r3, r5\n\t"
  9935. "ldr r4, [%[a], #44]\n\t"
  9936. "str r3, [%[r], #52]\n\t"
  9937. "lsr r5, r4, #1\n\t"
  9938. "lsl r4, r4, %[n]\n\t"
  9939. "lsr r5, r5, r6\n\t"
  9940. "orr r2, r2, r5\n\t"
  9941. "ldr r3, [%[a], #40]\n\t"
  9942. "str r2, [%[r], #48]\n\t"
  9943. "lsr r5, r3, #1\n\t"
  9944. "lsl r3, r3, %[n]\n\t"
  9945. "lsr r5, r5, r6\n\t"
  9946. "orr r4, r4, r5\n\t"
  9947. "ldr r2, [%[a], #36]\n\t"
  9948. "str r4, [%[r], #44]\n\t"
  9949. "lsr r5, r2, #1\n\t"
  9950. "lsl r2, r2, %[n]\n\t"
  9951. "lsr r5, r5, r6\n\t"
  9952. "orr r3, r3, r5\n\t"
  9953. "ldr r4, [%[a], #32]\n\t"
  9954. "str r3, [%[r], #40]\n\t"
  9955. "lsr r5, r4, #1\n\t"
  9956. "lsl r4, r4, %[n]\n\t"
  9957. "lsr r5, r5, r6\n\t"
  9958. "orr r2, r2, r5\n\t"
  9959. "ldr r3, [%[a], #28]\n\t"
  9960. "str r2, [%[r], #36]\n\t"
  9961. "lsr r5, r3, #1\n\t"
  9962. "lsl r3, r3, %[n]\n\t"
  9963. "lsr r5, r5, r6\n\t"
  9964. "orr r4, r4, r5\n\t"
  9965. "ldr r2, [%[a], #24]\n\t"
  9966. "str r4, [%[r], #32]\n\t"
  9967. "lsr r5, r2, #1\n\t"
  9968. "lsl r2, r2, %[n]\n\t"
  9969. "lsr r5, r5, r6\n\t"
  9970. "orr r3, r3, r5\n\t"
  9971. "ldr r4, [%[a], #20]\n\t"
  9972. "str r3, [%[r], #28]\n\t"
  9973. "lsr r5, r4, #1\n\t"
  9974. "lsl r4, r4, %[n]\n\t"
  9975. "lsr r5, r5, r6\n\t"
  9976. "orr r2, r2, r5\n\t"
  9977. "ldr r3, [%[a], #16]\n\t"
  9978. "str r2, [%[r], #24]\n\t"
  9979. "lsr r5, r3, #1\n\t"
  9980. "lsl r3, r3, %[n]\n\t"
  9981. "lsr r5, r5, r6\n\t"
  9982. "orr r4, r4, r5\n\t"
  9983. "ldr r2, [%[a], #12]\n\t"
  9984. "str r4, [%[r], #20]\n\t"
  9985. "lsr r5, r2, #1\n\t"
  9986. "lsl r2, r2, %[n]\n\t"
  9987. "lsr r5, r5, r6\n\t"
  9988. "orr r3, r3, r5\n\t"
  9989. "ldr r4, [%[a], #8]\n\t"
  9990. "str r3, [%[r], #16]\n\t"
  9991. "lsr r5, r4, #1\n\t"
  9992. "lsl r4, r4, %[n]\n\t"
  9993. "lsr r5, r5, r6\n\t"
  9994. "orr r2, r2, r5\n\t"
  9995. "ldr r3, [%[a], #4]\n\t"
  9996. "str r2, [%[r], #12]\n\t"
  9997. "lsr r5, r3, #1\n\t"
  9998. "lsl r3, r3, %[n]\n\t"
  9999. "lsr r5, r5, r6\n\t"
  10000. "orr r4, r4, r5\n\t"
  10001. "ldr r2, [%[a], #0]\n\t"
  10002. "str r4, [%[r], #8]\n\t"
  10003. "lsr r5, r2, #1\n\t"
  10004. "lsl r2, r2, %[n]\n\t"
  10005. "lsr r5, r5, r6\n\t"
  10006. "orr r3, r3, r5\n\t"
  10007. "sub %[a], %[a], #64\n\t"
  10008. "sub %[r], %[r], #64\n\t"
  10009. "ldr r4, [%[a], #60]\n\t"
  10010. "str r3, [%[r], #68]\n\t"
  10011. "lsr r5, r4, #1\n\t"
  10012. "lsl r4, r4, %[n]\n\t"
  10013. "lsr r5, r5, r6\n\t"
  10014. "orr r2, r2, r5\n\t"
  10015. "ldr r3, [%[a], #56]\n\t"
  10016. "str r2, [%[r], #64]\n\t"
  10017. "lsr r5, r3, #1\n\t"
  10018. "lsl r3, r3, %[n]\n\t"
  10019. "lsr r5, r5, r6\n\t"
  10020. "orr r4, r4, r5\n\t"
  10021. "ldr r2, [%[a], #52]\n\t"
  10022. "str r4, [%[r], #60]\n\t"
  10023. "lsr r5, r2, #1\n\t"
  10024. "lsl r2, r2, %[n]\n\t"
  10025. "lsr r5, r5, r6\n\t"
  10026. "orr r3, r3, r5\n\t"
  10027. "ldr r4, [%[a], #48]\n\t"
  10028. "str r3, [%[r], #56]\n\t"
  10029. "lsr r5, r4, #1\n\t"
  10030. "lsl r4, r4, %[n]\n\t"
  10031. "lsr r5, r5, r6\n\t"
  10032. "orr r2, r2, r5\n\t"
  10033. "ldr r3, [%[a], #44]\n\t"
  10034. "str r2, [%[r], #52]\n\t"
  10035. "lsr r5, r3, #1\n\t"
  10036. "lsl r3, r3, %[n]\n\t"
  10037. "lsr r5, r5, r6\n\t"
  10038. "orr r4, r4, r5\n\t"
  10039. "ldr r2, [%[a], #40]\n\t"
  10040. "str r4, [%[r], #48]\n\t"
  10041. "lsr r5, r2, #1\n\t"
  10042. "lsl r2, r2, %[n]\n\t"
  10043. "lsr r5, r5, r6\n\t"
  10044. "orr r3, r3, r5\n\t"
  10045. "ldr r4, [%[a], #36]\n\t"
  10046. "str r3, [%[r], #44]\n\t"
  10047. "lsr r5, r4, #1\n\t"
  10048. "lsl r4, r4, %[n]\n\t"
  10049. "lsr r5, r5, r6\n\t"
  10050. "orr r2, r2, r5\n\t"
  10051. "ldr r3, [%[a], #32]\n\t"
  10052. "str r2, [%[r], #40]\n\t"
  10053. "lsr r5, r3, #1\n\t"
  10054. "lsl r3, r3, %[n]\n\t"
  10055. "lsr r5, r5, r6\n\t"
  10056. "orr r4, r4, r5\n\t"
  10057. "ldr r2, [%[a], #28]\n\t"
  10058. "str r4, [%[r], #36]\n\t"
  10059. "lsr r5, r2, #1\n\t"
  10060. "lsl r2, r2, %[n]\n\t"
  10061. "lsr r5, r5, r6\n\t"
  10062. "orr r3, r3, r5\n\t"
  10063. "ldr r4, [%[a], #24]\n\t"
  10064. "str r3, [%[r], #32]\n\t"
  10065. "lsr r5, r4, #1\n\t"
  10066. "lsl r4, r4, %[n]\n\t"
  10067. "lsr r5, r5, r6\n\t"
  10068. "orr r2, r2, r5\n\t"
  10069. "ldr r3, [%[a], #20]\n\t"
  10070. "str r2, [%[r], #28]\n\t"
  10071. "lsr r5, r3, #1\n\t"
  10072. "lsl r3, r3, %[n]\n\t"
  10073. "lsr r5, r5, r6\n\t"
  10074. "orr r4, r4, r5\n\t"
  10075. "ldr r2, [%[a], #16]\n\t"
  10076. "str r4, [%[r], #24]\n\t"
  10077. "lsr r5, r2, #1\n\t"
  10078. "lsl r2, r2, %[n]\n\t"
  10079. "lsr r5, r5, r6\n\t"
  10080. "orr r3, r3, r5\n\t"
  10081. "ldr r4, [%[a], #12]\n\t"
  10082. "str r3, [%[r], #20]\n\t"
  10083. "lsr r5, r4, #1\n\t"
  10084. "lsl r4, r4, %[n]\n\t"
  10085. "lsr r5, r5, r6\n\t"
  10086. "orr r2, r2, r5\n\t"
  10087. "ldr r3, [%[a], #8]\n\t"
  10088. "str r2, [%[r], #16]\n\t"
  10089. "lsr r5, r3, #1\n\t"
  10090. "lsl r3, r3, %[n]\n\t"
  10091. "lsr r5, r5, r6\n\t"
  10092. "orr r4, r4, r5\n\t"
  10093. "ldr r2, [%[a], #4]\n\t"
  10094. "str r4, [%[r], #12]\n\t"
  10095. "lsr r5, r2, #1\n\t"
  10096. "lsl r2, r2, %[n]\n\t"
  10097. "lsr r5, r5, r6\n\t"
  10098. "orr r3, r3, r5\n\t"
  10099. "ldr r4, [%[a], #0]\n\t"
  10100. "str r3, [%[r], #8]\n\t"
  10101. "lsr r5, r4, #1\n\t"
  10102. "lsl r4, r4, %[n]\n\t"
  10103. "lsr r5, r5, r6\n\t"
  10104. "orr r2, r2, r5\n\t"
  10105. "sub %[a], %[a], #64\n\t"
  10106. "sub %[r], %[r], #64\n\t"
  10107. "ldr r3, [%[a], #60]\n\t"
  10108. "str r2, [%[r], #68]\n\t"
  10109. "lsr r5, r3, #1\n\t"
  10110. "lsl r3, r3, %[n]\n\t"
  10111. "lsr r5, r5, r6\n\t"
  10112. "orr r4, r4, r5\n\t"
  10113. "ldr r2, [%[a], #56]\n\t"
  10114. "str r4, [%[r], #64]\n\t"
  10115. "lsr r5, r2, #1\n\t"
  10116. "lsl r2, r2, %[n]\n\t"
  10117. "lsr r5, r5, r6\n\t"
  10118. "orr r3, r3, r5\n\t"
  10119. "ldr r4, [%[a], #52]\n\t"
  10120. "str r3, [%[r], #60]\n\t"
  10121. "lsr r5, r4, #1\n\t"
  10122. "lsl r4, r4, %[n]\n\t"
  10123. "lsr r5, r5, r6\n\t"
  10124. "orr r2, r2, r5\n\t"
  10125. "ldr r3, [%[a], #48]\n\t"
  10126. "str r2, [%[r], #56]\n\t"
  10127. "lsr r5, r3, #1\n\t"
  10128. "lsl r3, r3, %[n]\n\t"
  10129. "lsr r5, r5, r6\n\t"
  10130. "orr r4, r4, r5\n\t"
  10131. "ldr r2, [%[a], #44]\n\t"
  10132. "str r4, [%[r], #52]\n\t"
  10133. "lsr r5, r2, #1\n\t"
  10134. "lsl r2, r2, %[n]\n\t"
  10135. "lsr r5, r5, r6\n\t"
  10136. "orr r3, r3, r5\n\t"
  10137. "ldr r4, [%[a], #40]\n\t"
  10138. "str r3, [%[r], #48]\n\t"
  10139. "lsr r5, r4, #1\n\t"
  10140. "lsl r4, r4, %[n]\n\t"
  10141. "lsr r5, r5, r6\n\t"
  10142. "orr r2, r2, r5\n\t"
  10143. "ldr r3, [%[a], #36]\n\t"
  10144. "str r2, [%[r], #44]\n\t"
  10145. "lsr r5, r3, #1\n\t"
  10146. "lsl r3, r3, %[n]\n\t"
  10147. "lsr r5, r5, r6\n\t"
  10148. "orr r4, r4, r5\n\t"
  10149. "ldr r2, [%[a], #32]\n\t"
  10150. "str r4, [%[r], #40]\n\t"
  10151. "lsr r5, r2, #1\n\t"
  10152. "lsl r2, r2, %[n]\n\t"
  10153. "lsr r5, r5, r6\n\t"
  10154. "orr r3, r3, r5\n\t"
  10155. "ldr r4, [%[a], #28]\n\t"
  10156. "str r3, [%[r], #36]\n\t"
  10157. "lsr r5, r4, #1\n\t"
  10158. "lsl r4, r4, %[n]\n\t"
  10159. "lsr r5, r5, r6\n\t"
  10160. "orr r2, r2, r5\n\t"
  10161. "ldr r3, [%[a], #24]\n\t"
  10162. "str r2, [%[r], #32]\n\t"
  10163. "lsr r5, r3, #1\n\t"
  10164. "lsl r3, r3, %[n]\n\t"
  10165. "lsr r5, r5, r6\n\t"
  10166. "orr r4, r4, r5\n\t"
  10167. "ldr r2, [%[a], #20]\n\t"
  10168. "str r4, [%[r], #28]\n\t"
  10169. "lsr r5, r2, #1\n\t"
  10170. "lsl r2, r2, %[n]\n\t"
  10171. "lsr r5, r5, r6\n\t"
  10172. "orr r3, r3, r5\n\t"
  10173. "ldr r4, [%[a], #16]\n\t"
  10174. "str r3, [%[r], #24]\n\t"
  10175. "lsr r5, r4, #1\n\t"
  10176. "lsl r4, r4, %[n]\n\t"
  10177. "lsr r5, r5, r6\n\t"
  10178. "orr r2, r2, r5\n\t"
  10179. "ldr r3, [%[a], #12]\n\t"
  10180. "str r2, [%[r], #20]\n\t"
  10181. "lsr r5, r3, #1\n\t"
  10182. "lsl r3, r3, %[n]\n\t"
  10183. "lsr r5, r5, r6\n\t"
  10184. "orr r4, r4, r5\n\t"
  10185. "ldr r2, [%[a], #8]\n\t"
  10186. "str r4, [%[r], #16]\n\t"
  10187. "lsr r5, r2, #1\n\t"
  10188. "lsl r2, r2, %[n]\n\t"
  10189. "lsr r5, r5, r6\n\t"
  10190. "orr r3, r3, r5\n\t"
  10191. "ldr r4, [%[a], #4]\n\t"
  10192. "str r3, [%[r], #12]\n\t"
  10193. "lsr r5, r4, #1\n\t"
  10194. "lsl r4, r4, %[n]\n\t"
  10195. "lsr r5, r5, r6\n\t"
  10196. "orr r2, r2, r5\n\t"
  10197. "ldr r3, [%[a], #0]\n\t"
  10198. "str r2, [%[r], #8]\n\t"
  10199. "lsr r5, r3, #1\n\t"
  10200. "lsl r3, r3, %[n]\n\t"
  10201. "lsr r5, r5, r6\n\t"
  10202. "orr r4, r4, r5\n\t"
  10203. "sub %[a], %[a], #64\n\t"
  10204. "sub %[r], %[r], #64\n\t"
  10205. "ldr r2, [%[a], #60]\n\t"
  10206. "str r4, [%[r], #68]\n\t"
  10207. "lsr r5, r2, #1\n\t"
  10208. "lsl r2, r2, %[n]\n\t"
  10209. "lsr r5, r5, r6\n\t"
  10210. "orr r3, r3, r5\n\t"
  10211. "ldr r4, [%[a], #56]\n\t"
  10212. "str r3, [%[r], #64]\n\t"
  10213. "lsr r5, r4, #1\n\t"
  10214. "lsl r4, r4, %[n]\n\t"
  10215. "lsr r5, r5, r6\n\t"
  10216. "orr r2, r2, r5\n\t"
  10217. "ldr r3, [%[a], #52]\n\t"
  10218. "str r2, [%[r], #60]\n\t"
  10219. "lsr r5, r3, #1\n\t"
  10220. "lsl r3, r3, %[n]\n\t"
  10221. "lsr r5, r5, r6\n\t"
  10222. "orr r4, r4, r5\n\t"
  10223. "ldr r2, [%[a], #48]\n\t"
  10224. "str r4, [%[r], #56]\n\t"
  10225. "lsr r5, r2, #1\n\t"
  10226. "lsl r2, r2, %[n]\n\t"
  10227. "lsr r5, r5, r6\n\t"
  10228. "orr r3, r3, r5\n\t"
  10229. "ldr r4, [%[a], #44]\n\t"
  10230. "str r3, [%[r], #52]\n\t"
  10231. "lsr r5, r4, #1\n\t"
  10232. "lsl r4, r4, %[n]\n\t"
  10233. "lsr r5, r5, r6\n\t"
  10234. "orr r2, r2, r5\n\t"
  10235. "ldr r3, [%[a], #40]\n\t"
  10236. "str r2, [%[r], #48]\n\t"
  10237. "lsr r5, r3, #1\n\t"
  10238. "lsl r3, r3, %[n]\n\t"
  10239. "lsr r5, r5, r6\n\t"
  10240. "orr r4, r4, r5\n\t"
  10241. "ldr r2, [%[a], #36]\n\t"
  10242. "str r4, [%[r], #44]\n\t"
  10243. "lsr r5, r2, #1\n\t"
  10244. "lsl r2, r2, %[n]\n\t"
  10245. "lsr r5, r5, r6\n\t"
  10246. "orr r3, r3, r5\n\t"
  10247. "ldr r4, [%[a], #32]\n\t"
  10248. "str r3, [%[r], #40]\n\t"
  10249. "lsr r5, r4, #1\n\t"
  10250. "lsl r4, r4, %[n]\n\t"
  10251. "lsr r5, r5, r6\n\t"
  10252. "orr r2, r2, r5\n\t"
  10253. "ldr r3, [%[a], #28]\n\t"
  10254. "str r2, [%[r], #36]\n\t"
  10255. "lsr r5, r3, #1\n\t"
  10256. "lsl r3, r3, %[n]\n\t"
  10257. "lsr r5, r5, r6\n\t"
  10258. "orr r4, r4, r5\n\t"
  10259. "ldr r2, [%[a], #24]\n\t"
  10260. "str r4, [%[r], #32]\n\t"
  10261. "lsr r5, r2, #1\n\t"
  10262. "lsl r2, r2, %[n]\n\t"
  10263. "lsr r5, r5, r6\n\t"
  10264. "orr r3, r3, r5\n\t"
  10265. "ldr r4, [%[a], #20]\n\t"
  10266. "str r3, [%[r], #28]\n\t"
  10267. "lsr r5, r4, #1\n\t"
  10268. "lsl r4, r4, %[n]\n\t"
  10269. "lsr r5, r5, r6\n\t"
  10270. "orr r2, r2, r5\n\t"
  10271. "ldr r3, [%[a], #16]\n\t"
  10272. "str r2, [%[r], #24]\n\t"
  10273. "lsr r5, r3, #1\n\t"
  10274. "lsl r3, r3, %[n]\n\t"
  10275. "lsr r5, r5, r6\n\t"
  10276. "orr r4, r4, r5\n\t"
  10277. "ldr r2, [%[a], #12]\n\t"
  10278. "str r4, [%[r], #20]\n\t"
  10279. "lsr r5, r2, #1\n\t"
  10280. "lsl r2, r2, %[n]\n\t"
  10281. "lsr r5, r5, r6\n\t"
  10282. "orr r3, r3, r5\n\t"
  10283. "ldr r4, [%[a], #8]\n\t"
  10284. "str r3, [%[r], #16]\n\t"
  10285. "lsr r5, r4, #1\n\t"
  10286. "lsl r4, r4, %[n]\n\t"
  10287. "lsr r5, r5, r6\n\t"
  10288. "orr r2, r2, r5\n\t"
  10289. "ldr r3, [%[a], #4]\n\t"
  10290. "str r2, [%[r], #12]\n\t"
  10291. "lsr r5, r3, #1\n\t"
  10292. "lsl r3, r3, %[n]\n\t"
  10293. "lsr r5, r5, r6\n\t"
  10294. "orr r4, r4, r5\n\t"
  10295. "ldr r2, [%[a], #0]\n\t"
  10296. "str r4, [%[r], #8]\n\t"
  10297. "lsr r5, r2, #1\n\t"
  10298. "lsl r2, r2, %[n]\n\t"
  10299. "lsr r5, r5, r6\n\t"
  10300. "orr r3, r3, r5\n\t"
  10301. "sub %[a], %[a], #64\n\t"
  10302. "sub %[r], %[r], #64\n\t"
  10303. "ldr r4, [%[a], #60]\n\t"
  10304. "str r3, [%[r], #68]\n\t"
  10305. "lsr r5, r4, #1\n\t"
  10306. "lsl r4, r4, %[n]\n\t"
  10307. "lsr r5, r5, r6\n\t"
  10308. "orr r2, r2, r5\n\t"
  10309. "ldr r3, [%[a], #56]\n\t"
  10310. "str r2, [%[r], #64]\n\t"
  10311. "lsr r5, r3, #1\n\t"
  10312. "lsl r3, r3, %[n]\n\t"
  10313. "lsr r5, r5, r6\n\t"
  10314. "orr r4, r4, r5\n\t"
  10315. "ldr r2, [%[a], #52]\n\t"
  10316. "str r4, [%[r], #60]\n\t"
  10317. "lsr r5, r2, #1\n\t"
  10318. "lsl r2, r2, %[n]\n\t"
  10319. "lsr r5, r5, r6\n\t"
  10320. "orr r3, r3, r5\n\t"
  10321. "ldr r4, [%[a], #48]\n\t"
  10322. "str r3, [%[r], #56]\n\t"
  10323. "lsr r5, r4, #1\n\t"
  10324. "lsl r4, r4, %[n]\n\t"
  10325. "lsr r5, r5, r6\n\t"
  10326. "orr r2, r2, r5\n\t"
  10327. "ldr r3, [%[a], #44]\n\t"
  10328. "str r2, [%[r], #52]\n\t"
  10329. "lsr r5, r3, #1\n\t"
  10330. "lsl r3, r3, %[n]\n\t"
  10331. "lsr r5, r5, r6\n\t"
  10332. "orr r4, r4, r5\n\t"
  10333. "ldr r2, [%[a], #40]\n\t"
  10334. "str r4, [%[r], #48]\n\t"
  10335. "lsr r5, r2, #1\n\t"
  10336. "lsl r2, r2, %[n]\n\t"
  10337. "lsr r5, r5, r6\n\t"
  10338. "orr r3, r3, r5\n\t"
  10339. "ldr r4, [%[a], #36]\n\t"
  10340. "str r3, [%[r], #44]\n\t"
  10341. "lsr r5, r4, #1\n\t"
  10342. "lsl r4, r4, %[n]\n\t"
  10343. "lsr r5, r5, r6\n\t"
  10344. "orr r2, r2, r5\n\t"
  10345. "ldr r3, [%[a], #32]\n\t"
  10346. "str r2, [%[r], #40]\n\t"
  10347. "lsr r5, r3, #1\n\t"
  10348. "lsl r3, r3, %[n]\n\t"
  10349. "lsr r5, r5, r6\n\t"
  10350. "orr r4, r4, r5\n\t"
  10351. "ldr r2, [%[a], #28]\n\t"
  10352. "str r4, [%[r], #36]\n\t"
  10353. "lsr r5, r2, #1\n\t"
  10354. "lsl r2, r2, %[n]\n\t"
  10355. "lsr r5, r5, r6\n\t"
  10356. "orr r3, r3, r5\n\t"
  10357. "ldr r4, [%[a], #24]\n\t"
  10358. "str r3, [%[r], #32]\n\t"
  10359. "lsr r5, r4, #1\n\t"
  10360. "lsl r4, r4, %[n]\n\t"
  10361. "lsr r5, r5, r6\n\t"
  10362. "orr r2, r2, r5\n\t"
  10363. "ldr r3, [%[a], #20]\n\t"
  10364. "str r2, [%[r], #28]\n\t"
  10365. "lsr r5, r3, #1\n\t"
  10366. "lsl r3, r3, %[n]\n\t"
  10367. "lsr r5, r5, r6\n\t"
  10368. "orr r4, r4, r5\n\t"
  10369. "ldr r2, [%[a], #16]\n\t"
  10370. "str r4, [%[r], #24]\n\t"
  10371. "lsr r5, r2, #1\n\t"
  10372. "lsl r2, r2, %[n]\n\t"
  10373. "lsr r5, r5, r6\n\t"
  10374. "orr r3, r3, r5\n\t"
  10375. "ldr r4, [%[a], #12]\n\t"
  10376. "str r3, [%[r], #20]\n\t"
  10377. "lsr r5, r4, #1\n\t"
  10378. "lsl r4, r4, %[n]\n\t"
  10379. "lsr r5, r5, r6\n\t"
  10380. "orr r2, r2, r5\n\t"
  10381. "ldr r3, [%[a], #8]\n\t"
  10382. "str r2, [%[r], #16]\n\t"
  10383. "lsr r5, r3, #1\n\t"
  10384. "lsl r3, r3, %[n]\n\t"
  10385. "lsr r5, r5, r6\n\t"
  10386. "orr r4, r4, r5\n\t"
  10387. "ldr r2, [%[a], #4]\n\t"
  10388. "str r4, [%[r], #12]\n\t"
  10389. "lsr r5, r2, #1\n\t"
  10390. "lsl r2, r2, %[n]\n\t"
  10391. "lsr r5, r5, r6\n\t"
  10392. "orr r3, r3, r5\n\t"
  10393. "ldr r4, [%[a], #0]\n\t"
  10394. "str r3, [%[r], #8]\n\t"
  10395. "lsr r5, r4, #1\n\t"
  10396. "lsl r4, r4, %[n]\n\t"
  10397. "lsr r5, r5, r6\n\t"
  10398. "orr r2, r2, r5\n\t"
  10399. "str r4, [%[r]]\n\t"
  10400. "str r2, [%[r], #4]\n\t"
  10401. :
  10402. : [r] "r" (r), [a] "r" (a), [n] "r" (n)
  10403. : "memory", "r2", "r3", "r4", "r5", "r6"
  10404. );
  10405. }
  10406. /* Modular exponentiate 2 to the e mod m. (r = 2^e mod m)
  10407. *
  10408. * r A single precision number that is the result of the operation.
  10409. * e A single precision number that is the exponent.
  10410. * bits The number of bits in the exponent.
  10411. * m A single precision number that is the modulus.
  10412. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  10413. */
  10414. static int sp_3072_mod_exp_2_96(sp_digit* r, const sp_digit* e, int bits,
  10415. const sp_digit* m)
  10416. {
  10417. #ifndef WOLFSSL_SMALL_STACK
  10418. sp_digit nd[192];
  10419. sp_digit td[97];
  10420. #else
  10421. sp_digit* td;
  10422. #endif
  10423. sp_digit* norm;
  10424. sp_digit* tmp;
  10425. sp_digit mp = 1;
  10426. sp_digit n, o;
  10427. sp_digit mask;
  10428. int i;
  10429. int c, y;
  10430. int err = MP_OKAY;
  10431. #ifdef WOLFSSL_SMALL_STACK
  10432. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 289, NULL,
  10433. DYNAMIC_TYPE_TMP_BUFFER);
  10434. if (td == NULL) {
  10435. err = MEMORY_E;
  10436. }
  10437. #endif
  10438. if (err == MP_OKAY) {
  10439. #ifdef WOLFSSL_SMALL_STACK
  10440. norm = td;
  10441. tmp = td + 192;
  10442. #else
  10443. norm = nd;
  10444. tmp = td;
  10445. #endif
  10446. sp_3072_mont_setup(m, &mp);
  10447. sp_3072_mont_norm_96(norm, m);
  10448. i = (bits - 1) / 32;
  10449. n = e[i--];
  10450. c = bits & 31;
  10451. if (c == 0) {
  10452. c = 32;
  10453. }
  10454. c -= bits % 5;
  10455. if (c == 32) {
  10456. c = 27;
  10457. }
  10458. y = (int)(n >> c);
  10459. n <<= 32 - c;
  10460. sp_3072_lshift_96(r, norm, y);
  10461. for (; i>=0 || c>=5; ) {
  10462. if (c == 0) {
  10463. n = e[i--];
  10464. y = n >> 27;
  10465. n <<= 5;
  10466. c = 27;
  10467. }
  10468. else if (c < 5) {
  10469. y = n >> 27;
  10470. n = e[i--];
  10471. c = 5 - c;
  10472. y |= n >> (32 - c);
  10473. n <<= c;
  10474. c = 32 - c;
  10475. }
  10476. else {
  10477. y = (n >> 27) & 0x1f;
  10478. n <<= 5;
  10479. c -= 5;
  10480. }
  10481. sp_3072_mont_sqr_96(r, r, m, mp);
  10482. sp_3072_mont_sqr_96(r, r, m, mp);
  10483. sp_3072_mont_sqr_96(r, r, m, mp);
  10484. sp_3072_mont_sqr_96(r, r, m, mp);
  10485. sp_3072_mont_sqr_96(r, r, m, mp);
  10486. sp_3072_lshift_96(r, r, y);
  10487. sp_3072_mul_d_96(tmp, norm, r[96]);
  10488. r[96] = 0;
  10489. o = sp_3072_add_96(r, r, tmp);
  10490. sp_3072_cond_sub_96(r, r, m, (sp_digit)0 - o);
  10491. }
  10492. XMEMSET(&r[96], 0, sizeof(sp_digit) * 96U);
  10493. sp_3072_mont_reduce_96(r, m, mp);
  10494. mask = 0 - (sp_3072_cmp_96(r, m) >= 0);
  10495. sp_3072_cond_sub_96(r, r, m, mask);
  10496. }
  10497. #ifdef WOLFSSL_SMALL_STACK
  10498. if (td != NULL) {
  10499. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  10500. }
  10501. #endif
  10502. return err;
  10503. }
  10504. #endif /* HAVE_FFDHE_3072 */
  10505. /* Perform the modular exponentiation for Diffie-Hellman.
  10506. *
  10507. * base Base.
  10508. * exp Array of bytes that is the exponent.
  10509. * expLen Length of data, in bytes, in exponent.
  10510. * mod Modulus.
  10511. * out Buffer to hold big-endian bytes of exponentiation result.
  10512. * Must be at least 384 bytes long.
  10513. * outLen Length, in bytes, of exponentiation result.
  10514. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  10515. * and MEMORY_E if memory allocation fails.
  10516. */
  10517. int sp_DhExp_3072(mp_int* base, const byte* exp, word32 expLen,
  10518. mp_int* mod, byte* out, word32* outLen)
  10519. {
  10520. int err = MP_OKAY;
  10521. sp_digit b[192], e[96], m[96];
  10522. sp_digit* r = b;
  10523. word32 i;
  10524. if (mp_count_bits(base) > 3072) {
  10525. err = MP_READ_E;
  10526. }
  10527. if (err == MP_OKAY) {
  10528. if (expLen > 384) {
  10529. err = MP_READ_E;
  10530. }
  10531. }
  10532. if (err == MP_OKAY) {
  10533. if (mp_count_bits(mod) != 3072) {
  10534. err = MP_READ_E;
  10535. }
  10536. }
  10537. if (err == MP_OKAY) {
  10538. sp_3072_from_mp(b, 96, base);
  10539. sp_3072_from_bin(e, 96, exp, expLen);
  10540. sp_3072_from_mp(m, 96, mod);
  10541. #ifdef HAVE_FFDHE_3072
  10542. if (base->used == 1 && base->dp[0] == 2 && m[95] == (sp_digit)-1)
  10543. err = sp_3072_mod_exp_2_96(r, e, expLen * 8, m);
  10544. else
  10545. #endif
  10546. err = sp_3072_mod_exp_96(r, b, e, expLen * 8, m, 0);
  10547. }
  10548. if (err == MP_OKAY) {
  10549. sp_3072_to_bin(r, out);
  10550. *outLen = 384;
  10551. for (i=0; i<384 && out[i] == 0; i++) {
  10552. }
  10553. *outLen -= i;
  10554. XMEMMOVE(out, out + i, *outLen);
  10555. }
  10556. XMEMSET(e, 0, sizeof(e));
  10557. return err;
  10558. }
  10559. #endif /* WOLFSSL_HAVE_SP_DH */
  10560. /* Perform the modular exponentiation for Diffie-Hellman.
  10561. *
  10562. * base Base. MP integer.
  10563. * exp Exponent. MP integer.
  10564. * mod Modulus. MP integer.
  10565. * res Result. MP integer.
  10566. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  10567. * and MEMORY_E if memory allocation fails.
  10568. */
  10569. int sp_ModExp_1536(mp_int* base, mp_int* exp, mp_int* mod, mp_int* res)
  10570. {
  10571. int err = MP_OKAY;
  10572. sp_digit b[96], e[48], m[48];
  10573. sp_digit* r = b;
  10574. int expBits = mp_count_bits(exp);
  10575. if (mp_count_bits(base) > 1536) {
  10576. err = MP_READ_E;
  10577. }
  10578. if (err == MP_OKAY) {
  10579. if (expBits > 1536) {
  10580. err = MP_READ_E;
  10581. }
  10582. }
  10583. if (err == MP_OKAY) {
  10584. if (mp_count_bits(mod) != 1536) {
  10585. err = MP_READ_E;
  10586. }
  10587. }
  10588. if (err == MP_OKAY) {
  10589. sp_3072_from_mp(b, 48, base);
  10590. sp_3072_from_mp(e, 48, exp);
  10591. sp_3072_from_mp(m, 48, mod);
  10592. err = sp_3072_mod_exp_48(r, b, e, expBits, m, 0);
  10593. }
  10594. if (err == MP_OKAY) {
  10595. XMEMSET(r + 48, 0, sizeof(*r) * 48U);
  10596. err = sp_3072_to_mp(r, res);
  10597. res->used = mod->used;
  10598. mp_clamp(res);
  10599. }
  10600. XMEMSET(e, 0, sizeof(e));
  10601. return err;
  10602. }
  10603. #endif /* WOLFSSL_HAVE_SP_DH || (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) */
  10604. #endif /* !WOLFSSL_SP_NO_3072 */
  10605. #ifdef WOLFSSL_SP_4096
  10606. /* Read big endian unsigned byte array into r.
  10607. *
  10608. * r A single precision integer.
  10609. * size Maximum number of bytes to convert
  10610. * a Byte array.
  10611. * n Number of bytes in array to read.
  10612. */
  10613. static void sp_4096_from_bin(sp_digit* r, int size, const byte* a, int n)
  10614. {
  10615. int i, j = 0;
  10616. word32 s = 0;
  10617. r[0] = 0;
  10618. for (i = n-1; i >= 0; i--) {
  10619. r[j] |= (((sp_digit)a[i]) << s);
  10620. if (s >= 24U) {
  10621. r[j] &= 0xffffffff;
  10622. s = 32U - s;
  10623. if (j + 1 >= size) {
  10624. break;
  10625. }
  10626. r[++j] = (sp_digit)a[i] >> s;
  10627. s = 8U - s;
  10628. }
  10629. else {
  10630. s += 8U;
  10631. }
  10632. }
  10633. for (j++; j < size; j++) {
  10634. r[j] = 0;
  10635. }
  10636. }
  10637. /* Convert an mp_int to an array of sp_digit.
  10638. *
  10639. * r A single precision integer.
  10640. * size Maximum number of bytes to convert
  10641. * a A multi-precision integer.
  10642. */
  10643. static void sp_4096_from_mp(sp_digit* r, int size, const mp_int* a)
  10644. {
  10645. #if DIGIT_BIT == 32
  10646. int j;
  10647. XMEMCPY(r, a->dp, sizeof(sp_digit) * a->used);
  10648. for (j = a->used; j < size; j++) {
  10649. r[j] = 0;
  10650. }
  10651. #elif DIGIT_BIT > 32
  10652. int i, j = 0;
  10653. word32 s = 0;
  10654. r[0] = 0;
  10655. for (i = 0; i < a->used && j < size; i++) {
  10656. r[j] |= ((sp_digit)a->dp[i] << s);
  10657. r[j] &= 0xffffffff;
  10658. s = 32U - s;
  10659. if (j + 1 >= size) {
  10660. break;
  10661. }
  10662. /* lint allow cast of mismatch word32 and mp_digit */
  10663. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  10664. while ((s + 32U) <= (word32)DIGIT_BIT) {
  10665. s += 32U;
  10666. r[j] &= 0xffffffff;
  10667. if (j + 1 >= size) {
  10668. break;
  10669. }
  10670. if (s < (word32)DIGIT_BIT) {
  10671. /* lint allow cast of mismatch word32 and mp_digit */
  10672. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  10673. }
  10674. else {
  10675. r[++j] = 0L;
  10676. }
  10677. }
  10678. s = (word32)DIGIT_BIT - s;
  10679. }
  10680. for (j++; j < size; j++) {
  10681. r[j] = 0;
  10682. }
  10683. #else
  10684. int i, j = 0, s = 0;
  10685. r[0] = 0;
  10686. for (i = 0; i < a->used && j < size; i++) {
  10687. r[j] |= ((sp_digit)a->dp[i]) << s;
  10688. if (s + DIGIT_BIT >= 32) {
  10689. r[j] &= 0xffffffff;
  10690. if (j + 1 >= size) {
  10691. break;
  10692. }
  10693. s = 32 - s;
  10694. if (s == DIGIT_BIT) {
  10695. r[++j] = 0;
  10696. s = 0;
  10697. }
  10698. else {
  10699. r[++j] = a->dp[i] >> s;
  10700. s = DIGIT_BIT - s;
  10701. }
  10702. }
  10703. else {
  10704. s += DIGIT_BIT;
  10705. }
  10706. }
  10707. for (j++; j < size; j++) {
  10708. r[j] = 0;
  10709. }
  10710. #endif
  10711. }
  10712. /* Write r as big endian to byte array.
  10713. * Fixed length number of bytes written: 512
  10714. *
  10715. * r A single precision integer.
  10716. * a Byte array.
  10717. */
  10718. static void sp_4096_to_bin(sp_digit* r, byte* a)
  10719. {
  10720. int i, j, s = 0, b;
  10721. j = 4096 / 8 - 1;
  10722. a[j] = 0;
  10723. for (i=0; i<128 && j>=0; i++) {
  10724. b = 0;
  10725. /* lint allow cast of mismatch sp_digit and int */
  10726. a[j--] |= (byte)(r[i] << s); b += 8 - s; /*lint !e9033*/
  10727. if (j < 0) {
  10728. break;
  10729. }
  10730. while (b < 32) {
  10731. a[j--] = r[i] >> b; b += 8;
  10732. if (j < 0) {
  10733. break;
  10734. }
  10735. }
  10736. s = 8 - (b - 32);
  10737. if (j >= 0) {
  10738. a[j] = 0;
  10739. }
  10740. if (s != 0) {
  10741. j++;
  10742. }
  10743. }
  10744. }
  10745. #ifndef WOLFSSL_SP_SMALL
  10746. /* Add b to a into r. (r = a + b)
  10747. *
  10748. * r A single precision integer.
  10749. * a A single precision integer.
  10750. * b A single precision integer.
  10751. */
  10752. SP_NOINLINE static sp_digit sp_4096_add_64(sp_digit* r, const sp_digit* a,
  10753. const sp_digit* b)
  10754. {
  10755. sp_digit c = 0;
  10756. __asm__ __volatile__ (
  10757. "mov r7, #0\n\t"
  10758. "mvn r7, r7\n\t"
  10759. "ldr r4, [%[a], #0]\n\t"
  10760. "ldr r5, [%[b], #0]\n\t"
  10761. "add r4, r5\n\t"
  10762. "str r4, [%[r], #0]\n\t"
  10763. "ldr r4, [%[a], #4]\n\t"
  10764. "ldr r5, [%[b], #4]\n\t"
  10765. "adc r4, r5\n\t"
  10766. "str r4, [%[r], #4]\n\t"
  10767. "ldr r4, [%[a], #8]\n\t"
  10768. "ldr r5, [%[b], #8]\n\t"
  10769. "adc r4, r5\n\t"
  10770. "str r4, [%[r], #8]\n\t"
  10771. "ldr r4, [%[a], #12]\n\t"
  10772. "ldr r5, [%[b], #12]\n\t"
  10773. "adc r4, r5\n\t"
  10774. "str r4, [%[r], #12]\n\t"
  10775. "ldr r4, [%[a], #16]\n\t"
  10776. "ldr r5, [%[b], #16]\n\t"
  10777. "adc r4, r5\n\t"
  10778. "str r4, [%[r], #16]\n\t"
  10779. "ldr r4, [%[a], #20]\n\t"
  10780. "ldr r5, [%[b], #20]\n\t"
  10781. "adc r4, r5\n\t"
  10782. "str r4, [%[r], #20]\n\t"
  10783. "ldr r4, [%[a], #24]\n\t"
  10784. "ldr r5, [%[b], #24]\n\t"
  10785. "adc r4, r5\n\t"
  10786. "str r4, [%[r], #24]\n\t"
  10787. "ldr r4, [%[a], #28]\n\t"
  10788. "ldr r5, [%[b], #28]\n\t"
  10789. "adc r4, r5\n\t"
  10790. "str r4, [%[r], #28]\n\t"
  10791. "ldr r4, [%[a], #32]\n\t"
  10792. "ldr r5, [%[b], #32]\n\t"
  10793. "adc r4, r5\n\t"
  10794. "str r4, [%[r], #32]\n\t"
  10795. "ldr r4, [%[a], #36]\n\t"
  10796. "ldr r5, [%[b], #36]\n\t"
  10797. "adc r4, r5\n\t"
  10798. "str r4, [%[r], #36]\n\t"
  10799. "ldr r4, [%[a], #40]\n\t"
  10800. "ldr r5, [%[b], #40]\n\t"
  10801. "adc r4, r5\n\t"
  10802. "str r4, [%[r], #40]\n\t"
  10803. "ldr r4, [%[a], #44]\n\t"
  10804. "ldr r5, [%[b], #44]\n\t"
  10805. "adc r4, r5\n\t"
  10806. "str r4, [%[r], #44]\n\t"
  10807. "ldr r4, [%[a], #48]\n\t"
  10808. "ldr r5, [%[b], #48]\n\t"
  10809. "adc r4, r5\n\t"
  10810. "str r4, [%[r], #48]\n\t"
  10811. "ldr r4, [%[a], #52]\n\t"
  10812. "ldr r5, [%[b], #52]\n\t"
  10813. "adc r4, r5\n\t"
  10814. "str r4, [%[r], #52]\n\t"
  10815. "ldr r4, [%[a], #56]\n\t"
  10816. "ldr r5, [%[b], #56]\n\t"
  10817. "adc r4, r5\n\t"
  10818. "str r4, [%[r], #56]\n\t"
  10819. "ldr r4, [%[a], #60]\n\t"
  10820. "ldr r5, [%[b], #60]\n\t"
  10821. "adc r4, r5\n\t"
  10822. "str r4, [%[r], #60]\n\t"
  10823. "ldr r4, [%[a], #64]\n\t"
  10824. "ldr r5, [%[b], #64]\n\t"
  10825. "adc r4, r5\n\t"
  10826. "str r4, [%[r], #64]\n\t"
  10827. "ldr r4, [%[a], #68]\n\t"
  10828. "ldr r5, [%[b], #68]\n\t"
  10829. "adc r4, r5\n\t"
  10830. "str r4, [%[r], #68]\n\t"
  10831. "ldr r4, [%[a], #72]\n\t"
  10832. "ldr r5, [%[b], #72]\n\t"
  10833. "adc r4, r5\n\t"
  10834. "str r4, [%[r], #72]\n\t"
  10835. "ldr r4, [%[a], #76]\n\t"
  10836. "ldr r5, [%[b], #76]\n\t"
  10837. "adc r4, r5\n\t"
  10838. "str r4, [%[r], #76]\n\t"
  10839. "ldr r4, [%[a], #80]\n\t"
  10840. "ldr r5, [%[b], #80]\n\t"
  10841. "adc r4, r5\n\t"
  10842. "str r4, [%[r], #80]\n\t"
  10843. "ldr r4, [%[a], #84]\n\t"
  10844. "ldr r5, [%[b], #84]\n\t"
  10845. "adc r4, r5\n\t"
  10846. "str r4, [%[r], #84]\n\t"
  10847. "ldr r4, [%[a], #88]\n\t"
  10848. "ldr r5, [%[b], #88]\n\t"
  10849. "adc r4, r5\n\t"
  10850. "str r4, [%[r], #88]\n\t"
  10851. "ldr r4, [%[a], #92]\n\t"
  10852. "ldr r5, [%[b], #92]\n\t"
  10853. "adc r4, r5\n\t"
  10854. "str r4, [%[r], #92]\n\t"
  10855. "ldr r4, [%[a], #96]\n\t"
  10856. "ldr r5, [%[b], #96]\n\t"
  10857. "adc r4, r5\n\t"
  10858. "str r4, [%[r], #96]\n\t"
  10859. "ldr r4, [%[a], #100]\n\t"
  10860. "ldr r5, [%[b], #100]\n\t"
  10861. "adc r4, r5\n\t"
  10862. "str r4, [%[r], #100]\n\t"
  10863. "ldr r4, [%[a], #104]\n\t"
  10864. "ldr r5, [%[b], #104]\n\t"
  10865. "adc r4, r5\n\t"
  10866. "str r4, [%[r], #104]\n\t"
  10867. "ldr r4, [%[a], #108]\n\t"
  10868. "ldr r5, [%[b], #108]\n\t"
  10869. "adc r4, r5\n\t"
  10870. "str r4, [%[r], #108]\n\t"
  10871. "ldr r4, [%[a], #112]\n\t"
  10872. "ldr r5, [%[b], #112]\n\t"
  10873. "adc r4, r5\n\t"
  10874. "str r4, [%[r], #112]\n\t"
  10875. "ldr r4, [%[a], #116]\n\t"
  10876. "ldr r5, [%[b], #116]\n\t"
  10877. "adc r4, r5\n\t"
  10878. "str r4, [%[r], #116]\n\t"
  10879. "ldr r4, [%[a], #120]\n\t"
  10880. "ldr r5, [%[b], #120]\n\t"
  10881. "adc r4, r5\n\t"
  10882. "str r4, [%[r], #120]\n\t"
  10883. "ldr r4, [%[a], #124]\n\t"
  10884. "ldr r5, [%[b], #124]\n\t"
  10885. "adc r4, r5\n\t"
  10886. "str r4, [%[r], #124]\n\t"
  10887. "mov %[c], #0\n\t"
  10888. "adc %[c], %[c]\n\t"
  10889. "add %[a], #0x80\n\t"
  10890. "add %[b], #0x80\n\t"
  10891. "add %[r], #0x80\n\t"
  10892. "add %[c], r7\n\t"
  10893. "ldr r4, [%[a], #0]\n\t"
  10894. "ldr r5, [%[b], #0]\n\t"
  10895. "adc r4, r5\n\t"
  10896. "str r4, [%[r], #0]\n\t"
  10897. "ldr r4, [%[a], #4]\n\t"
  10898. "ldr r5, [%[b], #4]\n\t"
  10899. "adc r4, r5\n\t"
  10900. "str r4, [%[r], #4]\n\t"
  10901. "ldr r4, [%[a], #8]\n\t"
  10902. "ldr r5, [%[b], #8]\n\t"
  10903. "adc r4, r5\n\t"
  10904. "str r4, [%[r], #8]\n\t"
  10905. "ldr r4, [%[a], #12]\n\t"
  10906. "ldr r5, [%[b], #12]\n\t"
  10907. "adc r4, r5\n\t"
  10908. "str r4, [%[r], #12]\n\t"
  10909. "ldr r4, [%[a], #16]\n\t"
  10910. "ldr r5, [%[b], #16]\n\t"
  10911. "adc r4, r5\n\t"
  10912. "str r4, [%[r], #16]\n\t"
  10913. "ldr r4, [%[a], #20]\n\t"
  10914. "ldr r5, [%[b], #20]\n\t"
  10915. "adc r4, r5\n\t"
  10916. "str r4, [%[r], #20]\n\t"
  10917. "ldr r4, [%[a], #24]\n\t"
  10918. "ldr r5, [%[b], #24]\n\t"
  10919. "adc r4, r5\n\t"
  10920. "str r4, [%[r], #24]\n\t"
  10921. "ldr r4, [%[a], #28]\n\t"
  10922. "ldr r5, [%[b], #28]\n\t"
  10923. "adc r4, r5\n\t"
  10924. "str r4, [%[r], #28]\n\t"
  10925. "ldr r4, [%[a], #32]\n\t"
  10926. "ldr r5, [%[b], #32]\n\t"
  10927. "adc r4, r5\n\t"
  10928. "str r4, [%[r], #32]\n\t"
  10929. "ldr r4, [%[a], #36]\n\t"
  10930. "ldr r5, [%[b], #36]\n\t"
  10931. "adc r4, r5\n\t"
  10932. "str r4, [%[r], #36]\n\t"
  10933. "ldr r4, [%[a], #40]\n\t"
  10934. "ldr r5, [%[b], #40]\n\t"
  10935. "adc r4, r5\n\t"
  10936. "str r4, [%[r], #40]\n\t"
  10937. "ldr r4, [%[a], #44]\n\t"
  10938. "ldr r5, [%[b], #44]\n\t"
  10939. "adc r4, r5\n\t"
  10940. "str r4, [%[r], #44]\n\t"
  10941. "ldr r4, [%[a], #48]\n\t"
  10942. "ldr r5, [%[b], #48]\n\t"
  10943. "adc r4, r5\n\t"
  10944. "str r4, [%[r], #48]\n\t"
  10945. "ldr r4, [%[a], #52]\n\t"
  10946. "ldr r5, [%[b], #52]\n\t"
  10947. "adc r4, r5\n\t"
  10948. "str r4, [%[r], #52]\n\t"
  10949. "ldr r4, [%[a], #56]\n\t"
  10950. "ldr r5, [%[b], #56]\n\t"
  10951. "adc r4, r5\n\t"
  10952. "str r4, [%[r], #56]\n\t"
  10953. "ldr r4, [%[a], #60]\n\t"
  10954. "ldr r5, [%[b], #60]\n\t"
  10955. "adc r4, r5\n\t"
  10956. "str r4, [%[r], #60]\n\t"
  10957. "ldr r4, [%[a], #64]\n\t"
  10958. "ldr r5, [%[b], #64]\n\t"
  10959. "adc r4, r5\n\t"
  10960. "str r4, [%[r], #64]\n\t"
  10961. "ldr r4, [%[a], #68]\n\t"
  10962. "ldr r5, [%[b], #68]\n\t"
  10963. "adc r4, r5\n\t"
  10964. "str r4, [%[r], #68]\n\t"
  10965. "ldr r4, [%[a], #72]\n\t"
  10966. "ldr r5, [%[b], #72]\n\t"
  10967. "adc r4, r5\n\t"
  10968. "str r4, [%[r], #72]\n\t"
  10969. "ldr r4, [%[a], #76]\n\t"
  10970. "ldr r5, [%[b], #76]\n\t"
  10971. "adc r4, r5\n\t"
  10972. "str r4, [%[r], #76]\n\t"
  10973. "ldr r4, [%[a], #80]\n\t"
  10974. "ldr r5, [%[b], #80]\n\t"
  10975. "adc r4, r5\n\t"
  10976. "str r4, [%[r], #80]\n\t"
  10977. "ldr r4, [%[a], #84]\n\t"
  10978. "ldr r5, [%[b], #84]\n\t"
  10979. "adc r4, r5\n\t"
  10980. "str r4, [%[r], #84]\n\t"
  10981. "ldr r4, [%[a], #88]\n\t"
  10982. "ldr r5, [%[b], #88]\n\t"
  10983. "adc r4, r5\n\t"
  10984. "str r4, [%[r], #88]\n\t"
  10985. "ldr r4, [%[a], #92]\n\t"
  10986. "ldr r5, [%[b], #92]\n\t"
  10987. "adc r4, r5\n\t"
  10988. "str r4, [%[r], #92]\n\t"
  10989. "ldr r4, [%[a], #96]\n\t"
  10990. "ldr r5, [%[b], #96]\n\t"
  10991. "adc r4, r5\n\t"
  10992. "str r4, [%[r], #96]\n\t"
  10993. "ldr r4, [%[a], #100]\n\t"
  10994. "ldr r5, [%[b], #100]\n\t"
  10995. "adc r4, r5\n\t"
  10996. "str r4, [%[r], #100]\n\t"
  10997. "ldr r4, [%[a], #104]\n\t"
  10998. "ldr r5, [%[b], #104]\n\t"
  10999. "adc r4, r5\n\t"
  11000. "str r4, [%[r], #104]\n\t"
  11001. "ldr r4, [%[a], #108]\n\t"
  11002. "ldr r5, [%[b], #108]\n\t"
  11003. "adc r4, r5\n\t"
  11004. "str r4, [%[r], #108]\n\t"
  11005. "ldr r4, [%[a], #112]\n\t"
  11006. "ldr r5, [%[b], #112]\n\t"
  11007. "adc r4, r5\n\t"
  11008. "str r4, [%[r], #112]\n\t"
  11009. "ldr r4, [%[a], #116]\n\t"
  11010. "ldr r5, [%[b], #116]\n\t"
  11011. "adc r4, r5\n\t"
  11012. "str r4, [%[r], #116]\n\t"
  11013. "ldr r4, [%[a], #120]\n\t"
  11014. "ldr r5, [%[b], #120]\n\t"
  11015. "adc r4, r5\n\t"
  11016. "str r4, [%[r], #120]\n\t"
  11017. "ldr r4, [%[a], #124]\n\t"
  11018. "ldr r5, [%[b], #124]\n\t"
  11019. "adc r4, r5\n\t"
  11020. "str r4, [%[r], #124]\n\t"
  11021. "mov %[c], #0\n\t"
  11022. "adc %[c], %[c]\n\t"
  11023. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  11024. :
  11025. : "memory", "r4", "r5", "r7"
  11026. );
  11027. return c;
  11028. }
  11029. /* Sub b from a into r. (r = a - b)
  11030. *
  11031. * r A single precision integer.
  11032. * a A single precision integer.
  11033. * b A single precision integer.
  11034. */
  11035. SP_NOINLINE static sp_digit sp_4096_sub_in_place_128(sp_digit* a,
  11036. const sp_digit* b)
  11037. {
  11038. sp_digit c = 0;
  11039. __asm__ __volatile__ (
  11040. "ldr r3, [%[a], #0]\n\t"
  11041. "ldr r4, [%[a], #4]\n\t"
  11042. "ldr r5, [%[b], #0]\n\t"
  11043. "ldr r6, [%[b], #4]\n\t"
  11044. "sub r3, r5\n\t"
  11045. "sbc r4, r6\n\t"
  11046. "str r3, [%[a], #0]\n\t"
  11047. "str r4, [%[a], #4]\n\t"
  11048. "ldr r3, [%[a], #8]\n\t"
  11049. "ldr r4, [%[a], #12]\n\t"
  11050. "ldr r5, [%[b], #8]\n\t"
  11051. "ldr r6, [%[b], #12]\n\t"
  11052. "sbc r3, r5\n\t"
  11053. "sbc r4, r6\n\t"
  11054. "str r3, [%[a], #8]\n\t"
  11055. "str r4, [%[a], #12]\n\t"
  11056. "ldr r3, [%[a], #16]\n\t"
  11057. "ldr r4, [%[a], #20]\n\t"
  11058. "ldr r5, [%[b], #16]\n\t"
  11059. "ldr r6, [%[b], #20]\n\t"
  11060. "sbc r3, r5\n\t"
  11061. "sbc r4, r6\n\t"
  11062. "str r3, [%[a], #16]\n\t"
  11063. "str r4, [%[a], #20]\n\t"
  11064. "ldr r3, [%[a], #24]\n\t"
  11065. "ldr r4, [%[a], #28]\n\t"
  11066. "ldr r5, [%[b], #24]\n\t"
  11067. "ldr r6, [%[b], #28]\n\t"
  11068. "sbc r3, r5\n\t"
  11069. "sbc r4, r6\n\t"
  11070. "str r3, [%[a], #24]\n\t"
  11071. "str r4, [%[a], #28]\n\t"
  11072. "ldr r3, [%[a], #32]\n\t"
  11073. "ldr r4, [%[a], #36]\n\t"
  11074. "ldr r5, [%[b], #32]\n\t"
  11075. "ldr r6, [%[b], #36]\n\t"
  11076. "sbc r3, r5\n\t"
  11077. "sbc r4, r6\n\t"
  11078. "str r3, [%[a], #32]\n\t"
  11079. "str r4, [%[a], #36]\n\t"
  11080. "ldr r3, [%[a], #40]\n\t"
  11081. "ldr r4, [%[a], #44]\n\t"
  11082. "ldr r5, [%[b], #40]\n\t"
  11083. "ldr r6, [%[b], #44]\n\t"
  11084. "sbc r3, r5\n\t"
  11085. "sbc r4, r6\n\t"
  11086. "str r3, [%[a], #40]\n\t"
  11087. "str r4, [%[a], #44]\n\t"
  11088. "ldr r3, [%[a], #48]\n\t"
  11089. "ldr r4, [%[a], #52]\n\t"
  11090. "ldr r5, [%[b], #48]\n\t"
  11091. "ldr r6, [%[b], #52]\n\t"
  11092. "sbc r3, r5\n\t"
  11093. "sbc r4, r6\n\t"
  11094. "str r3, [%[a], #48]\n\t"
  11095. "str r4, [%[a], #52]\n\t"
  11096. "ldr r3, [%[a], #56]\n\t"
  11097. "ldr r4, [%[a], #60]\n\t"
  11098. "ldr r5, [%[b], #56]\n\t"
  11099. "ldr r6, [%[b], #60]\n\t"
  11100. "sbc r3, r5\n\t"
  11101. "sbc r4, r6\n\t"
  11102. "str r3, [%[a], #56]\n\t"
  11103. "str r4, [%[a], #60]\n\t"
  11104. "ldr r3, [%[a], #64]\n\t"
  11105. "ldr r4, [%[a], #68]\n\t"
  11106. "ldr r5, [%[b], #64]\n\t"
  11107. "ldr r6, [%[b], #68]\n\t"
  11108. "sbc r3, r5\n\t"
  11109. "sbc r4, r6\n\t"
  11110. "str r3, [%[a], #64]\n\t"
  11111. "str r4, [%[a], #68]\n\t"
  11112. "ldr r3, [%[a], #72]\n\t"
  11113. "ldr r4, [%[a], #76]\n\t"
  11114. "ldr r5, [%[b], #72]\n\t"
  11115. "ldr r6, [%[b], #76]\n\t"
  11116. "sbc r3, r5\n\t"
  11117. "sbc r4, r6\n\t"
  11118. "str r3, [%[a], #72]\n\t"
  11119. "str r4, [%[a], #76]\n\t"
  11120. "ldr r3, [%[a], #80]\n\t"
  11121. "ldr r4, [%[a], #84]\n\t"
  11122. "ldr r5, [%[b], #80]\n\t"
  11123. "ldr r6, [%[b], #84]\n\t"
  11124. "sbc r3, r5\n\t"
  11125. "sbc r4, r6\n\t"
  11126. "str r3, [%[a], #80]\n\t"
  11127. "str r4, [%[a], #84]\n\t"
  11128. "ldr r3, [%[a], #88]\n\t"
  11129. "ldr r4, [%[a], #92]\n\t"
  11130. "ldr r5, [%[b], #88]\n\t"
  11131. "ldr r6, [%[b], #92]\n\t"
  11132. "sbc r3, r5\n\t"
  11133. "sbc r4, r6\n\t"
  11134. "str r3, [%[a], #88]\n\t"
  11135. "str r4, [%[a], #92]\n\t"
  11136. "ldr r3, [%[a], #96]\n\t"
  11137. "ldr r4, [%[a], #100]\n\t"
  11138. "ldr r5, [%[b], #96]\n\t"
  11139. "ldr r6, [%[b], #100]\n\t"
  11140. "sbc r3, r5\n\t"
  11141. "sbc r4, r6\n\t"
  11142. "str r3, [%[a], #96]\n\t"
  11143. "str r4, [%[a], #100]\n\t"
  11144. "ldr r3, [%[a], #104]\n\t"
  11145. "ldr r4, [%[a], #108]\n\t"
  11146. "ldr r5, [%[b], #104]\n\t"
  11147. "ldr r6, [%[b], #108]\n\t"
  11148. "sbc r3, r5\n\t"
  11149. "sbc r4, r6\n\t"
  11150. "str r3, [%[a], #104]\n\t"
  11151. "str r4, [%[a], #108]\n\t"
  11152. "ldr r3, [%[a], #112]\n\t"
  11153. "ldr r4, [%[a], #116]\n\t"
  11154. "ldr r5, [%[b], #112]\n\t"
  11155. "ldr r6, [%[b], #116]\n\t"
  11156. "sbc r3, r5\n\t"
  11157. "sbc r4, r6\n\t"
  11158. "str r3, [%[a], #112]\n\t"
  11159. "str r4, [%[a], #116]\n\t"
  11160. "ldr r3, [%[a], #120]\n\t"
  11161. "ldr r4, [%[a], #124]\n\t"
  11162. "ldr r5, [%[b], #120]\n\t"
  11163. "ldr r6, [%[b], #124]\n\t"
  11164. "sbc r3, r5\n\t"
  11165. "sbc r4, r6\n\t"
  11166. "str r3, [%[a], #120]\n\t"
  11167. "str r4, [%[a], #124]\n\t"
  11168. "sbc %[c], %[c]\n\t"
  11169. "add %[a], #0x80\n\t"
  11170. "add %[b], #0x80\n\t"
  11171. "mov r5, #0\n\t"
  11172. "sub r5, %[c]\n\t"
  11173. "ldr r3, [%[a], #0]\n\t"
  11174. "ldr r4, [%[a], #4]\n\t"
  11175. "ldr r5, [%[b], #0]\n\t"
  11176. "ldr r6, [%[b], #4]\n\t"
  11177. "sbc r3, r5\n\t"
  11178. "sbc r4, r6\n\t"
  11179. "str r3, [%[a], #0]\n\t"
  11180. "str r4, [%[a], #4]\n\t"
  11181. "ldr r3, [%[a], #8]\n\t"
  11182. "ldr r4, [%[a], #12]\n\t"
  11183. "ldr r5, [%[b], #8]\n\t"
  11184. "ldr r6, [%[b], #12]\n\t"
  11185. "sbc r3, r5\n\t"
  11186. "sbc r4, r6\n\t"
  11187. "str r3, [%[a], #8]\n\t"
  11188. "str r4, [%[a], #12]\n\t"
  11189. "ldr r3, [%[a], #16]\n\t"
  11190. "ldr r4, [%[a], #20]\n\t"
  11191. "ldr r5, [%[b], #16]\n\t"
  11192. "ldr r6, [%[b], #20]\n\t"
  11193. "sbc r3, r5\n\t"
  11194. "sbc r4, r6\n\t"
  11195. "str r3, [%[a], #16]\n\t"
  11196. "str r4, [%[a], #20]\n\t"
  11197. "ldr r3, [%[a], #24]\n\t"
  11198. "ldr r4, [%[a], #28]\n\t"
  11199. "ldr r5, [%[b], #24]\n\t"
  11200. "ldr r6, [%[b], #28]\n\t"
  11201. "sbc r3, r5\n\t"
  11202. "sbc r4, r6\n\t"
  11203. "str r3, [%[a], #24]\n\t"
  11204. "str r4, [%[a], #28]\n\t"
  11205. "ldr r3, [%[a], #32]\n\t"
  11206. "ldr r4, [%[a], #36]\n\t"
  11207. "ldr r5, [%[b], #32]\n\t"
  11208. "ldr r6, [%[b], #36]\n\t"
  11209. "sbc r3, r5\n\t"
  11210. "sbc r4, r6\n\t"
  11211. "str r3, [%[a], #32]\n\t"
  11212. "str r4, [%[a], #36]\n\t"
  11213. "ldr r3, [%[a], #40]\n\t"
  11214. "ldr r4, [%[a], #44]\n\t"
  11215. "ldr r5, [%[b], #40]\n\t"
  11216. "ldr r6, [%[b], #44]\n\t"
  11217. "sbc r3, r5\n\t"
  11218. "sbc r4, r6\n\t"
  11219. "str r3, [%[a], #40]\n\t"
  11220. "str r4, [%[a], #44]\n\t"
  11221. "ldr r3, [%[a], #48]\n\t"
  11222. "ldr r4, [%[a], #52]\n\t"
  11223. "ldr r5, [%[b], #48]\n\t"
  11224. "ldr r6, [%[b], #52]\n\t"
  11225. "sbc r3, r5\n\t"
  11226. "sbc r4, r6\n\t"
  11227. "str r3, [%[a], #48]\n\t"
  11228. "str r4, [%[a], #52]\n\t"
  11229. "ldr r3, [%[a], #56]\n\t"
  11230. "ldr r4, [%[a], #60]\n\t"
  11231. "ldr r5, [%[b], #56]\n\t"
  11232. "ldr r6, [%[b], #60]\n\t"
  11233. "sbc r3, r5\n\t"
  11234. "sbc r4, r6\n\t"
  11235. "str r3, [%[a], #56]\n\t"
  11236. "str r4, [%[a], #60]\n\t"
  11237. "ldr r3, [%[a], #64]\n\t"
  11238. "ldr r4, [%[a], #68]\n\t"
  11239. "ldr r5, [%[b], #64]\n\t"
  11240. "ldr r6, [%[b], #68]\n\t"
  11241. "sbc r3, r5\n\t"
  11242. "sbc r4, r6\n\t"
  11243. "str r3, [%[a], #64]\n\t"
  11244. "str r4, [%[a], #68]\n\t"
  11245. "ldr r3, [%[a], #72]\n\t"
  11246. "ldr r4, [%[a], #76]\n\t"
  11247. "ldr r5, [%[b], #72]\n\t"
  11248. "ldr r6, [%[b], #76]\n\t"
  11249. "sbc r3, r5\n\t"
  11250. "sbc r4, r6\n\t"
  11251. "str r3, [%[a], #72]\n\t"
  11252. "str r4, [%[a], #76]\n\t"
  11253. "ldr r3, [%[a], #80]\n\t"
  11254. "ldr r4, [%[a], #84]\n\t"
  11255. "ldr r5, [%[b], #80]\n\t"
  11256. "ldr r6, [%[b], #84]\n\t"
  11257. "sbc r3, r5\n\t"
  11258. "sbc r4, r6\n\t"
  11259. "str r3, [%[a], #80]\n\t"
  11260. "str r4, [%[a], #84]\n\t"
  11261. "ldr r3, [%[a], #88]\n\t"
  11262. "ldr r4, [%[a], #92]\n\t"
  11263. "ldr r5, [%[b], #88]\n\t"
  11264. "ldr r6, [%[b], #92]\n\t"
  11265. "sbc r3, r5\n\t"
  11266. "sbc r4, r6\n\t"
  11267. "str r3, [%[a], #88]\n\t"
  11268. "str r4, [%[a], #92]\n\t"
  11269. "ldr r3, [%[a], #96]\n\t"
  11270. "ldr r4, [%[a], #100]\n\t"
  11271. "ldr r5, [%[b], #96]\n\t"
  11272. "ldr r6, [%[b], #100]\n\t"
  11273. "sbc r3, r5\n\t"
  11274. "sbc r4, r6\n\t"
  11275. "str r3, [%[a], #96]\n\t"
  11276. "str r4, [%[a], #100]\n\t"
  11277. "ldr r3, [%[a], #104]\n\t"
  11278. "ldr r4, [%[a], #108]\n\t"
  11279. "ldr r5, [%[b], #104]\n\t"
  11280. "ldr r6, [%[b], #108]\n\t"
  11281. "sbc r3, r5\n\t"
  11282. "sbc r4, r6\n\t"
  11283. "str r3, [%[a], #104]\n\t"
  11284. "str r4, [%[a], #108]\n\t"
  11285. "ldr r3, [%[a], #112]\n\t"
  11286. "ldr r4, [%[a], #116]\n\t"
  11287. "ldr r5, [%[b], #112]\n\t"
  11288. "ldr r6, [%[b], #116]\n\t"
  11289. "sbc r3, r5\n\t"
  11290. "sbc r4, r6\n\t"
  11291. "str r3, [%[a], #112]\n\t"
  11292. "str r4, [%[a], #116]\n\t"
  11293. "ldr r3, [%[a], #120]\n\t"
  11294. "ldr r4, [%[a], #124]\n\t"
  11295. "ldr r5, [%[b], #120]\n\t"
  11296. "ldr r6, [%[b], #124]\n\t"
  11297. "sbc r3, r5\n\t"
  11298. "sbc r4, r6\n\t"
  11299. "str r3, [%[a], #120]\n\t"
  11300. "str r4, [%[a], #124]\n\t"
  11301. "sbc %[c], %[c]\n\t"
  11302. "add %[a], #0x80\n\t"
  11303. "add %[b], #0x80\n\t"
  11304. "mov r5, #0\n\t"
  11305. "sub r5, %[c]\n\t"
  11306. "ldr r3, [%[a], #0]\n\t"
  11307. "ldr r4, [%[a], #4]\n\t"
  11308. "ldr r5, [%[b], #0]\n\t"
  11309. "ldr r6, [%[b], #4]\n\t"
  11310. "sbc r3, r5\n\t"
  11311. "sbc r4, r6\n\t"
  11312. "str r3, [%[a], #0]\n\t"
  11313. "str r4, [%[a], #4]\n\t"
  11314. "ldr r3, [%[a], #8]\n\t"
  11315. "ldr r4, [%[a], #12]\n\t"
  11316. "ldr r5, [%[b], #8]\n\t"
  11317. "ldr r6, [%[b], #12]\n\t"
  11318. "sbc r3, r5\n\t"
  11319. "sbc r4, r6\n\t"
  11320. "str r3, [%[a], #8]\n\t"
  11321. "str r4, [%[a], #12]\n\t"
  11322. "ldr r3, [%[a], #16]\n\t"
  11323. "ldr r4, [%[a], #20]\n\t"
  11324. "ldr r5, [%[b], #16]\n\t"
  11325. "ldr r6, [%[b], #20]\n\t"
  11326. "sbc r3, r5\n\t"
  11327. "sbc r4, r6\n\t"
  11328. "str r3, [%[a], #16]\n\t"
  11329. "str r4, [%[a], #20]\n\t"
  11330. "ldr r3, [%[a], #24]\n\t"
  11331. "ldr r4, [%[a], #28]\n\t"
  11332. "ldr r5, [%[b], #24]\n\t"
  11333. "ldr r6, [%[b], #28]\n\t"
  11334. "sbc r3, r5\n\t"
  11335. "sbc r4, r6\n\t"
  11336. "str r3, [%[a], #24]\n\t"
  11337. "str r4, [%[a], #28]\n\t"
  11338. "ldr r3, [%[a], #32]\n\t"
  11339. "ldr r4, [%[a], #36]\n\t"
  11340. "ldr r5, [%[b], #32]\n\t"
  11341. "ldr r6, [%[b], #36]\n\t"
  11342. "sbc r3, r5\n\t"
  11343. "sbc r4, r6\n\t"
  11344. "str r3, [%[a], #32]\n\t"
  11345. "str r4, [%[a], #36]\n\t"
  11346. "ldr r3, [%[a], #40]\n\t"
  11347. "ldr r4, [%[a], #44]\n\t"
  11348. "ldr r5, [%[b], #40]\n\t"
  11349. "ldr r6, [%[b], #44]\n\t"
  11350. "sbc r3, r5\n\t"
  11351. "sbc r4, r6\n\t"
  11352. "str r3, [%[a], #40]\n\t"
  11353. "str r4, [%[a], #44]\n\t"
  11354. "ldr r3, [%[a], #48]\n\t"
  11355. "ldr r4, [%[a], #52]\n\t"
  11356. "ldr r5, [%[b], #48]\n\t"
  11357. "ldr r6, [%[b], #52]\n\t"
  11358. "sbc r3, r5\n\t"
  11359. "sbc r4, r6\n\t"
  11360. "str r3, [%[a], #48]\n\t"
  11361. "str r4, [%[a], #52]\n\t"
  11362. "ldr r3, [%[a], #56]\n\t"
  11363. "ldr r4, [%[a], #60]\n\t"
  11364. "ldr r5, [%[b], #56]\n\t"
  11365. "ldr r6, [%[b], #60]\n\t"
  11366. "sbc r3, r5\n\t"
  11367. "sbc r4, r6\n\t"
  11368. "str r3, [%[a], #56]\n\t"
  11369. "str r4, [%[a], #60]\n\t"
  11370. "ldr r3, [%[a], #64]\n\t"
  11371. "ldr r4, [%[a], #68]\n\t"
  11372. "ldr r5, [%[b], #64]\n\t"
  11373. "ldr r6, [%[b], #68]\n\t"
  11374. "sbc r3, r5\n\t"
  11375. "sbc r4, r6\n\t"
  11376. "str r3, [%[a], #64]\n\t"
  11377. "str r4, [%[a], #68]\n\t"
  11378. "ldr r3, [%[a], #72]\n\t"
  11379. "ldr r4, [%[a], #76]\n\t"
  11380. "ldr r5, [%[b], #72]\n\t"
  11381. "ldr r6, [%[b], #76]\n\t"
  11382. "sbc r3, r5\n\t"
  11383. "sbc r4, r6\n\t"
  11384. "str r3, [%[a], #72]\n\t"
  11385. "str r4, [%[a], #76]\n\t"
  11386. "ldr r3, [%[a], #80]\n\t"
  11387. "ldr r4, [%[a], #84]\n\t"
  11388. "ldr r5, [%[b], #80]\n\t"
  11389. "ldr r6, [%[b], #84]\n\t"
  11390. "sbc r3, r5\n\t"
  11391. "sbc r4, r6\n\t"
  11392. "str r3, [%[a], #80]\n\t"
  11393. "str r4, [%[a], #84]\n\t"
  11394. "ldr r3, [%[a], #88]\n\t"
  11395. "ldr r4, [%[a], #92]\n\t"
  11396. "ldr r5, [%[b], #88]\n\t"
  11397. "ldr r6, [%[b], #92]\n\t"
  11398. "sbc r3, r5\n\t"
  11399. "sbc r4, r6\n\t"
  11400. "str r3, [%[a], #88]\n\t"
  11401. "str r4, [%[a], #92]\n\t"
  11402. "ldr r3, [%[a], #96]\n\t"
  11403. "ldr r4, [%[a], #100]\n\t"
  11404. "ldr r5, [%[b], #96]\n\t"
  11405. "ldr r6, [%[b], #100]\n\t"
  11406. "sbc r3, r5\n\t"
  11407. "sbc r4, r6\n\t"
  11408. "str r3, [%[a], #96]\n\t"
  11409. "str r4, [%[a], #100]\n\t"
  11410. "ldr r3, [%[a], #104]\n\t"
  11411. "ldr r4, [%[a], #108]\n\t"
  11412. "ldr r5, [%[b], #104]\n\t"
  11413. "ldr r6, [%[b], #108]\n\t"
  11414. "sbc r3, r5\n\t"
  11415. "sbc r4, r6\n\t"
  11416. "str r3, [%[a], #104]\n\t"
  11417. "str r4, [%[a], #108]\n\t"
  11418. "ldr r3, [%[a], #112]\n\t"
  11419. "ldr r4, [%[a], #116]\n\t"
  11420. "ldr r5, [%[b], #112]\n\t"
  11421. "ldr r6, [%[b], #116]\n\t"
  11422. "sbc r3, r5\n\t"
  11423. "sbc r4, r6\n\t"
  11424. "str r3, [%[a], #112]\n\t"
  11425. "str r4, [%[a], #116]\n\t"
  11426. "ldr r3, [%[a], #120]\n\t"
  11427. "ldr r4, [%[a], #124]\n\t"
  11428. "ldr r5, [%[b], #120]\n\t"
  11429. "ldr r6, [%[b], #124]\n\t"
  11430. "sbc r3, r5\n\t"
  11431. "sbc r4, r6\n\t"
  11432. "str r3, [%[a], #120]\n\t"
  11433. "str r4, [%[a], #124]\n\t"
  11434. "sbc %[c], %[c]\n\t"
  11435. "add %[a], #0x80\n\t"
  11436. "add %[b], #0x80\n\t"
  11437. "mov r5, #0\n\t"
  11438. "sub r5, %[c]\n\t"
  11439. "ldr r3, [%[a], #0]\n\t"
  11440. "ldr r4, [%[a], #4]\n\t"
  11441. "ldr r5, [%[b], #0]\n\t"
  11442. "ldr r6, [%[b], #4]\n\t"
  11443. "sbc r3, r5\n\t"
  11444. "sbc r4, r6\n\t"
  11445. "str r3, [%[a], #0]\n\t"
  11446. "str r4, [%[a], #4]\n\t"
  11447. "ldr r3, [%[a], #8]\n\t"
  11448. "ldr r4, [%[a], #12]\n\t"
  11449. "ldr r5, [%[b], #8]\n\t"
  11450. "ldr r6, [%[b], #12]\n\t"
  11451. "sbc r3, r5\n\t"
  11452. "sbc r4, r6\n\t"
  11453. "str r3, [%[a], #8]\n\t"
  11454. "str r4, [%[a], #12]\n\t"
  11455. "ldr r3, [%[a], #16]\n\t"
  11456. "ldr r4, [%[a], #20]\n\t"
  11457. "ldr r5, [%[b], #16]\n\t"
  11458. "ldr r6, [%[b], #20]\n\t"
  11459. "sbc r3, r5\n\t"
  11460. "sbc r4, r6\n\t"
  11461. "str r3, [%[a], #16]\n\t"
  11462. "str r4, [%[a], #20]\n\t"
  11463. "ldr r3, [%[a], #24]\n\t"
  11464. "ldr r4, [%[a], #28]\n\t"
  11465. "ldr r5, [%[b], #24]\n\t"
  11466. "ldr r6, [%[b], #28]\n\t"
  11467. "sbc r3, r5\n\t"
  11468. "sbc r4, r6\n\t"
  11469. "str r3, [%[a], #24]\n\t"
  11470. "str r4, [%[a], #28]\n\t"
  11471. "ldr r3, [%[a], #32]\n\t"
  11472. "ldr r4, [%[a], #36]\n\t"
  11473. "ldr r5, [%[b], #32]\n\t"
  11474. "ldr r6, [%[b], #36]\n\t"
  11475. "sbc r3, r5\n\t"
  11476. "sbc r4, r6\n\t"
  11477. "str r3, [%[a], #32]\n\t"
  11478. "str r4, [%[a], #36]\n\t"
  11479. "ldr r3, [%[a], #40]\n\t"
  11480. "ldr r4, [%[a], #44]\n\t"
  11481. "ldr r5, [%[b], #40]\n\t"
  11482. "ldr r6, [%[b], #44]\n\t"
  11483. "sbc r3, r5\n\t"
  11484. "sbc r4, r6\n\t"
  11485. "str r3, [%[a], #40]\n\t"
  11486. "str r4, [%[a], #44]\n\t"
  11487. "ldr r3, [%[a], #48]\n\t"
  11488. "ldr r4, [%[a], #52]\n\t"
  11489. "ldr r5, [%[b], #48]\n\t"
  11490. "ldr r6, [%[b], #52]\n\t"
  11491. "sbc r3, r5\n\t"
  11492. "sbc r4, r6\n\t"
  11493. "str r3, [%[a], #48]\n\t"
  11494. "str r4, [%[a], #52]\n\t"
  11495. "ldr r3, [%[a], #56]\n\t"
  11496. "ldr r4, [%[a], #60]\n\t"
  11497. "ldr r5, [%[b], #56]\n\t"
  11498. "ldr r6, [%[b], #60]\n\t"
  11499. "sbc r3, r5\n\t"
  11500. "sbc r4, r6\n\t"
  11501. "str r3, [%[a], #56]\n\t"
  11502. "str r4, [%[a], #60]\n\t"
  11503. "ldr r3, [%[a], #64]\n\t"
  11504. "ldr r4, [%[a], #68]\n\t"
  11505. "ldr r5, [%[b], #64]\n\t"
  11506. "ldr r6, [%[b], #68]\n\t"
  11507. "sbc r3, r5\n\t"
  11508. "sbc r4, r6\n\t"
  11509. "str r3, [%[a], #64]\n\t"
  11510. "str r4, [%[a], #68]\n\t"
  11511. "ldr r3, [%[a], #72]\n\t"
  11512. "ldr r4, [%[a], #76]\n\t"
  11513. "ldr r5, [%[b], #72]\n\t"
  11514. "ldr r6, [%[b], #76]\n\t"
  11515. "sbc r3, r5\n\t"
  11516. "sbc r4, r6\n\t"
  11517. "str r3, [%[a], #72]\n\t"
  11518. "str r4, [%[a], #76]\n\t"
  11519. "ldr r3, [%[a], #80]\n\t"
  11520. "ldr r4, [%[a], #84]\n\t"
  11521. "ldr r5, [%[b], #80]\n\t"
  11522. "ldr r6, [%[b], #84]\n\t"
  11523. "sbc r3, r5\n\t"
  11524. "sbc r4, r6\n\t"
  11525. "str r3, [%[a], #80]\n\t"
  11526. "str r4, [%[a], #84]\n\t"
  11527. "ldr r3, [%[a], #88]\n\t"
  11528. "ldr r4, [%[a], #92]\n\t"
  11529. "ldr r5, [%[b], #88]\n\t"
  11530. "ldr r6, [%[b], #92]\n\t"
  11531. "sbc r3, r5\n\t"
  11532. "sbc r4, r6\n\t"
  11533. "str r3, [%[a], #88]\n\t"
  11534. "str r4, [%[a], #92]\n\t"
  11535. "ldr r3, [%[a], #96]\n\t"
  11536. "ldr r4, [%[a], #100]\n\t"
  11537. "ldr r5, [%[b], #96]\n\t"
  11538. "ldr r6, [%[b], #100]\n\t"
  11539. "sbc r3, r5\n\t"
  11540. "sbc r4, r6\n\t"
  11541. "str r3, [%[a], #96]\n\t"
  11542. "str r4, [%[a], #100]\n\t"
  11543. "ldr r3, [%[a], #104]\n\t"
  11544. "ldr r4, [%[a], #108]\n\t"
  11545. "ldr r5, [%[b], #104]\n\t"
  11546. "ldr r6, [%[b], #108]\n\t"
  11547. "sbc r3, r5\n\t"
  11548. "sbc r4, r6\n\t"
  11549. "str r3, [%[a], #104]\n\t"
  11550. "str r4, [%[a], #108]\n\t"
  11551. "ldr r3, [%[a], #112]\n\t"
  11552. "ldr r4, [%[a], #116]\n\t"
  11553. "ldr r5, [%[b], #112]\n\t"
  11554. "ldr r6, [%[b], #116]\n\t"
  11555. "sbc r3, r5\n\t"
  11556. "sbc r4, r6\n\t"
  11557. "str r3, [%[a], #112]\n\t"
  11558. "str r4, [%[a], #116]\n\t"
  11559. "ldr r3, [%[a], #120]\n\t"
  11560. "ldr r4, [%[a], #124]\n\t"
  11561. "ldr r5, [%[b], #120]\n\t"
  11562. "ldr r6, [%[b], #124]\n\t"
  11563. "sbc r3, r5\n\t"
  11564. "sbc r4, r6\n\t"
  11565. "str r3, [%[a], #120]\n\t"
  11566. "str r4, [%[a], #124]\n\t"
  11567. "sbc %[c], %[c]\n\t"
  11568. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  11569. :
  11570. : "memory", "r3", "r4", "r5", "r6"
  11571. );
  11572. return c;
  11573. }
  11574. /* Add b to a into r. (r = a + b)
  11575. *
  11576. * r A single precision integer.
  11577. * a A single precision integer.
  11578. * b A single precision integer.
  11579. */
  11580. SP_NOINLINE static sp_digit sp_4096_add_128(sp_digit* r, const sp_digit* a,
  11581. const sp_digit* b)
  11582. {
  11583. sp_digit c = 0;
  11584. __asm__ __volatile__ (
  11585. "mov r7, #0\n\t"
  11586. "mvn r7, r7\n\t"
  11587. "ldr r4, [%[a], #0]\n\t"
  11588. "ldr r5, [%[b], #0]\n\t"
  11589. "add r4, r5\n\t"
  11590. "str r4, [%[r], #0]\n\t"
  11591. "ldr r4, [%[a], #4]\n\t"
  11592. "ldr r5, [%[b], #4]\n\t"
  11593. "adc r4, r5\n\t"
  11594. "str r4, [%[r], #4]\n\t"
  11595. "ldr r4, [%[a], #8]\n\t"
  11596. "ldr r5, [%[b], #8]\n\t"
  11597. "adc r4, r5\n\t"
  11598. "str r4, [%[r], #8]\n\t"
  11599. "ldr r4, [%[a], #12]\n\t"
  11600. "ldr r5, [%[b], #12]\n\t"
  11601. "adc r4, r5\n\t"
  11602. "str r4, [%[r], #12]\n\t"
  11603. "ldr r4, [%[a], #16]\n\t"
  11604. "ldr r5, [%[b], #16]\n\t"
  11605. "adc r4, r5\n\t"
  11606. "str r4, [%[r], #16]\n\t"
  11607. "ldr r4, [%[a], #20]\n\t"
  11608. "ldr r5, [%[b], #20]\n\t"
  11609. "adc r4, r5\n\t"
  11610. "str r4, [%[r], #20]\n\t"
  11611. "ldr r4, [%[a], #24]\n\t"
  11612. "ldr r5, [%[b], #24]\n\t"
  11613. "adc r4, r5\n\t"
  11614. "str r4, [%[r], #24]\n\t"
  11615. "ldr r4, [%[a], #28]\n\t"
  11616. "ldr r5, [%[b], #28]\n\t"
  11617. "adc r4, r5\n\t"
  11618. "str r4, [%[r], #28]\n\t"
  11619. "ldr r4, [%[a], #32]\n\t"
  11620. "ldr r5, [%[b], #32]\n\t"
  11621. "adc r4, r5\n\t"
  11622. "str r4, [%[r], #32]\n\t"
  11623. "ldr r4, [%[a], #36]\n\t"
  11624. "ldr r5, [%[b], #36]\n\t"
  11625. "adc r4, r5\n\t"
  11626. "str r4, [%[r], #36]\n\t"
  11627. "ldr r4, [%[a], #40]\n\t"
  11628. "ldr r5, [%[b], #40]\n\t"
  11629. "adc r4, r5\n\t"
  11630. "str r4, [%[r], #40]\n\t"
  11631. "ldr r4, [%[a], #44]\n\t"
  11632. "ldr r5, [%[b], #44]\n\t"
  11633. "adc r4, r5\n\t"
  11634. "str r4, [%[r], #44]\n\t"
  11635. "ldr r4, [%[a], #48]\n\t"
  11636. "ldr r5, [%[b], #48]\n\t"
  11637. "adc r4, r5\n\t"
  11638. "str r4, [%[r], #48]\n\t"
  11639. "ldr r4, [%[a], #52]\n\t"
  11640. "ldr r5, [%[b], #52]\n\t"
  11641. "adc r4, r5\n\t"
  11642. "str r4, [%[r], #52]\n\t"
  11643. "ldr r4, [%[a], #56]\n\t"
  11644. "ldr r5, [%[b], #56]\n\t"
  11645. "adc r4, r5\n\t"
  11646. "str r4, [%[r], #56]\n\t"
  11647. "ldr r4, [%[a], #60]\n\t"
  11648. "ldr r5, [%[b], #60]\n\t"
  11649. "adc r4, r5\n\t"
  11650. "str r4, [%[r], #60]\n\t"
  11651. "ldr r4, [%[a], #64]\n\t"
  11652. "ldr r5, [%[b], #64]\n\t"
  11653. "adc r4, r5\n\t"
  11654. "str r4, [%[r], #64]\n\t"
  11655. "ldr r4, [%[a], #68]\n\t"
  11656. "ldr r5, [%[b], #68]\n\t"
  11657. "adc r4, r5\n\t"
  11658. "str r4, [%[r], #68]\n\t"
  11659. "ldr r4, [%[a], #72]\n\t"
  11660. "ldr r5, [%[b], #72]\n\t"
  11661. "adc r4, r5\n\t"
  11662. "str r4, [%[r], #72]\n\t"
  11663. "ldr r4, [%[a], #76]\n\t"
  11664. "ldr r5, [%[b], #76]\n\t"
  11665. "adc r4, r5\n\t"
  11666. "str r4, [%[r], #76]\n\t"
  11667. "ldr r4, [%[a], #80]\n\t"
  11668. "ldr r5, [%[b], #80]\n\t"
  11669. "adc r4, r5\n\t"
  11670. "str r4, [%[r], #80]\n\t"
  11671. "ldr r4, [%[a], #84]\n\t"
  11672. "ldr r5, [%[b], #84]\n\t"
  11673. "adc r4, r5\n\t"
  11674. "str r4, [%[r], #84]\n\t"
  11675. "ldr r4, [%[a], #88]\n\t"
  11676. "ldr r5, [%[b], #88]\n\t"
  11677. "adc r4, r5\n\t"
  11678. "str r4, [%[r], #88]\n\t"
  11679. "ldr r4, [%[a], #92]\n\t"
  11680. "ldr r5, [%[b], #92]\n\t"
  11681. "adc r4, r5\n\t"
  11682. "str r4, [%[r], #92]\n\t"
  11683. "ldr r4, [%[a], #96]\n\t"
  11684. "ldr r5, [%[b], #96]\n\t"
  11685. "adc r4, r5\n\t"
  11686. "str r4, [%[r], #96]\n\t"
  11687. "ldr r4, [%[a], #100]\n\t"
  11688. "ldr r5, [%[b], #100]\n\t"
  11689. "adc r4, r5\n\t"
  11690. "str r4, [%[r], #100]\n\t"
  11691. "ldr r4, [%[a], #104]\n\t"
  11692. "ldr r5, [%[b], #104]\n\t"
  11693. "adc r4, r5\n\t"
  11694. "str r4, [%[r], #104]\n\t"
  11695. "ldr r4, [%[a], #108]\n\t"
  11696. "ldr r5, [%[b], #108]\n\t"
  11697. "adc r4, r5\n\t"
  11698. "str r4, [%[r], #108]\n\t"
  11699. "ldr r4, [%[a], #112]\n\t"
  11700. "ldr r5, [%[b], #112]\n\t"
  11701. "adc r4, r5\n\t"
  11702. "str r4, [%[r], #112]\n\t"
  11703. "ldr r4, [%[a], #116]\n\t"
  11704. "ldr r5, [%[b], #116]\n\t"
  11705. "adc r4, r5\n\t"
  11706. "str r4, [%[r], #116]\n\t"
  11707. "ldr r4, [%[a], #120]\n\t"
  11708. "ldr r5, [%[b], #120]\n\t"
  11709. "adc r4, r5\n\t"
  11710. "str r4, [%[r], #120]\n\t"
  11711. "ldr r4, [%[a], #124]\n\t"
  11712. "ldr r5, [%[b], #124]\n\t"
  11713. "adc r4, r5\n\t"
  11714. "str r4, [%[r], #124]\n\t"
  11715. "mov %[c], #0\n\t"
  11716. "adc %[c], %[c]\n\t"
  11717. "add %[a], #0x80\n\t"
  11718. "add %[b], #0x80\n\t"
  11719. "add %[r], #0x80\n\t"
  11720. "add %[c], r7\n\t"
  11721. "ldr r4, [%[a], #0]\n\t"
  11722. "ldr r5, [%[b], #0]\n\t"
  11723. "adc r4, r5\n\t"
  11724. "str r4, [%[r], #0]\n\t"
  11725. "ldr r4, [%[a], #4]\n\t"
  11726. "ldr r5, [%[b], #4]\n\t"
  11727. "adc r4, r5\n\t"
  11728. "str r4, [%[r], #4]\n\t"
  11729. "ldr r4, [%[a], #8]\n\t"
  11730. "ldr r5, [%[b], #8]\n\t"
  11731. "adc r4, r5\n\t"
  11732. "str r4, [%[r], #8]\n\t"
  11733. "ldr r4, [%[a], #12]\n\t"
  11734. "ldr r5, [%[b], #12]\n\t"
  11735. "adc r4, r5\n\t"
  11736. "str r4, [%[r], #12]\n\t"
  11737. "ldr r4, [%[a], #16]\n\t"
  11738. "ldr r5, [%[b], #16]\n\t"
  11739. "adc r4, r5\n\t"
  11740. "str r4, [%[r], #16]\n\t"
  11741. "ldr r4, [%[a], #20]\n\t"
  11742. "ldr r5, [%[b], #20]\n\t"
  11743. "adc r4, r5\n\t"
  11744. "str r4, [%[r], #20]\n\t"
  11745. "ldr r4, [%[a], #24]\n\t"
  11746. "ldr r5, [%[b], #24]\n\t"
  11747. "adc r4, r5\n\t"
  11748. "str r4, [%[r], #24]\n\t"
  11749. "ldr r4, [%[a], #28]\n\t"
  11750. "ldr r5, [%[b], #28]\n\t"
  11751. "adc r4, r5\n\t"
  11752. "str r4, [%[r], #28]\n\t"
  11753. "ldr r4, [%[a], #32]\n\t"
  11754. "ldr r5, [%[b], #32]\n\t"
  11755. "adc r4, r5\n\t"
  11756. "str r4, [%[r], #32]\n\t"
  11757. "ldr r4, [%[a], #36]\n\t"
  11758. "ldr r5, [%[b], #36]\n\t"
  11759. "adc r4, r5\n\t"
  11760. "str r4, [%[r], #36]\n\t"
  11761. "ldr r4, [%[a], #40]\n\t"
  11762. "ldr r5, [%[b], #40]\n\t"
  11763. "adc r4, r5\n\t"
  11764. "str r4, [%[r], #40]\n\t"
  11765. "ldr r4, [%[a], #44]\n\t"
  11766. "ldr r5, [%[b], #44]\n\t"
  11767. "adc r4, r5\n\t"
  11768. "str r4, [%[r], #44]\n\t"
  11769. "ldr r4, [%[a], #48]\n\t"
  11770. "ldr r5, [%[b], #48]\n\t"
  11771. "adc r4, r5\n\t"
  11772. "str r4, [%[r], #48]\n\t"
  11773. "ldr r4, [%[a], #52]\n\t"
  11774. "ldr r5, [%[b], #52]\n\t"
  11775. "adc r4, r5\n\t"
  11776. "str r4, [%[r], #52]\n\t"
  11777. "ldr r4, [%[a], #56]\n\t"
  11778. "ldr r5, [%[b], #56]\n\t"
  11779. "adc r4, r5\n\t"
  11780. "str r4, [%[r], #56]\n\t"
  11781. "ldr r4, [%[a], #60]\n\t"
  11782. "ldr r5, [%[b], #60]\n\t"
  11783. "adc r4, r5\n\t"
  11784. "str r4, [%[r], #60]\n\t"
  11785. "ldr r4, [%[a], #64]\n\t"
  11786. "ldr r5, [%[b], #64]\n\t"
  11787. "adc r4, r5\n\t"
  11788. "str r4, [%[r], #64]\n\t"
  11789. "ldr r4, [%[a], #68]\n\t"
  11790. "ldr r5, [%[b], #68]\n\t"
  11791. "adc r4, r5\n\t"
  11792. "str r4, [%[r], #68]\n\t"
  11793. "ldr r4, [%[a], #72]\n\t"
  11794. "ldr r5, [%[b], #72]\n\t"
  11795. "adc r4, r5\n\t"
  11796. "str r4, [%[r], #72]\n\t"
  11797. "ldr r4, [%[a], #76]\n\t"
  11798. "ldr r5, [%[b], #76]\n\t"
  11799. "adc r4, r5\n\t"
  11800. "str r4, [%[r], #76]\n\t"
  11801. "ldr r4, [%[a], #80]\n\t"
  11802. "ldr r5, [%[b], #80]\n\t"
  11803. "adc r4, r5\n\t"
  11804. "str r4, [%[r], #80]\n\t"
  11805. "ldr r4, [%[a], #84]\n\t"
  11806. "ldr r5, [%[b], #84]\n\t"
  11807. "adc r4, r5\n\t"
  11808. "str r4, [%[r], #84]\n\t"
  11809. "ldr r4, [%[a], #88]\n\t"
  11810. "ldr r5, [%[b], #88]\n\t"
  11811. "adc r4, r5\n\t"
  11812. "str r4, [%[r], #88]\n\t"
  11813. "ldr r4, [%[a], #92]\n\t"
  11814. "ldr r5, [%[b], #92]\n\t"
  11815. "adc r4, r5\n\t"
  11816. "str r4, [%[r], #92]\n\t"
  11817. "ldr r4, [%[a], #96]\n\t"
  11818. "ldr r5, [%[b], #96]\n\t"
  11819. "adc r4, r5\n\t"
  11820. "str r4, [%[r], #96]\n\t"
  11821. "ldr r4, [%[a], #100]\n\t"
  11822. "ldr r5, [%[b], #100]\n\t"
  11823. "adc r4, r5\n\t"
  11824. "str r4, [%[r], #100]\n\t"
  11825. "ldr r4, [%[a], #104]\n\t"
  11826. "ldr r5, [%[b], #104]\n\t"
  11827. "adc r4, r5\n\t"
  11828. "str r4, [%[r], #104]\n\t"
  11829. "ldr r4, [%[a], #108]\n\t"
  11830. "ldr r5, [%[b], #108]\n\t"
  11831. "adc r4, r5\n\t"
  11832. "str r4, [%[r], #108]\n\t"
  11833. "ldr r4, [%[a], #112]\n\t"
  11834. "ldr r5, [%[b], #112]\n\t"
  11835. "adc r4, r5\n\t"
  11836. "str r4, [%[r], #112]\n\t"
  11837. "ldr r4, [%[a], #116]\n\t"
  11838. "ldr r5, [%[b], #116]\n\t"
  11839. "adc r4, r5\n\t"
  11840. "str r4, [%[r], #116]\n\t"
  11841. "ldr r4, [%[a], #120]\n\t"
  11842. "ldr r5, [%[b], #120]\n\t"
  11843. "adc r4, r5\n\t"
  11844. "str r4, [%[r], #120]\n\t"
  11845. "ldr r4, [%[a], #124]\n\t"
  11846. "ldr r5, [%[b], #124]\n\t"
  11847. "adc r4, r5\n\t"
  11848. "str r4, [%[r], #124]\n\t"
  11849. "mov %[c], #0\n\t"
  11850. "adc %[c], %[c]\n\t"
  11851. "add %[a], #0x80\n\t"
  11852. "add %[b], #0x80\n\t"
  11853. "add %[r], #0x80\n\t"
  11854. "add %[c], r7\n\t"
  11855. "ldr r4, [%[a], #0]\n\t"
  11856. "ldr r5, [%[b], #0]\n\t"
  11857. "adc r4, r5\n\t"
  11858. "str r4, [%[r], #0]\n\t"
  11859. "ldr r4, [%[a], #4]\n\t"
  11860. "ldr r5, [%[b], #4]\n\t"
  11861. "adc r4, r5\n\t"
  11862. "str r4, [%[r], #4]\n\t"
  11863. "ldr r4, [%[a], #8]\n\t"
  11864. "ldr r5, [%[b], #8]\n\t"
  11865. "adc r4, r5\n\t"
  11866. "str r4, [%[r], #8]\n\t"
  11867. "ldr r4, [%[a], #12]\n\t"
  11868. "ldr r5, [%[b], #12]\n\t"
  11869. "adc r4, r5\n\t"
  11870. "str r4, [%[r], #12]\n\t"
  11871. "ldr r4, [%[a], #16]\n\t"
  11872. "ldr r5, [%[b], #16]\n\t"
  11873. "adc r4, r5\n\t"
  11874. "str r4, [%[r], #16]\n\t"
  11875. "ldr r4, [%[a], #20]\n\t"
  11876. "ldr r5, [%[b], #20]\n\t"
  11877. "adc r4, r5\n\t"
  11878. "str r4, [%[r], #20]\n\t"
  11879. "ldr r4, [%[a], #24]\n\t"
  11880. "ldr r5, [%[b], #24]\n\t"
  11881. "adc r4, r5\n\t"
  11882. "str r4, [%[r], #24]\n\t"
  11883. "ldr r4, [%[a], #28]\n\t"
  11884. "ldr r5, [%[b], #28]\n\t"
  11885. "adc r4, r5\n\t"
  11886. "str r4, [%[r], #28]\n\t"
  11887. "ldr r4, [%[a], #32]\n\t"
  11888. "ldr r5, [%[b], #32]\n\t"
  11889. "adc r4, r5\n\t"
  11890. "str r4, [%[r], #32]\n\t"
  11891. "ldr r4, [%[a], #36]\n\t"
  11892. "ldr r5, [%[b], #36]\n\t"
  11893. "adc r4, r5\n\t"
  11894. "str r4, [%[r], #36]\n\t"
  11895. "ldr r4, [%[a], #40]\n\t"
  11896. "ldr r5, [%[b], #40]\n\t"
  11897. "adc r4, r5\n\t"
  11898. "str r4, [%[r], #40]\n\t"
  11899. "ldr r4, [%[a], #44]\n\t"
  11900. "ldr r5, [%[b], #44]\n\t"
  11901. "adc r4, r5\n\t"
  11902. "str r4, [%[r], #44]\n\t"
  11903. "ldr r4, [%[a], #48]\n\t"
  11904. "ldr r5, [%[b], #48]\n\t"
  11905. "adc r4, r5\n\t"
  11906. "str r4, [%[r], #48]\n\t"
  11907. "ldr r4, [%[a], #52]\n\t"
  11908. "ldr r5, [%[b], #52]\n\t"
  11909. "adc r4, r5\n\t"
  11910. "str r4, [%[r], #52]\n\t"
  11911. "ldr r4, [%[a], #56]\n\t"
  11912. "ldr r5, [%[b], #56]\n\t"
  11913. "adc r4, r5\n\t"
  11914. "str r4, [%[r], #56]\n\t"
  11915. "ldr r4, [%[a], #60]\n\t"
  11916. "ldr r5, [%[b], #60]\n\t"
  11917. "adc r4, r5\n\t"
  11918. "str r4, [%[r], #60]\n\t"
  11919. "ldr r4, [%[a], #64]\n\t"
  11920. "ldr r5, [%[b], #64]\n\t"
  11921. "adc r4, r5\n\t"
  11922. "str r4, [%[r], #64]\n\t"
  11923. "ldr r4, [%[a], #68]\n\t"
  11924. "ldr r5, [%[b], #68]\n\t"
  11925. "adc r4, r5\n\t"
  11926. "str r4, [%[r], #68]\n\t"
  11927. "ldr r4, [%[a], #72]\n\t"
  11928. "ldr r5, [%[b], #72]\n\t"
  11929. "adc r4, r5\n\t"
  11930. "str r4, [%[r], #72]\n\t"
  11931. "ldr r4, [%[a], #76]\n\t"
  11932. "ldr r5, [%[b], #76]\n\t"
  11933. "adc r4, r5\n\t"
  11934. "str r4, [%[r], #76]\n\t"
  11935. "ldr r4, [%[a], #80]\n\t"
  11936. "ldr r5, [%[b], #80]\n\t"
  11937. "adc r4, r5\n\t"
  11938. "str r4, [%[r], #80]\n\t"
  11939. "ldr r4, [%[a], #84]\n\t"
  11940. "ldr r5, [%[b], #84]\n\t"
  11941. "adc r4, r5\n\t"
  11942. "str r4, [%[r], #84]\n\t"
  11943. "ldr r4, [%[a], #88]\n\t"
  11944. "ldr r5, [%[b], #88]\n\t"
  11945. "adc r4, r5\n\t"
  11946. "str r4, [%[r], #88]\n\t"
  11947. "ldr r4, [%[a], #92]\n\t"
  11948. "ldr r5, [%[b], #92]\n\t"
  11949. "adc r4, r5\n\t"
  11950. "str r4, [%[r], #92]\n\t"
  11951. "ldr r4, [%[a], #96]\n\t"
  11952. "ldr r5, [%[b], #96]\n\t"
  11953. "adc r4, r5\n\t"
  11954. "str r4, [%[r], #96]\n\t"
  11955. "ldr r4, [%[a], #100]\n\t"
  11956. "ldr r5, [%[b], #100]\n\t"
  11957. "adc r4, r5\n\t"
  11958. "str r4, [%[r], #100]\n\t"
  11959. "ldr r4, [%[a], #104]\n\t"
  11960. "ldr r5, [%[b], #104]\n\t"
  11961. "adc r4, r5\n\t"
  11962. "str r4, [%[r], #104]\n\t"
  11963. "ldr r4, [%[a], #108]\n\t"
  11964. "ldr r5, [%[b], #108]\n\t"
  11965. "adc r4, r5\n\t"
  11966. "str r4, [%[r], #108]\n\t"
  11967. "ldr r4, [%[a], #112]\n\t"
  11968. "ldr r5, [%[b], #112]\n\t"
  11969. "adc r4, r5\n\t"
  11970. "str r4, [%[r], #112]\n\t"
  11971. "ldr r4, [%[a], #116]\n\t"
  11972. "ldr r5, [%[b], #116]\n\t"
  11973. "adc r4, r5\n\t"
  11974. "str r4, [%[r], #116]\n\t"
  11975. "ldr r4, [%[a], #120]\n\t"
  11976. "ldr r5, [%[b], #120]\n\t"
  11977. "adc r4, r5\n\t"
  11978. "str r4, [%[r], #120]\n\t"
  11979. "ldr r4, [%[a], #124]\n\t"
  11980. "ldr r5, [%[b], #124]\n\t"
  11981. "adc r4, r5\n\t"
  11982. "str r4, [%[r], #124]\n\t"
  11983. "mov %[c], #0\n\t"
  11984. "adc %[c], %[c]\n\t"
  11985. "add %[a], #0x80\n\t"
  11986. "add %[b], #0x80\n\t"
  11987. "add %[r], #0x80\n\t"
  11988. "add %[c], r7\n\t"
  11989. "ldr r4, [%[a], #0]\n\t"
  11990. "ldr r5, [%[b], #0]\n\t"
  11991. "adc r4, r5\n\t"
  11992. "str r4, [%[r], #0]\n\t"
  11993. "ldr r4, [%[a], #4]\n\t"
  11994. "ldr r5, [%[b], #4]\n\t"
  11995. "adc r4, r5\n\t"
  11996. "str r4, [%[r], #4]\n\t"
  11997. "ldr r4, [%[a], #8]\n\t"
  11998. "ldr r5, [%[b], #8]\n\t"
  11999. "adc r4, r5\n\t"
  12000. "str r4, [%[r], #8]\n\t"
  12001. "ldr r4, [%[a], #12]\n\t"
  12002. "ldr r5, [%[b], #12]\n\t"
  12003. "adc r4, r5\n\t"
  12004. "str r4, [%[r], #12]\n\t"
  12005. "ldr r4, [%[a], #16]\n\t"
  12006. "ldr r5, [%[b], #16]\n\t"
  12007. "adc r4, r5\n\t"
  12008. "str r4, [%[r], #16]\n\t"
  12009. "ldr r4, [%[a], #20]\n\t"
  12010. "ldr r5, [%[b], #20]\n\t"
  12011. "adc r4, r5\n\t"
  12012. "str r4, [%[r], #20]\n\t"
  12013. "ldr r4, [%[a], #24]\n\t"
  12014. "ldr r5, [%[b], #24]\n\t"
  12015. "adc r4, r5\n\t"
  12016. "str r4, [%[r], #24]\n\t"
  12017. "ldr r4, [%[a], #28]\n\t"
  12018. "ldr r5, [%[b], #28]\n\t"
  12019. "adc r4, r5\n\t"
  12020. "str r4, [%[r], #28]\n\t"
  12021. "ldr r4, [%[a], #32]\n\t"
  12022. "ldr r5, [%[b], #32]\n\t"
  12023. "adc r4, r5\n\t"
  12024. "str r4, [%[r], #32]\n\t"
  12025. "ldr r4, [%[a], #36]\n\t"
  12026. "ldr r5, [%[b], #36]\n\t"
  12027. "adc r4, r5\n\t"
  12028. "str r4, [%[r], #36]\n\t"
  12029. "ldr r4, [%[a], #40]\n\t"
  12030. "ldr r5, [%[b], #40]\n\t"
  12031. "adc r4, r5\n\t"
  12032. "str r4, [%[r], #40]\n\t"
  12033. "ldr r4, [%[a], #44]\n\t"
  12034. "ldr r5, [%[b], #44]\n\t"
  12035. "adc r4, r5\n\t"
  12036. "str r4, [%[r], #44]\n\t"
  12037. "ldr r4, [%[a], #48]\n\t"
  12038. "ldr r5, [%[b], #48]\n\t"
  12039. "adc r4, r5\n\t"
  12040. "str r4, [%[r], #48]\n\t"
  12041. "ldr r4, [%[a], #52]\n\t"
  12042. "ldr r5, [%[b], #52]\n\t"
  12043. "adc r4, r5\n\t"
  12044. "str r4, [%[r], #52]\n\t"
  12045. "ldr r4, [%[a], #56]\n\t"
  12046. "ldr r5, [%[b], #56]\n\t"
  12047. "adc r4, r5\n\t"
  12048. "str r4, [%[r], #56]\n\t"
  12049. "ldr r4, [%[a], #60]\n\t"
  12050. "ldr r5, [%[b], #60]\n\t"
  12051. "adc r4, r5\n\t"
  12052. "str r4, [%[r], #60]\n\t"
  12053. "ldr r4, [%[a], #64]\n\t"
  12054. "ldr r5, [%[b], #64]\n\t"
  12055. "adc r4, r5\n\t"
  12056. "str r4, [%[r], #64]\n\t"
  12057. "ldr r4, [%[a], #68]\n\t"
  12058. "ldr r5, [%[b], #68]\n\t"
  12059. "adc r4, r5\n\t"
  12060. "str r4, [%[r], #68]\n\t"
  12061. "ldr r4, [%[a], #72]\n\t"
  12062. "ldr r5, [%[b], #72]\n\t"
  12063. "adc r4, r5\n\t"
  12064. "str r4, [%[r], #72]\n\t"
  12065. "ldr r4, [%[a], #76]\n\t"
  12066. "ldr r5, [%[b], #76]\n\t"
  12067. "adc r4, r5\n\t"
  12068. "str r4, [%[r], #76]\n\t"
  12069. "ldr r4, [%[a], #80]\n\t"
  12070. "ldr r5, [%[b], #80]\n\t"
  12071. "adc r4, r5\n\t"
  12072. "str r4, [%[r], #80]\n\t"
  12073. "ldr r4, [%[a], #84]\n\t"
  12074. "ldr r5, [%[b], #84]\n\t"
  12075. "adc r4, r5\n\t"
  12076. "str r4, [%[r], #84]\n\t"
  12077. "ldr r4, [%[a], #88]\n\t"
  12078. "ldr r5, [%[b], #88]\n\t"
  12079. "adc r4, r5\n\t"
  12080. "str r4, [%[r], #88]\n\t"
  12081. "ldr r4, [%[a], #92]\n\t"
  12082. "ldr r5, [%[b], #92]\n\t"
  12083. "adc r4, r5\n\t"
  12084. "str r4, [%[r], #92]\n\t"
  12085. "ldr r4, [%[a], #96]\n\t"
  12086. "ldr r5, [%[b], #96]\n\t"
  12087. "adc r4, r5\n\t"
  12088. "str r4, [%[r], #96]\n\t"
  12089. "ldr r4, [%[a], #100]\n\t"
  12090. "ldr r5, [%[b], #100]\n\t"
  12091. "adc r4, r5\n\t"
  12092. "str r4, [%[r], #100]\n\t"
  12093. "ldr r4, [%[a], #104]\n\t"
  12094. "ldr r5, [%[b], #104]\n\t"
  12095. "adc r4, r5\n\t"
  12096. "str r4, [%[r], #104]\n\t"
  12097. "ldr r4, [%[a], #108]\n\t"
  12098. "ldr r5, [%[b], #108]\n\t"
  12099. "adc r4, r5\n\t"
  12100. "str r4, [%[r], #108]\n\t"
  12101. "ldr r4, [%[a], #112]\n\t"
  12102. "ldr r5, [%[b], #112]\n\t"
  12103. "adc r4, r5\n\t"
  12104. "str r4, [%[r], #112]\n\t"
  12105. "ldr r4, [%[a], #116]\n\t"
  12106. "ldr r5, [%[b], #116]\n\t"
  12107. "adc r4, r5\n\t"
  12108. "str r4, [%[r], #116]\n\t"
  12109. "ldr r4, [%[a], #120]\n\t"
  12110. "ldr r5, [%[b], #120]\n\t"
  12111. "adc r4, r5\n\t"
  12112. "str r4, [%[r], #120]\n\t"
  12113. "ldr r4, [%[a], #124]\n\t"
  12114. "ldr r5, [%[b], #124]\n\t"
  12115. "adc r4, r5\n\t"
  12116. "str r4, [%[r], #124]\n\t"
  12117. "mov %[c], #0\n\t"
  12118. "adc %[c], %[c]\n\t"
  12119. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  12120. :
  12121. : "memory", "r4", "r5", "r7"
  12122. );
  12123. return c;
  12124. }
  12125. /* Multiply a and b into r. (r = a * b)
  12126. *
  12127. * r A single precision integer.
  12128. * a A single precision integer.
  12129. * b A single precision integer.
  12130. */
  12131. SP_NOINLINE static void sp_4096_mul_64(sp_digit* r, const sp_digit* a,
  12132. const sp_digit* b)
  12133. {
  12134. sp_digit tmp[64 * 2];
  12135. __asm__ __volatile__ (
  12136. "mov r3, #0\n\t"
  12137. "mov r4, #0\n\t"
  12138. "mov r8, r3\n\t"
  12139. "mov r11, %[r]\n\t"
  12140. "mov r9, %[a]\n\t"
  12141. "mov r10, %[b]\n\t"
  12142. "mov r6, #1\n\t"
  12143. "lsl r6, r6, #8\n\t"
  12144. "add r6, r9\n\t"
  12145. "mov r12, r6\n\t"
  12146. "\n1:\n\t"
  12147. "mov %[r], #0\n\t"
  12148. "mov r5, #0\n\t"
  12149. "mov r6, #252\n\t"
  12150. "mov %[a], r8\n\t"
  12151. "sub %[a], r6\n\t"
  12152. "sbc r6, r6\n\t"
  12153. "mvn r6, r6\n\t"
  12154. "and %[a], r6\n\t"
  12155. "mov %[b], r8\n\t"
  12156. "sub %[b], %[a]\n\t"
  12157. "add %[a], r9\n\t"
  12158. "add %[b], r10\n\t"
  12159. "\n2:\n\t"
  12160. "# Multiply Start\n\t"
  12161. "ldr r6, [%[a]]\n\t"
  12162. "ldr r7, [%[b]]\n\t"
  12163. "lsl r6, r6, #16\n\t"
  12164. "lsl r7, r7, #16\n\t"
  12165. "lsr r6, r6, #16\n\t"
  12166. "lsr r7, r7, #16\n\t"
  12167. "mul r7, r6\n\t"
  12168. "add r3, r7\n\t"
  12169. "adc r4, %[r]\n\t"
  12170. "adc r5, %[r]\n\t"
  12171. "ldr r7, [%[b]]\n\t"
  12172. "lsr r7, r7, #16\n\t"
  12173. "mul r6, r7\n\t"
  12174. "lsr r7, r6, #16\n\t"
  12175. "lsl r6, r6, #16\n\t"
  12176. "add r3, r6\n\t"
  12177. "adc r4, r7\n\t"
  12178. "adc r5, %[r]\n\t"
  12179. "ldr r6, [%[a]]\n\t"
  12180. "ldr r7, [%[b]]\n\t"
  12181. "lsr r6, r6, #16\n\t"
  12182. "lsr r7, r7, #16\n\t"
  12183. "mul r7, r6\n\t"
  12184. "add r4, r7\n\t"
  12185. "adc r5, %[r]\n\t"
  12186. "ldr r7, [%[b]]\n\t"
  12187. "lsl r7, r7, #16\n\t"
  12188. "lsr r7, r7, #16\n\t"
  12189. "mul r6, r7\n\t"
  12190. "lsr r7, r6, #16\n\t"
  12191. "lsl r6, r6, #16\n\t"
  12192. "add r3, r6\n\t"
  12193. "adc r4, r7\n\t"
  12194. "adc r5, %[r]\n\t"
  12195. "# Multiply Done\n\t"
  12196. "add %[a], #4\n\t"
  12197. "sub %[b], #4\n\t"
  12198. "cmp %[a], r12\n\t"
  12199. "beq 3f\n\t"
  12200. "mov r6, r8\n\t"
  12201. "add r6, r9\n\t"
  12202. "cmp %[a], r6\n\t"
  12203. "ble 2b\n\t"
  12204. "\n3:\n\t"
  12205. "mov %[r], r11\n\t"
  12206. "mov r7, r8\n\t"
  12207. "str r3, [%[r], r7]\n\t"
  12208. "mov r3, r4\n\t"
  12209. "mov r4, r5\n\t"
  12210. "add r7, #4\n\t"
  12211. "mov r8, r7\n\t"
  12212. "mov r6, #1\n\t"
  12213. "lsl r6, r6, #8\n\t"
  12214. "add r6, #248\n\t"
  12215. "cmp r7, r6\n\t"
  12216. "ble 1b\n\t"
  12217. "str r3, [%[r], r7]\n\t"
  12218. "mov %[a], r9\n\t"
  12219. "mov %[b], r10\n\t"
  12220. :
  12221. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  12222. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  12223. );
  12224. XMEMCPY(r, tmp, sizeof(tmp));
  12225. }
  12226. /* AND m into each word of a and store in r.
  12227. *
  12228. * r A single precision integer.
  12229. * a A single precision integer.
  12230. * m Mask to AND against each digit.
  12231. */
  12232. static void sp_4096_mask_64(sp_digit* r, const sp_digit* a, sp_digit m)
  12233. {
  12234. #ifdef WOLFSSL_SP_SMALL
  12235. int i;
  12236. for (i=0; i<64; i++) {
  12237. r[i] = a[i] & m;
  12238. }
  12239. #else
  12240. int i;
  12241. for (i = 0; i < 64; i += 8) {
  12242. r[i+0] = a[i+0] & m;
  12243. r[i+1] = a[i+1] & m;
  12244. r[i+2] = a[i+2] & m;
  12245. r[i+3] = a[i+3] & m;
  12246. r[i+4] = a[i+4] & m;
  12247. r[i+5] = a[i+5] & m;
  12248. r[i+6] = a[i+6] & m;
  12249. r[i+7] = a[i+7] & m;
  12250. }
  12251. #endif
  12252. }
  12253. /* Multiply a and b into r. (r = a * b)
  12254. *
  12255. * r A single precision integer.
  12256. * a A single precision integer.
  12257. * b A single precision integer.
  12258. */
  12259. SP_NOINLINE static void sp_4096_mul_128(sp_digit* r, const sp_digit* a,
  12260. const sp_digit* b)
  12261. {
  12262. sp_digit* z0 = r;
  12263. sp_digit z1[128];
  12264. sp_digit a1[64];
  12265. sp_digit b1[64];
  12266. sp_digit z2[128];
  12267. sp_digit u, ca, cb;
  12268. ca = sp_2048_add_64(a1, a, &a[64]);
  12269. cb = sp_2048_add_64(b1, b, &b[64]);
  12270. u = ca & cb;
  12271. sp_2048_mul_64(z1, a1, b1);
  12272. sp_2048_mul_64(z2, &a[64], &b[64]);
  12273. sp_2048_mul_64(z0, a, b);
  12274. sp_2048_mask_64(r + 128, a1, 0 - cb);
  12275. sp_2048_mask_64(b1, b1, 0 - ca);
  12276. u += sp_2048_add_64(r + 128, r + 128, b1);
  12277. u += sp_4096_sub_in_place_128(z1, z2);
  12278. u += sp_4096_sub_in_place_128(z1, z0);
  12279. u += sp_4096_add_128(r + 64, r + 64, z1);
  12280. r[192] = u;
  12281. XMEMSET(r + 192 + 1, 0, sizeof(sp_digit) * (64 - 1));
  12282. (void)sp_4096_add_128(r + 128, r + 128, z2);
  12283. }
  12284. /* Square a and put result in r. (r = a * a)
  12285. *
  12286. * r A single precision integer.
  12287. * a A single precision integer.
  12288. */
  12289. SP_NOINLINE static void sp_4096_sqr_64(sp_digit* r, const sp_digit* a)
  12290. {
  12291. __asm__ __volatile__ (
  12292. "mov r3, #0\n\t"
  12293. "mov r4, #0\n\t"
  12294. "mov r5, #0\n\t"
  12295. "mov r8, r3\n\t"
  12296. "mov r11, %[r]\n\t"
  12297. "mov r6, #2\n\t"
  12298. "lsl r6, r6, #8\n\t"
  12299. "neg r6, r6\n\t"
  12300. "add sp, r6\n\t"
  12301. "mov r10, sp\n\t"
  12302. "mov r9, %[a]\n\t"
  12303. "\n1:\n\t"
  12304. "mov %[r], #0\n\t"
  12305. "mov r6, #252\n\t"
  12306. "mov %[a], r8\n\t"
  12307. "sub %[a], r6\n\t"
  12308. "sbc r6, r6\n\t"
  12309. "mvn r6, r6\n\t"
  12310. "and %[a], r6\n\t"
  12311. "mov r2, r8\n\t"
  12312. "sub r2, %[a]\n\t"
  12313. "add %[a], r9\n\t"
  12314. "add r2, r9\n\t"
  12315. "\n2:\n\t"
  12316. "cmp r2, %[a]\n\t"
  12317. "beq 4f\n\t"
  12318. "# Multiply * 2: Start\n\t"
  12319. "ldr r6, [%[a]]\n\t"
  12320. "ldr r7, [r2]\n\t"
  12321. "lsl r6, r6, #16\n\t"
  12322. "lsl r7, r7, #16\n\t"
  12323. "lsr r6, r6, #16\n\t"
  12324. "lsr r7, r7, #16\n\t"
  12325. "mul r7, r6\n\t"
  12326. "add r3, r7\n\t"
  12327. "adc r4, %[r]\n\t"
  12328. "adc r5, %[r]\n\t"
  12329. "add r3, r7\n\t"
  12330. "adc r4, %[r]\n\t"
  12331. "adc r5, %[r]\n\t"
  12332. "ldr r7, [r2]\n\t"
  12333. "lsr r7, r7, #16\n\t"
  12334. "mul r6, r7\n\t"
  12335. "lsr r7, r6, #16\n\t"
  12336. "lsl r6, r6, #16\n\t"
  12337. "add r3, r6\n\t"
  12338. "adc r4, r7\n\t"
  12339. "adc r5, %[r]\n\t"
  12340. "add r3, r6\n\t"
  12341. "adc r4, r7\n\t"
  12342. "adc r5, %[r]\n\t"
  12343. "ldr r6, [%[a]]\n\t"
  12344. "ldr r7, [r2]\n\t"
  12345. "lsr r6, r6, #16\n\t"
  12346. "lsr r7, r7, #16\n\t"
  12347. "mul r7, r6\n\t"
  12348. "add r4, r7\n\t"
  12349. "adc r5, %[r]\n\t"
  12350. "add r4, r7\n\t"
  12351. "adc r5, %[r]\n\t"
  12352. "ldr r7, [r2]\n\t"
  12353. "lsl r7, r7, #16\n\t"
  12354. "lsr r7, r7, #16\n\t"
  12355. "mul r6, r7\n\t"
  12356. "lsr r7, r6, #16\n\t"
  12357. "lsl r6, r6, #16\n\t"
  12358. "add r3, r6\n\t"
  12359. "adc r4, r7\n\t"
  12360. "adc r5, %[r]\n\t"
  12361. "add r3, r6\n\t"
  12362. "adc r4, r7\n\t"
  12363. "adc r5, %[r]\n\t"
  12364. "# Multiply * 2: Done\n\t"
  12365. "bal 5f\n\t"
  12366. "\n4:\n\t"
  12367. "# Square: Start\n\t"
  12368. "ldr r6, [%[a]]\n\t"
  12369. "lsr r7, r6, #16\n\t"
  12370. "lsl r6, r6, #16\n\t"
  12371. "lsr r6, r6, #16\n\t"
  12372. "mul r6, r6\n\t"
  12373. "add r3, r6\n\t"
  12374. "adc r4, %[r]\n\t"
  12375. "adc r5, %[r]\n\t"
  12376. "mul r7, r7\n\t"
  12377. "add r4, r7\n\t"
  12378. "adc r5, %[r]\n\t"
  12379. "ldr r6, [%[a]]\n\t"
  12380. "lsr r7, r6, #16\n\t"
  12381. "lsl r6, r6, #16\n\t"
  12382. "lsr r6, r6, #16\n\t"
  12383. "mul r6, r7\n\t"
  12384. "lsr r7, r6, #15\n\t"
  12385. "lsl r6, r6, #17\n\t"
  12386. "add r3, r6\n\t"
  12387. "adc r4, r7\n\t"
  12388. "adc r5, %[r]\n\t"
  12389. "# Square: Done\n\t"
  12390. "\n5:\n\t"
  12391. "add %[a], #4\n\t"
  12392. "sub r2, #4\n\t"
  12393. "mov r6, #1\n\t"
  12394. "lsl r6, r6, #8\n\t"
  12395. "add r6, r9\n\t"
  12396. "cmp %[a], r6\n\t"
  12397. "beq 3f\n\t"
  12398. "cmp %[a], r2\n\t"
  12399. "bgt 3f\n\t"
  12400. "mov r7, r8\n\t"
  12401. "add r7, r9\n\t"
  12402. "cmp %[a], r7\n\t"
  12403. "ble 2b\n\t"
  12404. "\n3:\n\t"
  12405. "mov %[r], r10\n\t"
  12406. "mov r7, r8\n\t"
  12407. "str r3, [%[r], r7]\n\t"
  12408. "mov r3, r4\n\t"
  12409. "mov r4, r5\n\t"
  12410. "mov r5, #0\n\t"
  12411. "add r7, #4\n\t"
  12412. "mov r8, r7\n\t"
  12413. "mov r6, #1\n\t"
  12414. "lsl r6, r6, #8\n\t"
  12415. "add r6, #248\n\t"
  12416. "cmp r7, r6\n\t"
  12417. "ble 1b\n\t"
  12418. "mov %[a], r9\n\t"
  12419. "str r3, [%[r], r7]\n\t"
  12420. "mov %[r], r11\n\t"
  12421. "mov %[a], r10\n\t"
  12422. "mov r3, #1\n\t"
  12423. "lsl r3, r3, #8\n\t"
  12424. "add r3, #252\n\t"
  12425. "\n4:\n\t"
  12426. "ldr r6, [%[a], r3]\n\t"
  12427. "str r6, [%[r], r3]\n\t"
  12428. "sub r3, #4\n\t"
  12429. "bge 4b\n\t"
  12430. "mov r6, #2\n\t"
  12431. "lsl r6, r6, #8\n\t"
  12432. "add sp, r6\n\t"
  12433. :
  12434. : [r] "r" (r), [a] "r" (a)
  12435. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  12436. );
  12437. }
  12438. /* Square a and put result in r. (r = a * a)
  12439. *
  12440. * r A single precision integer.
  12441. * a A single precision integer.
  12442. */
  12443. SP_NOINLINE static void sp_4096_sqr_128(sp_digit* r, const sp_digit* a)
  12444. {
  12445. sp_digit* z0 = r;
  12446. sp_digit z2[128];
  12447. sp_digit z1[128];
  12448. sp_digit a1[64];
  12449. sp_digit u;
  12450. u = sp_2048_add_64(a1, a, &a[64]);
  12451. sp_2048_sqr_64(z1, a1);
  12452. sp_2048_sqr_64(z2, &a[64]);
  12453. sp_2048_sqr_64(z0, a);
  12454. sp_2048_mask_64(r + 128, a1, 0 - u);
  12455. u += sp_2048_add_64(r + 128, r + 128, r + 128);
  12456. u += sp_4096_sub_in_place_128(z1, z2);
  12457. u += sp_4096_sub_in_place_128(z1, z0);
  12458. u += sp_4096_add_128(r + 64, r + 64, z1);
  12459. r[192] = u;
  12460. XMEMSET(r + 192 + 1, 0, sizeof(sp_digit) * (64 - 1));
  12461. (void)sp_4096_add_128(r + 128, r + 128, z2);
  12462. }
  12463. #endif /* !WOLFSSL_SP_SMALL */
  12464. #ifdef WOLFSSL_SP_SMALL
  12465. /* Add b to a into r. (r = a + b)
  12466. *
  12467. * r A single precision integer.
  12468. * a A single precision integer.
  12469. * b A single precision integer.
  12470. */
  12471. SP_NOINLINE static sp_digit sp_4096_add_128(sp_digit* r, const sp_digit* a,
  12472. const sp_digit* b)
  12473. {
  12474. sp_digit c = 0;
  12475. __asm__ __volatile__ (
  12476. "mov r6, %[a]\n\t"
  12477. "mov r7, #0\n\t"
  12478. "mov r4, #2\n\t"
  12479. "lsl r4, #8\n\t"
  12480. "sub r7, #1\n\t"
  12481. "add r6, r4\n\t"
  12482. "\n1:\n\t"
  12483. "add %[c], r7\n\t"
  12484. "ldr r4, [%[a]]\n\t"
  12485. "ldr r5, [%[b]]\n\t"
  12486. "adc r4, r5\n\t"
  12487. "str r4, [%[r]]\n\t"
  12488. "mov %[c], #0\n\t"
  12489. "adc %[c], %[c]\n\t"
  12490. "add %[a], #4\n\t"
  12491. "add %[b], #4\n\t"
  12492. "add %[r], #4\n\t"
  12493. "cmp %[a], r6\n\t"
  12494. "bne 1b\n\t"
  12495. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  12496. :
  12497. : "memory", "r4", "r5", "r6", "r7"
  12498. );
  12499. return c;
  12500. }
  12501. #endif /* WOLFSSL_SP_SMALL */
  12502. #ifdef WOLFSSL_SP_SMALL
  12503. /* Sub b from a into a. (a -= b)
  12504. *
  12505. * a A single precision integer.
  12506. * b A single precision integer.
  12507. */
  12508. SP_NOINLINE static sp_digit sp_4096_sub_in_place_128(sp_digit* a,
  12509. const sp_digit* b)
  12510. {
  12511. sp_digit c = 0;
  12512. __asm__ __volatile__ (
  12513. "mov r7, %[a]\n\t"
  12514. "mov r5, #2\n\t"
  12515. "lsl r5, #8\n\t"
  12516. "add r7, r5\n\t"
  12517. "\n1:\n\t"
  12518. "mov r5, #0\n\t"
  12519. "sub r5, %[c]\n\t"
  12520. "ldr r3, [%[a]]\n\t"
  12521. "ldr r4, [%[a], #4]\n\t"
  12522. "ldr r5, [%[b]]\n\t"
  12523. "ldr r6, [%[b], #4]\n\t"
  12524. "sbc r3, r5\n\t"
  12525. "sbc r4, r6\n\t"
  12526. "str r3, [%[a]]\n\t"
  12527. "str r4, [%[a], #4]\n\t"
  12528. "sbc %[c], %[c]\n\t"
  12529. "add %[a], #8\n\t"
  12530. "add %[b], #8\n\t"
  12531. "cmp %[a], r7\n\t"
  12532. "bne 1b\n\t"
  12533. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  12534. :
  12535. : "memory", "r3", "r4", "r5", "r6", "r7"
  12536. );
  12537. return c;
  12538. }
  12539. #endif /* WOLFSSL_SP_SMALL */
  12540. #ifdef WOLFSSL_SP_SMALL
  12541. /* Multiply a and b into r. (r = a * b)
  12542. *
  12543. * r A single precision integer.
  12544. * a A single precision integer.
  12545. * b A single precision integer.
  12546. */
  12547. SP_NOINLINE static void sp_4096_mul_128(sp_digit* r, const sp_digit* a,
  12548. const sp_digit* b)
  12549. {
  12550. sp_digit tmp[128 * 2];
  12551. __asm__ __volatile__ (
  12552. "mov r3, #0\n\t"
  12553. "mov r4, #0\n\t"
  12554. "mov r8, r3\n\t"
  12555. "mov r11, %[r]\n\t"
  12556. "mov r9, %[a]\n\t"
  12557. "mov r10, %[b]\n\t"
  12558. "mov r6, #2\n\t"
  12559. "lsl r6, r6, #8\n\t"
  12560. "add r6, r9\n\t"
  12561. "mov r12, r6\n\t"
  12562. "\n1:\n\t"
  12563. "mov %[r], #0\n\t"
  12564. "mov r5, #0\n\t"
  12565. "mov r6, #1\n\t"
  12566. "lsl r6, r6, #8\n\t"
  12567. "add r6, #252\n\t"
  12568. "mov %[a], r8\n\t"
  12569. "sub %[a], r6\n\t"
  12570. "sbc r6, r6\n\t"
  12571. "mvn r6, r6\n\t"
  12572. "and %[a], r6\n\t"
  12573. "mov %[b], r8\n\t"
  12574. "sub %[b], %[a]\n\t"
  12575. "add %[a], r9\n\t"
  12576. "add %[b], r10\n\t"
  12577. "\n2:\n\t"
  12578. "# Multiply Start\n\t"
  12579. "ldr r6, [%[a]]\n\t"
  12580. "ldr r7, [%[b]]\n\t"
  12581. "lsl r6, r6, #16\n\t"
  12582. "lsl r7, r7, #16\n\t"
  12583. "lsr r6, r6, #16\n\t"
  12584. "lsr r7, r7, #16\n\t"
  12585. "mul r7, r6\n\t"
  12586. "add r3, r7\n\t"
  12587. "adc r4, %[r]\n\t"
  12588. "adc r5, %[r]\n\t"
  12589. "ldr r7, [%[b]]\n\t"
  12590. "lsr r7, r7, #16\n\t"
  12591. "mul r6, r7\n\t"
  12592. "lsr r7, r6, #16\n\t"
  12593. "lsl r6, r6, #16\n\t"
  12594. "add r3, r6\n\t"
  12595. "adc r4, r7\n\t"
  12596. "adc r5, %[r]\n\t"
  12597. "ldr r6, [%[a]]\n\t"
  12598. "ldr r7, [%[b]]\n\t"
  12599. "lsr r6, r6, #16\n\t"
  12600. "lsr r7, r7, #16\n\t"
  12601. "mul r7, r6\n\t"
  12602. "add r4, r7\n\t"
  12603. "adc r5, %[r]\n\t"
  12604. "ldr r7, [%[b]]\n\t"
  12605. "lsl r7, r7, #16\n\t"
  12606. "lsr r7, r7, #16\n\t"
  12607. "mul r6, r7\n\t"
  12608. "lsr r7, r6, #16\n\t"
  12609. "lsl r6, r6, #16\n\t"
  12610. "add r3, r6\n\t"
  12611. "adc r4, r7\n\t"
  12612. "adc r5, %[r]\n\t"
  12613. "# Multiply Done\n\t"
  12614. "add %[a], #4\n\t"
  12615. "sub %[b], #4\n\t"
  12616. "cmp %[a], r12\n\t"
  12617. "beq 3f\n\t"
  12618. "mov r6, r8\n\t"
  12619. "add r6, r9\n\t"
  12620. "cmp %[a], r6\n\t"
  12621. "ble 2b\n\t"
  12622. "\n3:\n\t"
  12623. "mov %[r], r11\n\t"
  12624. "mov r7, r8\n\t"
  12625. "str r3, [%[r], r7]\n\t"
  12626. "mov r3, r4\n\t"
  12627. "mov r4, r5\n\t"
  12628. "add r7, #4\n\t"
  12629. "mov r8, r7\n\t"
  12630. "mov r6, #3\n\t"
  12631. "lsl r6, r6, #8\n\t"
  12632. "add r6, #248\n\t"
  12633. "cmp r7, r6\n\t"
  12634. "ble 1b\n\t"
  12635. "str r3, [%[r], r7]\n\t"
  12636. "mov %[a], r9\n\t"
  12637. "mov %[b], r10\n\t"
  12638. :
  12639. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  12640. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  12641. );
  12642. XMEMCPY(r, tmp, sizeof(tmp));
  12643. }
  12644. /* Square a and put result in r. (r = a * a)
  12645. *
  12646. * r A single precision integer.
  12647. * a A single precision integer.
  12648. */
  12649. SP_NOINLINE static void sp_4096_sqr_128(sp_digit* r, const sp_digit* a)
  12650. {
  12651. __asm__ __volatile__ (
  12652. "mov r3, #0\n\t"
  12653. "mov r4, #0\n\t"
  12654. "mov r5, #0\n\t"
  12655. "mov r8, r3\n\t"
  12656. "mov r11, %[r]\n\t"
  12657. "mov r6, #4\n\t"
  12658. "lsl r6, r6, #8\n\t"
  12659. "neg r6, r6\n\t"
  12660. "add sp, r6\n\t"
  12661. "mov r10, sp\n\t"
  12662. "mov r9, %[a]\n\t"
  12663. "\n1:\n\t"
  12664. "mov %[r], #0\n\t"
  12665. "mov r6, #1\n\t"
  12666. "lsl r6, r6, #8\n\t"
  12667. "add r6, #252\n\t"
  12668. "mov %[a], r8\n\t"
  12669. "sub %[a], r6\n\t"
  12670. "sbc r6, r6\n\t"
  12671. "mvn r6, r6\n\t"
  12672. "and %[a], r6\n\t"
  12673. "mov r2, r8\n\t"
  12674. "sub r2, %[a]\n\t"
  12675. "add %[a], r9\n\t"
  12676. "add r2, r9\n\t"
  12677. "\n2:\n\t"
  12678. "cmp r2, %[a]\n\t"
  12679. "beq 4f\n\t"
  12680. "# Multiply * 2: Start\n\t"
  12681. "ldr r6, [%[a]]\n\t"
  12682. "ldr r7, [r2]\n\t"
  12683. "lsl r6, r6, #16\n\t"
  12684. "lsl r7, r7, #16\n\t"
  12685. "lsr r6, r6, #16\n\t"
  12686. "lsr r7, r7, #16\n\t"
  12687. "mul r7, r6\n\t"
  12688. "add r3, r7\n\t"
  12689. "adc r4, %[r]\n\t"
  12690. "adc r5, %[r]\n\t"
  12691. "add r3, r7\n\t"
  12692. "adc r4, %[r]\n\t"
  12693. "adc r5, %[r]\n\t"
  12694. "ldr r7, [r2]\n\t"
  12695. "lsr r7, r7, #16\n\t"
  12696. "mul r6, r7\n\t"
  12697. "lsr r7, r6, #16\n\t"
  12698. "lsl r6, r6, #16\n\t"
  12699. "add r3, r6\n\t"
  12700. "adc r4, r7\n\t"
  12701. "adc r5, %[r]\n\t"
  12702. "add r3, r6\n\t"
  12703. "adc r4, r7\n\t"
  12704. "adc r5, %[r]\n\t"
  12705. "ldr r6, [%[a]]\n\t"
  12706. "ldr r7, [r2]\n\t"
  12707. "lsr r6, r6, #16\n\t"
  12708. "lsr r7, r7, #16\n\t"
  12709. "mul r7, r6\n\t"
  12710. "add r4, r7\n\t"
  12711. "adc r5, %[r]\n\t"
  12712. "add r4, r7\n\t"
  12713. "adc r5, %[r]\n\t"
  12714. "ldr r7, [r2]\n\t"
  12715. "lsl r7, r7, #16\n\t"
  12716. "lsr r7, r7, #16\n\t"
  12717. "mul r6, r7\n\t"
  12718. "lsr r7, r6, #16\n\t"
  12719. "lsl r6, r6, #16\n\t"
  12720. "add r3, r6\n\t"
  12721. "adc r4, r7\n\t"
  12722. "adc r5, %[r]\n\t"
  12723. "add r3, r6\n\t"
  12724. "adc r4, r7\n\t"
  12725. "adc r5, %[r]\n\t"
  12726. "# Multiply * 2: Done\n\t"
  12727. "bal 5f\n\t"
  12728. "\n4:\n\t"
  12729. "# Square: Start\n\t"
  12730. "ldr r6, [%[a]]\n\t"
  12731. "lsr r7, r6, #16\n\t"
  12732. "lsl r6, r6, #16\n\t"
  12733. "lsr r6, r6, #16\n\t"
  12734. "mul r6, r6\n\t"
  12735. "add r3, r6\n\t"
  12736. "adc r4, %[r]\n\t"
  12737. "adc r5, %[r]\n\t"
  12738. "mul r7, r7\n\t"
  12739. "add r4, r7\n\t"
  12740. "adc r5, %[r]\n\t"
  12741. "ldr r6, [%[a]]\n\t"
  12742. "lsr r7, r6, #16\n\t"
  12743. "lsl r6, r6, #16\n\t"
  12744. "lsr r6, r6, #16\n\t"
  12745. "mul r6, r7\n\t"
  12746. "lsr r7, r6, #15\n\t"
  12747. "lsl r6, r6, #17\n\t"
  12748. "add r3, r6\n\t"
  12749. "adc r4, r7\n\t"
  12750. "adc r5, %[r]\n\t"
  12751. "# Square: Done\n\t"
  12752. "\n5:\n\t"
  12753. "add %[a], #4\n\t"
  12754. "sub r2, #4\n\t"
  12755. "mov r6, #2\n\t"
  12756. "lsl r6, r6, #8\n\t"
  12757. "add r6, r9\n\t"
  12758. "cmp %[a], r6\n\t"
  12759. "beq 3f\n\t"
  12760. "cmp %[a], r2\n\t"
  12761. "bgt 3f\n\t"
  12762. "mov r7, r8\n\t"
  12763. "add r7, r9\n\t"
  12764. "cmp %[a], r7\n\t"
  12765. "ble 2b\n\t"
  12766. "\n3:\n\t"
  12767. "mov %[r], r10\n\t"
  12768. "mov r7, r8\n\t"
  12769. "str r3, [%[r], r7]\n\t"
  12770. "mov r3, r4\n\t"
  12771. "mov r4, r5\n\t"
  12772. "mov r5, #0\n\t"
  12773. "add r7, #4\n\t"
  12774. "mov r8, r7\n\t"
  12775. "mov r6, #3\n\t"
  12776. "lsl r6, r6, #8\n\t"
  12777. "add r6, #248\n\t"
  12778. "cmp r7, r6\n\t"
  12779. "ble 1b\n\t"
  12780. "mov %[a], r9\n\t"
  12781. "str r3, [%[r], r7]\n\t"
  12782. "mov %[r], r11\n\t"
  12783. "mov %[a], r10\n\t"
  12784. "mov r3, #3\n\t"
  12785. "lsl r3, r3, #8\n\t"
  12786. "add r3, #252\n\t"
  12787. "\n4:\n\t"
  12788. "ldr r6, [%[a], r3]\n\t"
  12789. "str r6, [%[r], r3]\n\t"
  12790. "sub r3, #4\n\t"
  12791. "bge 4b\n\t"
  12792. "mov r6, #4\n\t"
  12793. "lsl r6, r6, #8\n\t"
  12794. "add sp, r6\n\t"
  12795. :
  12796. : [r] "r" (r), [a] "r" (a)
  12797. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  12798. );
  12799. }
  12800. #endif /* WOLFSSL_SP_SMALL */
  12801. /* Caclulate the bottom digit of -1/a mod 2^n.
  12802. *
  12803. * a A single precision number.
  12804. * rho Bottom word of inverse.
  12805. */
  12806. static void sp_4096_mont_setup(const sp_digit* a, sp_digit* rho)
  12807. {
  12808. sp_digit x, b;
  12809. b = a[0];
  12810. x = (((b + 2) & 4) << 1) + b; /* here x*a==1 mod 2**4 */
  12811. x *= 2 - b * x; /* here x*a==1 mod 2**8 */
  12812. x *= 2 - b * x; /* here x*a==1 mod 2**16 */
  12813. x *= 2 - b * x; /* here x*a==1 mod 2**32 */
  12814. /* rho = -1/m mod b */
  12815. *rho = -x;
  12816. }
  12817. /* Mul a by digit b into r. (r = a * b)
  12818. *
  12819. * r A single precision integer.
  12820. * a A single precision integer.
  12821. * b A single precision digit.
  12822. */
  12823. SP_NOINLINE static void sp_4096_mul_d_128(sp_digit* r, const sp_digit* a,
  12824. sp_digit b)
  12825. {
  12826. __asm__ __volatile__ (
  12827. "mov r6, #2\n\t"
  12828. "lsl r6, r6, #8\n\t"
  12829. "add r6, %[a]\n\t"
  12830. "mov r8, %[r]\n\t"
  12831. "mov r9, r6\n\t"
  12832. "mov r3, #0\n\t"
  12833. "mov r4, #0\n\t"
  12834. "1:\n\t"
  12835. "mov %[r], #0\n\t"
  12836. "mov r5, #0\n\t"
  12837. "# A[] * B\n\t"
  12838. "ldr r6, [%[a]]\n\t"
  12839. "lsl r6, r6, #16\n\t"
  12840. "lsl r7, %[b], #16\n\t"
  12841. "lsr r6, r6, #16\n\t"
  12842. "lsr r7, r7, #16\n\t"
  12843. "mul r7, r6\n\t"
  12844. "add r3, r7\n\t"
  12845. "adc r4, %[r]\n\t"
  12846. "adc r5, %[r]\n\t"
  12847. "lsr r7, %[b], #16\n\t"
  12848. "mul r6, r7\n\t"
  12849. "lsr r7, r6, #16\n\t"
  12850. "lsl r6, r6, #16\n\t"
  12851. "add r3, r6\n\t"
  12852. "adc r4, r7\n\t"
  12853. "adc r5, %[r]\n\t"
  12854. "ldr r6, [%[a]]\n\t"
  12855. "lsr r6, r6, #16\n\t"
  12856. "lsr r7, %[b], #16\n\t"
  12857. "mul r7, r6\n\t"
  12858. "add r4, r7\n\t"
  12859. "adc r5, %[r]\n\t"
  12860. "lsl r7, %[b], #16\n\t"
  12861. "lsr r7, r7, #16\n\t"
  12862. "mul r6, r7\n\t"
  12863. "lsr r7, r6, #16\n\t"
  12864. "lsl r6, r6, #16\n\t"
  12865. "add r3, r6\n\t"
  12866. "adc r4, r7\n\t"
  12867. "adc r5, %[r]\n\t"
  12868. "# A[] * B - Done\n\t"
  12869. "mov %[r], r8\n\t"
  12870. "str r3, [%[r]]\n\t"
  12871. "mov r3, r4\n\t"
  12872. "mov r4, r5\n\t"
  12873. "add %[r], #4\n\t"
  12874. "add %[a], #4\n\t"
  12875. "mov r8, %[r]\n\t"
  12876. "cmp %[a], r9\n\t"
  12877. "blt 1b\n\t"
  12878. "str r3, [%[r]]\n\t"
  12879. : [r] "+r" (r), [a] "+r" (a)
  12880. : [b] "r" (b)
  12881. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  12882. );
  12883. }
  12884. #if defined(WOLFSSL_HAVE_SP_RSA) || defined(WOLFSSL_HAVE_SP_DH)
  12885. /* r = 2^n mod m where n is the number of bits to reduce by.
  12886. * Given m must be 4096 bits, just need to subtract.
  12887. *
  12888. * r A single precision number.
  12889. * m A single precision number.
  12890. */
  12891. static void sp_4096_mont_norm_128(sp_digit* r, const sp_digit* m)
  12892. {
  12893. XMEMSET(r, 0, sizeof(sp_digit) * 128);
  12894. /* r = 2^n mod m */
  12895. sp_4096_sub_in_place_128(r, m);
  12896. }
  12897. #endif /* WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH */
  12898. /* Conditionally subtract b from a using the mask m.
  12899. * m is -1 to subtract and 0 when not copying.
  12900. *
  12901. * r A single precision number representing condition subtract result.
  12902. * a A single precision number to subtract from.
  12903. * b A single precision number to subtract.
  12904. * m Mask value to apply.
  12905. */
  12906. SP_NOINLINE static sp_digit sp_4096_cond_sub_128(sp_digit* r, const sp_digit* a,
  12907. const sp_digit* b, sp_digit m)
  12908. {
  12909. sp_digit c = 0;
  12910. __asm__ __volatile__ (
  12911. "mov r5, #2\n\t"
  12912. "lsl r5, r5, #8\n\t"
  12913. "mov r8, r5\n\t"
  12914. "mov r7, #0\n\t"
  12915. "1:\n\t"
  12916. "ldr r6, [%[b], r7]\n\t"
  12917. "and r6, %[m]\n\t"
  12918. "mov r5, #0\n\t"
  12919. "sub r5, %[c]\n\t"
  12920. "ldr r5, [%[a], r7]\n\t"
  12921. "sbc r5, r6\n\t"
  12922. "sbc %[c], %[c]\n\t"
  12923. "str r5, [%[r], r7]\n\t"
  12924. "add r7, #4\n\t"
  12925. "cmp r7, r8\n\t"
  12926. "blt 1b\n\t"
  12927. : [c] "+r" (c)
  12928. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  12929. : "memory", "r5", "r6", "r7", "r8"
  12930. );
  12931. return c;
  12932. }
  12933. /* Reduce the number back to 4096 bits using Montgomery reduction.
  12934. *
  12935. * a A single precision number to reduce in place.
  12936. * m The single precision number representing the modulus.
  12937. * mp The digit representing the negative inverse of m mod 2^n.
  12938. */
  12939. SP_NOINLINE static void sp_4096_mont_reduce_128(sp_digit* a, const sp_digit* m,
  12940. sp_digit mp)
  12941. {
  12942. sp_digit ca = 0;
  12943. __asm__ __volatile__ (
  12944. "mov r8, %[mp]\n\t"
  12945. "mov r12, %[ca]\n\t"
  12946. "mov r14, %[m]\n\t"
  12947. "mov r9, %[a]\n\t"
  12948. "mov r4, #0\n\t"
  12949. "# i = 0\n\t"
  12950. "mov r11, r4\n\t"
  12951. "\n1:\n\t"
  12952. "mov r5, #0\n\t"
  12953. "mov %[ca], #0\n\t"
  12954. "# mu = a[i] * mp\n\t"
  12955. "mov %[mp], r8\n\t"
  12956. "ldr %[a], [%[a]]\n\t"
  12957. "mul %[mp], %[a]\n\t"
  12958. "mov %[m], r14\n\t"
  12959. "mov r10, r9\n\t"
  12960. "\n2:\n\t"
  12961. "# a[i+j] += m[j] * mu\n\t"
  12962. "mov %[a], r10\n\t"
  12963. "ldr %[a], [%[a]]\n\t"
  12964. "mov %[ca], #0\n\t"
  12965. "mov r4, r5\n\t"
  12966. "mov r5, #0\n\t"
  12967. "# Multiply m[j] and mu - Start\n\t"
  12968. "ldr r7, [%[m]]\n\t"
  12969. "lsl r6, %[mp], #16\n\t"
  12970. "lsl r7, r7, #16\n\t"
  12971. "lsr r6, r6, #16\n\t"
  12972. "lsr r7, r7, #16\n\t"
  12973. "mul r7, r6\n\t"
  12974. "add %[a], r7\n\t"
  12975. "adc r5, %[ca]\n\t"
  12976. "ldr r7, [%[m]]\n\t"
  12977. "lsr r7, r7, #16\n\t"
  12978. "mul r6, r7\n\t"
  12979. "lsr r7, r6, #16\n\t"
  12980. "lsl r6, r6, #16\n\t"
  12981. "add %[a], r6\n\t"
  12982. "adc r5, r7\n\t"
  12983. "ldr r7, [%[m]]\n\t"
  12984. "lsr r6, %[mp], #16\n\t"
  12985. "lsr r7, r7, #16\n\t"
  12986. "mul r7, r6\n\t"
  12987. "add r5, r7\n\t"
  12988. "ldr r7, [%[m]]\n\t"
  12989. "lsl r7, r7, #16\n\t"
  12990. "lsr r7, r7, #16\n\t"
  12991. "mul r6, r7\n\t"
  12992. "lsr r7, r6, #16\n\t"
  12993. "lsl r6, r6, #16\n\t"
  12994. "add %[a], r6\n\t"
  12995. "adc r5, r7\n\t"
  12996. "# Multiply m[j] and mu - Done\n\t"
  12997. "add r4, %[a]\n\t"
  12998. "adc r5, %[ca]\n\t"
  12999. "mov %[a], r10\n\t"
  13000. "str r4, [%[a]]\n\t"
  13001. "mov r6, #4\n\t"
  13002. "add %[m], #4\n\t"
  13003. "add r10, r6\n\t"
  13004. "mov r4, #1\n\t"
  13005. "lsl r4, r4, #8\n\t"
  13006. "add r4, #252\n\t"
  13007. "add r4, r9\n\t"
  13008. "cmp r10, r4\n\t"
  13009. "blt 2b\n\t"
  13010. "# a[i+127] += m[127] * mu\n\t"
  13011. "mov %[ca], #0\n\t"
  13012. "mov r4, r12\n\t"
  13013. "mov %[a], #0\n\t"
  13014. "# Multiply m[127] and mu - Start\n\t"
  13015. "ldr r7, [%[m]]\n\t"
  13016. "lsl r6, %[mp], #16\n\t"
  13017. "lsl r7, r7, #16\n\t"
  13018. "lsr r6, r6, #16\n\t"
  13019. "lsr r7, r7, #16\n\t"
  13020. "mul r7, r6\n\t"
  13021. "add r5, r7\n\t"
  13022. "adc r4, %[ca]\n\t"
  13023. "adc %[a], %[ca]\n\t"
  13024. "ldr r7, [%[m]]\n\t"
  13025. "lsr r7, r7, #16\n\t"
  13026. "mul r6, r7\n\t"
  13027. "lsr r7, r6, #16\n\t"
  13028. "lsl r6, r6, #16\n\t"
  13029. "add r5, r6\n\t"
  13030. "adc r4, r7\n\t"
  13031. "adc %[a], %[ca]\n\t"
  13032. "ldr r7, [%[m]]\n\t"
  13033. "lsr r6, %[mp], #16\n\t"
  13034. "lsr r7, r7, #16\n\t"
  13035. "mul r7, r6\n\t"
  13036. "add r4, r7\n\t"
  13037. "adc %[a], %[ca]\n\t"
  13038. "ldr r7, [%[m]]\n\t"
  13039. "lsl r7, r7, #16\n\t"
  13040. "lsr r7, r7, #16\n\t"
  13041. "mul r6, r7\n\t"
  13042. "lsr r7, r6, #16\n\t"
  13043. "lsl r6, r6, #16\n\t"
  13044. "add r5, r6\n\t"
  13045. "adc r4, r7\n\t"
  13046. "adc %[a], %[ca]\n\t"
  13047. "# Multiply m[127] and mu - Done\n\t"
  13048. "mov %[ca], %[a]\n\t"
  13049. "mov %[a], r10\n\t"
  13050. "ldr r7, [%[a], #4]\n\t"
  13051. "ldr %[a], [%[a]]\n\t"
  13052. "mov r6, #0\n\t"
  13053. "add r5, %[a]\n\t"
  13054. "adc r7, r4\n\t"
  13055. "adc %[ca], r6\n\t"
  13056. "mov %[a], r10\n\t"
  13057. "str r5, [%[a]]\n\t"
  13058. "str r7, [%[a], #4]\n\t"
  13059. "# i += 1\n\t"
  13060. "mov r6, #4\n\t"
  13061. "add r9, r6\n\t"
  13062. "add r11, r6\n\t"
  13063. "mov r12, %[ca]\n\t"
  13064. "mov %[a], r9\n\t"
  13065. "mov r4, #2\n\t"
  13066. "lsl r4, r4, #8\n\t"
  13067. "cmp r11, r4\n\t"
  13068. "blt 1b\n\t"
  13069. "mov %[m], r14\n\t"
  13070. : [ca] "+r" (ca), [a] "+r" (a)
  13071. : [m] "r" (m), [mp] "r" (mp)
  13072. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  13073. );
  13074. sp_4096_cond_sub_128(a - 128, a, m, (sp_digit)0 - ca);
  13075. }
  13076. /* Multiply two Montogmery form numbers mod the modulus (prime).
  13077. * (r = a * b mod m)
  13078. *
  13079. * r Result of multiplication.
  13080. * a First number to multiply in Montogmery form.
  13081. * b Second number to multiply in Montogmery form.
  13082. * m Modulus (prime).
  13083. * mp Montogmery mulitplier.
  13084. */
  13085. static void sp_4096_mont_mul_128(sp_digit* r, const sp_digit* a, const sp_digit* b,
  13086. const sp_digit* m, sp_digit mp)
  13087. {
  13088. sp_4096_mul_128(r, a, b);
  13089. sp_4096_mont_reduce_128(r, m, mp);
  13090. }
  13091. /* Square the Montgomery form number. (r = a * a mod m)
  13092. *
  13093. * r Result of squaring.
  13094. * a Number to square in Montogmery form.
  13095. * m Modulus (prime).
  13096. * mp Montogmery mulitplier.
  13097. */
  13098. static void sp_4096_mont_sqr_128(sp_digit* r, const sp_digit* a, const sp_digit* m,
  13099. sp_digit mp)
  13100. {
  13101. sp_4096_sqr_128(r, a);
  13102. sp_4096_mont_reduce_128(r, m, mp);
  13103. }
  13104. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  13105. *
  13106. * d1 The high order half of the number to divide.
  13107. * d0 The low order half of the number to divide.
  13108. * div The dividend.
  13109. * returns the result of the division.
  13110. *
  13111. * Note that this is an approximate div. It may give an answer 1 larger.
  13112. */
  13113. SP_NOINLINE static sp_digit div_4096_word_128(sp_digit d1, sp_digit d0,
  13114. sp_digit div)
  13115. {
  13116. sp_digit r = 0;
  13117. __asm__ __volatile__ (
  13118. "lsr r5, %[div], #1\n\t"
  13119. "add r5, #1\n\t"
  13120. "mov r8, %[d0]\n\t"
  13121. "mov r9, %[d1]\n\t"
  13122. "# Do top 32\n\t"
  13123. "mov r6, r5\n\t"
  13124. "sub r6, %[d1]\n\t"
  13125. "sbc r6, r6\n\t"
  13126. "add %[r], %[r]\n\t"
  13127. "sub %[r], r6\n\t"
  13128. "and r6, r5\n\t"
  13129. "sub %[d1], r6\n\t"
  13130. "# Next 30 bits\n\t"
  13131. "mov r4, #29\n\t"
  13132. "1:\n\t"
  13133. "lsl %[d0], %[d0], #1\n\t"
  13134. "adc %[d1], %[d1]\n\t"
  13135. "mov r6, r5\n\t"
  13136. "sub r6, %[d1]\n\t"
  13137. "sbc r6, r6\n\t"
  13138. "add %[r], %[r]\n\t"
  13139. "sub %[r], r6\n\t"
  13140. "and r6, r5\n\t"
  13141. "sub %[d1], r6\n\t"
  13142. "sub r4, #1\n\t"
  13143. "bpl 1b\n\t"
  13144. "mov r7, #0\n\t"
  13145. "add %[r], %[r]\n\t"
  13146. "add %[r], #1\n\t"
  13147. "# r * div - Start\n\t"
  13148. "lsl %[d1], %[r], #16\n\t"
  13149. "lsl r4, %[div], #16\n\t"
  13150. "lsr %[d1], %[d1], #16\n\t"
  13151. "lsr r4, r4, #16\n\t"
  13152. "mul r4, %[d1]\n\t"
  13153. "lsr r6, %[div], #16\n\t"
  13154. "mul %[d1], r6\n\t"
  13155. "lsr r5, %[d1], #16\n\t"
  13156. "lsl %[d1], %[d1], #16\n\t"
  13157. "add r4, %[d1]\n\t"
  13158. "adc r5, r7\n\t"
  13159. "lsr %[d1], %[r], #16\n\t"
  13160. "mul r6, %[d1]\n\t"
  13161. "add r5, r6\n\t"
  13162. "lsl r6, %[div], #16\n\t"
  13163. "lsr r6, r6, #16\n\t"
  13164. "mul %[d1], r6\n\t"
  13165. "lsr r6, %[d1], #16\n\t"
  13166. "lsl %[d1], %[d1], #16\n\t"
  13167. "add r4, %[d1]\n\t"
  13168. "adc r5, r6\n\t"
  13169. "# r * div - Done\n\t"
  13170. "mov %[d1], r8\n\t"
  13171. "sub %[d1], r4\n\t"
  13172. "mov r4, %[d1]\n\t"
  13173. "mov %[d1], r9\n\t"
  13174. "sbc %[d1], r5\n\t"
  13175. "mov r5, %[d1]\n\t"
  13176. "add %[r], r5\n\t"
  13177. "# r * div - Start\n\t"
  13178. "lsl %[d1], %[r], #16\n\t"
  13179. "lsl r4, %[div], #16\n\t"
  13180. "lsr %[d1], %[d1], #16\n\t"
  13181. "lsr r4, r4, #16\n\t"
  13182. "mul r4, %[d1]\n\t"
  13183. "lsr r6, %[div], #16\n\t"
  13184. "mul %[d1], r6\n\t"
  13185. "lsr r5, %[d1], #16\n\t"
  13186. "lsl %[d1], %[d1], #16\n\t"
  13187. "add r4, %[d1]\n\t"
  13188. "adc r5, r7\n\t"
  13189. "lsr %[d1], %[r], #16\n\t"
  13190. "mul r6, %[d1]\n\t"
  13191. "add r5, r6\n\t"
  13192. "lsl r6, %[div], #16\n\t"
  13193. "lsr r6, r6, #16\n\t"
  13194. "mul %[d1], r6\n\t"
  13195. "lsr r6, %[d1], #16\n\t"
  13196. "lsl %[d1], %[d1], #16\n\t"
  13197. "add r4, %[d1]\n\t"
  13198. "adc r5, r6\n\t"
  13199. "# r * div - Done\n\t"
  13200. "mov %[d1], r8\n\t"
  13201. "mov r6, r9\n\t"
  13202. "sub r4, %[d1], r4\n\t"
  13203. "sbc r6, r5\n\t"
  13204. "mov r5, r6\n\t"
  13205. "add %[r], r5\n\t"
  13206. "# r * div - Start\n\t"
  13207. "lsl %[d1], %[r], #16\n\t"
  13208. "lsl r4, %[div], #16\n\t"
  13209. "lsr %[d1], %[d1], #16\n\t"
  13210. "lsr r4, r4, #16\n\t"
  13211. "mul r4, %[d1]\n\t"
  13212. "lsr r6, %[div], #16\n\t"
  13213. "mul %[d1], r6\n\t"
  13214. "lsr r5, %[d1], #16\n\t"
  13215. "lsl %[d1], %[d1], #16\n\t"
  13216. "add r4, %[d1]\n\t"
  13217. "adc r5, r7\n\t"
  13218. "lsr %[d1], %[r], #16\n\t"
  13219. "mul r6, %[d1]\n\t"
  13220. "add r5, r6\n\t"
  13221. "lsl r6, %[div], #16\n\t"
  13222. "lsr r6, r6, #16\n\t"
  13223. "mul %[d1], r6\n\t"
  13224. "lsr r6, %[d1], #16\n\t"
  13225. "lsl %[d1], %[d1], #16\n\t"
  13226. "add r4, %[d1]\n\t"
  13227. "adc r5, r6\n\t"
  13228. "# r * div - Done\n\t"
  13229. "mov %[d1], r8\n\t"
  13230. "mov r6, r9\n\t"
  13231. "sub r4, %[d1], r4\n\t"
  13232. "sbc r6, r5\n\t"
  13233. "mov r5, r6\n\t"
  13234. "add %[r], r5\n\t"
  13235. "mov r6, %[div]\n\t"
  13236. "sub r6, r4\n\t"
  13237. "sbc r6, r6\n\t"
  13238. "sub %[r], r6\n\t"
  13239. : [r] "+r" (r)
  13240. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  13241. : "r4", "r5", "r7", "r6", "r8", "r9"
  13242. );
  13243. return r;
  13244. }
  13245. /* AND m into each word of a and store in r.
  13246. *
  13247. * r A single precision integer.
  13248. * a A single precision integer.
  13249. * m Mask to AND against each digit.
  13250. */
  13251. static void sp_4096_mask_128(sp_digit* r, const sp_digit* a, sp_digit m)
  13252. {
  13253. #ifdef WOLFSSL_SP_SMALL
  13254. int i;
  13255. for (i=0; i<128; i++) {
  13256. r[i] = a[i] & m;
  13257. }
  13258. #else
  13259. int i;
  13260. for (i = 0; i < 128; i += 8) {
  13261. r[i+0] = a[i+0] & m;
  13262. r[i+1] = a[i+1] & m;
  13263. r[i+2] = a[i+2] & m;
  13264. r[i+3] = a[i+3] & m;
  13265. r[i+4] = a[i+4] & m;
  13266. r[i+5] = a[i+5] & m;
  13267. r[i+6] = a[i+6] & m;
  13268. r[i+7] = a[i+7] & m;
  13269. }
  13270. #endif
  13271. }
  13272. /* Compare a with b in constant time.
  13273. *
  13274. * a A single precision integer.
  13275. * b A single precision integer.
  13276. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  13277. * respectively.
  13278. */
  13279. SP_NOINLINE static int32_t sp_4096_cmp_128(const sp_digit* a, const sp_digit* b)
  13280. {
  13281. sp_digit r = 0;
  13282. __asm__ __volatile__ (
  13283. "mov r3, #0\n\t"
  13284. "mvn r3, r3\n\t"
  13285. "mov r6, #1\n\t"
  13286. "lsl r6, r6, #8\n\t"
  13287. "add r6, #252\n\t"
  13288. "1:\n\t"
  13289. "ldr r7, [%[a], r6]\n\t"
  13290. "ldr r5, [%[b], r6]\n\t"
  13291. "and r7, r3\n\t"
  13292. "and r5, r3\n\t"
  13293. "mov r4, r7\n\t"
  13294. "sub r7, r5\n\t"
  13295. "sbc r7, r7\n\t"
  13296. "add %[r], r7\n\t"
  13297. "mvn r7, r7\n\t"
  13298. "and r3, r7\n\t"
  13299. "sub r5, r4\n\t"
  13300. "sbc r7, r7\n\t"
  13301. "sub %[r], r7\n\t"
  13302. "mvn r7, r7\n\t"
  13303. "and r3, r7\n\t"
  13304. "sub r6, #4\n\t"
  13305. "cmp r6, #0\n\t"
  13306. "bge 1b\n\t"
  13307. : [r] "+r" (r)
  13308. : [a] "r" (a), [b] "r" (b)
  13309. : "r3", "r4", "r5", "r6", "r7"
  13310. );
  13311. return r;
  13312. }
  13313. /* Divide d in a and put remainder into r (m*d + r = a)
  13314. * m is not calculated as it is not needed at this time.
  13315. *
  13316. * a Nmber to be divided.
  13317. * d Number to divide with.
  13318. * m Multiplier result.
  13319. * r Remainder from the division.
  13320. * returns MP_OKAY indicating success.
  13321. */
  13322. static WC_INLINE int sp_4096_div_128(const sp_digit* a, const sp_digit* d, sp_digit* m,
  13323. sp_digit* r)
  13324. {
  13325. sp_digit t1[256], t2[129];
  13326. sp_digit div, r1;
  13327. int i;
  13328. (void)m;
  13329. div = d[127];
  13330. XMEMCPY(t1, a, sizeof(*t1) * 2 * 128);
  13331. for (i=127; i>=0; i--) {
  13332. r1 = div_4096_word_128(t1[128 + i], t1[128 + i - 1], div);
  13333. sp_4096_mul_d_128(t2, d, r1);
  13334. t1[128 + i] += sp_4096_sub_in_place_128(&t1[i], t2);
  13335. t1[128 + i] -= t2[128];
  13336. sp_4096_mask_128(t2, d, t1[128 + i]);
  13337. t1[128 + i] += sp_4096_add_128(&t1[i], &t1[i], t2);
  13338. sp_4096_mask_128(t2, d, t1[128 + i]);
  13339. t1[128 + i] += sp_4096_add_128(&t1[i], &t1[i], t2);
  13340. }
  13341. r1 = sp_4096_cmp_128(t1, d) >= 0;
  13342. sp_4096_cond_sub_128(r, t1, d, (sp_digit)0 - r1);
  13343. return MP_OKAY;
  13344. }
  13345. /* Reduce a modulo m into r. (r = a mod m)
  13346. *
  13347. * r A single precision number that is the reduced result.
  13348. * a A single precision number that is to be reduced.
  13349. * m A single precision number that is the modulus to reduce with.
  13350. * returns MP_OKAY indicating success.
  13351. */
  13352. static WC_INLINE int sp_4096_mod_128(sp_digit* r, const sp_digit* a, const sp_digit* m)
  13353. {
  13354. return sp_4096_div_128(a, m, NULL, r);
  13355. }
  13356. /* Divide d in a and put remainder into r (m*d + r = a)
  13357. * m is not calculated as it is not needed at this time.
  13358. *
  13359. * a Nmber to be divided.
  13360. * d Number to divide with.
  13361. * m Multiplier result.
  13362. * r Remainder from the division.
  13363. * returns MP_OKAY indicating success.
  13364. */
  13365. static WC_INLINE int sp_4096_div_128_cond(const sp_digit* a, const sp_digit* d, sp_digit* m,
  13366. sp_digit* r)
  13367. {
  13368. sp_digit t1[256], t2[129];
  13369. sp_digit div, r1;
  13370. int i;
  13371. (void)m;
  13372. div = d[127];
  13373. XMEMCPY(t1, a, sizeof(*t1) * 2 * 128);
  13374. for (i=127; i>=0; i--) {
  13375. r1 = div_4096_word_128(t1[128 + i], t1[128 + i - 1], div);
  13376. sp_4096_mul_d_128(t2, d, r1);
  13377. t1[128 + i] += sp_4096_sub_in_place_128(&t1[i], t2);
  13378. t1[128 + i] -= t2[128];
  13379. if (t1[128 + i] != 0) {
  13380. t1[128 + i] += sp_4096_add_128(&t1[i], &t1[i], d);
  13381. if (t1[128 + i] != 0)
  13382. t1[128 + i] += sp_4096_add_128(&t1[i], &t1[i], d);
  13383. }
  13384. }
  13385. r1 = sp_4096_cmp_128(t1, d) >= 0;
  13386. sp_4096_cond_sub_128(r, t1, d, (sp_digit)0 - r1);
  13387. return MP_OKAY;
  13388. }
  13389. /* Reduce a modulo m into r. (r = a mod m)
  13390. *
  13391. * r A single precision number that is the reduced result.
  13392. * a A single precision number that is to be reduced.
  13393. * m A single precision number that is the modulus to reduce with.
  13394. * returns MP_OKAY indicating success.
  13395. */
  13396. static WC_INLINE int sp_4096_mod_128_cond(sp_digit* r, const sp_digit* a, const sp_digit* m)
  13397. {
  13398. return sp_4096_div_128_cond(a, m, NULL, r);
  13399. }
  13400. #if (defined(WOLFSSL_HAVE_SP_RSA) && !defined(WOLFSSL_RSA_PUBLIC_ONLY)) || \
  13401. defined(WOLFSSL_HAVE_SP_DH)
  13402. #ifdef WOLFSSL_SP_SMALL
  13403. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  13404. *
  13405. * r A single precision number that is the result of the operation.
  13406. * a A single precision number being exponentiated.
  13407. * e A single precision number that is the exponent.
  13408. * bits The number of bits in the exponent.
  13409. * m A single precision number that is the modulus.
  13410. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  13411. */
  13412. static int sp_4096_mod_exp_128(sp_digit* r, const sp_digit* a, const sp_digit* e,
  13413. int bits, const sp_digit* m, int reduceA)
  13414. {
  13415. #ifndef WOLFSSL_SMALL_STACK
  13416. sp_digit t[16][256];
  13417. #else
  13418. sp_digit* t[16];
  13419. sp_digit* td;
  13420. #endif
  13421. sp_digit* norm;
  13422. sp_digit mp = 1;
  13423. sp_digit n;
  13424. sp_digit mask;
  13425. int i;
  13426. int c, y;
  13427. int err = MP_OKAY;
  13428. #ifdef WOLFSSL_SMALL_STACK
  13429. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 256, NULL,
  13430. DYNAMIC_TYPE_TMP_BUFFER);
  13431. if (td == NULL) {
  13432. err = MEMORY_E;
  13433. }
  13434. #endif
  13435. if (err == MP_OKAY) {
  13436. #ifdef WOLFSSL_SMALL_STACK
  13437. for (i=0; i<16; i++) {
  13438. t[i] = td + i * 256;
  13439. }
  13440. #endif
  13441. norm = t[0];
  13442. sp_4096_mont_setup(m, &mp);
  13443. sp_4096_mont_norm_128(norm, m);
  13444. XMEMSET(t[1], 0, sizeof(sp_digit) * 128U);
  13445. if (reduceA != 0) {
  13446. err = sp_4096_mod_128(t[1] + 128, a, m);
  13447. if (err == MP_OKAY) {
  13448. err = sp_4096_mod_128(t[1], t[1], m);
  13449. }
  13450. }
  13451. else {
  13452. XMEMCPY(t[1] + 128, a, sizeof(sp_digit) * 128);
  13453. err = sp_4096_mod_128(t[1], t[1], m);
  13454. }
  13455. }
  13456. if (err == MP_OKAY) {
  13457. sp_4096_mont_sqr_128(t[ 2], t[ 1], m, mp);
  13458. sp_4096_mont_mul_128(t[ 3], t[ 2], t[ 1], m, mp);
  13459. sp_4096_mont_sqr_128(t[ 4], t[ 2], m, mp);
  13460. sp_4096_mont_mul_128(t[ 5], t[ 3], t[ 2], m, mp);
  13461. sp_4096_mont_sqr_128(t[ 6], t[ 3], m, mp);
  13462. sp_4096_mont_mul_128(t[ 7], t[ 4], t[ 3], m, mp);
  13463. sp_4096_mont_sqr_128(t[ 8], t[ 4], m, mp);
  13464. sp_4096_mont_mul_128(t[ 9], t[ 5], t[ 4], m, mp);
  13465. sp_4096_mont_sqr_128(t[10], t[ 5], m, mp);
  13466. sp_4096_mont_mul_128(t[11], t[ 6], t[ 5], m, mp);
  13467. sp_4096_mont_sqr_128(t[12], t[ 6], m, mp);
  13468. sp_4096_mont_mul_128(t[13], t[ 7], t[ 6], m, mp);
  13469. sp_4096_mont_sqr_128(t[14], t[ 7], m, mp);
  13470. sp_4096_mont_mul_128(t[15], t[ 8], t[ 7], m, mp);
  13471. i = (bits - 1) / 32;
  13472. n = e[i--];
  13473. c = bits & 31;
  13474. if (c == 0) {
  13475. c = 32;
  13476. }
  13477. c -= bits % 4;
  13478. if (c == 32) {
  13479. c = 28;
  13480. }
  13481. y = (int)(n >> c);
  13482. n <<= 32 - c;
  13483. XMEMCPY(r, t[y], sizeof(sp_digit) * 128);
  13484. for (; i>=0 || c>=4; ) {
  13485. if (c == 0) {
  13486. n = e[i--];
  13487. y = n >> 28;
  13488. n <<= 4;
  13489. c = 28;
  13490. }
  13491. else if (c < 4) {
  13492. y = n >> 28;
  13493. n = e[i--];
  13494. c = 4 - c;
  13495. y |= n >> (32 - c);
  13496. n <<= c;
  13497. c = 32 - c;
  13498. }
  13499. else {
  13500. y = (n >> 28) & 0xf;
  13501. n <<= 4;
  13502. c -= 4;
  13503. }
  13504. sp_4096_mont_sqr_128(r, r, m, mp);
  13505. sp_4096_mont_sqr_128(r, r, m, mp);
  13506. sp_4096_mont_sqr_128(r, r, m, mp);
  13507. sp_4096_mont_sqr_128(r, r, m, mp);
  13508. sp_4096_mont_mul_128(r, r, t[y], m, mp);
  13509. }
  13510. XMEMSET(&r[128], 0, sizeof(sp_digit) * 128U);
  13511. sp_4096_mont_reduce_128(r, m, mp);
  13512. mask = 0 - (sp_4096_cmp_128(r, m) >= 0);
  13513. sp_4096_cond_sub_128(r, r, m, mask);
  13514. }
  13515. #ifdef WOLFSSL_SMALL_STACK
  13516. if (td != NULL) {
  13517. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  13518. }
  13519. #endif
  13520. return err;
  13521. }
  13522. #else
  13523. /* Modular exponentiate a to the e mod m. (r = a^e mod m)
  13524. *
  13525. * r A single precision number that is the result of the operation.
  13526. * a A single precision number being exponentiated.
  13527. * e A single precision number that is the exponent.
  13528. * bits The number of bits in the exponent.
  13529. * m A single precision number that is the modulus.
  13530. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  13531. */
  13532. static int sp_4096_mod_exp_128(sp_digit* r, const sp_digit* a, const sp_digit* e,
  13533. int bits, const sp_digit* m, int reduceA)
  13534. {
  13535. #ifndef WOLFSSL_SMALL_STACK
  13536. sp_digit t[32][256];
  13537. #else
  13538. sp_digit* t[32];
  13539. sp_digit* td;
  13540. #endif
  13541. sp_digit* norm;
  13542. sp_digit mp = 1;
  13543. sp_digit n;
  13544. sp_digit mask;
  13545. int i;
  13546. int c, y;
  13547. int err = MP_OKAY;
  13548. #ifdef WOLFSSL_SMALL_STACK
  13549. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 32 * 256, NULL,
  13550. DYNAMIC_TYPE_TMP_BUFFER);
  13551. if (td == NULL) {
  13552. err = MEMORY_E;
  13553. }
  13554. #endif
  13555. if (err == MP_OKAY) {
  13556. #ifdef WOLFSSL_SMALL_STACK
  13557. for (i=0; i<32; i++) {
  13558. t[i] = td + i * 256;
  13559. }
  13560. #endif
  13561. norm = t[0];
  13562. sp_4096_mont_setup(m, &mp);
  13563. sp_4096_mont_norm_128(norm, m);
  13564. XMEMSET(t[1], 0, sizeof(sp_digit) * 128U);
  13565. if (reduceA != 0) {
  13566. err = sp_4096_mod_128(t[1] + 128, a, m);
  13567. if (err == MP_OKAY) {
  13568. err = sp_4096_mod_128(t[1], t[1], m);
  13569. }
  13570. }
  13571. else {
  13572. XMEMCPY(t[1] + 128, a, sizeof(sp_digit) * 128);
  13573. err = sp_4096_mod_128(t[1], t[1], m);
  13574. }
  13575. }
  13576. if (err == MP_OKAY) {
  13577. sp_4096_mont_sqr_128(t[ 2], t[ 1], m, mp);
  13578. sp_4096_mont_mul_128(t[ 3], t[ 2], t[ 1], m, mp);
  13579. sp_4096_mont_sqr_128(t[ 4], t[ 2], m, mp);
  13580. sp_4096_mont_mul_128(t[ 5], t[ 3], t[ 2], m, mp);
  13581. sp_4096_mont_sqr_128(t[ 6], t[ 3], m, mp);
  13582. sp_4096_mont_mul_128(t[ 7], t[ 4], t[ 3], m, mp);
  13583. sp_4096_mont_sqr_128(t[ 8], t[ 4], m, mp);
  13584. sp_4096_mont_mul_128(t[ 9], t[ 5], t[ 4], m, mp);
  13585. sp_4096_mont_sqr_128(t[10], t[ 5], m, mp);
  13586. sp_4096_mont_mul_128(t[11], t[ 6], t[ 5], m, mp);
  13587. sp_4096_mont_sqr_128(t[12], t[ 6], m, mp);
  13588. sp_4096_mont_mul_128(t[13], t[ 7], t[ 6], m, mp);
  13589. sp_4096_mont_sqr_128(t[14], t[ 7], m, mp);
  13590. sp_4096_mont_mul_128(t[15], t[ 8], t[ 7], m, mp);
  13591. sp_4096_mont_sqr_128(t[16], t[ 8], m, mp);
  13592. sp_4096_mont_mul_128(t[17], t[ 9], t[ 8], m, mp);
  13593. sp_4096_mont_sqr_128(t[18], t[ 9], m, mp);
  13594. sp_4096_mont_mul_128(t[19], t[10], t[ 9], m, mp);
  13595. sp_4096_mont_sqr_128(t[20], t[10], m, mp);
  13596. sp_4096_mont_mul_128(t[21], t[11], t[10], m, mp);
  13597. sp_4096_mont_sqr_128(t[22], t[11], m, mp);
  13598. sp_4096_mont_mul_128(t[23], t[12], t[11], m, mp);
  13599. sp_4096_mont_sqr_128(t[24], t[12], m, mp);
  13600. sp_4096_mont_mul_128(t[25], t[13], t[12], m, mp);
  13601. sp_4096_mont_sqr_128(t[26], t[13], m, mp);
  13602. sp_4096_mont_mul_128(t[27], t[14], t[13], m, mp);
  13603. sp_4096_mont_sqr_128(t[28], t[14], m, mp);
  13604. sp_4096_mont_mul_128(t[29], t[15], t[14], m, mp);
  13605. sp_4096_mont_sqr_128(t[30], t[15], m, mp);
  13606. sp_4096_mont_mul_128(t[31], t[16], t[15], m, mp);
  13607. i = (bits - 1) / 32;
  13608. n = e[i--];
  13609. c = bits & 31;
  13610. if (c == 0) {
  13611. c = 32;
  13612. }
  13613. c -= bits % 5;
  13614. if (c == 32) {
  13615. c = 27;
  13616. }
  13617. y = (int)(n >> c);
  13618. n <<= 32 - c;
  13619. XMEMCPY(r, t[y], sizeof(sp_digit) * 128);
  13620. for (; i>=0 || c>=5; ) {
  13621. if (c == 0) {
  13622. n = e[i--];
  13623. y = n >> 27;
  13624. n <<= 5;
  13625. c = 27;
  13626. }
  13627. else if (c < 5) {
  13628. y = n >> 27;
  13629. n = e[i--];
  13630. c = 5 - c;
  13631. y |= n >> (32 - c);
  13632. n <<= c;
  13633. c = 32 - c;
  13634. }
  13635. else {
  13636. y = (n >> 27) & 0x1f;
  13637. n <<= 5;
  13638. c -= 5;
  13639. }
  13640. sp_4096_mont_sqr_128(r, r, m, mp);
  13641. sp_4096_mont_sqr_128(r, r, m, mp);
  13642. sp_4096_mont_sqr_128(r, r, m, mp);
  13643. sp_4096_mont_sqr_128(r, r, m, mp);
  13644. sp_4096_mont_sqr_128(r, r, m, mp);
  13645. sp_4096_mont_mul_128(r, r, t[y], m, mp);
  13646. }
  13647. XMEMSET(&r[128], 0, sizeof(sp_digit) * 128U);
  13648. sp_4096_mont_reduce_128(r, m, mp);
  13649. mask = 0 - (sp_4096_cmp_128(r, m) >= 0);
  13650. sp_4096_cond_sub_128(r, r, m, mask);
  13651. }
  13652. #ifdef WOLFSSL_SMALL_STACK
  13653. if (td != NULL) {
  13654. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  13655. }
  13656. #endif
  13657. return err;
  13658. }
  13659. #endif /* WOLFSSL_SP_SMALL */
  13660. #endif /* (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) || WOLFSSL_HAVE_SP_DH */
  13661. #ifdef WOLFSSL_HAVE_SP_RSA
  13662. /* RSA public key operation.
  13663. *
  13664. * in Array of bytes representing the number to exponentiate, base.
  13665. * inLen Number of bytes in base.
  13666. * em Public exponent.
  13667. * mm Modulus.
  13668. * out Buffer to hold big-endian bytes of exponentiation result.
  13669. * Must be at least 512 bytes long.
  13670. * outLen Number of bytes in result.
  13671. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  13672. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  13673. */
  13674. int sp_RsaPublic_4096(const byte* in, word32 inLen, mp_int* em, mp_int* mm,
  13675. byte* out, word32* outLen)
  13676. {
  13677. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  13678. sp_digit ad[256], md[128], rd[256];
  13679. #else
  13680. sp_digit* d = NULL;
  13681. #endif
  13682. sp_digit* a;
  13683. sp_digit *ah;
  13684. sp_digit* m;
  13685. sp_digit* r;
  13686. sp_digit e[1];
  13687. int err = MP_OKAY;
  13688. if (*outLen < 512)
  13689. err = MP_TO_E;
  13690. if (err == MP_OKAY && (mp_count_bits(em) > 32 || inLen > 512 ||
  13691. mp_count_bits(mm) != 4096))
  13692. err = MP_READ_E;
  13693. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  13694. if (err == MP_OKAY) {
  13695. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 128 * 5, NULL,
  13696. DYNAMIC_TYPE_RSA);
  13697. if (d == NULL)
  13698. err = MEMORY_E;
  13699. }
  13700. if (err == MP_OKAY) {
  13701. a = d;
  13702. r = a + 128 * 2;
  13703. m = r + 128 * 2;
  13704. ah = a + 128;
  13705. }
  13706. #else
  13707. a = ad;
  13708. m = md;
  13709. r = rd;
  13710. ah = a + 128;
  13711. #endif
  13712. if (err == MP_OKAY) {
  13713. sp_4096_from_bin(ah, 128, in, inLen);
  13714. #if DIGIT_BIT >= 32
  13715. e[0] = em->dp[0];
  13716. #else
  13717. e[0] = em->dp[0];
  13718. if (em->used > 1)
  13719. e[0] |= ((sp_digit)em->dp[1]) << DIGIT_BIT;
  13720. #endif
  13721. if (e[0] == 0)
  13722. err = MP_EXPTMOD_E;
  13723. }
  13724. if (err == MP_OKAY) {
  13725. sp_4096_from_mp(m, 128, mm);
  13726. if (e[0] == 0x3) {
  13727. if (err == MP_OKAY) {
  13728. sp_4096_sqr_128(r, ah);
  13729. err = sp_4096_mod_128_cond(r, r, m);
  13730. }
  13731. if (err == MP_OKAY) {
  13732. sp_4096_mul_128(r, ah, r);
  13733. err = sp_4096_mod_128_cond(r, r, m);
  13734. }
  13735. }
  13736. else {
  13737. int i;
  13738. sp_digit mp;
  13739. sp_4096_mont_setup(m, &mp);
  13740. /* Convert to Montgomery form. */
  13741. XMEMSET(a, 0, sizeof(sp_digit) * 128);
  13742. err = sp_4096_mod_128_cond(a, a, m);
  13743. if (err == MP_OKAY) {
  13744. for (i=31; i>=0; i--)
  13745. if (e[0] >> i)
  13746. break;
  13747. XMEMCPY(r, a, sizeof(sp_digit) * 128);
  13748. for (i--; i>=0; i--) {
  13749. sp_4096_mont_sqr_128(r, r, m, mp);
  13750. if (((e[0] >> i) & 1) == 1)
  13751. sp_4096_mont_mul_128(r, r, a, m, mp);
  13752. }
  13753. XMEMSET(&r[128], 0, sizeof(sp_digit) * 128);
  13754. sp_4096_mont_reduce_128(r, m, mp);
  13755. for (i = 127; i > 0; i--) {
  13756. if (r[i] != m[i])
  13757. break;
  13758. }
  13759. if (r[i] >= m[i])
  13760. sp_4096_sub_in_place_128(r, m);
  13761. }
  13762. }
  13763. }
  13764. if (err == MP_OKAY) {
  13765. sp_4096_to_bin(r, out);
  13766. *outLen = 512;
  13767. }
  13768. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  13769. if (d != NULL)
  13770. XFREE(d, NULL, DYNAMIC_TYPE_RSA);
  13771. #endif
  13772. return err;
  13773. }
  13774. /* RSA private key operation.
  13775. *
  13776. * in Array of bytes representing the number to exponentiate, base.
  13777. * inLen Number of bytes in base.
  13778. * dm Private exponent.
  13779. * pm First prime.
  13780. * qm Second prime.
  13781. * dpm First prime's CRT exponent.
  13782. * dqm Second prime's CRT exponent.
  13783. * qim Inverse of second prime mod p.
  13784. * mm Modulus.
  13785. * out Buffer to hold big-endian bytes of exponentiation result.
  13786. * Must be at least 512 bytes long.
  13787. * outLen Number of bytes in result.
  13788. * returns 0 on success, MP_TO_E when the outLen is too small, MP_READ_E when
  13789. * an array is too long and MEMORY_E when dynamic memory allocation fails.
  13790. */
  13791. int sp_RsaPrivate_4096(const byte* in, word32 inLen, mp_int* dm,
  13792. mp_int* pm, mp_int* qm, mp_int* dpm, mp_int* dqm, mp_int* qim, mp_int* mm,
  13793. byte* out, word32* outLen)
  13794. {
  13795. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  13796. sp_digit ad[128 * 2];
  13797. sp_digit pd[64], qd[64], dpd[64];
  13798. sp_digit tmpad[128], tmpbd[128];
  13799. #else
  13800. sp_digit* t = NULL;
  13801. #endif
  13802. sp_digit* a;
  13803. sp_digit* p;
  13804. sp_digit* q;
  13805. sp_digit* dp;
  13806. sp_digit* dq;
  13807. sp_digit* qi;
  13808. sp_digit* tmp;
  13809. sp_digit* tmpa;
  13810. sp_digit* tmpb;
  13811. sp_digit* r;
  13812. sp_digit c;
  13813. int err = MP_OKAY;
  13814. (void)dm;
  13815. (void)mm;
  13816. if (*outLen < 512)
  13817. err = MP_TO_E;
  13818. if (err == MP_OKAY && (inLen > 512 || mp_count_bits(mm) != 4096))
  13819. err = MP_READ_E;
  13820. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  13821. if (err == MP_OKAY) {
  13822. t = (sp_digit*)XMALLOC(sizeof(sp_digit) * 64 * 11, NULL,
  13823. DYNAMIC_TYPE_RSA);
  13824. if (t == NULL)
  13825. err = MEMORY_E;
  13826. }
  13827. if (err == MP_OKAY) {
  13828. a = t;
  13829. p = a + 128 * 2;
  13830. q = p + 64;
  13831. qi = dq = dp = q + 64;
  13832. tmpa = qi + 64;
  13833. tmpb = tmpa + 128;
  13834. tmp = t;
  13835. r = tmp + 128;
  13836. }
  13837. #else
  13838. r = a = ad;
  13839. p = pd;
  13840. q = qd;
  13841. qi = dq = dp = dpd;
  13842. tmpa = tmpad;
  13843. tmpb = tmpbd;
  13844. tmp = a + 128;
  13845. #endif
  13846. if (err == MP_OKAY) {
  13847. sp_4096_from_bin(a, 128, in, inLen);
  13848. sp_4096_from_mp(p, 64, pm);
  13849. sp_4096_from_mp(q, 64, qm);
  13850. sp_4096_from_mp(dp, 64, dpm);
  13851. err = sp_4096_mod_exp_64(tmpa, a, dp, 2048, p, 1);
  13852. }
  13853. if (err == MP_OKAY) {
  13854. sp_4096_from_mp(dq, 64, dqm);
  13855. err = sp_4096_mod_exp_64(tmpb, a, dq, 2048, q, 1);
  13856. }
  13857. if (err == MP_OKAY) {
  13858. c = sp_4096_sub_in_place_64(tmpa, tmpb);
  13859. sp_4096_mask_64(tmp, p, c);
  13860. sp_4096_add_64(tmpa, tmpa, tmp);
  13861. sp_4096_from_mp(qi, 64, qim);
  13862. sp_4096_mul_64(tmpa, tmpa, qi);
  13863. err = sp_4096_mod_64(tmpa, tmpa, p);
  13864. }
  13865. if (err == MP_OKAY) {
  13866. sp_4096_mul_64(tmpa, q, tmpa);
  13867. XMEMSET(&tmpb[64], 0, sizeof(sp_digit) * 64);
  13868. sp_4096_add_128(r, tmpb, tmpa);
  13869. sp_4096_to_bin(r, out);
  13870. *outLen = 512;
  13871. }
  13872. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  13873. if (t != NULL) {
  13874. XMEMSET(t, 0, sizeof(sp_digit) * 64 * 11);
  13875. XFREE(t, NULL, DYNAMIC_TYPE_RSA);
  13876. }
  13877. #else
  13878. XMEMSET(tmpad, 0, sizeof(tmpad));
  13879. XMEMSET(tmpbd, 0, sizeof(tmpbd));
  13880. XMEMSET(pd, 0, sizeof(pd));
  13881. XMEMSET(qd, 0, sizeof(qd));
  13882. XMEMSET(dpd, 0, sizeof(dpd));
  13883. #endif
  13884. return err;
  13885. }
  13886. #endif /* WOLFSSL_HAVE_SP_RSA */
  13887. #if defined(WOLFSSL_HAVE_SP_DH) || (defined(WOLFSSL_HAVE_SP_RSA) && \
  13888. !defined(WOLFSSL_RSA_PUBLIC_ONLY))
  13889. /* Convert an array of sp_digit to an mp_int.
  13890. *
  13891. * a A single precision integer.
  13892. * r A multi-precision integer.
  13893. */
  13894. static int sp_4096_to_mp(const sp_digit* a, mp_int* r)
  13895. {
  13896. int err;
  13897. err = mp_grow(r, (4096 + DIGIT_BIT - 1) / DIGIT_BIT);
  13898. if (err == MP_OKAY) { /*lint !e774 case where err is always MP_OKAY*/
  13899. #if DIGIT_BIT == 32
  13900. XMEMCPY(r->dp, a, sizeof(sp_digit) * 128);
  13901. r->used = 128;
  13902. mp_clamp(r);
  13903. #elif DIGIT_BIT < 32
  13904. int i, j = 0, s = 0;
  13905. r->dp[0] = 0;
  13906. for (i = 0; i < 128; i++) {
  13907. r->dp[j] |= a[i] << s;
  13908. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  13909. s = DIGIT_BIT - s;
  13910. r->dp[++j] = a[i] >> s;
  13911. while (s + DIGIT_BIT <= 32) {
  13912. s += DIGIT_BIT;
  13913. r->dp[j++] &= (1L << DIGIT_BIT) - 1;
  13914. if (s == SP_WORD_SIZE) {
  13915. r->dp[j] = 0;
  13916. }
  13917. else {
  13918. r->dp[j] = a[i] >> s;
  13919. }
  13920. }
  13921. s = 32 - s;
  13922. }
  13923. r->used = (4096 + DIGIT_BIT - 1) / DIGIT_BIT;
  13924. mp_clamp(r);
  13925. #else
  13926. int i, j = 0, s = 0;
  13927. r->dp[0] = 0;
  13928. for (i = 0; i < 128; i++) {
  13929. r->dp[j] |= ((mp_digit)a[i]) << s;
  13930. if (s + 32 >= DIGIT_BIT) {
  13931. #if DIGIT_BIT != 32 && DIGIT_BIT != 64
  13932. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  13933. #endif
  13934. s = DIGIT_BIT - s;
  13935. r->dp[++j] = a[i] >> s;
  13936. s = 32 - s;
  13937. }
  13938. else {
  13939. s += 32;
  13940. }
  13941. }
  13942. r->used = (4096 + DIGIT_BIT - 1) / DIGIT_BIT;
  13943. mp_clamp(r);
  13944. #endif
  13945. }
  13946. return err;
  13947. }
  13948. /* Perform the modular exponentiation for Diffie-Hellman.
  13949. *
  13950. * base Base. MP integer.
  13951. * exp Exponent. MP integer.
  13952. * mod Modulus. MP integer.
  13953. * res Result. MP integer.
  13954. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  13955. * and MEMORY_E if memory allocation fails.
  13956. */
  13957. int sp_ModExp_4096(mp_int* base, mp_int* exp, mp_int* mod, mp_int* res)
  13958. {
  13959. int err = MP_OKAY;
  13960. sp_digit b[256], e[128], m[128];
  13961. sp_digit* r = b;
  13962. int expBits = mp_count_bits(exp);
  13963. if (mp_count_bits(base) > 4096) {
  13964. err = MP_READ_E;
  13965. }
  13966. if (err == MP_OKAY) {
  13967. if (expBits > 4096) {
  13968. err = MP_READ_E;
  13969. }
  13970. }
  13971. if (err == MP_OKAY) {
  13972. if (mp_count_bits(mod) != 4096) {
  13973. err = MP_READ_E;
  13974. }
  13975. }
  13976. if (err == MP_OKAY) {
  13977. sp_4096_from_mp(b, 128, base);
  13978. sp_4096_from_mp(e, 128, exp);
  13979. sp_4096_from_mp(m, 128, mod);
  13980. err = sp_4096_mod_exp_128(r, b, e, expBits, m, 0);
  13981. }
  13982. if (err == MP_OKAY) {
  13983. err = sp_4096_to_mp(r, res);
  13984. }
  13985. XMEMSET(e, 0, sizeof(e));
  13986. return err;
  13987. }
  13988. #ifdef WOLFSSL_HAVE_SP_DH
  13989. #ifdef HAVE_FFDHE_4096
  13990. static void sp_4096_lshift_128(sp_digit* r, sp_digit* a, byte n)
  13991. {
  13992. __asm__ __volatile__ (
  13993. "mov r6, #31\n\t"
  13994. "sub r6, r6, %[n]\n\t"
  13995. "add %[a], %[a], #255\n\t"
  13996. "add %[r], %[r], #255\n\t"
  13997. "add %[a], %[a], #193\n\t"
  13998. "add %[r], %[r], #193\n\t"
  13999. "ldr r3, [%[a], #60]\n\t"
  14000. "lsr r4, r3, #1\n\t"
  14001. "lsl r3, r3, %[n]\n\t"
  14002. "lsr r4, r4, r6\n\t"
  14003. "ldr r2, [%[a], #56]\n\t"
  14004. "str r4, [%[r], #64]\n\t"
  14005. "lsr r5, r2, #1\n\t"
  14006. "lsl r2, r2, %[n]\n\t"
  14007. "lsr r5, r5, r6\n\t"
  14008. "orr r3, r3, r5\n\t"
  14009. "ldr r4, [%[a], #52]\n\t"
  14010. "str r3, [%[r], #60]\n\t"
  14011. "lsr r5, r4, #1\n\t"
  14012. "lsl r4, r4, %[n]\n\t"
  14013. "lsr r5, r5, r6\n\t"
  14014. "orr r2, r2, r5\n\t"
  14015. "ldr r3, [%[a], #48]\n\t"
  14016. "str r2, [%[r], #56]\n\t"
  14017. "lsr r5, r3, #1\n\t"
  14018. "lsl r3, r3, %[n]\n\t"
  14019. "lsr r5, r5, r6\n\t"
  14020. "orr r4, r4, r5\n\t"
  14021. "ldr r2, [%[a], #44]\n\t"
  14022. "str r4, [%[r], #52]\n\t"
  14023. "lsr r5, r2, #1\n\t"
  14024. "lsl r2, r2, %[n]\n\t"
  14025. "lsr r5, r5, r6\n\t"
  14026. "orr r3, r3, r5\n\t"
  14027. "ldr r4, [%[a], #40]\n\t"
  14028. "str r3, [%[r], #48]\n\t"
  14029. "lsr r5, r4, #1\n\t"
  14030. "lsl r4, r4, %[n]\n\t"
  14031. "lsr r5, r5, r6\n\t"
  14032. "orr r2, r2, r5\n\t"
  14033. "ldr r3, [%[a], #36]\n\t"
  14034. "str r2, [%[r], #44]\n\t"
  14035. "lsr r5, r3, #1\n\t"
  14036. "lsl r3, r3, %[n]\n\t"
  14037. "lsr r5, r5, r6\n\t"
  14038. "orr r4, r4, r5\n\t"
  14039. "ldr r2, [%[a], #32]\n\t"
  14040. "str r4, [%[r], #40]\n\t"
  14041. "lsr r5, r2, #1\n\t"
  14042. "lsl r2, r2, %[n]\n\t"
  14043. "lsr r5, r5, r6\n\t"
  14044. "orr r3, r3, r5\n\t"
  14045. "ldr r4, [%[a], #28]\n\t"
  14046. "str r3, [%[r], #36]\n\t"
  14047. "lsr r5, r4, #1\n\t"
  14048. "lsl r4, r4, %[n]\n\t"
  14049. "lsr r5, r5, r6\n\t"
  14050. "orr r2, r2, r5\n\t"
  14051. "ldr r3, [%[a], #24]\n\t"
  14052. "str r2, [%[r], #32]\n\t"
  14053. "lsr r5, r3, #1\n\t"
  14054. "lsl r3, r3, %[n]\n\t"
  14055. "lsr r5, r5, r6\n\t"
  14056. "orr r4, r4, r5\n\t"
  14057. "ldr r2, [%[a], #20]\n\t"
  14058. "str r4, [%[r], #28]\n\t"
  14059. "lsr r5, r2, #1\n\t"
  14060. "lsl r2, r2, %[n]\n\t"
  14061. "lsr r5, r5, r6\n\t"
  14062. "orr r3, r3, r5\n\t"
  14063. "ldr r4, [%[a], #16]\n\t"
  14064. "str r3, [%[r], #24]\n\t"
  14065. "lsr r5, r4, #1\n\t"
  14066. "lsl r4, r4, %[n]\n\t"
  14067. "lsr r5, r5, r6\n\t"
  14068. "orr r2, r2, r5\n\t"
  14069. "ldr r3, [%[a], #12]\n\t"
  14070. "str r2, [%[r], #20]\n\t"
  14071. "lsr r5, r3, #1\n\t"
  14072. "lsl r3, r3, %[n]\n\t"
  14073. "lsr r5, r5, r6\n\t"
  14074. "orr r4, r4, r5\n\t"
  14075. "ldr r2, [%[a], #8]\n\t"
  14076. "str r4, [%[r], #16]\n\t"
  14077. "lsr r5, r2, #1\n\t"
  14078. "lsl r2, r2, %[n]\n\t"
  14079. "lsr r5, r5, r6\n\t"
  14080. "orr r3, r3, r5\n\t"
  14081. "ldr r4, [%[a], #4]\n\t"
  14082. "str r3, [%[r], #12]\n\t"
  14083. "lsr r5, r4, #1\n\t"
  14084. "lsl r4, r4, %[n]\n\t"
  14085. "lsr r5, r5, r6\n\t"
  14086. "orr r2, r2, r5\n\t"
  14087. "ldr r3, [%[a], #0]\n\t"
  14088. "str r2, [%[r], #8]\n\t"
  14089. "lsr r5, r3, #1\n\t"
  14090. "lsl r3, r3, %[n]\n\t"
  14091. "lsr r5, r5, r6\n\t"
  14092. "orr r4, r4, r5\n\t"
  14093. "sub %[a], %[a], #64\n\t"
  14094. "sub %[r], %[r], #64\n\t"
  14095. "ldr r2, [%[a], #60]\n\t"
  14096. "str r4, [%[r], #68]\n\t"
  14097. "lsr r5, r2, #1\n\t"
  14098. "lsl r2, r2, %[n]\n\t"
  14099. "lsr r5, r5, r6\n\t"
  14100. "orr r3, r3, r5\n\t"
  14101. "ldr r4, [%[a], #56]\n\t"
  14102. "str r3, [%[r], #64]\n\t"
  14103. "lsr r5, r4, #1\n\t"
  14104. "lsl r4, r4, %[n]\n\t"
  14105. "lsr r5, r5, r6\n\t"
  14106. "orr r2, r2, r5\n\t"
  14107. "ldr r3, [%[a], #52]\n\t"
  14108. "str r2, [%[r], #60]\n\t"
  14109. "lsr r5, r3, #1\n\t"
  14110. "lsl r3, r3, %[n]\n\t"
  14111. "lsr r5, r5, r6\n\t"
  14112. "orr r4, r4, r5\n\t"
  14113. "ldr r2, [%[a], #48]\n\t"
  14114. "str r4, [%[r], #56]\n\t"
  14115. "lsr r5, r2, #1\n\t"
  14116. "lsl r2, r2, %[n]\n\t"
  14117. "lsr r5, r5, r6\n\t"
  14118. "orr r3, r3, r5\n\t"
  14119. "ldr r4, [%[a], #44]\n\t"
  14120. "str r3, [%[r], #52]\n\t"
  14121. "lsr r5, r4, #1\n\t"
  14122. "lsl r4, r4, %[n]\n\t"
  14123. "lsr r5, r5, r6\n\t"
  14124. "orr r2, r2, r5\n\t"
  14125. "ldr r3, [%[a], #40]\n\t"
  14126. "str r2, [%[r], #48]\n\t"
  14127. "lsr r5, r3, #1\n\t"
  14128. "lsl r3, r3, %[n]\n\t"
  14129. "lsr r5, r5, r6\n\t"
  14130. "orr r4, r4, r5\n\t"
  14131. "ldr r2, [%[a], #36]\n\t"
  14132. "str r4, [%[r], #44]\n\t"
  14133. "lsr r5, r2, #1\n\t"
  14134. "lsl r2, r2, %[n]\n\t"
  14135. "lsr r5, r5, r6\n\t"
  14136. "orr r3, r3, r5\n\t"
  14137. "ldr r4, [%[a], #32]\n\t"
  14138. "str r3, [%[r], #40]\n\t"
  14139. "lsr r5, r4, #1\n\t"
  14140. "lsl r4, r4, %[n]\n\t"
  14141. "lsr r5, r5, r6\n\t"
  14142. "orr r2, r2, r5\n\t"
  14143. "ldr r3, [%[a], #28]\n\t"
  14144. "str r2, [%[r], #36]\n\t"
  14145. "lsr r5, r3, #1\n\t"
  14146. "lsl r3, r3, %[n]\n\t"
  14147. "lsr r5, r5, r6\n\t"
  14148. "orr r4, r4, r5\n\t"
  14149. "ldr r2, [%[a], #24]\n\t"
  14150. "str r4, [%[r], #32]\n\t"
  14151. "lsr r5, r2, #1\n\t"
  14152. "lsl r2, r2, %[n]\n\t"
  14153. "lsr r5, r5, r6\n\t"
  14154. "orr r3, r3, r5\n\t"
  14155. "ldr r4, [%[a], #20]\n\t"
  14156. "str r3, [%[r], #28]\n\t"
  14157. "lsr r5, r4, #1\n\t"
  14158. "lsl r4, r4, %[n]\n\t"
  14159. "lsr r5, r5, r6\n\t"
  14160. "orr r2, r2, r5\n\t"
  14161. "ldr r3, [%[a], #16]\n\t"
  14162. "str r2, [%[r], #24]\n\t"
  14163. "lsr r5, r3, #1\n\t"
  14164. "lsl r3, r3, %[n]\n\t"
  14165. "lsr r5, r5, r6\n\t"
  14166. "orr r4, r4, r5\n\t"
  14167. "ldr r2, [%[a], #12]\n\t"
  14168. "str r4, [%[r], #20]\n\t"
  14169. "lsr r5, r2, #1\n\t"
  14170. "lsl r2, r2, %[n]\n\t"
  14171. "lsr r5, r5, r6\n\t"
  14172. "orr r3, r3, r5\n\t"
  14173. "ldr r4, [%[a], #8]\n\t"
  14174. "str r3, [%[r], #16]\n\t"
  14175. "lsr r5, r4, #1\n\t"
  14176. "lsl r4, r4, %[n]\n\t"
  14177. "lsr r5, r5, r6\n\t"
  14178. "orr r2, r2, r5\n\t"
  14179. "ldr r3, [%[a], #4]\n\t"
  14180. "str r2, [%[r], #12]\n\t"
  14181. "lsr r5, r3, #1\n\t"
  14182. "lsl r3, r3, %[n]\n\t"
  14183. "lsr r5, r5, r6\n\t"
  14184. "orr r4, r4, r5\n\t"
  14185. "ldr r2, [%[a], #0]\n\t"
  14186. "str r4, [%[r], #8]\n\t"
  14187. "lsr r5, r2, #1\n\t"
  14188. "lsl r2, r2, %[n]\n\t"
  14189. "lsr r5, r5, r6\n\t"
  14190. "orr r3, r3, r5\n\t"
  14191. "sub %[a], %[a], #64\n\t"
  14192. "sub %[r], %[r], #64\n\t"
  14193. "ldr r4, [%[a], #60]\n\t"
  14194. "str r3, [%[r], #68]\n\t"
  14195. "lsr r5, r4, #1\n\t"
  14196. "lsl r4, r4, %[n]\n\t"
  14197. "lsr r5, r5, r6\n\t"
  14198. "orr r2, r2, r5\n\t"
  14199. "ldr r3, [%[a], #56]\n\t"
  14200. "str r2, [%[r], #64]\n\t"
  14201. "lsr r5, r3, #1\n\t"
  14202. "lsl r3, r3, %[n]\n\t"
  14203. "lsr r5, r5, r6\n\t"
  14204. "orr r4, r4, r5\n\t"
  14205. "ldr r2, [%[a], #52]\n\t"
  14206. "str r4, [%[r], #60]\n\t"
  14207. "lsr r5, r2, #1\n\t"
  14208. "lsl r2, r2, %[n]\n\t"
  14209. "lsr r5, r5, r6\n\t"
  14210. "orr r3, r3, r5\n\t"
  14211. "ldr r4, [%[a], #48]\n\t"
  14212. "str r3, [%[r], #56]\n\t"
  14213. "lsr r5, r4, #1\n\t"
  14214. "lsl r4, r4, %[n]\n\t"
  14215. "lsr r5, r5, r6\n\t"
  14216. "orr r2, r2, r5\n\t"
  14217. "ldr r3, [%[a], #44]\n\t"
  14218. "str r2, [%[r], #52]\n\t"
  14219. "lsr r5, r3, #1\n\t"
  14220. "lsl r3, r3, %[n]\n\t"
  14221. "lsr r5, r5, r6\n\t"
  14222. "orr r4, r4, r5\n\t"
  14223. "ldr r2, [%[a], #40]\n\t"
  14224. "str r4, [%[r], #48]\n\t"
  14225. "lsr r5, r2, #1\n\t"
  14226. "lsl r2, r2, %[n]\n\t"
  14227. "lsr r5, r5, r6\n\t"
  14228. "orr r3, r3, r5\n\t"
  14229. "ldr r4, [%[a], #36]\n\t"
  14230. "str r3, [%[r], #44]\n\t"
  14231. "lsr r5, r4, #1\n\t"
  14232. "lsl r4, r4, %[n]\n\t"
  14233. "lsr r5, r5, r6\n\t"
  14234. "orr r2, r2, r5\n\t"
  14235. "ldr r3, [%[a], #32]\n\t"
  14236. "str r2, [%[r], #40]\n\t"
  14237. "lsr r5, r3, #1\n\t"
  14238. "lsl r3, r3, %[n]\n\t"
  14239. "lsr r5, r5, r6\n\t"
  14240. "orr r4, r4, r5\n\t"
  14241. "ldr r2, [%[a], #28]\n\t"
  14242. "str r4, [%[r], #36]\n\t"
  14243. "lsr r5, r2, #1\n\t"
  14244. "lsl r2, r2, %[n]\n\t"
  14245. "lsr r5, r5, r6\n\t"
  14246. "orr r3, r3, r5\n\t"
  14247. "ldr r4, [%[a], #24]\n\t"
  14248. "str r3, [%[r], #32]\n\t"
  14249. "lsr r5, r4, #1\n\t"
  14250. "lsl r4, r4, %[n]\n\t"
  14251. "lsr r5, r5, r6\n\t"
  14252. "orr r2, r2, r5\n\t"
  14253. "ldr r3, [%[a], #20]\n\t"
  14254. "str r2, [%[r], #28]\n\t"
  14255. "lsr r5, r3, #1\n\t"
  14256. "lsl r3, r3, %[n]\n\t"
  14257. "lsr r5, r5, r6\n\t"
  14258. "orr r4, r4, r5\n\t"
  14259. "ldr r2, [%[a], #16]\n\t"
  14260. "str r4, [%[r], #24]\n\t"
  14261. "lsr r5, r2, #1\n\t"
  14262. "lsl r2, r2, %[n]\n\t"
  14263. "lsr r5, r5, r6\n\t"
  14264. "orr r3, r3, r5\n\t"
  14265. "ldr r4, [%[a], #12]\n\t"
  14266. "str r3, [%[r], #20]\n\t"
  14267. "lsr r5, r4, #1\n\t"
  14268. "lsl r4, r4, %[n]\n\t"
  14269. "lsr r5, r5, r6\n\t"
  14270. "orr r2, r2, r5\n\t"
  14271. "ldr r3, [%[a], #8]\n\t"
  14272. "str r2, [%[r], #16]\n\t"
  14273. "lsr r5, r3, #1\n\t"
  14274. "lsl r3, r3, %[n]\n\t"
  14275. "lsr r5, r5, r6\n\t"
  14276. "orr r4, r4, r5\n\t"
  14277. "ldr r2, [%[a], #4]\n\t"
  14278. "str r4, [%[r], #12]\n\t"
  14279. "lsr r5, r2, #1\n\t"
  14280. "lsl r2, r2, %[n]\n\t"
  14281. "lsr r5, r5, r6\n\t"
  14282. "orr r3, r3, r5\n\t"
  14283. "ldr r4, [%[a], #0]\n\t"
  14284. "str r3, [%[r], #8]\n\t"
  14285. "lsr r5, r4, #1\n\t"
  14286. "lsl r4, r4, %[n]\n\t"
  14287. "lsr r5, r5, r6\n\t"
  14288. "orr r2, r2, r5\n\t"
  14289. "sub %[a], %[a], #64\n\t"
  14290. "sub %[r], %[r], #64\n\t"
  14291. "ldr r3, [%[a], #60]\n\t"
  14292. "str r2, [%[r], #68]\n\t"
  14293. "lsr r5, r3, #1\n\t"
  14294. "lsl r3, r3, %[n]\n\t"
  14295. "lsr r5, r5, r6\n\t"
  14296. "orr r4, r4, r5\n\t"
  14297. "ldr r2, [%[a], #56]\n\t"
  14298. "str r4, [%[r], #64]\n\t"
  14299. "lsr r5, r2, #1\n\t"
  14300. "lsl r2, r2, %[n]\n\t"
  14301. "lsr r5, r5, r6\n\t"
  14302. "orr r3, r3, r5\n\t"
  14303. "ldr r4, [%[a], #52]\n\t"
  14304. "str r3, [%[r], #60]\n\t"
  14305. "lsr r5, r4, #1\n\t"
  14306. "lsl r4, r4, %[n]\n\t"
  14307. "lsr r5, r5, r6\n\t"
  14308. "orr r2, r2, r5\n\t"
  14309. "ldr r3, [%[a], #48]\n\t"
  14310. "str r2, [%[r], #56]\n\t"
  14311. "lsr r5, r3, #1\n\t"
  14312. "lsl r3, r3, %[n]\n\t"
  14313. "lsr r5, r5, r6\n\t"
  14314. "orr r4, r4, r5\n\t"
  14315. "ldr r2, [%[a], #44]\n\t"
  14316. "str r4, [%[r], #52]\n\t"
  14317. "lsr r5, r2, #1\n\t"
  14318. "lsl r2, r2, %[n]\n\t"
  14319. "lsr r5, r5, r6\n\t"
  14320. "orr r3, r3, r5\n\t"
  14321. "ldr r4, [%[a], #40]\n\t"
  14322. "str r3, [%[r], #48]\n\t"
  14323. "lsr r5, r4, #1\n\t"
  14324. "lsl r4, r4, %[n]\n\t"
  14325. "lsr r5, r5, r6\n\t"
  14326. "orr r2, r2, r5\n\t"
  14327. "ldr r3, [%[a], #36]\n\t"
  14328. "str r2, [%[r], #44]\n\t"
  14329. "lsr r5, r3, #1\n\t"
  14330. "lsl r3, r3, %[n]\n\t"
  14331. "lsr r5, r5, r6\n\t"
  14332. "orr r4, r4, r5\n\t"
  14333. "ldr r2, [%[a], #32]\n\t"
  14334. "str r4, [%[r], #40]\n\t"
  14335. "lsr r5, r2, #1\n\t"
  14336. "lsl r2, r2, %[n]\n\t"
  14337. "lsr r5, r5, r6\n\t"
  14338. "orr r3, r3, r5\n\t"
  14339. "ldr r4, [%[a], #28]\n\t"
  14340. "str r3, [%[r], #36]\n\t"
  14341. "lsr r5, r4, #1\n\t"
  14342. "lsl r4, r4, %[n]\n\t"
  14343. "lsr r5, r5, r6\n\t"
  14344. "orr r2, r2, r5\n\t"
  14345. "ldr r3, [%[a], #24]\n\t"
  14346. "str r2, [%[r], #32]\n\t"
  14347. "lsr r5, r3, #1\n\t"
  14348. "lsl r3, r3, %[n]\n\t"
  14349. "lsr r5, r5, r6\n\t"
  14350. "orr r4, r4, r5\n\t"
  14351. "ldr r2, [%[a], #20]\n\t"
  14352. "str r4, [%[r], #28]\n\t"
  14353. "lsr r5, r2, #1\n\t"
  14354. "lsl r2, r2, %[n]\n\t"
  14355. "lsr r5, r5, r6\n\t"
  14356. "orr r3, r3, r5\n\t"
  14357. "ldr r4, [%[a], #16]\n\t"
  14358. "str r3, [%[r], #24]\n\t"
  14359. "lsr r5, r4, #1\n\t"
  14360. "lsl r4, r4, %[n]\n\t"
  14361. "lsr r5, r5, r6\n\t"
  14362. "orr r2, r2, r5\n\t"
  14363. "ldr r3, [%[a], #12]\n\t"
  14364. "str r2, [%[r], #20]\n\t"
  14365. "lsr r5, r3, #1\n\t"
  14366. "lsl r3, r3, %[n]\n\t"
  14367. "lsr r5, r5, r6\n\t"
  14368. "orr r4, r4, r5\n\t"
  14369. "ldr r2, [%[a], #8]\n\t"
  14370. "str r4, [%[r], #16]\n\t"
  14371. "lsr r5, r2, #1\n\t"
  14372. "lsl r2, r2, %[n]\n\t"
  14373. "lsr r5, r5, r6\n\t"
  14374. "orr r3, r3, r5\n\t"
  14375. "ldr r4, [%[a], #4]\n\t"
  14376. "str r3, [%[r], #12]\n\t"
  14377. "lsr r5, r4, #1\n\t"
  14378. "lsl r4, r4, %[n]\n\t"
  14379. "lsr r5, r5, r6\n\t"
  14380. "orr r2, r2, r5\n\t"
  14381. "ldr r3, [%[a], #0]\n\t"
  14382. "str r2, [%[r], #8]\n\t"
  14383. "lsr r5, r3, #1\n\t"
  14384. "lsl r3, r3, %[n]\n\t"
  14385. "lsr r5, r5, r6\n\t"
  14386. "orr r4, r4, r5\n\t"
  14387. "sub %[a], %[a], #64\n\t"
  14388. "sub %[r], %[r], #64\n\t"
  14389. "ldr r2, [%[a], #60]\n\t"
  14390. "str r4, [%[r], #68]\n\t"
  14391. "lsr r5, r2, #1\n\t"
  14392. "lsl r2, r2, %[n]\n\t"
  14393. "lsr r5, r5, r6\n\t"
  14394. "orr r3, r3, r5\n\t"
  14395. "ldr r4, [%[a], #56]\n\t"
  14396. "str r3, [%[r], #64]\n\t"
  14397. "lsr r5, r4, #1\n\t"
  14398. "lsl r4, r4, %[n]\n\t"
  14399. "lsr r5, r5, r6\n\t"
  14400. "orr r2, r2, r5\n\t"
  14401. "ldr r3, [%[a], #52]\n\t"
  14402. "str r2, [%[r], #60]\n\t"
  14403. "lsr r5, r3, #1\n\t"
  14404. "lsl r3, r3, %[n]\n\t"
  14405. "lsr r5, r5, r6\n\t"
  14406. "orr r4, r4, r5\n\t"
  14407. "ldr r2, [%[a], #48]\n\t"
  14408. "str r4, [%[r], #56]\n\t"
  14409. "lsr r5, r2, #1\n\t"
  14410. "lsl r2, r2, %[n]\n\t"
  14411. "lsr r5, r5, r6\n\t"
  14412. "orr r3, r3, r5\n\t"
  14413. "ldr r4, [%[a], #44]\n\t"
  14414. "str r3, [%[r], #52]\n\t"
  14415. "lsr r5, r4, #1\n\t"
  14416. "lsl r4, r4, %[n]\n\t"
  14417. "lsr r5, r5, r6\n\t"
  14418. "orr r2, r2, r5\n\t"
  14419. "ldr r3, [%[a], #40]\n\t"
  14420. "str r2, [%[r], #48]\n\t"
  14421. "lsr r5, r3, #1\n\t"
  14422. "lsl r3, r3, %[n]\n\t"
  14423. "lsr r5, r5, r6\n\t"
  14424. "orr r4, r4, r5\n\t"
  14425. "ldr r2, [%[a], #36]\n\t"
  14426. "str r4, [%[r], #44]\n\t"
  14427. "lsr r5, r2, #1\n\t"
  14428. "lsl r2, r2, %[n]\n\t"
  14429. "lsr r5, r5, r6\n\t"
  14430. "orr r3, r3, r5\n\t"
  14431. "ldr r4, [%[a], #32]\n\t"
  14432. "str r3, [%[r], #40]\n\t"
  14433. "lsr r5, r4, #1\n\t"
  14434. "lsl r4, r4, %[n]\n\t"
  14435. "lsr r5, r5, r6\n\t"
  14436. "orr r2, r2, r5\n\t"
  14437. "ldr r3, [%[a], #28]\n\t"
  14438. "str r2, [%[r], #36]\n\t"
  14439. "lsr r5, r3, #1\n\t"
  14440. "lsl r3, r3, %[n]\n\t"
  14441. "lsr r5, r5, r6\n\t"
  14442. "orr r4, r4, r5\n\t"
  14443. "ldr r2, [%[a], #24]\n\t"
  14444. "str r4, [%[r], #32]\n\t"
  14445. "lsr r5, r2, #1\n\t"
  14446. "lsl r2, r2, %[n]\n\t"
  14447. "lsr r5, r5, r6\n\t"
  14448. "orr r3, r3, r5\n\t"
  14449. "ldr r4, [%[a], #20]\n\t"
  14450. "str r3, [%[r], #28]\n\t"
  14451. "lsr r5, r4, #1\n\t"
  14452. "lsl r4, r4, %[n]\n\t"
  14453. "lsr r5, r5, r6\n\t"
  14454. "orr r2, r2, r5\n\t"
  14455. "ldr r3, [%[a], #16]\n\t"
  14456. "str r2, [%[r], #24]\n\t"
  14457. "lsr r5, r3, #1\n\t"
  14458. "lsl r3, r3, %[n]\n\t"
  14459. "lsr r5, r5, r6\n\t"
  14460. "orr r4, r4, r5\n\t"
  14461. "ldr r2, [%[a], #12]\n\t"
  14462. "str r4, [%[r], #20]\n\t"
  14463. "lsr r5, r2, #1\n\t"
  14464. "lsl r2, r2, %[n]\n\t"
  14465. "lsr r5, r5, r6\n\t"
  14466. "orr r3, r3, r5\n\t"
  14467. "ldr r4, [%[a], #8]\n\t"
  14468. "str r3, [%[r], #16]\n\t"
  14469. "lsr r5, r4, #1\n\t"
  14470. "lsl r4, r4, %[n]\n\t"
  14471. "lsr r5, r5, r6\n\t"
  14472. "orr r2, r2, r5\n\t"
  14473. "ldr r3, [%[a], #4]\n\t"
  14474. "str r2, [%[r], #12]\n\t"
  14475. "lsr r5, r3, #1\n\t"
  14476. "lsl r3, r3, %[n]\n\t"
  14477. "lsr r5, r5, r6\n\t"
  14478. "orr r4, r4, r5\n\t"
  14479. "ldr r2, [%[a], #0]\n\t"
  14480. "str r4, [%[r], #8]\n\t"
  14481. "lsr r5, r2, #1\n\t"
  14482. "lsl r2, r2, %[n]\n\t"
  14483. "lsr r5, r5, r6\n\t"
  14484. "orr r3, r3, r5\n\t"
  14485. "sub %[a], %[a], #64\n\t"
  14486. "sub %[r], %[r], #64\n\t"
  14487. "ldr r4, [%[a], #60]\n\t"
  14488. "str r3, [%[r], #68]\n\t"
  14489. "lsr r5, r4, #1\n\t"
  14490. "lsl r4, r4, %[n]\n\t"
  14491. "lsr r5, r5, r6\n\t"
  14492. "orr r2, r2, r5\n\t"
  14493. "ldr r3, [%[a], #56]\n\t"
  14494. "str r2, [%[r], #64]\n\t"
  14495. "lsr r5, r3, #1\n\t"
  14496. "lsl r3, r3, %[n]\n\t"
  14497. "lsr r5, r5, r6\n\t"
  14498. "orr r4, r4, r5\n\t"
  14499. "ldr r2, [%[a], #52]\n\t"
  14500. "str r4, [%[r], #60]\n\t"
  14501. "lsr r5, r2, #1\n\t"
  14502. "lsl r2, r2, %[n]\n\t"
  14503. "lsr r5, r5, r6\n\t"
  14504. "orr r3, r3, r5\n\t"
  14505. "ldr r4, [%[a], #48]\n\t"
  14506. "str r3, [%[r], #56]\n\t"
  14507. "lsr r5, r4, #1\n\t"
  14508. "lsl r4, r4, %[n]\n\t"
  14509. "lsr r5, r5, r6\n\t"
  14510. "orr r2, r2, r5\n\t"
  14511. "ldr r3, [%[a], #44]\n\t"
  14512. "str r2, [%[r], #52]\n\t"
  14513. "lsr r5, r3, #1\n\t"
  14514. "lsl r3, r3, %[n]\n\t"
  14515. "lsr r5, r5, r6\n\t"
  14516. "orr r4, r4, r5\n\t"
  14517. "ldr r2, [%[a], #40]\n\t"
  14518. "str r4, [%[r], #48]\n\t"
  14519. "lsr r5, r2, #1\n\t"
  14520. "lsl r2, r2, %[n]\n\t"
  14521. "lsr r5, r5, r6\n\t"
  14522. "orr r3, r3, r5\n\t"
  14523. "ldr r4, [%[a], #36]\n\t"
  14524. "str r3, [%[r], #44]\n\t"
  14525. "lsr r5, r4, #1\n\t"
  14526. "lsl r4, r4, %[n]\n\t"
  14527. "lsr r5, r5, r6\n\t"
  14528. "orr r2, r2, r5\n\t"
  14529. "ldr r3, [%[a], #32]\n\t"
  14530. "str r2, [%[r], #40]\n\t"
  14531. "lsr r5, r3, #1\n\t"
  14532. "lsl r3, r3, %[n]\n\t"
  14533. "lsr r5, r5, r6\n\t"
  14534. "orr r4, r4, r5\n\t"
  14535. "ldr r2, [%[a], #28]\n\t"
  14536. "str r4, [%[r], #36]\n\t"
  14537. "lsr r5, r2, #1\n\t"
  14538. "lsl r2, r2, %[n]\n\t"
  14539. "lsr r5, r5, r6\n\t"
  14540. "orr r3, r3, r5\n\t"
  14541. "ldr r4, [%[a], #24]\n\t"
  14542. "str r3, [%[r], #32]\n\t"
  14543. "lsr r5, r4, #1\n\t"
  14544. "lsl r4, r4, %[n]\n\t"
  14545. "lsr r5, r5, r6\n\t"
  14546. "orr r2, r2, r5\n\t"
  14547. "ldr r3, [%[a], #20]\n\t"
  14548. "str r2, [%[r], #28]\n\t"
  14549. "lsr r5, r3, #1\n\t"
  14550. "lsl r3, r3, %[n]\n\t"
  14551. "lsr r5, r5, r6\n\t"
  14552. "orr r4, r4, r5\n\t"
  14553. "ldr r2, [%[a], #16]\n\t"
  14554. "str r4, [%[r], #24]\n\t"
  14555. "lsr r5, r2, #1\n\t"
  14556. "lsl r2, r2, %[n]\n\t"
  14557. "lsr r5, r5, r6\n\t"
  14558. "orr r3, r3, r5\n\t"
  14559. "ldr r4, [%[a], #12]\n\t"
  14560. "str r3, [%[r], #20]\n\t"
  14561. "lsr r5, r4, #1\n\t"
  14562. "lsl r4, r4, %[n]\n\t"
  14563. "lsr r5, r5, r6\n\t"
  14564. "orr r2, r2, r5\n\t"
  14565. "ldr r3, [%[a], #8]\n\t"
  14566. "str r2, [%[r], #16]\n\t"
  14567. "lsr r5, r3, #1\n\t"
  14568. "lsl r3, r3, %[n]\n\t"
  14569. "lsr r5, r5, r6\n\t"
  14570. "orr r4, r4, r5\n\t"
  14571. "ldr r2, [%[a], #4]\n\t"
  14572. "str r4, [%[r], #12]\n\t"
  14573. "lsr r5, r2, #1\n\t"
  14574. "lsl r2, r2, %[n]\n\t"
  14575. "lsr r5, r5, r6\n\t"
  14576. "orr r3, r3, r5\n\t"
  14577. "ldr r4, [%[a], #0]\n\t"
  14578. "str r3, [%[r], #8]\n\t"
  14579. "lsr r5, r4, #1\n\t"
  14580. "lsl r4, r4, %[n]\n\t"
  14581. "lsr r5, r5, r6\n\t"
  14582. "orr r2, r2, r5\n\t"
  14583. "sub %[a], %[a], #64\n\t"
  14584. "sub %[r], %[r], #64\n\t"
  14585. "ldr r3, [%[a], #60]\n\t"
  14586. "str r2, [%[r], #68]\n\t"
  14587. "lsr r5, r3, #1\n\t"
  14588. "lsl r3, r3, %[n]\n\t"
  14589. "lsr r5, r5, r6\n\t"
  14590. "orr r4, r4, r5\n\t"
  14591. "ldr r2, [%[a], #56]\n\t"
  14592. "str r4, [%[r], #64]\n\t"
  14593. "lsr r5, r2, #1\n\t"
  14594. "lsl r2, r2, %[n]\n\t"
  14595. "lsr r5, r5, r6\n\t"
  14596. "orr r3, r3, r5\n\t"
  14597. "ldr r4, [%[a], #52]\n\t"
  14598. "str r3, [%[r], #60]\n\t"
  14599. "lsr r5, r4, #1\n\t"
  14600. "lsl r4, r4, %[n]\n\t"
  14601. "lsr r5, r5, r6\n\t"
  14602. "orr r2, r2, r5\n\t"
  14603. "ldr r3, [%[a], #48]\n\t"
  14604. "str r2, [%[r], #56]\n\t"
  14605. "lsr r5, r3, #1\n\t"
  14606. "lsl r3, r3, %[n]\n\t"
  14607. "lsr r5, r5, r6\n\t"
  14608. "orr r4, r4, r5\n\t"
  14609. "ldr r2, [%[a], #44]\n\t"
  14610. "str r4, [%[r], #52]\n\t"
  14611. "lsr r5, r2, #1\n\t"
  14612. "lsl r2, r2, %[n]\n\t"
  14613. "lsr r5, r5, r6\n\t"
  14614. "orr r3, r3, r5\n\t"
  14615. "ldr r4, [%[a], #40]\n\t"
  14616. "str r3, [%[r], #48]\n\t"
  14617. "lsr r5, r4, #1\n\t"
  14618. "lsl r4, r4, %[n]\n\t"
  14619. "lsr r5, r5, r6\n\t"
  14620. "orr r2, r2, r5\n\t"
  14621. "ldr r3, [%[a], #36]\n\t"
  14622. "str r2, [%[r], #44]\n\t"
  14623. "lsr r5, r3, #1\n\t"
  14624. "lsl r3, r3, %[n]\n\t"
  14625. "lsr r5, r5, r6\n\t"
  14626. "orr r4, r4, r5\n\t"
  14627. "ldr r2, [%[a], #32]\n\t"
  14628. "str r4, [%[r], #40]\n\t"
  14629. "lsr r5, r2, #1\n\t"
  14630. "lsl r2, r2, %[n]\n\t"
  14631. "lsr r5, r5, r6\n\t"
  14632. "orr r3, r3, r5\n\t"
  14633. "ldr r4, [%[a], #28]\n\t"
  14634. "str r3, [%[r], #36]\n\t"
  14635. "lsr r5, r4, #1\n\t"
  14636. "lsl r4, r4, %[n]\n\t"
  14637. "lsr r5, r5, r6\n\t"
  14638. "orr r2, r2, r5\n\t"
  14639. "ldr r3, [%[a], #24]\n\t"
  14640. "str r2, [%[r], #32]\n\t"
  14641. "lsr r5, r3, #1\n\t"
  14642. "lsl r3, r3, %[n]\n\t"
  14643. "lsr r5, r5, r6\n\t"
  14644. "orr r4, r4, r5\n\t"
  14645. "ldr r2, [%[a], #20]\n\t"
  14646. "str r4, [%[r], #28]\n\t"
  14647. "lsr r5, r2, #1\n\t"
  14648. "lsl r2, r2, %[n]\n\t"
  14649. "lsr r5, r5, r6\n\t"
  14650. "orr r3, r3, r5\n\t"
  14651. "ldr r4, [%[a], #16]\n\t"
  14652. "str r3, [%[r], #24]\n\t"
  14653. "lsr r5, r4, #1\n\t"
  14654. "lsl r4, r4, %[n]\n\t"
  14655. "lsr r5, r5, r6\n\t"
  14656. "orr r2, r2, r5\n\t"
  14657. "ldr r3, [%[a], #12]\n\t"
  14658. "str r2, [%[r], #20]\n\t"
  14659. "lsr r5, r3, #1\n\t"
  14660. "lsl r3, r3, %[n]\n\t"
  14661. "lsr r5, r5, r6\n\t"
  14662. "orr r4, r4, r5\n\t"
  14663. "ldr r2, [%[a], #8]\n\t"
  14664. "str r4, [%[r], #16]\n\t"
  14665. "lsr r5, r2, #1\n\t"
  14666. "lsl r2, r2, %[n]\n\t"
  14667. "lsr r5, r5, r6\n\t"
  14668. "orr r3, r3, r5\n\t"
  14669. "ldr r4, [%[a], #4]\n\t"
  14670. "str r3, [%[r], #12]\n\t"
  14671. "lsr r5, r4, #1\n\t"
  14672. "lsl r4, r4, %[n]\n\t"
  14673. "lsr r5, r5, r6\n\t"
  14674. "orr r2, r2, r5\n\t"
  14675. "ldr r3, [%[a], #0]\n\t"
  14676. "str r2, [%[r], #8]\n\t"
  14677. "lsr r5, r3, #1\n\t"
  14678. "lsl r3, r3, %[n]\n\t"
  14679. "lsr r5, r5, r6\n\t"
  14680. "orr r4, r4, r5\n\t"
  14681. "sub %[a], %[a], #64\n\t"
  14682. "sub %[r], %[r], #64\n\t"
  14683. "ldr r2, [%[a], #60]\n\t"
  14684. "str r4, [%[r], #68]\n\t"
  14685. "lsr r5, r2, #1\n\t"
  14686. "lsl r2, r2, %[n]\n\t"
  14687. "lsr r5, r5, r6\n\t"
  14688. "orr r3, r3, r5\n\t"
  14689. "ldr r4, [%[a], #56]\n\t"
  14690. "str r3, [%[r], #64]\n\t"
  14691. "lsr r5, r4, #1\n\t"
  14692. "lsl r4, r4, %[n]\n\t"
  14693. "lsr r5, r5, r6\n\t"
  14694. "orr r2, r2, r5\n\t"
  14695. "ldr r3, [%[a], #52]\n\t"
  14696. "str r2, [%[r], #60]\n\t"
  14697. "lsr r5, r3, #1\n\t"
  14698. "lsl r3, r3, %[n]\n\t"
  14699. "lsr r5, r5, r6\n\t"
  14700. "orr r4, r4, r5\n\t"
  14701. "ldr r2, [%[a], #48]\n\t"
  14702. "str r4, [%[r], #56]\n\t"
  14703. "lsr r5, r2, #1\n\t"
  14704. "lsl r2, r2, %[n]\n\t"
  14705. "lsr r5, r5, r6\n\t"
  14706. "orr r3, r3, r5\n\t"
  14707. "ldr r4, [%[a], #44]\n\t"
  14708. "str r3, [%[r], #52]\n\t"
  14709. "lsr r5, r4, #1\n\t"
  14710. "lsl r4, r4, %[n]\n\t"
  14711. "lsr r5, r5, r6\n\t"
  14712. "orr r2, r2, r5\n\t"
  14713. "ldr r3, [%[a], #40]\n\t"
  14714. "str r2, [%[r], #48]\n\t"
  14715. "lsr r5, r3, #1\n\t"
  14716. "lsl r3, r3, %[n]\n\t"
  14717. "lsr r5, r5, r6\n\t"
  14718. "orr r4, r4, r5\n\t"
  14719. "ldr r2, [%[a], #36]\n\t"
  14720. "str r4, [%[r], #44]\n\t"
  14721. "lsr r5, r2, #1\n\t"
  14722. "lsl r2, r2, %[n]\n\t"
  14723. "lsr r5, r5, r6\n\t"
  14724. "orr r3, r3, r5\n\t"
  14725. "ldr r4, [%[a], #32]\n\t"
  14726. "str r3, [%[r], #40]\n\t"
  14727. "lsr r5, r4, #1\n\t"
  14728. "lsl r4, r4, %[n]\n\t"
  14729. "lsr r5, r5, r6\n\t"
  14730. "orr r2, r2, r5\n\t"
  14731. "ldr r3, [%[a], #28]\n\t"
  14732. "str r2, [%[r], #36]\n\t"
  14733. "lsr r5, r3, #1\n\t"
  14734. "lsl r3, r3, %[n]\n\t"
  14735. "lsr r5, r5, r6\n\t"
  14736. "orr r4, r4, r5\n\t"
  14737. "ldr r2, [%[a], #24]\n\t"
  14738. "str r4, [%[r], #32]\n\t"
  14739. "lsr r5, r2, #1\n\t"
  14740. "lsl r2, r2, %[n]\n\t"
  14741. "lsr r5, r5, r6\n\t"
  14742. "orr r3, r3, r5\n\t"
  14743. "ldr r4, [%[a], #20]\n\t"
  14744. "str r3, [%[r], #28]\n\t"
  14745. "lsr r5, r4, #1\n\t"
  14746. "lsl r4, r4, %[n]\n\t"
  14747. "lsr r5, r5, r6\n\t"
  14748. "orr r2, r2, r5\n\t"
  14749. "ldr r3, [%[a], #16]\n\t"
  14750. "str r2, [%[r], #24]\n\t"
  14751. "lsr r5, r3, #1\n\t"
  14752. "lsl r3, r3, %[n]\n\t"
  14753. "lsr r5, r5, r6\n\t"
  14754. "orr r4, r4, r5\n\t"
  14755. "ldr r2, [%[a], #12]\n\t"
  14756. "str r4, [%[r], #20]\n\t"
  14757. "lsr r5, r2, #1\n\t"
  14758. "lsl r2, r2, %[n]\n\t"
  14759. "lsr r5, r5, r6\n\t"
  14760. "orr r3, r3, r5\n\t"
  14761. "ldr r4, [%[a], #8]\n\t"
  14762. "str r3, [%[r], #16]\n\t"
  14763. "lsr r5, r4, #1\n\t"
  14764. "lsl r4, r4, %[n]\n\t"
  14765. "lsr r5, r5, r6\n\t"
  14766. "orr r2, r2, r5\n\t"
  14767. "ldr r3, [%[a], #4]\n\t"
  14768. "str r2, [%[r], #12]\n\t"
  14769. "lsr r5, r3, #1\n\t"
  14770. "lsl r3, r3, %[n]\n\t"
  14771. "lsr r5, r5, r6\n\t"
  14772. "orr r4, r4, r5\n\t"
  14773. "ldr r2, [%[a], #0]\n\t"
  14774. "str r4, [%[r], #8]\n\t"
  14775. "lsr r5, r2, #1\n\t"
  14776. "lsl r2, r2, %[n]\n\t"
  14777. "lsr r5, r5, r6\n\t"
  14778. "orr r3, r3, r5\n\t"
  14779. "str r2, [%[r]]\n\t"
  14780. "str r3, [%[r], #4]\n\t"
  14781. :
  14782. : [r] "r" (r), [a] "r" (a), [n] "r" (n)
  14783. : "memory", "r2", "r3", "r4", "r5", "r6"
  14784. );
  14785. }
  14786. /* Modular exponentiate 2 to the e mod m. (r = 2^e mod m)
  14787. *
  14788. * r A single precision number that is the result of the operation.
  14789. * e A single precision number that is the exponent.
  14790. * bits The number of bits in the exponent.
  14791. * m A single precision number that is the modulus.
  14792. * returns 0 on success and MEMORY_E on dynamic memory allocation failure.
  14793. */
  14794. static int sp_4096_mod_exp_2_128(sp_digit* r, const sp_digit* e, int bits,
  14795. const sp_digit* m)
  14796. {
  14797. #ifndef WOLFSSL_SMALL_STACK
  14798. sp_digit nd[256];
  14799. sp_digit td[129];
  14800. #else
  14801. sp_digit* td;
  14802. #endif
  14803. sp_digit* norm;
  14804. sp_digit* tmp;
  14805. sp_digit mp = 1;
  14806. sp_digit n, o;
  14807. sp_digit mask;
  14808. int i;
  14809. int c, y;
  14810. int err = MP_OKAY;
  14811. #ifdef WOLFSSL_SMALL_STACK
  14812. td = (sp_digit*)XMALLOC(sizeof(sp_digit) * 385, NULL,
  14813. DYNAMIC_TYPE_TMP_BUFFER);
  14814. if (td == NULL) {
  14815. err = MEMORY_E;
  14816. }
  14817. #endif
  14818. if (err == MP_OKAY) {
  14819. #ifdef WOLFSSL_SMALL_STACK
  14820. norm = td;
  14821. tmp = td + 256;
  14822. #else
  14823. norm = nd;
  14824. tmp = td;
  14825. #endif
  14826. sp_4096_mont_setup(m, &mp);
  14827. sp_4096_mont_norm_128(norm, m);
  14828. i = (bits - 1) / 32;
  14829. n = e[i--];
  14830. c = bits & 31;
  14831. if (c == 0) {
  14832. c = 32;
  14833. }
  14834. c -= bits % 5;
  14835. if (c == 32) {
  14836. c = 27;
  14837. }
  14838. y = (int)(n >> c);
  14839. n <<= 32 - c;
  14840. sp_4096_lshift_128(r, norm, y);
  14841. for (; i>=0 || c>=5; ) {
  14842. if (c == 0) {
  14843. n = e[i--];
  14844. y = n >> 27;
  14845. n <<= 5;
  14846. c = 27;
  14847. }
  14848. else if (c < 5) {
  14849. y = n >> 27;
  14850. n = e[i--];
  14851. c = 5 - c;
  14852. y |= n >> (32 - c);
  14853. n <<= c;
  14854. c = 32 - c;
  14855. }
  14856. else {
  14857. y = (n >> 27) & 0x1f;
  14858. n <<= 5;
  14859. c -= 5;
  14860. }
  14861. sp_4096_mont_sqr_128(r, r, m, mp);
  14862. sp_4096_mont_sqr_128(r, r, m, mp);
  14863. sp_4096_mont_sqr_128(r, r, m, mp);
  14864. sp_4096_mont_sqr_128(r, r, m, mp);
  14865. sp_4096_mont_sqr_128(r, r, m, mp);
  14866. sp_4096_lshift_128(r, r, y);
  14867. sp_4096_mul_d_128(tmp, norm, r[128]);
  14868. r[128] = 0;
  14869. o = sp_4096_add_128(r, r, tmp);
  14870. sp_4096_cond_sub_128(r, r, m, (sp_digit)0 - o);
  14871. }
  14872. XMEMSET(&r[128], 0, sizeof(sp_digit) * 128U);
  14873. sp_4096_mont_reduce_128(r, m, mp);
  14874. mask = 0 - (sp_4096_cmp_128(r, m) >= 0);
  14875. sp_4096_cond_sub_128(r, r, m, mask);
  14876. }
  14877. #ifdef WOLFSSL_SMALL_STACK
  14878. if (td != NULL) {
  14879. XFREE(td, NULL, DYNAMIC_TYPE_TMP_BUFFER);
  14880. }
  14881. #endif
  14882. return err;
  14883. }
  14884. #endif /* HAVE_FFDHE_4096 */
  14885. /* Perform the modular exponentiation for Diffie-Hellman.
  14886. *
  14887. * base Base.
  14888. * exp Array of bytes that is the exponent.
  14889. * expLen Length of data, in bytes, in exponent.
  14890. * mod Modulus.
  14891. * out Buffer to hold big-endian bytes of exponentiation result.
  14892. * Must be at least 512 bytes long.
  14893. * outLen Length, in bytes, of exponentiation result.
  14894. * returns 0 on success, MP_READ_E if there are too many bytes in an array
  14895. * and MEMORY_E if memory allocation fails.
  14896. */
  14897. int sp_DhExp_4096(mp_int* base, const byte* exp, word32 expLen,
  14898. mp_int* mod, byte* out, word32* outLen)
  14899. {
  14900. int err = MP_OKAY;
  14901. sp_digit b[256], e[128], m[128];
  14902. sp_digit* r = b;
  14903. word32 i;
  14904. if (mp_count_bits(base) > 4096) {
  14905. err = MP_READ_E;
  14906. }
  14907. if (err == MP_OKAY) {
  14908. if (expLen > 512) {
  14909. err = MP_READ_E;
  14910. }
  14911. }
  14912. if (err == MP_OKAY) {
  14913. if (mp_count_bits(mod) != 4096) {
  14914. err = MP_READ_E;
  14915. }
  14916. }
  14917. if (err == MP_OKAY) {
  14918. sp_4096_from_mp(b, 128, base);
  14919. sp_4096_from_bin(e, 128, exp, expLen);
  14920. sp_4096_from_mp(m, 128, mod);
  14921. #ifdef HAVE_FFDHE_4096
  14922. if (base->used == 1 && base->dp[0] == 2 && m[127] == (sp_digit)-1)
  14923. err = sp_4096_mod_exp_2_128(r, e, expLen * 8, m);
  14924. else
  14925. #endif
  14926. err = sp_4096_mod_exp_128(r, b, e, expLen * 8, m, 0);
  14927. }
  14928. if (err == MP_OKAY) {
  14929. sp_4096_to_bin(r, out);
  14930. *outLen = 512;
  14931. for (i=0; i<512 && out[i] == 0; i++) {
  14932. }
  14933. *outLen -= i;
  14934. XMEMMOVE(out, out + i, *outLen);
  14935. }
  14936. XMEMSET(e, 0, sizeof(e));
  14937. return err;
  14938. }
  14939. #endif /* WOLFSSL_HAVE_SP_DH */
  14940. #endif /* WOLFSSL_HAVE_SP_DH || (WOLFSSL_HAVE_SP_RSA && !WOLFSSL_RSA_PUBLIC_ONLY) */
  14941. #endif /* WOLFSSL_SP_4096 */
  14942. #endif /* WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH */
  14943. #ifdef WOLFSSL_HAVE_SP_ECC
  14944. #ifndef WOLFSSL_SP_NO_256
  14945. /* Point structure to use. */
  14946. typedef struct sp_point {
  14947. sp_digit x[2 * 8];
  14948. sp_digit y[2 * 8];
  14949. sp_digit z[2 * 8];
  14950. int infinity;
  14951. } sp_point;
  14952. /* The modulus (prime) of the curve P256. */
  14953. static const sp_digit p256_mod[8] = {
  14954. 0xffffffff,0xffffffff,0xffffffff,0x00000000,0x00000000,0x00000000,
  14955. 0x00000001,0xffffffff
  14956. };
  14957. /* The Montogmery normalizer for modulus of the curve P256. */
  14958. static const sp_digit p256_norm_mod[8] = {
  14959. 0x00000001,0x00000000,0x00000000,0xffffffff,0xffffffff,0xffffffff,
  14960. 0xfffffffe,0x00000000
  14961. };
  14962. /* The Montogmery multiplier for modulus of the curve P256. */
  14963. static const sp_digit p256_mp_mod = 0x00000001;
  14964. #if defined(WOLFSSL_VALIDATE_ECC_KEYGEN) || defined(HAVE_ECC_SIGN) || \
  14965. defined(HAVE_ECC_VERIFY)
  14966. /* The order of the curve P256. */
  14967. static const sp_digit p256_order[8] = {
  14968. 0xfc632551,0xf3b9cac2,0xa7179e84,0xbce6faad,0xffffffff,0xffffffff,
  14969. 0x00000000,0xffffffff
  14970. };
  14971. #endif
  14972. /* The order of the curve P256 minus 2. */
  14973. static const sp_digit p256_order2[8] = {
  14974. 0xfc63254f,0xf3b9cac2,0xa7179e84,0xbce6faad,0xffffffff,0xffffffff,
  14975. 0x00000000,0xffffffff
  14976. };
  14977. #if defined(HAVE_ECC_SIGN) || defined(HAVE_ECC_VERIFY)
  14978. /* The Montogmery normalizer for order of the curve P256. */
  14979. static const sp_digit p256_norm_order[8] = {
  14980. 0x039cdaaf,0x0c46353d,0x58e8617b,0x43190552,0x00000000,0x00000000,
  14981. 0xffffffff,0x00000000
  14982. };
  14983. #endif
  14984. #if defined(HAVE_ECC_SIGN) || defined(HAVE_ECC_VERIFY)
  14985. /* The Montogmery multiplier for order of the curve P256. */
  14986. static const sp_digit p256_mp_order = 0xee00bc4f;
  14987. #endif
  14988. /* The base point of curve P256. */
  14989. static const sp_point p256_base = {
  14990. /* X ordinate */
  14991. {
  14992. 0xd898c296,0xf4a13945,0x2deb33a0,0x77037d81,0x63a440f2,0xf8bce6e5,
  14993. 0xe12c4247,0x6b17d1f2, 0L, 0L, 0L, 0L, 0L, 0L, 0L, 0L
  14994. },
  14995. /* Y ordinate */
  14996. {
  14997. 0x37bf51f5,0xcbb64068,0x6b315ece,0x2bce3357,0x7c0f9e16,0x8ee7eb4a,
  14998. 0xfe1a7f9b,0x4fe342e2, 0L, 0L, 0L, 0L, 0L, 0L, 0L, 0L
  14999. },
  15000. /* Z ordinate */
  15001. {
  15002. 0x00000001,0x00000000,0x00000000,0x00000000,0x00000000,0x00000000,
  15003. 0x00000000,0x00000000, 0L, 0L, 0L, 0L, 0L, 0L, 0L, 0L
  15004. },
  15005. /* infinity */
  15006. 0
  15007. };
  15008. #if defined(HAVE_ECC_CHECK_KEY) || defined(HAVE_COMP_KEY)
  15009. static const sp_digit p256_b[8] = {
  15010. 0x27d2604b,0x3bce3c3e,0xcc53b0f6,0x651d06b0,0x769886bc,0xb3ebbd55,
  15011. 0xaa3a93e7,0x5ac635d8
  15012. };
  15013. #endif
  15014. static int sp_ecc_point_new_ex(void* heap, sp_point* sp, sp_point** p)
  15015. {
  15016. int ret = MP_OKAY;
  15017. if (p == NULL) {
  15018. ret = MEMORY_E;
  15019. }
  15020. else {
  15021. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  15022. *p = (sp_point*)XMALLOC(sizeof(sp_point), heap, DYNAMIC_TYPE_ECC);
  15023. (void)sp;
  15024. #else
  15025. *p = sp;
  15026. (void)heap;
  15027. #endif
  15028. }
  15029. return ret;
  15030. }
  15031. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  15032. /* Allocate memory for point and return error. */
  15033. #define sp_ecc_point_new(heap, sp, p) sp_ecc_point_new_ex((heap), NULL, &(p))
  15034. #else
  15035. /* Set pointer to data and return no error. */
  15036. #define sp_ecc_point_new(heap, sp, p) sp_ecc_point_new_ex((heap), &(sp), &(p))
  15037. #endif
  15038. static void sp_ecc_point_free(sp_point* p, int clear, void* heap)
  15039. {
  15040. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  15041. /* If valid pointer then clear point data if requested and free data. */
  15042. if (p != NULL) {
  15043. if (clear != 0) {
  15044. XMEMSET(p, 0, sizeof(*p));
  15045. }
  15046. XFREE(p, heap, DYNAMIC_TYPE_ECC);
  15047. }
  15048. #else
  15049. /* Clear point data if requested. */
  15050. if (clear != 0) {
  15051. XMEMSET(p, 0, sizeof(*p));
  15052. }
  15053. #endif
  15054. (void)heap;
  15055. }
  15056. /* Multiply a number by Montogmery normalizer mod modulus (prime).
  15057. *
  15058. * r The resulting Montgomery form number.
  15059. * a The number to convert.
  15060. * m The modulus (prime).
  15061. */
  15062. static int sp_256_mod_mul_norm_8(sp_digit* r, const sp_digit* a, const sp_digit* m)
  15063. {
  15064. int64_t t[8];
  15065. int64_t a64[8];
  15066. int64_t o;
  15067. (void)m;
  15068. a64[0] = a[0];
  15069. a64[1] = a[1];
  15070. a64[2] = a[2];
  15071. a64[3] = a[3];
  15072. a64[4] = a[4];
  15073. a64[5] = a[5];
  15074. a64[6] = a[6];
  15075. a64[7] = a[7];
  15076. /* 1 1 0 -1 -1 -1 -1 0 */
  15077. t[0] = 0 + a64[0] + a64[1] - a64[3] - a64[4] - a64[5] - a64[6];
  15078. /* 0 1 1 0 -1 -1 -1 -1 */
  15079. t[1] = 0 + a64[1] + a64[2] - a64[4] - a64[5] - a64[6] - a64[7];
  15080. /* 0 0 1 1 0 -1 -1 -1 */
  15081. t[2] = 0 + a64[2] + a64[3] - a64[5] - a64[6] - a64[7];
  15082. /* -1 -1 0 2 2 1 0 -1 */
  15083. t[3] = 0 - a64[0] - a64[1] + 2 * a64[3] + 2 * a64[4] + a64[5] - a64[7];
  15084. /* 0 -1 -1 0 2 2 1 0 */
  15085. t[4] = 0 - a64[1] - a64[2] + 2 * a64[4] + 2 * a64[5] + a64[6];
  15086. /* 0 0 -1 -1 0 2 2 1 */
  15087. t[5] = 0 - a64[2] - a64[3] + 2 * a64[5] + 2 * a64[6] + a64[7];
  15088. /* -1 -1 0 0 0 1 3 2 */
  15089. t[6] = 0 - a64[0] - a64[1] + a64[5] + 3 * a64[6] + 2 * a64[7];
  15090. /* 1 0 -1 -1 -1 -1 0 3 */
  15091. t[7] = 0 + a64[0] - a64[2] - a64[3] - a64[4] - a64[5] + 3 * a64[7];
  15092. t[1] += t[0] >> 32; t[0] &= 0xffffffff;
  15093. t[2] += t[1] >> 32; t[1] &= 0xffffffff;
  15094. t[3] += t[2] >> 32; t[2] &= 0xffffffff;
  15095. t[4] += t[3] >> 32; t[3] &= 0xffffffff;
  15096. t[5] += t[4] >> 32; t[4] &= 0xffffffff;
  15097. t[6] += t[5] >> 32; t[5] &= 0xffffffff;
  15098. t[7] += t[6] >> 32; t[6] &= 0xffffffff;
  15099. o = t[7] >> 32; t[7] &= 0xffffffff;
  15100. t[0] += o;
  15101. t[3] -= o;
  15102. t[6] -= o;
  15103. t[7] += o;
  15104. t[1] += t[0] >> 32; t[0] &= 0xffffffff;
  15105. t[2] += t[1] >> 32; t[1] &= 0xffffffff;
  15106. t[3] += t[2] >> 32; t[2] &= 0xffffffff;
  15107. t[4] += t[3] >> 32; t[3] &= 0xffffffff;
  15108. t[5] += t[4] >> 32; t[4] &= 0xffffffff;
  15109. t[6] += t[5] >> 32; t[5] &= 0xffffffff;
  15110. t[7] += t[6] >> 32; t[6] &= 0xffffffff;
  15111. r[0] = t[0];
  15112. r[1] = t[1];
  15113. r[2] = t[2];
  15114. r[3] = t[3];
  15115. r[4] = t[4];
  15116. r[5] = t[5];
  15117. r[6] = t[6];
  15118. r[7] = t[7];
  15119. return MP_OKAY;
  15120. }
  15121. /* Convert an mp_int to an array of sp_digit.
  15122. *
  15123. * r A single precision integer.
  15124. * size Maximum number of bytes to convert
  15125. * a A multi-precision integer.
  15126. */
  15127. static void sp_256_from_mp(sp_digit* r, int size, const mp_int* a)
  15128. {
  15129. #if DIGIT_BIT == 32
  15130. int j;
  15131. XMEMCPY(r, a->dp, sizeof(sp_digit) * a->used);
  15132. for (j = a->used; j < size; j++) {
  15133. r[j] = 0;
  15134. }
  15135. #elif DIGIT_BIT > 32
  15136. int i, j = 0;
  15137. word32 s = 0;
  15138. r[0] = 0;
  15139. for (i = 0; i < a->used && j < size; i++) {
  15140. r[j] |= ((sp_digit)a->dp[i] << s);
  15141. r[j] &= 0xffffffff;
  15142. s = 32U - s;
  15143. if (j + 1 >= size) {
  15144. break;
  15145. }
  15146. /* lint allow cast of mismatch word32 and mp_digit */
  15147. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  15148. while ((s + 32U) <= (word32)DIGIT_BIT) {
  15149. s += 32U;
  15150. r[j] &= 0xffffffff;
  15151. if (j + 1 >= size) {
  15152. break;
  15153. }
  15154. if (s < (word32)DIGIT_BIT) {
  15155. /* lint allow cast of mismatch word32 and mp_digit */
  15156. r[++j] = (sp_digit)(a->dp[i] >> s); /*lint !e9033*/
  15157. }
  15158. else {
  15159. r[++j] = 0L;
  15160. }
  15161. }
  15162. s = (word32)DIGIT_BIT - s;
  15163. }
  15164. for (j++; j < size; j++) {
  15165. r[j] = 0;
  15166. }
  15167. #else
  15168. int i, j = 0, s = 0;
  15169. r[0] = 0;
  15170. for (i = 0; i < a->used && j < size; i++) {
  15171. r[j] |= ((sp_digit)a->dp[i]) << s;
  15172. if (s + DIGIT_BIT >= 32) {
  15173. r[j] &= 0xffffffff;
  15174. if (j + 1 >= size) {
  15175. break;
  15176. }
  15177. s = 32 - s;
  15178. if (s == DIGIT_BIT) {
  15179. r[++j] = 0;
  15180. s = 0;
  15181. }
  15182. else {
  15183. r[++j] = a->dp[i] >> s;
  15184. s = DIGIT_BIT - s;
  15185. }
  15186. }
  15187. else {
  15188. s += DIGIT_BIT;
  15189. }
  15190. }
  15191. for (j++; j < size; j++) {
  15192. r[j] = 0;
  15193. }
  15194. #endif
  15195. }
  15196. /* Convert a point of type ecc_point to type sp_point.
  15197. *
  15198. * p Point of type sp_point (result).
  15199. * pm Point of type ecc_point.
  15200. */
  15201. static void sp_256_point_from_ecc_point_8(sp_point* p, const ecc_point* pm)
  15202. {
  15203. XMEMSET(p->x, 0, sizeof(p->x));
  15204. XMEMSET(p->y, 0, sizeof(p->y));
  15205. XMEMSET(p->z, 0, sizeof(p->z));
  15206. sp_256_from_mp(p->x, 8, pm->x);
  15207. sp_256_from_mp(p->y, 8, pm->y);
  15208. sp_256_from_mp(p->z, 8, pm->z);
  15209. p->infinity = 0;
  15210. }
  15211. /* Convert an array of sp_digit to an mp_int.
  15212. *
  15213. * a A single precision integer.
  15214. * r A multi-precision integer.
  15215. */
  15216. static int sp_256_to_mp(const sp_digit* a, mp_int* r)
  15217. {
  15218. int err;
  15219. err = mp_grow(r, (256 + DIGIT_BIT - 1) / DIGIT_BIT);
  15220. if (err == MP_OKAY) { /*lint !e774 case where err is always MP_OKAY*/
  15221. #if DIGIT_BIT == 32
  15222. XMEMCPY(r->dp, a, sizeof(sp_digit) * 8);
  15223. r->used = 8;
  15224. mp_clamp(r);
  15225. #elif DIGIT_BIT < 32
  15226. int i, j = 0, s = 0;
  15227. r->dp[0] = 0;
  15228. for (i = 0; i < 8; i++) {
  15229. r->dp[j] |= a[i] << s;
  15230. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  15231. s = DIGIT_BIT - s;
  15232. r->dp[++j] = a[i] >> s;
  15233. while (s + DIGIT_BIT <= 32) {
  15234. s += DIGIT_BIT;
  15235. r->dp[j++] &= (1L << DIGIT_BIT) - 1;
  15236. if (s == SP_WORD_SIZE) {
  15237. r->dp[j] = 0;
  15238. }
  15239. else {
  15240. r->dp[j] = a[i] >> s;
  15241. }
  15242. }
  15243. s = 32 - s;
  15244. }
  15245. r->used = (256 + DIGIT_BIT - 1) / DIGIT_BIT;
  15246. mp_clamp(r);
  15247. #else
  15248. int i, j = 0, s = 0;
  15249. r->dp[0] = 0;
  15250. for (i = 0; i < 8; i++) {
  15251. r->dp[j] |= ((mp_digit)a[i]) << s;
  15252. if (s + 32 >= DIGIT_BIT) {
  15253. #if DIGIT_BIT != 32 && DIGIT_BIT != 64
  15254. r->dp[j] &= (1L << DIGIT_BIT) - 1;
  15255. #endif
  15256. s = DIGIT_BIT - s;
  15257. r->dp[++j] = a[i] >> s;
  15258. s = 32 - s;
  15259. }
  15260. else {
  15261. s += 32;
  15262. }
  15263. }
  15264. r->used = (256 + DIGIT_BIT - 1) / DIGIT_BIT;
  15265. mp_clamp(r);
  15266. #endif
  15267. }
  15268. return err;
  15269. }
  15270. /* Convert a point of type sp_point to type ecc_point.
  15271. *
  15272. * p Point of type sp_point.
  15273. * pm Point of type ecc_point (result).
  15274. * returns MEMORY_E when allocation of memory in ecc_point fails otherwise
  15275. * MP_OKAY.
  15276. */
  15277. static int sp_256_point_to_ecc_point_8(const sp_point* p, ecc_point* pm)
  15278. {
  15279. int err;
  15280. err = sp_256_to_mp(p->x, pm->x);
  15281. if (err == MP_OKAY) {
  15282. err = sp_256_to_mp(p->y, pm->y);
  15283. }
  15284. if (err == MP_OKAY) {
  15285. err = sp_256_to_mp(p->z, pm->z);
  15286. }
  15287. return err;
  15288. }
  15289. /* Compare a with b in constant time.
  15290. *
  15291. * a A single precision integer.
  15292. * b A single precision integer.
  15293. * return -ve, 0 or +ve if a is less than, equal to or greater than b
  15294. * respectively.
  15295. */
  15296. SP_NOINLINE static int32_t sp_256_cmp_8(const sp_digit* a, const sp_digit* b)
  15297. {
  15298. sp_digit r = 0;
  15299. __asm__ __volatile__ (
  15300. "mov r3, #0\n\t"
  15301. "mvn r3, r3\n\t"
  15302. "mov r6, #28\n\t"
  15303. "1:\n\t"
  15304. "ldr r7, [%[a], r6]\n\t"
  15305. "ldr r5, [%[b], r6]\n\t"
  15306. "and r7, r3\n\t"
  15307. "and r5, r3\n\t"
  15308. "mov r4, r7\n\t"
  15309. "sub r7, r5\n\t"
  15310. "sbc r7, r7\n\t"
  15311. "add %[r], r7\n\t"
  15312. "mvn r7, r7\n\t"
  15313. "and r3, r7\n\t"
  15314. "sub r5, r4\n\t"
  15315. "sbc r7, r7\n\t"
  15316. "sub %[r], r7\n\t"
  15317. "mvn r7, r7\n\t"
  15318. "and r3, r7\n\t"
  15319. "sub r6, #4\n\t"
  15320. "cmp r6, #0\n\t"
  15321. "bge 1b\n\t"
  15322. : [r] "+r" (r)
  15323. : [a] "r" (a), [b] "r" (b)
  15324. : "r3", "r4", "r5", "r6", "r7"
  15325. );
  15326. return r;
  15327. }
  15328. /* Normalize the values in each word to 32.
  15329. *
  15330. * a Array of sp_digit to normalize.
  15331. */
  15332. #define sp_256_norm_8(a)
  15333. /* Conditionally subtract b from a using the mask m.
  15334. * m is -1 to subtract and 0 when not copying.
  15335. *
  15336. * r A single precision number representing condition subtract result.
  15337. * a A single precision number to subtract from.
  15338. * b A single precision number to subtract.
  15339. * m Mask value to apply.
  15340. */
  15341. SP_NOINLINE static sp_digit sp_256_cond_sub_8(sp_digit* r, const sp_digit* a,
  15342. const sp_digit* b, sp_digit m)
  15343. {
  15344. sp_digit c = 0;
  15345. __asm__ __volatile__ (
  15346. "mov r5, #32\n\t"
  15347. "mov r8, r5\n\t"
  15348. "mov r7, #0\n\t"
  15349. "1:\n\t"
  15350. "ldr r6, [%[b], r7]\n\t"
  15351. "and r6, %[m]\n\t"
  15352. "mov r5, #0\n\t"
  15353. "sub r5, %[c]\n\t"
  15354. "ldr r5, [%[a], r7]\n\t"
  15355. "sbc r5, r6\n\t"
  15356. "sbc %[c], %[c]\n\t"
  15357. "str r5, [%[r], r7]\n\t"
  15358. "add r7, #4\n\t"
  15359. "cmp r7, r8\n\t"
  15360. "blt 1b\n\t"
  15361. : [c] "+r" (c)
  15362. : [r] "r" (r), [a] "r" (a), [b] "r" (b), [m] "r" (m)
  15363. : "memory", "r5", "r6", "r7", "r8"
  15364. );
  15365. return c;
  15366. }
  15367. /* Reduce the number back to 256 bits using Montgomery reduction.
  15368. *
  15369. * a A single precision number to reduce in place.
  15370. * m The single precision number representing the modulus.
  15371. * mp The digit representing the negative inverse of m mod 2^n.
  15372. */
  15373. SP_NOINLINE static void sp_256_mont_reduce_8(sp_digit* a, const sp_digit* m,
  15374. sp_digit mp)
  15375. {
  15376. (void)mp;
  15377. (void)m;
  15378. __asm__ __volatile__ (
  15379. "mov r2, #0\n\t"
  15380. "mov r1, #0\n\t"
  15381. "# i = 0\n\t"
  15382. "mov r8, r2\n\t"
  15383. "\n1:\n\t"
  15384. "mov r4, #0\n\t"
  15385. "# mu = a[i] * 1 (mp) = a[i]\n\t"
  15386. "ldr r3, [%[a]]\n\t"
  15387. "# a[i+0] += -1 * mu\n\t"
  15388. "mov r5, r3\n\t"
  15389. "str r4, [%[a], #0]\n\t"
  15390. "# a[i+1] += -1 * mu\n\t"
  15391. "ldr r6, [%[a], #4]\n\t"
  15392. "mov r4, r3\n\t"
  15393. "sub r5, r3\n\t"
  15394. "sbc r4, r2\n\t"
  15395. "add r5, r6\n\t"
  15396. "adc r4, r2\n\t"
  15397. "str r5, [%[a], #4]\n\t"
  15398. "# a[i+2] += -1 * mu\n\t"
  15399. "ldr r6, [%[a], #8]\n\t"
  15400. "mov r5, r3\n\t"
  15401. "sub r4, r3\n\t"
  15402. "sbc r5, r2\n\t"
  15403. "add r4, r6\n\t"
  15404. "adc r5, r2\n\t"
  15405. "str r4, [%[a], #8]\n\t"
  15406. "# a[i+3] += 0 * mu\n\t"
  15407. "ldr r6, [%[a], #12]\n\t"
  15408. "mov r4, #0\n\t"
  15409. "add r5, r6\n\t"
  15410. "adc r4, r2\n\t"
  15411. "str r5, [%[a], #12]\n\t"
  15412. "# a[i+4] += 0 * mu\n\t"
  15413. "ldr r6, [%[a], #16]\n\t"
  15414. "mov r5, #0\n\t"
  15415. "add r4, r6\n\t"
  15416. "adc r5, r2\n\t"
  15417. "str r4, [%[a], #16]\n\t"
  15418. "# a[i+5] += 0 * mu\n\t"
  15419. "ldr r6, [%[a], #20]\n\t"
  15420. "mov r4, #0\n\t"
  15421. "add r5, r6\n\t"
  15422. "adc r4, r2\n\t"
  15423. "str r5, [%[a], #20]\n\t"
  15424. "# a[i+6] += 1 * mu\n\t"
  15425. "ldr r6, [%[a], #24]\n\t"
  15426. "mov r5, #0\n\t"
  15427. "add r4, r3\n\t"
  15428. "adc r5, r2\n\t"
  15429. "add r4, r6\n\t"
  15430. "adc r5, r2\n\t"
  15431. "str r4, [%[a], #24]\n\t"
  15432. "# a[i+7] += -1 * mu\n\t"
  15433. "ldr r6, [%[a], #28]\n\t"
  15434. "ldr r7, [%[a], #32]\n\t"
  15435. "add r4, r1, r3\n\t"
  15436. "mov r1, #0\n\t"
  15437. "adc r1, r2\n\t"
  15438. "sub r5, r3\n\t"
  15439. "sbc r4, r2\n\t"
  15440. "sbc r1, r2\n\t"
  15441. "add r5, r6\n\t"
  15442. "adc r4, r7\n\t"
  15443. "adc r1, r2\n\t"
  15444. "str r5, [%[a], #28]\n\t"
  15445. "str r4, [%[a], #32]\n\t"
  15446. "# i += 1\n\t"
  15447. "mov r6, #4\n\t"
  15448. "add r8, r6\n\t"
  15449. "add %[a], #4\n\t"
  15450. "mov r6, #32\n\t"
  15451. "cmp r8, r6\n\t"
  15452. "blt 1b\n\t"
  15453. "sub %[a], #32\n\t"
  15454. "mov r3, r1\n\t"
  15455. "sub r1, #1\n\t"
  15456. "mvn r1, r1\n\t"
  15457. "ldr r5, [%[a],#32]\n\t"
  15458. "ldr r4, [%[a],#36]\n\t"
  15459. "ldr r6, [%[a],#40]\n\t"
  15460. "ldr r7, [%[a],#44]\n\t"
  15461. "sub r5, r1\n\t"
  15462. "sbc r4, r1\n\t"
  15463. "sbc r6, r1\n\t"
  15464. "sbc r7, r2\n\t"
  15465. "str r5, [%[a],#0]\n\t"
  15466. "str r4, [%[a],#4]\n\t"
  15467. "str r6, [%[a],#8]\n\t"
  15468. "str r7, [%[a],#12]\n\t"
  15469. "ldr r5, [%[a],#48]\n\t"
  15470. "ldr r4, [%[a],#52]\n\t"
  15471. "ldr r6, [%[a],#56]\n\t"
  15472. "ldr r7, [%[a],#60]\n\t"
  15473. "sbc r5, r2\n\t"
  15474. "sbc r4, r2\n\t"
  15475. "sbc r6, r3\n\t"
  15476. "sbc r7, r1\n\t"
  15477. "str r5, [%[a],#16]\n\t"
  15478. "str r4, [%[a],#20]\n\t"
  15479. "str r6, [%[a],#24]\n\t"
  15480. "str r7, [%[a],#28]\n\t"
  15481. : [a] "+r" (a)
  15482. :
  15483. : "memory", "r1", "r2", "r3", "r4", "r5", "r6", "r7", "r8"
  15484. );
  15485. (void)m;
  15486. (void)mp;
  15487. }
  15488. /* Reduce the number back to 256 bits using Montgomery reduction.
  15489. *
  15490. * a A single precision number to reduce in place.
  15491. * m The single precision number representing the modulus.
  15492. * mp The digit representing the negative inverse of m mod 2^n.
  15493. */
  15494. SP_NOINLINE static void sp_256_mont_reduce_order_8(sp_digit* a, const sp_digit* m,
  15495. sp_digit mp)
  15496. {
  15497. sp_digit ca = 0;
  15498. __asm__ __volatile__ (
  15499. "mov r8, %[mp]\n\t"
  15500. "mov r12, %[ca]\n\t"
  15501. "mov r14, %[m]\n\t"
  15502. "mov r9, %[a]\n\t"
  15503. "mov r4, #0\n\t"
  15504. "# i = 0\n\t"
  15505. "mov r11, r4\n\t"
  15506. "\n1:\n\t"
  15507. "mov r5, #0\n\t"
  15508. "mov %[ca], #0\n\t"
  15509. "# mu = a[i] * mp\n\t"
  15510. "mov %[mp], r8\n\t"
  15511. "ldr %[a], [%[a]]\n\t"
  15512. "mul %[mp], %[a]\n\t"
  15513. "mov %[m], r14\n\t"
  15514. "mov r10, r9\n\t"
  15515. "\n2:\n\t"
  15516. "# a[i+j] += m[j] * mu\n\t"
  15517. "mov %[a], r10\n\t"
  15518. "ldr %[a], [%[a]]\n\t"
  15519. "mov %[ca], #0\n\t"
  15520. "mov r4, r5\n\t"
  15521. "mov r5, #0\n\t"
  15522. "# Multiply m[j] and mu - Start\n\t"
  15523. "ldr r7, [%[m]]\n\t"
  15524. "lsl r6, %[mp], #16\n\t"
  15525. "lsl r7, r7, #16\n\t"
  15526. "lsr r6, r6, #16\n\t"
  15527. "lsr r7, r7, #16\n\t"
  15528. "mul r7, r6\n\t"
  15529. "add %[a], r7\n\t"
  15530. "adc r5, %[ca]\n\t"
  15531. "ldr r7, [%[m]]\n\t"
  15532. "lsr r7, r7, #16\n\t"
  15533. "mul r6, r7\n\t"
  15534. "lsr r7, r6, #16\n\t"
  15535. "lsl r6, r6, #16\n\t"
  15536. "add %[a], r6\n\t"
  15537. "adc r5, r7\n\t"
  15538. "ldr r7, [%[m]]\n\t"
  15539. "lsr r6, %[mp], #16\n\t"
  15540. "lsr r7, r7, #16\n\t"
  15541. "mul r7, r6\n\t"
  15542. "add r5, r7\n\t"
  15543. "ldr r7, [%[m]]\n\t"
  15544. "lsl r7, r7, #16\n\t"
  15545. "lsr r7, r7, #16\n\t"
  15546. "mul r6, r7\n\t"
  15547. "lsr r7, r6, #16\n\t"
  15548. "lsl r6, r6, #16\n\t"
  15549. "add %[a], r6\n\t"
  15550. "adc r5, r7\n\t"
  15551. "# Multiply m[j] and mu - Done\n\t"
  15552. "add r4, %[a]\n\t"
  15553. "adc r5, %[ca]\n\t"
  15554. "mov %[a], r10\n\t"
  15555. "str r4, [%[a]]\n\t"
  15556. "mov r6, #4\n\t"
  15557. "add %[m], #4\n\t"
  15558. "add r10, r6\n\t"
  15559. "mov r4, #28\n\t"
  15560. "add r4, r9\n\t"
  15561. "cmp r10, r4\n\t"
  15562. "blt 2b\n\t"
  15563. "# a[i+7] += m[7] * mu\n\t"
  15564. "mov %[ca], #0\n\t"
  15565. "mov r4, r12\n\t"
  15566. "mov %[a], #0\n\t"
  15567. "# Multiply m[7] and mu - Start\n\t"
  15568. "ldr r7, [%[m]]\n\t"
  15569. "lsl r6, %[mp], #16\n\t"
  15570. "lsl r7, r7, #16\n\t"
  15571. "lsr r6, r6, #16\n\t"
  15572. "lsr r7, r7, #16\n\t"
  15573. "mul r7, r6\n\t"
  15574. "add r5, r7\n\t"
  15575. "adc r4, %[ca]\n\t"
  15576. "adc %[a], %[ca]\n\t"
  15577. "ldr r7, [%[m]]\n\t"
  15578. "lsr r7, r7, #16\n\t"
  15579. "mul r6, r7\n\t"
  15580. "lsr r7, r6, #16\n\t"
  15581. "lsl r6, r6, #16\n\t"
  15582. "add r5, r6\n\t"
  15583. "adc r4, r7\n\t"
  15584. "adc %[a], %[ca]\n\t"
  15585. "ldr r7, [%[m]]\n\t"
  15586. "lsr r6, %[mp], #16\n\t"
  15587. "lsr r7, r7, #16\n\t"
  15588. "mul r7, r6\n\t"
  15589. "add r4, r7\n\t"
  15590. "adc %[a], %[ca]\n\t"
  15591. "ldr r7, [%[m]]\n\t"
  15592. "lsl r7, r7, #16\n\t"
  15593. "lsr r7, r7, #16\n\t"
  15594. "mul r6, r7\n\t"
  15595. "lsr r7, r6, #16\n\t"
  15596. "lsl r6, r6, #16\n\t"
  15597. "add r5, r6\n\t"
  15598. "adc r4, r7\n\t"
  15599. "adc %[a], %[ca]\n\t"
  15600. "# Multiply m[7] and mu - Done\n\t"
  15601. "mov %[ca], %[a]\n\t"
  15602. "mov %[a], r10\n\t"
  15603. "ldr r7, [%[a], #4]\n\t"
  15604. "ldr %[a], [%[a]]\n\t"
  15605. "mov r6, #0\n\t"
  15606. "add r5, %[a]\n\t"
  15607. "adc r7, r4\n\t"
  15608. "adc %[ca], r6\n\t"
  15609. "mov %[a], r10\n\t"
  15610. "str r5, [%[a]]\n\t"
  15611. "str r7, [%[a], #4]\n\t"
  15612. "# i += 1\n\t"
  15613. "mov r6, #4\n\t"
  15614. "add r9, r6\n\t"
  15615. "add r11, r6\n\t"
  15616. "mov r12, %[ca]\n\t"
  15617. "mov %[a], r9\n\t"
  15618. "mov r4, #32\n\t"
  15619. "cmp r11, r4\n\t"
  15620. "blt 1b\n\t"
  15621. "mov %[m], r14\n\t"
  15622. : [ca] "+r" (ca), [a] "+r" (a)
  15623. : [m] "r" (m), [mp] "r" (mp)
  15624. : "memory", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  15625. );
  15626. sp_256_cond_sub_8(a - 8, a, m, (sp_digit)0 - ca);
  15627. }
  15628. /* Multiply a and b into r. (r = a * b)
  15629. *
  15630. * r A single precision integer.
  15631. * a A single precision integer.
  15632. * b A single precision integer.
  15633. */
  15634. SP_NOINLINE static void sp_256_mul_8(sp_digit* r, const sp_digit* a,
  15635. const sp_digit* b)
  15636. {
  15637. sp_digit tmp[8 * 2];
  15638. __asm__ __volatile__ (
  15639. "mov r3, #0\n\t"
  15640. "mov r4, #0\n\t"
  15641. "mov r8, r3\n\t"
  15642. "mov r11, %[r]\n\t"
  15643. "mov r9, %[a]\n\t"
  15644. "mov r10, %[b]\n\t"
  15645. "mov r6, #32\n\t"
  15646. "add r6, r9\n\t"
  15647. "mov r12, r6\n\t"
  15648. "\n1:\n\t"
  15649. "mov %[r], #0\n\t"
  15650. "mov r5, #0\n\t"
  15651. "mov r6, #28\n\t"
  15652. "mov %[a], r8\n\t"
  15653. "sub %[a], r6\n\t"
  15654. "sbc r6, r6\n\t"
  15655. "mvn r6, r6\n\t"
  15656. "and %[a], r6\n\t"
  15657. "mov %[b], r8\n\t"
  15658. "sub %[b], %[a]\n\t"
  15659. "add %[a], r9\n\t"
  15660. "add %[b], r10\n\t"
  15661. "\n2:\n\t"
  15662. "# Multiply Start\n\t"
  15663. "ldr r6, [%[a]]\n\t"
  15664. "ldr r7, [%[b]]\n\t"
  15665. "lsl r6, r6, #16\n\t"
  15666. "lsl r7, r7, #16\n\t"
  15667. "lsr r6, r6, #16\n\t"
  15668. "lsr r7, r7, #16\n\t"
  15669. "mul r7, r6\n\t"
  15670. "add r3, r7\n\t"
  15671. "adc r4, %[r]\n\t"
  15672. "adc r5, %[r]\n\t"
  15673. "ldr r7, [%[b]]\n\t"
  15674. "lsr r7, r7, #16\n\t"
  15675. "mul r6, r7\n\t"
  15676. "lsr r7, r6, #16\n\t"
  15677. "lsl r6, r6, #16\n\t"
  15678. "add r3, r6\n\t"
  15679. "adc r4, r7\n\t"
  15680. "adc r5, %[r]\n\t"
  15681. "ldr r6, [%[a]]\n\t"
  15682. "ldr r7, [%[b]]\n\t"
  15683. "lsr r6, r6, #16\n\t"
  15684. "lsr r7, r7, #16\n\t"
  15685. "mul r7, r6\n\t"
  15686. "add r4, r7\n\t"
  15687. "adc r5, %[r]\n\t"
  15688. "ldr r7, [%[b]]\n\t"
  15689. "lsl r7, r7, #16\n\t"
  15690. "lsr r7, r7, #16\n\t"
  15691. "mul r6, r7\n\t"
  15692. "lsr r7, r6, #16\n\t"
  15693. "lsl r6, r6, #16\n\t"
  15694. "add r3, r6\n\t"
  15695. "adc r4, r7\n\t"
  15696. "adc r5, %[r]\n\t"
  15697. "# Multiply Done\n\t"
  15698. "add %[a], #4\n\t"
  15699. "sub %[b], #4\n\t"
  15700. "cmp %[a], r12\n\t"
  15701. "beq 3f\n\t"
  15702. "mov r6, r8\n\t"
  15703. "add r6, r9\n\t"
  15704. "cmp %[a], r6\n\t"
  15705. "ble 2b\n\t"
  15706. "\n3:\n\t"
  15707. "mov %[r], r11\n\t"
  15708. "mov r7, r8\n\t"
  15709. "str r3, [%[r], r7]\n\t"
  15710. "mov r3, r4\n\t"
  15711. "mov r4, r5\n\t"
  15712. "add r7, #4\n\t"
  15713. "mov r8, r7\n\t"
  15714. "mov r6, #56\n\t"
  15715. "cmp r7, r6\n\t"
  15716. "ble 1b\n\t"
  15717. "str r3, [%[r], r7]\n\t"
  15718. "mov %[a], r9\n\t"
  15719. "mov %[b], r10\n\t"
  15720. :
  15721. : [r] "r" (tmp), [a] "r" (a), [b] "r" (b)
  15722. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12"
  15723. );
  15724. XMEMCPY(r, tmp, sizeof(tmp));
  15725. }
  15726. /* Multiply two Montogmery form numbers mod the modulus (prime).
  15727. * (r = a * b mod m)
  15728. *
  15729. * r Result of multiplication.
  15730. * a First number to multiply in Montogmery form.
  15731. * b Second number to multiply in Montogmery form.
  15732. * m Modulus (prime).
  15733. * mp Montogmery mulitplier.
  15734. */
  15735. static void sp_256_mont_mul_8(sp_digit* r, const sp_digit* a, const sp_digit* b,
  15736. const sp_digit* m, sp_digit mp)
  15737. {
  15738. sp_256_mul_8(r, a, b);
  15739. sp_256_mont_reduce_8(r, m, mp);
  15740. }
  15741. /* Square a and put result in r. (r = a * a)
  15742. *
  15743. * r A single precision integer.
  15744. * a A single precision integer.
  15745. */
  15746. SP_NOINLINE static void sp_256_sqr_8(sp_digit* r, const sp_digit* a)
  15747. {
  15748. __asm__ __volatile__ (
  15749. "mov r3, #0\n\t"
  15750. "mov r4, #0\n\t"
  15751. "mov r5, #0\n\t"
  15752. "mov r8, r3\n\t"
  15753. "mov r11, %[r]\n\t"
  15754. "mov r6, #64\n\t"
  15755. "neg r6, r6\n\t"
  15756. "add sp, r6\n\t"
  15757. "mov r10, sp\n\t"
  15758. "mov r9, %[a]\n\t"
  15759. "\n1:\n\t"
  15760. "mov %[r], #0\n\t"
  15761. "mov r6, #28\n\t"
  15762. "mov %[a], r8\n\t"
  15763. "sub %[a], r6\n\t"
  15764. "sbc r6, r6\n\t"
  15765. "mvn r6, r6\n\t"
  15766. "and %[a], r6\n\t"
  15767. "mov r2, r8\n\t"
  15768. "sub r2, %[a]\n\t"
  15769. "add %[a], r9\n\t"
  15770. "add r2, r9\n\t"
  15771. "\n2:\n\t"
  15772. "cmp r2, %[a]\n\t"
  15773. "beq 4f\n\t"
  15774. "# Multiply * 2: Start\n\t"
  15775. "ldr r6, [%[a]]\n\t"
  15776. "ldr r7, [r2]\n\t"
  15777. "lsl r6, r6, #16\n\t"
  15778. "lsl r7, r7, #16\n\t"
  15779. "lsr r6, r6, #16\n\t"
  15780. "lsr r7, r7, #16\n\t"
  15781. "mul r7, r6\n\t"
  15782. "add r3, r7\n\t"
  15783. "adc r4, %[r]\n\t"
  15784. "adc r5, %[r]\n\t"
  15785. "add r3, r7\n\t"
  15786. "adc r4, %[r]\n\t"
  15787. "adc r5, %[r]\n\t"
  15788. "ldr r7, [r2]\n\t"
  15789. "lsr r7, r7, #16\n\t"
  15790. "mul r6, r7\n\t"
  15791. "lsr r7, r6, #16\n\t"
  15792. "lsl r6, r6, #16\n\t"
  15793. "add r3, r6\n\t"
  15794. "adc r4, r7\n\t"
  15795. "adc r5, %[r]\n\t"
  15796. "add r3, r6\n\t"
  15797. "adc r4, r7\n\t"
  15798. "adc r5, %[r]\n\t"
  15799. "ldr r6, [%[a]]\n\t"
  15800. "ldr r7, [r2]\n\t"
  15801. "lsr r6, r6, #16\n\t"
  15802. "lsr r7, r7, #16\n\t"
  15803. "mul r7, r6\n\t"
  15804. "add r4, r7\n\t"
  15805. "adc r5, %[r]\n\t"
  15806. "add r4, r7\n\t"
  15807. "adc r5, %[r]\n\t"
  15808. "ldr r7, [r2]\n\t"
  15809. "lsl r7, r7, #16\n\t"
  15810. "lsr r7, r7, #16\n\t"
  15811. "mul r6, r7\n\t"
  15812. "lsr r7, r6, #16\n\t"
  15813. "lsl r6, r6, #16\n\t"
  15814. "add r3, r6\n\t"
  15815. "adc r4, r7\n\t"
  15816. "adc r5, %[r]\n\t"
  15817. "add r3, r6\n\t"
  15818. "adc r4, r7\n\t"
  15819. "adc r5, %[r]\n\t"
  15820. "# Multiply * 2: Done\n\t"
  15821. "bal 5f\n\t"
  15822. "\n4:\n\t"
  15823. "# Square: Start\n\t"
  15824. "ldr r6, [%[a]]\n\t"
  15825. "lsr r7, r6, #16\n\t"
  15826. "lsl r6, r6, #16\n\t"
  15827. "lsr r6, r6, #16\n\t"
  15828. "mul r6, r6\n\t"
  15829. "add r3, r6\n\t"
  15830. "adc r4, %[r]\n\t"
  15831. "adc r5, %[r]\n\t"
  15832. "mul r7, r7\n\t"
  15833. "add r4, r7\n\t"
  15834. "adc r5, %[r]\n\t"
  15835. "ldr r6, [%[a]]\n\t"
  15836. "lsr r7, r6, #16\n\t"
  15837. "lsl r6, r6, #16\n\t"
  15838. "lsr r6, r6, #16\n\t"
  15839. "mul r6, r7\n\t"
  15840. "lsr r7, r6, #15\n\t"
  15841. "lsl r6, r6, #17\n\t"
  15842. "add r3, r6\n\t"
  15843. "adc r4, r7\n\t"
  15844. "adc r5, %[r]\n\t"
  15845. "# Square: Done\n\t"
  15846. "\n5:\n\t"
  15847. "add %[a], #4\n\t"
  15848. "sub r2, #4\n\t"
  15849. "mov r6, #32\n\t"
  15850. "add r6, r9\n\t"
  15851. "cmp %[a], r6\n\t"
  15852. "beq 3f\n\t"
  15853. "cmp %[a], r2\n\t"
  15854. "bgt 3f\n\t"
  15855. "mov r7, r8\n\t"
  15856. "add r7, r9\n\t"
  15857. "cmp %[a], r7\n\t"
  15858. "ble 2b\n\t"
  15859. "\n3:\n\t"
  15860. "mov %[r], r10\n\t"
  15861. "mov r7, r8\n\t"
  15862. "str r3, [%[r], r7]\n\t"
  15863. "mov r3, r4\n\t"
  15864. "mov r4, r5\n\t"
  15865. "mov r5, #0\n\t"
  15866. "add r7, #4\n\t"
  15867. "mov r8, r7\n\t"
  15868. "mov r6, #56\n\t"
  15869. "cmp r7, r6\n\t"
  15870. "ble 1b\n\t"
  15871. "mov %[a], r9\n\t"
  15872. "str r3, [%[r], r7]\n\t"
  15873. "mov %[r], r11\n\t"
  15874. "mov %[a], r10\n\t"
  15875. "mov r3, #60\n\t"
  15876. "\n4:\n\t"
  15877. "ldr r6, [%[a], r3]\n\t"
  15878. "str r6, [%[r], r3]\n\t"
  15879. "sub r3, #4\n\t"
  15880. "bge 4b\n\t"
  15881. "mov r6, #64\n\t"
  15882. "add sp, r6\n\t"
  15883. :
  15884. : [r] "r" (r), [a] "r" (a)
  15885. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  15886. );
  15887. }
  15888. /* Square the Montgomery form number. (r = a * a mod m)
  15889. *
  15890. * r Result of squaring.
  15891. * a Number to square in Montogmery form.
  15892. * m Modulus (prime).
  15893. * mp Montogmery mulitplier.
  15894. */
  15895. static void sp_256_mont_sqr_8(sp_digit* r, const sp_digit* a, const sp_digit* m,
  15896. sp_digit mp)
  15897. {
  15898. sp_256_sqr_8(r, a);
  15899. sp_256_mont_reduce_8(r, m, mp);
  15900. }
  15901. #if !defined(WOLFSSL_SP_SMALL) || defined(HAVE_COMP_KEY)
  15902. /* Square the Montgomery form number a number of times. (r = a ^ n mod m)
  15903. *
  15904. * r Result of squaring.
  15905. * a Number to square in Montogmery form.
  15906. * n Number of times to square.
  15907. * m Modulus (prime).
  15908. * mp Montogmery mulitplier.
  15909. */
  15910. static void sp_256_mont_sqr_n_8(sp_digit* r, const sp_digit* a, int n,
  15911. const sp_digit* m, sp_digit mp)
  15912. {
  15913. sp_256_mont_sqr_8(r, a, m, mp);
  15914. for (; n > 1; n--) {
  15915. sp_256_mont_sqr_8(r, r, m, mp);
  15916. }
  15917. }
  15918. #endif /* !WOLFSSL_SP_SMALL || HAVE_COMP_KEY */
  15919. #ifdef WOLFSSL_SP_SMALL
  15920. /* Mod-2 for the P256 curve. */
  15921. static const uint32_t p256_mod_2[8] = {
  15922. 0xfffffffdU,0xffffffffU,0xffffffffU,0x00000000U,0x00000000U,0x00000000U,
  15923. 0x00000001U,0xffffffffU
  15924. };
  15925. #endif /* !WOLFSSL_SP_SMALL */
  15926. /* Invert the number, in Montgomery form, modulo the modulus (prime) of the
  15927. * P256 curve. (r = 1 / a mod m)
  15928. *
  15929. * r Inverse result.
  15930. * a Number to invert.
  15931. * td Temporary data.
  15932. */
  15933. static void sp_256_mont_inv_8(sp_digit* r, const sp_digit* a, sp_digit* td)
  15934. {
  15935. #ifdef WOLFSSL_SP_SMALL
  15936. sp_digit* t = td;
  15937. int i;
  15938. XMEMCPY(t, a, sizeof(sp_digit) * 8);
  15939. for (i=254; i>=0; i--) {
  15940. sp_256_mont_sqr_8(t, t, p256_mod, p256_mp_mod);
  15941. if (p256_mod_2[i / 32] & ((sp_digit)1 << (i % 32)))
  15942. sp_256_mont_mul_8(t, t, a, p256_mod, p256_mp_mod);
  15943. }
  15944. XMEMCPY(r, t, sizeof(sp_digit) * 8);
  15945. #else
  15946. sp_digit* t = td;
  15947. sp_digit* t2 = td + 2 * 8;
  15948. sp_digit* t3 = td + 4 * 8;
  15949. /* t = a^2 */
  15950. sp_256_mont_sqr_8(t, a, p256_mod, p256_mp_mod);
  15951. /* t = a^3 = t * a */
  15952. sp_256_mont_mul_8(t, t, a, p256_mod, p256_mp_mod);
  15953. /* t2= a^c = t ^ 2 ^ 2 */
  15954. sp_256_mont_sqr_n_8(t2, t, 2, p256_mod, p256_mp_mod);
  15955. /* t3= a^d = t2 * a */
  15956. sp_256_mont_mul_8(t3, t2, a, p256_mod, p256_mp_mod);
  15957. /* t = a^f = t2 * t */
  15958. sp_256_mont_mul_8(t, t2, t, p256_mod, p256_mp_mod);
  15959. /* t2= a^f0 = t ^ 2 ^ 4 */
  15960. sp_256_mont_sqr_n_8(t2, t, 4, p256_mod, p256_mp_mod);
  15961. /* t3= a^fd = t2 * t3 */
  15962. sp_256_mont_mul_8(t3, t2, t3, p256_mod, p256_mp_mod);
  15963. /* t = a^ff = t2 * t */
  15964. sp_256_mont_mul_8(t, t2, t, p256_mod, p256_mp_mod);
  15965. /* t2= a^ff00 = t ^ 2 ^ 8 */
  15966. sp_256_mont_sqr_n_8(t2, t, 8, p256_mod, p256_mp_mod);
  15967. /* t3= a^fffd = t2 * t3 */
  15968. sp_256_mont_mul_8(t3, t2, t3, p256_mod, p256_mp_mod);
  15969. /* t = a^ffff = t2 * t */
  15970. sp_256_mont_mul_8(t, t2, t, p256_mod, p256_mp_mod);
  15971. /* t2= a^ffff0000 = t ^ 2 ^ 16 */
  15972. sp_256_mont_sqr_n_8(t2, t, 16, p256_mod, p256_mp_mod);
  15973. /* t3= a^fffffffd = t2 * t3 */
  15974. sp_256_mont_mul_8(t3, t2, t3, p256_mod, p256_mp_mod);
  15975. /* t = a^ffffffff = t2 * t */
  15976. sp_256_mont_mul_8(t, t2, t, p256_mod, p256_mp_mod);
  15977. /* t = a^ffffffff00000000 = t ^ 2 ^ 32 */
  15978. sp_256_mont_sqr_n_8(t2, t, 32, p256_mod, p256_mp_mod);
  15979. /* t2= a^ffffffffffffffff = t2 * t */
  15980. sp_256_mont_mul_8(t, t2, t, p256_mod, p256_mp_mod);
  15981. /* t2= a^ffffffff00000001 = t2 * a */
  15982. sp_256_mont_mul_8(t2, t2, a, p256_mod, p256_mp_mod);
  15983. /* t2= a^ffffffff000000010000000000000000000000000000000000000000
  15984. * = t2 ^ 2 ^ 160 */
  15985. sp_256_mont_sqr_n_8(t2, t2, 160, p256_mod, p256_mp_mod);
  15986. /* t2= a^ffffffff00000001000000000000000000000000ffffffffffffffff
  15987. * = t2 * t */
  15988. sp_256_mont_mul_8(t2, t2, t, p256_mod, p256_mp_mod);
  15989. /* t2= a^ffffffff00000001000000000000000000000000ffffffffffffffff00000000
  15990. * = t2 ^ 2 ^ 32 */
  15991. sp_256_mont_sqr_n_8(t2, t2, 32, p256_mod, p256_mp_mod);
  15992. /* r = a^ffffffff00000001000000000000000000000000fffffffffffffffffffffffd
  15993. * = t2 * t3 */
  15994. sp_256_mont_mul_8(r, t2, t3, p256_mod, p256_mp_mod);
  15995. #endif /* WOLFSSL_SP_SMALL */
  15996. }
  15997. /* Map the Montgomery form projective coordinate point to an affine point.
  15998. *
  15999. * r Resulting affine coordinate point.
  16000. * p Montgomery form projective coordinate point.
  16001. * t Temporary ordinate data.
  16002. */
  16003. static void sp_256_map_8(sp_point* r, const sp_point* p, sp_digit* t)
  16004. {
  16005. sp_digit* t1 = t;
  16006. sp_digit* t2 = t + 2*8;
  16007. int32_t n;
  16008. sp_256_mont_inv_8(t1, p->z, t + 2*8);
  16009. sp_256_mont_sqr_8(t2, t1, p256_mod, p256_mp_mod);
  16010. sp_256_mont_mul_8(t1, t2, t1, p256_mod, p256_mp_mod);
  16011. /* x /= z^2 */
  16012. sp_256_mont_mul_8(r->x, p->x, t2, p256_mod, p256_mp_mod);
  16013. XMEMSET(r->x + 8, 0, sizeof(r->x) / 2U);
  16014. sp_256_mont_reduce_8(r->x, p256_mod, p256_mp_mod);
  16015. /* Reduce x to less than modulus */
  16016. n = sp_256_cmp_8(r->x, p256_mod);
  16017. sp_256_cond_sub_8(r->x, r->x, p256_mod, 0 - ((n >= 0) ?
  16018. (sp_digit)1 : (sp_digit)0));
  16019. sp_256_norm_8(r->x);
  16020. /* y /= z^3 */
  16021. sp_256_mont_mul_8(r->y, p->y, t1, p256_mod, p256_mp_mod);
  16022. XMEMSET(r->y + 8, 0, sizeof(r->y) / 2U);
  16023. sp_256_mont_reduce_8(r->y, p256_mod, p256_mp_mod);
  16024. /* Reduce y to less than modulus */
  16025. n = sp_256_cmp_8(r->y, p256_mod);
  16026. sp_256_cond_sub_8(r->y, r->y, p256_mod, 0 - ((n >= 0) ?
  16027. (sp_digit)1 : (sp_digit)0));
  16028. sp_256_norm_8(r->y);
  16029. XMEMSET(r->z, 0, sizeof(r->z));
  16030. r->z[0] = 1;
  16031. }
  16032. #ifdef WOLFSSL_SP_SMALL
  16033. /* Add b to a into r. (r = a + b)
  16034. *
  16035. * r A single precision integer.
  16036. * a A single precision integer.
  16037. * b A single precision integer.
  16038. */
  16039. SP_NOINLINE static sp_digit sp_256_add_8(sp_digit* r, const sp_digit* a,
  16040. const sp_digit* b)
  16041. {
  16042. sp_digit c = 0;
  16043. __asm__ __volatile__ (
  16044. "mov r6, %[a]\n\t"
  16045. "mov r7, #0\n\t"
  16046. "add r6, #32\n\t"
  16047. "sub r7, #1\n\t"
  16048. "\n1:\n\t"
  16049. "add %[c], r7\n\t"
  16050. "ldr r4, [%[a]]\n\t"
  16051. "ldr r5, [%[b]]\n\t"
  16052. "adc r4, r5\n\t"
  16053. "str r4, [%[r]]\n\t"
  16054. "mov %[c], #0\n\t"
  16055. "adc %[c], %[c]\n\t"
  16056. "add %[a], #4\n\t"
  16057. "add %[b], #4\n\t"
  16058. "add %[r], #4\n\t"
  16059. "cmp %[a], r6\n\t"
  16060. "bne 1b\n\t"
  16061. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  16062. :
  16063. : "memory", "r4", "r5", "r6", "r7"
  16064. );
  16065. return c;
  16066. }
  16067. #else
  16068. /* Add b to a into r. (r = a + b)
  16069. *
  16070. * r A single precision integer.
  16071. * a A single precision integer.
  16072. * b A single precision integer.
  16073. */
  16074. SP_NOINLINE static sp_digit sp_256_add_8(sp_digit* r, const sp_digit* a,
  16075. const sp_digit* b)
  16076. {
  16077. sp_digit c = 0;
  16078. __asm__ __volatile__ (
  16079. "ldr r4, [%[a], #0]\n\t"
  16080. "ldr r5, [%[b], #0]\n\t"
  16081. "add r4, r5\n\t"
  16082. "str r4, [%[r], #0]\n\t"
  16083. "ldr r4, [%[a], #4]\n\t"
  16084. "ldr r5, [%[b], #4]\n\t"
  16085. "adc r4, r5\n\t"
  16086. "str r4, [%[r], #4]\n\t"
  16087. "ldr r4, [%[a], #8]\n\t"
  16088. "ldr r5, [%[b], #8]\n\t"
  16089. "adc r4, r5\n\t"
  16090. "str r4, [%[r], #8]\n\t"
  16091. "ldr r4, [%[a], #12]\n\t"
  16092. "ldr r5, [%[b], #12]\n\t"
  16093. "adc r4, r5\n\t"
  16094. "str r4, [%[r], #12]\n\t"
  16095. "ldr r4, [%[a], #16]\n\t"
  16096. "ldr r5, [%[b], #16]\n\t"
  16097. "adc r4, r5\n\t"
  16098. "str r4, [%[r], #16]\n\t"
  16099. "ldr r4, [%[a], #20]\n\t"
  16100. "ldr r5, [%[b], #20]\n\t"
  16101. "adc r4, r5\n\t"
  16102. "str r4, [%[r], #20]\n\t"
  16103. "ldr r4, [%[a], #24]\n\t"
  16104. "ldr r5, [%[b], #24]\n\t"
  16105. "adc r4, r5\n\t"
  16106. "str r4, [%[r], #24]\n\t"
  16107. "ldr r4, [%[a], #28]\n\t"
  16108. "ldr r5, [%[b], #28]\n\t"
  16109. "adc r4, r5\n\t"
  16110. "str r4, [%[r], #28]\n\t"
  16111. "mov %[c], #0\n\t"
  16112. "adc %[c], %[c]\n\t"
  16113. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  16114. :
  16115. : "memory", "r4", "r5"
  16116. );
  16117. return c;
  16118. }
  16119. #endif /* WOLFSSL_SP_SMALL */
  16120. /* Add two Montgomery form numbers (r = a + b % m).
  16121. *
  16122. * r Result of addition.
  16123. * a First number to add in Montogmery form.
  16124. * b Second number to add in Montogmery form.
  16125. * m Modulus (prime).
  16126. */
  16127. SP_NOINLINE static void sp_256_mont_add_8(sp_digit* r, const sp_digit* a, const sp_digit* b,
  16128. const sp_digit* m)
  16129. {
  16130. (void)m;
  16131. __asm__ __volatile__ (
  16132. "mov r3, #0\n\t"
  16133. "ldr r4, [%[a],#0]\n\t"
  16134. "ldr r5, [%[a],#4]\n\t"
  16135. "ldr r6, [%[b],#0]\n\t"
  16136. "ldr r7, [%[b],#4]\n\t"
  16137. "add r4, r6\n\t"
  16138. "adc r5, r7\n\t"
  16139. "str r4, [%[r],#0]\n\t"
  16140. "str r5, [%[r],#4]\n\t"
  16141. "ldr r4, [%[a],#8]\n\t"
  16142. "ldr r5, [%[a],#12]\n\t"
  16143. "ldr r6, [%[b],#8]\n\t"
  16144. "ldr r7, [%[b],#12]\n\t"
  16145. "adc r4, r6\n\t"
  16146. "adc r5, r7\n\t"
  16147. "str r4, [%[r],#8]\n\t"
  16148. "str r5, [%[r],#12]\n\t"
  16149. "ldr r4, [%[a],#16]\n\t"
  16150. "ldr r5, [%[a],#20]\n\t"
  16151. "ldr r6, [%[b],#16]\n\t"
  16152. "ldr r7, [%[b],#20]\n\t"
  16153. "adc r4, r6\n\t"
  16154. "adc r5, r7\n\t"
  16155. "mov r8, r4\n\t"
  16156. "mov r9, r5\n\t"
  16157. "ldr r4, [%[a],#24]\n\t"
  16158. "ldr r5, [%[a],#28]\n\t"
  16159. "ldr r6, [%[b],#24]\n\t"
  16160. "ldr r7, [%[b],#28]\n\t"
  16161. "adc r4, r6\n\t"
  16162. "adc r5, r7\n\t"
  16163. "mov r10, r4\n\t"
  16164. "mov r11, r5\n\t"
  16165. "adc r3, r3\n\t"
  16166. "mov r6, r3\n\t"
  16167. "sub r3, #1\n\t"
  16168. "mvn r3, r3\n\t"
  16169. "mov r7, #0\n\t"
  16170. "ldr r4, [%[r],#0]\n\t"
  16171. "ldr r5, [%[r],#4]\n\t"
  16172. "sub r4, r3\n\t"
  16173. "sbc r5, r3\n\t"
  16174. "str r4, [%[r],#0]\n\t"
  16175. "str r5, [%[r],#4]\n\t"
  16176. "ldr r4, [%[r],#8]\n\t"
  16177. "ldr r5, [%[r],#12]\n\t"
  16178. "sbc r4, r3\n\t"
  16179. "sbc r5, r7\n\t"
  16180. "str r4, [%[r],#8]\n\t"
  16181. "str r5, [%[r],#12]\n\t"
  16182. "mov r4, r8\n\t"
  16183. "mov r5, r9\n\t"
  16184. "sbc r4, r7\n\t"
  16185. "sbc r5, r7\n\t"
  16186. "str r4, [%[r],#16]\n\t"
  16187. "str r5, [%[r],#20]\n\t"
  16188. "mov r4, r10\n\t"
  16189. "mov r5, r11\n\t"
  16190. "sbc r4, r6\n\t"
  16191. "sbc r5, r3\n\t"
  16192. "str r4, [%[r],#24]\n\t"
  16193. "str r5, [%[r],#28]\n\t"
  16194. :
  16195. : [r] "r" (r), [a] "r" (a), [b] "r" (b)
  16196. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  16197. );
  16198. }
  16199. /* Double a Montgomery form number (r = a + a % m).
  16200. *
  16201. * r Result of doubling.
  16202. * a Number to double in Montogmery form.
  16203. * m Modulus (prime).
  16204. */
  16205. SP_NOINLINE static void sp_256_mont_dbl_8(sp_digit* r, const sp_digit* a, const sp_digit* m)
  16206. {
  16207. (void)m;
  16208. __asm__ __volatile__ (
  16209. "ldr r4, [%[a],#0]\n\t"
  16210. "ldr r5, [%[a],#4]\n\t"
  16211. "ldr r6, [%[a],#8]\n\t"
  16212. "ldr r7, [%[a],#12]\n\t"
  16213. "add r4, r4\n\t"
  16214. "adc r5, r5\n\t"
  16215. "adc r6, r6\n\t"
  16216. "adc r7, r7\n\t"
  16217. "str r4, [%[r],#0]\n\t"
  16218. "str r5, [%[r],#4]\n\t"
  16219. "str r6, [%[r],#8]\n\t"
  16220. "str r7, [%[r],#12]\n\t"
  16221. "ldr r4, [%[a],#16]\n\t"
  16222. "ldr r5, [%[a],#20]\n\t"
  16223. "ldr r6, [%[a],#24]\n\t"
  16224. "ldr r7, [%[a],#28]\n\t"
  16225. "adc r4, r4\n\t"
  16226. "adc r5, r5\n\t"
  16227. "adc r6, r6\n\t"
  16228. "adc r7, r7\n\t"
  16229. "mov r8, r4\n\t"
  16230. "mov r9, r5\n\t"
  16231. "mov r10, r6\n\t"
  16232. "mov r11, r7\n\t"
  16233. "mov r3, #0\n\t"
  16234. "mov r7, #0\n\t"
  16235. "adc r3, r3\n\t"
  16236. "mov r2, r3\n\t"
  16237. "sub r3, #1\n\t"
  16238. "mvn r3, r3\n\t"
  16239. "ldr r4, [%[r],#0]\n\t"
  16240. "ldr r5, [%[r],#4]\n\t"
  16241. "ldr r6, [%[r],#8]\n\t"
  16242. "sub r4, r3\n\t"
  16243. "sbc r5, r3\n\t"
  16244. "sbc r6, r3\n\t"
  16245. "str r4, [%[r],#0]\n\t"
  16246. "str r5, [%[r],#4]\n\t"
  16247. "str r6, [%[r],#8]\n\t"
  16248. "ldr r4, [%[r],#12]\n\t"
  16249. "mov r5, r8\n\t"
  16250. "mov r6, r9\n\t"
  16251. "sbc r4, r7\n\t"
  16252. "sbc r5, r7\n\t"
  16253. "sbc r6, r7\n\t"
  16254. "str r4, [%[r],#12]\n\t"
  16255. "str r5, [%[r],#16]\n\t"
  16256. "str r6, [%[r],#20]\n\t"
  16257. "mov r4, r10\n\t"
  16258. "mov r5, r11\n\t"
  16259. "sbc r4, r2\n\t"
  16260. "sbc r5, r3\n\t"
  16261. "str r4, [%[r],#24]\n\t"
  16262. "str r5, [%[r],#28]\n\t"
  16263. :
  16264. : [r] "r" (r), [a] "r" (a)
  16265. : "memory", "r3", "r2", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  16266. );
  16267. }
  16268. /* Triple a Montgomery form number (r = a + a + a % m).
  16269. *
  16270. * r Result of Tripling.
  16271. * a Number to triple in Montogmery form.
  16272. * m Modulus (prime).
  16273. */
  16274. SP_NOINLINE static void sp_256_mont_tpl_8(sp_digit* r, const sp_digit* a, const sp_digit* m)
  16275. {
  16276. (void)m;
  16277. __asm__ __volatile__ (
  16278. "ldr r6, [%[a],#0]\n\t"
  16279. "ldr r7, [%[a],#4]\n\t"
  16280. "ldr r4, [%[a],#8]\n\t"
  16281. "ldr r5, [%[a],#12]\n\t"
  16282. "add r6, r6\n\t"
  16283. "adc r7, r7\n\t"
  16284. "adc r4, r4\n\t"
  16285. "adc r5, r5\n\t"
  16286. "mov r8, r4\n\t"
  16287. "mov r9, r5\n\t"
  16288. "ldr r2, [%[a],#16]\n\t"
  16289. "ldr r3, [%[a],#20]\n\t"
  16290. "ldr r4, [%[a],#24]\n\t"
  16291. "ldr r5, [%[a],#28]\n\t"
  16292. "adc r2, r2\n\t"
  16293. "adc r3, r3\n\t"
  16294. "adc r4, r4\n\t"
  16295. "adc r5, r5\n\t"
  16296. "mov r10, r2\n\t"
  16297. "mov r11, r3\n\t"
  16298. "mov r12, r4\n\t"
  16299. "mov r14, r5\n\t"
  16300. "mov r3, #0\n\t"
  16301. "mov r5, #0\n\t"
  16302. "adc r3, r3\n\t"
  16303. "mov r4, r3\n\t"
  16304. "sub r3, #1\n\t"
  16305. "mvn r3, r3\n\t"
  16306. "sub r6, r3\n\t"
  16307. "sbc r7, r3\n\t"
  16308. "mov r2, r8\n\t"
  16309. "sbc r2, r3\n\t"
  16310. "mov r8, r2\n\t"
  16311. "mov r2, r9\n\t"
  16312. "sbc r2, r5\n\t"
  16313. "mov r9, r2\n\t"
  16314. "mov r2, r10\n\t"
  16315. "sbc r2, r5\n\t"
  16316. "mov r10, r2\n\t"
  16317. "mov r2, r11\n\t"
  16318. "sbc r2, r5\n\t"
  16319. "mov r11, r2\n\t"
  16320. "mov r2, r12\n\t"
  16321. "sbc r2, r4\n\t"
  16322. "mov r12, r2\n\t"
  16323. "mov r2, r14\n\t"
  16324. "sbc r2, r3\n\t"
  16325. "mov r14, r2\n\t"
  16326. "ldr r2, [%[a],#0]\n\t"
  16327. "ldr r3, [%[a],#4]\n\t"
  16328. "add r6, r2\n\t"
  16329. "adc r7, r3\n\t"
  16330. "ldr r2, [%[a],#8]\n\t"
  16331. "ldr r3, [%[a],#12]\n\t"
  16332. "mov r4, r8\n\t"
  16333. "mov r5, r9\n\t"
  16334. "adc r2, r4\n\t"
  16335. "adc r3, r5\n\t"
  16336. "mov r8, r2\n\t"
  16337. "mov r9, r3\n\t"
  16338. "ldr r2, [%[a],#16]\n\t"
  16339. "ldr r3, [%[a],#20]\n\t"
  16340. "mov r4, r10\n\t"
  16341. "mov r5, r11\n\t"
  16342. "adc r2, r4\n\t"
  16343. "adc r3, r5\n\t"
  16344. "mov r10, r2\n\t"
  16345. "mov r11, r3\n\t"
  16346. "ldr r2, [%[a],#24]\n\t"
  16347. "ldr r3, [%[a],#28]\n\t"
  16348. "mov r4, r12\n\t"
  16349. "mov r5, r14\n\t"
  16350. "adc r2, r4\n\t"
  16351. "adc r3, r5\n\t"
  16352. "mov r12, r2\n\t"
  16353. "mov r14, r3\n\t"
  16354. "mov r3, #0\n\t"
  16355. "mov r5, #0\n\t"
  16356. "adc r3, r3\n\t"
  16357. "mov r4, r3\n\t"
  16358. "sub r3, #1\n\t"
  16359. "mvn r3, r3\n\t"
  16360. "sub r6, r3\n\t"
  16361. "str r6, [%[r],#0]\n\t"
  16362. "sbc r7, r3\n\t"
  16363. "str r7, [%[r],#4]\n\t"
  16364. "mov r2, r8\n\t"
  16365. "sbc r2, r3\n\t"
  16366. "str r2, [%[r],#8]\n\t"
  16367. "mov r2, r9\n\t"
  16368. "sbc r2, r5\n\t"
  16369. "str r2, [%[r],#12]\n\t"
  16370. "mov r2, r10\n\t"
  16371. "sbc r2, r5\n\t"
  16372. "str r2, [%[r],#16]\n\t"
  16373. "mov r2, r11\n\t"
  16374. "sbc r2, r5\n\t"
  16375. "str r2, [%[r],#20]\n\t"
  16376. "mov r2, r12\n\t"
  16377. "sbc r2, r4\n\t"
  16378. "str r2, [%[r],#24]\n\t"
  16379. "mov r2, r14\n\t"
  16380. "sbc r2, r3\n\t"
  16381. "str r2, [%[r],#28]\n\t"
  16382. :
  16383. : [r] "r" (r), [a] "r" (a)
  16384. : "memory", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "r14"
  16385. );
  16386. }
  16387. /* Subtract two Montgomery form numbers (r = a - b % m).
  16388. *
  16389. * r Result of subtration.
  16390. * a Number to subtract from in Montogmery form.
  16391. * b Number to subtract with in Montogmery form.
  16392. * m Modulus (prime).
  16393. */
  16394. SP_NOINLINE static void sp_256_mont_sub_8(sp_digit* r, const sp_digit* a, const sp_digit* b,
  16395. const sp_digit* m)
  16396. {
  16397. (void)m;
  16398. __asm__ __volatile__ (
  16399. "ldr r4, [%[a],#0]\n\t"
  16400. "ldr r5, [%[a],#4]\n\t"
  16401. "ldr r6, [%[b],#0]\n\t"
  16402. "ldr r7, [%[b],#4]\n\t"
  16403. "sub r4, r6\n\t"
  16404. "sbc r5, r7\n\t"
  16405. "str r4, [%[r],#0]\n\t"
  16406. "str r5, [%[r],#4]\n\t"
  16407. "ldr r4, [%[a],#8]\n\t"
  16408. "ldr r5, [%[a],#12]\n\t"
  16409. "ldr r6, [%[b],#8]\n\t"
  16410. "ldr r7, [%[b],#12]\n\t"
  16411. "sbc r4, r6\n\t"
  16412. "sbc r5, r7\n\t"
  16413. "str r4, [%[r],#8]\n\t"
  16414. "str r5, [%[r],#12]\n\t"
  16415. "ldr r4, [%[a],#16]\n\t"
  16416. "ldr r5, [%[a],#20]\n\t"
  16417. "ldr r6, [%[b],#16]\n\t"
  16418. "ldr r7, [%[b],#20]\n\t"
  16419. "sbc r4, r6\n\t"
  16420. "sbc r5, r7\n\t"
  16421. "mov r8, r4\n\t"
  16422. "mov r9, r5\n\t"
  16423. "ldr r4, [%[a],#24]\n\t"
  16424. "ldr r5, [%[a],#28]\n\t"
  16425. "ldr r6, [%[b],#24]\n\t"
  16426. "ldr r7, [%[b],#28]\n\t"
  16427. "sbc r4, r6\n\t"
  16428. "sbc r5, r7\n\t"
  16429. "mov r10, r4\n\t"
  16430. "mov r11, r5\n\t"
  16431. "sbc r3, r3\n\t"
  16432. "lsr r7, r3, #31\n\t"
  16433. "mov r6, #0\n\t"
  16434. "ldr r4, [%[r],#0]\n\t"
  16435. "ldr r5, [%[r],#4]\n\t"
  16436. "add r4, r3\n\t"
  16437. "adc r5, r3\n\t"
  16438. "str r4, [%[r],#0]\n\t"
  16439. "str r5, [%[r],#4]\n\t"
  16440. "ldr r4, [%[r],#8]\n\t"
  16441. "ldr r5, [%[r],#12]\n\t"
  16442. "adc r4, r3\n\t"
  16443. "adc r5, r6\n\t"
  16444. "str r4, [%[r],#8]\n\t"
  16445. "str r5, [%[r],#12]\n\t"
  16446. "mov r4, r8\n\t"
  16447. "mov r5, r9\n\t"
  16448. "adc r4, r6\n\t"
  16449. "adc r5, r6\n\t"
  16450. "str r4, [%[r],#16]\n\t"
  16451. "str r5, [%[r],#20]\n\t"
  16452. "mov r4, r10\n\t"
  16453. "mov r5, r11\n\t"
  16454. "adc r4, r7\n\t"
  16455. "adc r5, r3\n\t"
  16456. "str r4, [%[r],#24]\n\t"
  16457. "str r5, [%[r],#28]\n\t"
  16458. :
  16459. : [r] "r" (r), [a] "r" (a), [b] "r" (b)
  16460. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11"
  16461. );
  16462. }
  16463. /* Divide the number by 2 mod the modulus (prime). (r = a / 2 % m)
  16464. *
  16465. * r Result of division by 2.
  16466. * a Number to divide.
  16467. * m Modulus (prime).
  16468. */
  16469. SP_NOINLINE static void sp_256_div2_8(sp_digit* r, const sp_digit* a, const sp_digit* m)
  16470. {
  16471. __asm__ __volatile__ (
  16472. "ldr r7, [%[a], #0]\n\t"
  16473. "lsl r7, r7, #31\n\t"
  16474. "lsr r7, r7, #31\n\t"
  16475. "mov r5, #0\n\t"
  16476. "sub r5, r7\n\t"
  16477. "mov r7, #0\n\t"
  16478. "lsl r6, r5, #31\n\t"
  16479. "lsr r6, r6, #31\n\t"
  16480. "ldr r3, [%[a], #0]\n\t"
  16481. "ldr r4, [%[a], #4]\n\t"
  16482. "add r3, r5\n\t"
  16483. "adc r4, r5\n\t"
  16484. "str r3, [%[r], #0]\n\t"
  16485. "str r4, [%[r], #4]\n\t"
  16486. "ldr r3, [%[a], #8]\n\t"
  16487. "ldr r4, [%[a], #12]\n\t"
  16488. "adc r3, r5\n\t"
  16489. "adc r4, r7\n\t"
  16490. "str r3, [%[r], #8]\n\t"
  16491. "str r4, [%[r], #12]\n\t"
  16492. "ldr r3, [%[a], #16]\n\t"
  16493. "ldr r4, [%[a], #20]\n\t"
  16494. "adc r3, r7\n\t"
  16495. "adc r4, r7\n\t"
  16496. "str r3, [%[r], #16]\n\t"
  16497. "str r4, [%[r], #20]\n\t"
  16498. "ldr r3, [%[a], #24]\n\t"
  16499. "ldr r4, [%[a], #28]\n\t"
  16500. "adc r3, r6\n\t"
  16501. "adc r4, r5\n\t"
  16502. "adc r7, r7\n\t"
  16503. "lsl r7, r7, #31\n\t"
  16504. "lsr r5, r3, #1\n\t"
  16505. "lsl r3, r3, #31\n\t"
  16506. "lsr r6, r4, #1\n\t"
  16507. "lsl r4, r4, #31\n\t"
  16508. "orr r5, r4\n\t"
  16509. "orr r6, r7\n\t"
  16510. "mov r7, r3\n\t"
  16511. "str r5, [%[r], #24]\n\t"
  16512. "str r6, [%[r], #28]\n\t"
  16513. "ldr r3, [%[a], #16]\n\t"
  16514. "ldr r4, [%[a], #20]\n\t"
  16515. "lsr r5, r3, #1\n\t"
  16516. "lsl r3, r3, #31\n\t"
  16517. "lsr r6, r4, #1\n\t"
  16518. "lsl r4, r4, #31\n\t"
  16519. "orr r5, r4\n\t"
  16520. "orr r6, r7\n\t"
  16521. "mov r7, r3\n\t"
  16522. "str r5, [%[r], #16]\n\t"
  16523. "str r6, [%[r], #20]\n\t"
  16524. "ldr r3, [%[a], #8]\n\t"
  16525. "ldr r4, [%[a], #12]\n\t"
  16526. "lsr r5, r3, #1\n\t"
  16527. "lsl r3, r3, #31\n\t"
  16528. "lsr r6, r4, #1\n\t"
  16529. "lsl r4, r4, #31\n\t"
  16530. "orr r5, r4\n\t"
  16531. "orr r6, r7\n\t"
  16532. "mov r7, r3\n\t"
  16533. "str r5, [%[r], #8]\n\t"
  16534. "str r6, [%[r], #12]\n\t"
  16535. "ldr r3, [%[r], #0]\n\t"
  16536. "ldr r4, [%[r], #4]\n\t"
  16537. "lsr r5, r3, #1\n\t"
  16538. "lsr r6, r4, #1\n\t"
  16539. "lsl r4, r4, #31\n\t"
  16540. "orr r5, r4\n\t"
  16541. "orr r6, r7\n\t"
  16542. "str r5, [%[r], #0]\n\t"
  16543. "str r6, [%[r], #4]\n\t"
  16544. :
  16545. : [r] "r" (r), [a] "r" (a), [m] "r" (m)
  16546. : "memory", "r3", "r4", "r5", "r6", "r7"
  16547. );
  16548. }
  16549. /* Double the Montgomery form projective point p.
  16550. *
  16551. * r Result of doubling point.
  16552. * p Point to double.
  16553. * t Temporary ordinate data.
  16554. */
  16555. static void sp_256_proj_point_dbl_8(sp_point* r, const sp_point* p, sp_digit* t)
  16556. {
  16557. sp_point* rp[2];
  16558. sp_digit* t1 = t;
  16559. sp_digit* t2 = t + 2*8;
  16560. sp_digit* x;
  16561. sp_digit* y;
  16562. sp_digit* z;
  16563. int i;
  16564. /* When infinity don't double point passed in - constant time. */
  16565. rp[0] = r;
  16566. /*lint allow cast to different type of pointer*/
  16567. rp[1] = (sp_point*)t; /*lint !e9087 !e740*/
  16568. XMEMSET(rp[1], 0, sizeof(sp_point));
  16569. x = rp[p->infinity]->x;
  16570. y = rp[p->infinity]->y;
  16571. z = rp[p->infinity]->z;
  16572. /* Put point to double into result - good for infinity. */
  16573. if (r != p) {
  16574. for (i=0; i<8; i++) {
  16575. r->x[i] = p->x[i];
  16576. }
  16577. for (i=0; i<8; i++) {
  16578. r->y[i] = p->y[i];
  16579. }
  16580. for (i=0; i<8; i++) {
  16581. r->z[i] = p->z[i];
  16582. }
  16583. r->infinity = p->infinity;
  16584. }
  16585. /* T1 = Z * Z */
  16586. sp_256_mont_sqr_8(t1, z, p256_mod, p256_mp_mod);
  16587. /* Z = Y * Z */
  16588. sp_256_mont_mul_8(z, y, z, p256_mod, p256_mp_mod);
  16589. /* Z = 2Z */
  16590. sp_256_mont_dbl_8(z, z, p256_mod);
  16591. /* T2 = X - T1 */
  16592. sp_256_mont_sub_8(t2, x, t1, p256_mod);
  16593. /* T1 = X + T1 */
  16594. sp_256_mont_add_8(t1, x, t1, p256_mod);
  16595. /* T2 = T1 * T2 */
  16596. sp_256_mont_mul_8(t2, t1, t2, p256_mod, p256_mp_mod);
  16597. /* T1 = 3T2 */
  16598. sp_256_mont_tpl_8(t1, t2, p256_mod);
  16599. /* Y = 2Y */
  16600. sp_256_mont_dbl_8(y, y, p256_mod);
  16601. /* Y = Y * Y */
  16602. sp_256_mont_sqr_8(y, y, p256_mod, p256_mp_mod);
  16603. /* T2 = Y * Y */
  16604. sp_256_mont_sqr_8(t2, y, p256_mod, p256_mp_mod);
  16605. /* T2 = T2/2 */
  16606. sp_256_div2_8(t2, t2, p256_mod);
  16607. /* Y = Y * X */
  16608. sp_256_mont_mul_8(y, y, x, p256_mod, p256_mp_mod);
  16609. /* X = T1 * T1 */
  16610. sp_256_mont_mul_8(x, t1, t1, p256_mod, p256_mp_mod);
  16611. /* X = X - Y */
  16612. sp_256_mont_sub_8(x, x, y, p256_mod);
  16613. /* X = X - Y */
  16614. sp_256_mont_sub_8(x, x, y, p256_mod);
  16615. /* Y = Y - X */
  16616. sp_256_mont_sub_8(y, y, x, p256_mod);
  16617. /* Y = Y * T1 */
  16618. sp_256_mont_mul_8(y, y, t1, p256_mod, p256_mp_mod);
  16619. /* Y = Y - T2 */
  16620. sp_256_mont_sub_8(y, y, t2, p256_mod);
  16621. }
  16622. #ifdef WOLFSSL_SP_SMALL
  16623. /* Sub b from a into r. (r = a - b)
  16624. *
  16625. * r A single precision integer.
  16626. * a A single precision integer.
  16627. * b A single precision integer.
  16628. */
  16629. SP_NOINLINE static sp_digit sp_256_sub_8(sp_digit* r, const sp_digit* a,
  16630. const sp_digit* b)
  16631. {
  16632. sp_digit c = 0;
  16633. __asm__ __volatile__ (
  16634. "mov r6, %[a]\n\t"
  16635. "add r6, #32\n\t"
  16636. "\n1:\n\t"
  16637. "mov r5, #0\n\t"
  16638. "sub r5, %[c]\n\t"
  16639. "ldr r4, [%[a]]\n\t"
  16640. "ldr r5, [%[b]]\n\t"
  16641. "sbc r4, r5\n\t"
  16642. "str r4, [%[r]]\n\t"
  16643. "sbc %[c], %[c]\n\t"
  16644. "add %[a], #4\n\t"
  16645. "add %[b], #4\n\t"
  16646. "add %[r], #4\n\t"
  16647. "cmp %[a], r6\n\t"
  16648. "bne 1b\n\t"
  16649. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  16650. :
  16651. : "memory", "r4", "r5", "r6"
  16652. );
  16653. return c;
  16654. }
  16655. #else
  16656. /* Sub b from a into r. (r = a - b)
  16657. *
  16658. * r A single precision integer.
  16659. * a A single precision integer.
  16660. * b A single precision integer.
  16661. */
  16662. SP_NOINLINE static sp_digit sp_256_sub_8(sp_digit* r, const sp_digit* a,
  16663. const sp_digit* b)
  16664. {
  16665. sp_digit c = 0;
  16666. __asm__ __volatile__ (
  16667. "ldr r4, [%[a], #0]\n\t"
  16668. "ldr r5, [%[a], #4]\n\t"
  16669. "ldr r6, [%[b], #0]\n\t"
  16670. "ldr r7, [%[b], #4]\n\t"
  16671. "sub r4, r6\n\t"
  16672. "sbc r5, r7\n\t"
  16673. "str r4, [%[r], #0]\n\t"
  16674. "str r5, [%[r], #4]\n\t"
  16675. "ldr r4, [%[a], #8]\n\t"
  16676. "ldr r5, [%[a], #12]\n\t"
  16677. "ldr r6, [%[b], #8]\n\t"
  16678. "ldr r7, [%[b], #12]\n\t"
  16679. "sbc r4, r6\n\t"
  16680. "sbc r5, r7\n\t"
  16681. "str r4, [%[r], #8]\n\t"
  16682. "str r5, [%[r], #12]\n\t"
  16683. "ldr r4, [%[a], #16]\n\t"
  16684. "ldr r5, [%[a], #20]\n\t"
  16685. "ldr r6, [%[b], #16]\n\t"
  16686. "ldr r7, [%[b], #20]\n\t"
  16687. "sbc r4, r6\n\t"
  16688. "sbc r5, r7\n\t"
  16689. "str r4, [%[r], #16]\n\t"
  16690. "str r5, [%[r], #20]\n\t"
  16691. "ldr r4, [%[a], #24]\n\t"
  16692. "ldr r5, [%[a], #28]\n\t"
  16693. "ldr r6, [%[b], #24]\n\t"
  16694. "ldr r7, [%[b], #28]\n\t"
  16695. "sbc r4, r6\n\t"
  16696. "sbc r5, r7\n\t"
  16697. "str r4, [%[r], #24]\n\t"
  16698. "str r5, [%[r], #28]\n\t"
  16699. "sbc %[c], %[c]\n\t"
  16700. : [c] "+r" (c), [r] "+r" (r), [a] "+r" (a), [b] "+r" (b)
  16701. :
  16702. : "memory", "r4", "r5", "r6", "r7"
  16703. );
  16704. return c;
  16705. }
  16706. #endif /* WOLFSSL_SP_SMALL */
  16707. /* Compare two numbers to determine if they are equal.
  16708. * Constant time implementation.
  16709. *
  16710. * a First number to compare.
  16711. * b Second number to compare.
  16712. * returns 1 when equal and 0 otherwise.
  16713. */
  16714. static int sp_256_cmp_equal_8(const sp_digit* a, const sp_digit* b)
  16715. {
  16716. return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2]) | (a[3] ^ b[3]) |
  16717. (a[4] ^ b[4]) | (a[5] ^ b[5]) | (a[6] ^ b[6]) | (a[7] ^ b[7])) == 0;
  16718. }
  16719. /* Add two Montgomery form projective points.
  16720. *
  16721. * r Result of addition.
  16722. * p First point to add.
  16723. * q Second point to add.
  16724. * t Temporary ordinate data.
  16725. */
  16726. static void sp_256_proj_point_add_8(sp_point* r, const sp_point* p, const sp_point* q,
  16727. sp_digit* t)
  16728. {
  16729. const sp_point* ap[2];
  16730. sp_point* rp[2];
  16731. sp_digit* t1 = t;
  16732. sp_digit* t2 = t + 2*8;
  16733. sp_digit* t3 = t + 4*8;
  16734. sp_digit* t4 = t + 6*8;
  16735. sp_digit* t5 = t + 8*8;
  16736. sp_digit* x;
  16737. sp_digit* y;
  16738. sp_digit* z;
  16739. int i;
  16740. /* Ensure only the first point is the same as the result. */
  16741. if (q == r) {
  16742. const sp_point* a = p;
  16743. p = q;
  16744. q = a;
  16745. }
  16746. /* Check double */
  16747. (void)sp_256_sub_8(t1, p256_mod, q->y);
  16748. sp_256_norm_8(t1);
  16749. if ((sp_256_cmp_equal_8(p->x, q->x) & sp_256_cmp_equal_8(p->z, q->z) &
  16750. (sp_256_cmp_equal_8(p->y, q->y) | sp_256_cmp_equal_8(p->y, t1))) != 0) {
  16751. sp_256_proj_point_dbl_8(r, p, t);
  16752. }
  16753. else {
  16754. rp[0] = r;
  16755. /*lint allow cast to different type of pointer*/
  16756. rp[1] = (sp_point*)t; /*lint !e9087 !e740*/
  16757. XMEMSET(rp[1], 0, sizeof(sp_point));
  16758. x = rp[p->infinity | q->infinity]->x;
  16759. y = rp[p->infinity | q->infinity]->y;
  16760. z = rp[p->infinity | q->infinity]->z;
  16761. ap[0] = p;
  16762. ap[1] = q;
  16763. for (i=0; i<8; i++) {
  16764. r->x[i] = ap[p->infinity]->x[i];
  16765. }
  16766. for (i=0; i<8; i++) {
  16767. r->y[i] = ap[p->infinity]->y[i];
  16768. }
  16769. for (i=0; i<8; i++) {
  16770. r->z[i] = ap[p->infinity]->z[i];
  16771. }
  16772. r->infinity = ap[p->infinity]->infinity;
  16773. /* U1 = X1*Z2^2 */
  16774. sp_256_mont_sqr_8(t1, q->z, p256_mod, p256_mp_mod);
  16775. sp_256_mont_mul_8(t3, t1, q->z, p256_mod, p256_mp_mod);
  16776. sp_256_mont_mul_8(t1, t1, x, p256_mod, p256_mp_mod);
  16777. /* U2 = X2*Z1^2 */
  16778. sp_256_mont_sqr_8(t2, z, p256_mod, p256_mp_mod);
  16779. sp_256_mont_mul_8(t4, t2, z, p256_mod, p256_mp_mod);
  16780. sp_256_mont_mul_8(t2, t2, q->x, p256_mod, p256_mp_mod);
  16781. /* S1 = Y1*Z2^3 */
  16782. sp_256_mont_mul_8(t3, t3, y, p256_mod, p256_mp_mod);
  16783. /* S2 = Y2*Z1^3 */
  16784. sp_256_mont_mul_8(t4, t4, q->y, p256_mod, p256_mp_mod);
  16785. /* H = U2 - U1 */
  16786. sp_256_mont_sub_8(t2, t2, t1, p256_mod);
  16787. /* R = S2 - S1 */
  16788. sp_256_mont_sub_8(t4, t4, t3, p256_mod);
  16789. /* Z3 = H*Z1*Z2 */
  16790. sp_256_mont_mul_8(z, z, q->z, p256_mod, p256_mp_mod);
  16791. sp_256_mont_mul_8(z, z, t2, p256_mod, p256_mp_mod);
  16792. /* X3 = R^2 - H^3 - 2*U1*H^2 */
  16793. sp_256_mont_sqr_8(x, t4, p256_mod, p256_mp_mod);
  16794. sp_256_mont_sqr_8(t5, t2, p256_mod, p256_mp_mod);
  16795. sp_256_mont_mul_8(y, t1, t5, p256_mod, p256_mp_mod);
  16796. sp_256_mont_mul_8(t5, t5, t2, p256_mod, p256_mp_mod);
  16797. sp_256_mont_sub_8(x, x, t5, p256_mod);
  16798. sp_256_mont_dbl_8(t1, y, p256_mod);
  16799. sp_256_mont_sub_8(x, x, t1, p256_mod);
  16800. /* Y3 = R*(U1*H^2 - X3) - S1*H^3 */
  16801. sp_256_mont_sub_8(y, y, x, p256_mod);
  16802. sp_256_mont_mul_8(y, y, t4, p256_mod, p256_mp_mod);
  16803. sp_256_mont_mul_8(t5, t5, t3, p256_mod, p256_mp_mod);
  16804. sp_256_mont_sub_8(y, y, t5, p256_mod);
  16805. }
  16806. }
  16807. /* Multiply the point by the scalar and return the result.
  16808. * If map is true then convert result to affine coordinates.
  16809. *
  16810. * r Resulting point.
  16811. * g Point to multiply.
  16812. * k Scalar to multiply by.
  16813. * map Indicates whether to convert result to affine.
  16814. * heap Heap to use for allocation.
  16815. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  16816. */
  16817. static int sp_256_ecc_mulmod_fast_8(sp_point* r, const sp_point* g, const sp_digit* k,
  16818. int map, void* heap)
  16819. {
  16820. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  16821. sp_point td[16];
  16822. sp_point rtd;
  16823. sp_digit tmpd[2 * 8 * 5];
  16824. #endif
  16825. sp_point* t;
  16826. sp_point* rt;
  16827. sp_digit* tmp;
  16828. sp_digit n;
  16829. int i;
  16830. int c, y;
  16831. int err;
  16832. (void)heap;
  16833. err = sp_ecc_point_new(heap, rtd, rt);
  16834. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  16835. t = (sp_point*)XMALLOC(sizeof(sp_point) * 16, heap, DYNAMIC_TYPE_ECC);
  16836. if (t == NULL)
  16837. err = MEMORY_E;
  16838. tmp = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 5, heap,
  16839. DYNAMIC_TYPE_ECC);
  16840. if (tmp == NULL)
  16841. err = MEMORY_E;
  16842. #else
  16843. t = td;
  16844. tmp = tmpd;
  16845. #endif
  16846. if (err == MP_OKAY) {
  16847. /* t[0] = {0, 0, 1} * norm */
  16848. XMEMSET(&t[0], 0, sizeof(t[0]));
  16849. t[0].infinity = 1;
  16850. /* t[1] = {g->x, g->y, g->z} * norm */
  16851. (void)sp_256_mod_mul_norm_8(t[1].x, g->x, p256_mod);
  16852. (void)sp_256_mod_mul_norm_8(t[1].y, g->y, p256_mod);
  16853. (void)sp_256_mod_mul_norm_8(t[1].z, g->z, p256_mod);
  16854. t[1].infinity = 0;
  16855. sp_256_proj_point_dbl_8(&t[ 2], &t[ 1], tmp);
  16856. t[ 2].infinity = 0;
  16857. sp_256_proj_point_add_8(&t[ 3], &t[ 2], &t[ 1], tmp);
  16858. t[ 3].infinity = 0;
  16859. sp_256_proj_point_dbl_8(&t[ 4], &t[ 2], tmp);
  16860. t[ 4].infinity = 0;
  16861. sp_256_proj_point_add_8(&t[ 5], &t[ 3], &t[ 2], tmp);
  16862. t[ 5].infinity = 0;
  16863. sp_256_proj_point_dbl_8(&t[ 6], &t[ 3], tmp);
  16864. t[ 6].infinity = 0;
  16865. sp_256_proj_point_add_8(&t[ 7], &t[ 4], &t[ 3], tmp);
  16866. t[ 7].infinity = 0;
  16867. sp_256_proj_point_dbl_8(&t[ 8], &t[ 4], tmp);
  16868. t[ 8].infinity = 0;
  16869. sp_256_proj_point_add_8(&t[ 9], &t[ 5], &t[ 4], tmp);
  16870. t[ 9].infinity = 0;
  16871. sp_256_proj_point_dbl_8(&t[10], &t[ 5], tmp);
  16872. t[10].infinity = 0;
  16873. sp_256_proj_point_add_8(&t[11], &t[ 6], &t[ 5], tmp);
  16874. t[11].infinity = 0;
  16875. sp_256_proj_point_dbl_8(&t[12], &t[ 6], tmp);
  16876. t[12].infinity = 0;
  16877. sp_256_proj_point_add_8(&t[13], &t[ 7], &t[ 6], tmp);
  16878. t[13].infinity = 0;
  16879. sp_256_proj_point_dbl_8(&t[14], &t[ 7], tmp);
  16880. t[14].infinity = 0;
  16881. sp_256_proj_point_add_8(&t[15], &t[ 8], &t[ 7], tmp);
  16882. t[15].infinity = 0;
  16883. i = 6;
  16884. n = k[i+1] << 0;
  16885. c = 28;
  16886. y = n >> 28;
  16887. XMEMCPY(rt, &t[y], sizeof(sp_point));
  16888. n <<= 4;
  16889. for (; i>=0 || c>=4; ) {
  16890. if (c < 4) {
  16891. n |= k[i--] << (0 - c);
  16892. c += 32;
  16893. }
  16894. y = (n >> 28) & 0xf;
  16895. n <<= 4;
  16896. c -= 4;
  16897. sp_256_proj_point_dbl_8(rt, rt, tmp);
  16898. sp_256_proj_point_dbl_8(rt, rt, tmp);
  16899. sp_256_proj_point_dbl_8(rt, rt, tmp);
  16900. sp_256_proj_point_dbl_8(rt, rt, tmp);
  16901. sp_256_proj_point_add_8(rt, rt, &t[y], tmp);
  16902. }
  16903. if (map != 0) {
  16904. sp_256_map_8(r, rt, tmp);
  16905. }
  16906. else {
  16907. XMEMCPY(r, rt, sizeof(sp_point));
  16908. }
  16909. }
  16910. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  16911. if (tmp != NULL) {
  16912. XMEMSET(tmp, 0, sizeof(sp_digit) * 2 * 8 * 5);
  16913. XFREE(tmp, heap, DYNAMIC_TYPE_ECC);
  16914. }
  16915. if (t != NULL) {
  16916. XMEMSET(t, 0, sizeof(sp_point) * 16);
  16917. XFREE(t, heap, DYNAMIC_TYPE_ECC);
  16918. }
  16919. #else
  16920. ForceZero(tmpd, sizeof(tmpd));
  16921. ForceZero(td, sizeof(td));
  16922. #endif
  16923. sp_ecc_point_free(rt, 1, heap);
  16924. return err;
  16925. }
  16926. /* A table entry for pre-computed points. */
  16927. typedef struct sp_table_entry {
  16928. sp_digit x[8];
  16929. sp_digit y[8];
  16930. } sp_table_entry;
  16931. #ifdef FP_ECC
  16932. /* Double the Montgomery form projective point p a number of times.
  16933. *
  16934. * r Result of repeated doubling of point.
  16935. * p Point to double.
  16936. * n Number of times to double
  16937. * t Temporary ordinate data.
  16938. */
  16939. static void sp_256_proj_point_dbl_n_8(sp_point* r, const sp_point* p, int n,
  16940. sp_digit* t)
  16941. {
  16942. sp_point* rp[2];
  16943. sp_digit* w = t;
  16944. sp_digit* a = t + 2*8;
  16945. sp_digit* b = t + 4*8;
  16946. sp_digit* t1 = t + 6*8;
  16947. sp_digit* t2 = t + 8*8;
  16948. sp_digit* x;
  16949. sp_digit* y;
  16950. sp_digit* z;
  16951. int i;
  16952. rp[0] = r;
  16953. /*lint allow cast to different type of pointer*/
  16954. rp[1] = (sp_point*)t; /*lint !e9087 !e740*/
  16955. XMEMSET(rp[1], 0, sizeof(sp_point));
  16956. x = rp[p->infinity]->x;
  16957. y = rp[p->infinity]->y;
  16958. z = rp[p->infinity]->z;
  16959. if (r != p) {
  16960. for (i=0; i<8; i++) {
  16961. r->x[i] = p->x[i];
  16962. }
  16963. for (i=0; i<8; i++) {
  16964. r->y[i] = p->y[i];
  16965. }
  16966. for (i=0; i<8; i++) {
  16967. r->z[i] = p->z[i];
  16968. }
  16969. r->infinity = p->infinity;
  16970. }
  16971. /* Y = 2*Y */
  16972. sp_256_mont_dbl_8(y, y, p256_mod);
  16973. /* W = Z^4 */
  16974. sp_256_mont_sqr_8(w, z, p256_mod, p256_mp_mod);
  16975. sp_256_mont_sqr_8(w, w, p256_mod, p256_mp_mod);
  16976. while (n-- > 0) {
  16977. /* A = 3*(X^2 - W) */
  16978. sp_256_mont_sqr_8(t1, x, p256_mod, p256_mp_mod);
  16979. sp_256_mont_sub_8(t1, t1, w, p256_mod);
  16980. sp_256_mont_tpl_8(a, t1, p256_mod);
  16981. /* B = X*Y^2 */
  16982. sp_256_mont_sqr_8(t2, y, p256_mod, p256_mp_mod);
  16983. sp_256_mont_mul_8(b, t2, x, p256_mod, p256_mp_mod);
  16984. /* X = A^2 - 2B */
  16985. sp_256_mont_sqr_8(x, a, p256_mod, p256_mp_mod);
  16986. sp_256_mont_dbl_8(t1, b, p256_mod);
  16987. sp_256_mont_sub_8(x, x, t1, p256_mod);
  16988. /* Z = Z*Y */
  16989. sp_256_mont_mul_8(z, z, y, p256_mod, p256_mp_mod);
  16990. /* t2 = Y^4 */
  16991. sp_256_mont_sqr_8(t2, t2, p256_mod, p256_mp_mod);
  16992. if (n != 0) {
  16993. /* W = W*Y^4 */
  16994. sp_256_mont_mul_8(w, w, t2, p256_mod, p256_mp_mod);
  16995. }
  16996. /* y = 2*A*(B - X) - Y^4 */
  16997. sp_256_mont_sub_8(y, b, x, p256_mod);
  16998. sp_256_mont_mul_8(y, y, a, p256_mod, p256_mp_mod);
  16999. sp_256_mont_dbl_8(y, y, p256_mod);
  17000. sp_256_mont_sub_8(y, y, t2, p256_mod);
  17001. }
  17002. /* Y = Y/2 */
  17003. sp_256_div2_8(y, y, p256_mod);
  17004. }
  17005. #endif /* FP_ECC */
  17006. /* Add two Montgomery form projective points. The second point has a q value of
  17007. * one.
  17008. * Only the first point can be the same pointer as the result point.
  17009. *
  17010. * r Result of addition.
  17011. * p First point to add.
  17012. * q Second point to add.
  17013. * t Temporary ordinate data.
  17014. */
  17015. static void sp_256_proj_point_add_qz1_8(sp_point* r, const sp_point* p,
  17016. const sp_point* q, sp_digit* t)
  17017. {
  17018. const sp_point* ap[2];
  17019. sp_point* rp[2];
  17020. sp_digit* t1 = t;
  17021. sp_digit* t2 = t + 2*8;
  17022. sp_digit* t3 = t + 4*8;
  17023. sp_digit* t4 = t + 6*8;
  17024. sp_digit* t5 = t + 8*8;
  17025. sp_digit* x;
  17026. sp_digit* y;
  17027. sp_digit* z;
  17028. int i;
  17029. /* Check double */
  17030. (void)sp_256_sub_8(t1, p256_mod, q->y);
  17031. sp_256_norm_8(t1);
  17032. if ((sp_256_cmp_equal_8(p->x, q->x) & sp_256_cmp_equal_8(p->z, q->z) &
  17033. (sp_256_cmp_equal_8(p->y, q->y) | sp_256_cmp_equal_8(p->y, t1))) != 0) {
  17034. sp_256_proj_point_dbl_8(r, p, t);
  17035. }
  17036. else {
  17037. rp[0] = r;
  17038. /*lint allow cast to different type of pointer*/
  17039. rp[1] = (sp_point*)t; /*lint !e9087 !e740*/
  17040. XMEMSET(rp[1], 0, sizeof(sp_point));
  17041. x = rp[p->infinity | q->infinity]->x;
  17042. y = rp[p->infinity | q->infinity]->y;
  17043. z = rp[p->infinity | q->infinity]->z;
  17044. ap[0] = p;
  17045. ap[1] = q;
  17046. for (i=0; i<8; i++) {
  17047. r->x[i] = ap[p->infinity]->x[i];
  17048. }
  17049. for (i=0; i<8; i++) {
  17050. r->y[i] = ap[p->infinity]->y[i];
  17051. }
  17052. for (i=0; i<8; i++) {
  17053. r->z[i] = ap[p->infinity]->z[i];
  17054. }
  17055. r->infinity = ap[p->infinity]->infinity;
  17056. /* U2 = X2*Z1^2 */
  17057. sp_256_mont_sqr_8(t2, z, p256_mod, p256_mp_mod);
  17058. sp_256_mont_mul_8(t4, t2, z, p256_mod, p256_mp_mod);
  17059. sp_256_mont_mul_8(t2, t2, q->x, p256_mod, p256_mp_mod);
  17060. /* S2 = Y2*Z1^3 */
  17061. sp_256_mont_mul_8(t4, t4, q->y, p256_mod, p256_mp_mod);
  17062. /* H = U2 - X1 */
  17063. sp_256_mont_sub_8(t2, t2, x, p256_mod);
  17064. /* R = S2 - Y1 */
  17065. sp_256_mont_sub_8(t4, t4, y, p256_mod);
  17066. /* Z3 = H*Z1 */
  17067. sp_256_mont_mul_8(z, z, t2, p256_mod, p256_mp_mod);
  17068. /* X3 = R^2 - H^3 - 2*X1*H^2 */
  17069. sp_256_mont_sqr_8(t1, t4, p256_mod, p256_mp_mod);
  17070. sp_256_mont_sqr_8(t5, t2, p256_mod, p256_mp_mod);
  17071. sp_256_mont_mul_8(t3, x, t5, p256_mod, p256_mp_mod);
  17072. sp_256_mont_mul_8(t5, t5, t2, p256_mod, p256_mp_mod);
  17073. sp_256_mont_sub_8(x, t1, t5, p256_mod);
  17074. sp_256_mont_dbl_8(t1, t3, p256_mod);
  17075. sp_256_mont_sub_8(x, x, t1, p256_mod);
  17076. /* Y3 = R*(X1*H^2 - X3) - Y1*H^3 */
  17077. sp_256_mont_sub_8(t3, t3, x, p256_mod);
  17078. sp_256_mont_mul_8(t3, t3, t4, p256_mod, p256_mp_mod);
  17079. sp_256_mont_mul_8(t5, t5, y, p256_mod, p256_mp_mod);
  17080. sp_256_mont_sub_8(y, t3, t5, p256_mod);
  17081. }
  17082. }
  17083. #ifdef WOLFSSL_SP_SMALL
  17084. #ifdef FP_ECC
  17085. /* Convert the projective point to affine.
  17086. * Ordinates are in Montgomery form.
  17087. *
  17088. * a Point to convert.
  17089. * t Temporary data.
  17090. */
  17091. static void sp_256_proj_to_affine_8(sp_point* a, sp_digit* t)
  17092. {
  17093. sp_digit* t1 = t;
  17094. sp_digit* t2 = t + 2 * 8;
  17095. sp_digit* tmp = t + 4 * 8;
  17096. sp_256_mont_inv_8(t1, a->z, tmp);
  17097. sp_256_mont_sqr_8(t2, t1, p256_mod, p256_mp_mod);
  17098. sp_256_mont_mul_8(t1, t2, t1, p256_mod, p256_mp_mod);
  17099. sp_256_mont_mul_8(a->x, a->x, t2, p256_mod, p256_mp_mod);
  17100. sp_256_mont_mul_8(a->y, a->y, t1, p256_mod, p256_mp_mod);
  17101. XMEMCPY(a->z, p256_norm_mod, sizeof(p256_norm_mod));
  17102. }
  17103. /* Generate the pre-computed table of points for the base point.
  17104. *
  17105. * a The base point.
  17106. * table Place to store generated point data.
  17107. * tmp Temporary data.
  17108. * heap Heap to use for allocation.
  17109. */
  17110. static int sp_256_gen_stripe_table_8(const sp_point* a,
  17111. sp_table_entry* table, sp_digit* tmp, void* heap)
  17112. {
  17113. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  17114. sp_point td, s1d, s2d;
  17115. #endif
  17116. sp_point* t;
  17117. sp_point* s1 = NULL;
  17118. sp_point* s2 = NULL;
  17119. int i, j;
  17120. int err;
  17121. (void)heap;
  17122. err = sp_ecc_point_new(heap, td, t);
  17123. if (err == MP_OKAY) {
  17124. err = sp_ecc_point_new(heap, s1d, s1);
  17125. }
  17126. if (err == MP_OKAY) {
  17127. err = sp_ecc_point_new(heap, s2d, s2);
  17128. }
  17129. if (err == MP_OKAY) {
  17130. err = sp_256_mod_mul_norm_8(t->x, a->x, p256_mod);
  17131. }
  17132. if (err == MP_OKAY) {
  17133. err = sp_256_mod_mul_norm_8(t->y, a->y, p256_mod);
  17134. }
  17135. if (err == MP_OKAY) {
  17136. err = sp_256_mod_mul_norm_8(t->z, a->z, p256_mod);
  17137. }
  17138. if (err == MP_OKAY) {
  17139. t->infinity = 0;
  17140. sp_256_proj_to_affine_8(t, tmp);
  17141. XMEMCPY(s1->z, p256_norm_mod, sizeof(p256_norm_mod));
  17142. s1->infinity = 0;
  17143. XMEMCPY(s2->z, p256_norm_mod, sizeof(p256_norm_mod));
  17144. s2->infinity = 0;
  17145. /* table[0] = {0, 0, infinity} */
  17146. XMEMSET(&table[0], 0, sizeof(sp_table_entry));
  17147. /* table[1] = Affine version of 'a' in Montgomery form */
  17148. XMEMCPY(table[1].x, t->x, sizeof(table->x));
  17149. XMEMCPY(table[1].y, t->y, sizeof(table->y));
  17150. for (i=1; i<4; i++) {
  17151. sp_256_proj_point_dbl_n_8(t, t, 64, tmp);
  17152. sp_256_proj_to_affine_8(t, tmp);
  17153. XMEMCPY(table[1<<i].x, t->x, sizeof(table->x));
  17154. XMEMCPY(table[1<<i].y, t->y, sizeof(table->y));
  17155. }
  17156. for (i=1; i<4; i++) {
  17157. XMEMCPY(s1->x, table[1<<i].x, sizeof(table->x));
  17158. XMEMCPY(s1->y, table[1<<i].y, sizeof(table->y));
  17159. for (j=(1<<i)+1; j<(1<<(i+1)); j++) {
  17160. XMEMCPY(s2->x, table[j-(1<<i)].x, sizeof(table->x));
  17161. XMEMCPY(s2->y, table[j-(1<<i)].y, sizeof(table->y));
  17162. sp_256_proj_point_add_qz1_8(t, s1, s2, tmp);
  17163. sp_256_proj_to_affine_8(t, tmp);
  17164. XMEMCPY(table[j].x, t->x, sizeof(table->x));
  17165. XMEMCPY(table[j].y, t->y, sizeof(table->y));
  17166. }
  17167. }
  17168. }
  17169. sp_ecc_point_free(s2, 0, heap);
  17170. sp_ecc_point_free(s1, 0, heap);
  17171. sp_ecc_point_free( t, 0, heap);
  17172. return err;
  17173. }
  17174. #endif /* FP_ECC */
  17175. /* Multiply the point by the scalar and return the result.
  17176. * If map is true then convert result to affine coordinates.
  17177. *
  17178. * r Resulting point.
  17179. * k Scalar to multiply by.
  17180. * map Indicates whether to convert result to affine.
  17181. * heap Heap to use for allocation.
  17182. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17183. */
  17184. static int sp_256_ecc_mulmod_stripe_8(sp_point* r, const sp_point* g,
  17185. const sp_table_entry* table, const sp_digit* k, int map, void* heap)
  17186. {
  17187. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  17188. sp_point rtd;
  17189. sp_point pd;
  17190. sp_digit td[2 * 8 * 5];
  17191. #endif
  17192. sp_point* rt;
  17193. sp_point* p = NULL;
  17194. sp_digit* t;
  17195. int i, j;
  17196. int y, x;
  17197. int err;
  17198. (void)g;
  17199. (void)heap;
  17200. err = sp_ecc_point_new(heap, rtd, rt);
  17201. if (err == MP_OKAY) {
  17202. err = sp_ecc_point_new(heap, pd, p);
  17203. }
  17204. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17205. t = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 5, heap,
  17206. DYNAMIC_TYPE_ECC);
  17207. if (t == NULL) {
  17208. err = MEMORY_E;
  17209. }
  17210. #else
  17211. t = td;
  17212. #endif
  17213. if (err == MP_OKAY) {
  17214. XMEMCPY(p->z, p256_norm_mod, sizeof(p256_norm_mod));
  17215. XMEMCPY(rt->z, p256_norm_mod, sizeof(p256_norm_mod));
  17216. y = 0;
  17217. for (j=0,x=63; j<4; j++,x+=64) {
  17218. y |= ((k[x / 32] >> (x % 32)) & 1) << j;
  17219. }
  17220. XMEMCPY(rt->x, table[y].x, sizeof(table[y].x));
  17221. XMEMCPY(rt->y, table[y].y, sizeof(table[y].y));
  17222. rt->infinity = !y;
  17223. for (i=62; i>=0; i--) {
  17224. y = 0;
  17225. for (j=0,x=i; j<4; j++,x+=64) {
  17226. y |= ((k[x / 32] >> (x % 32)) & 1) << j;
  17227. }
  17228. sp_256_proj_point_dbl_8(rt, rt, t);
  17229. XMEMCPY(p->x, table[y].x, sizeof(table[y].x));
  17230. XMEMCPY(p->y, table[y].y, sizeof(table[y].y));
  17231. p->infinity = !y;
  17232. sp_256_proj_point_add_qz1_8(rt, rt, p, t);
  17233. }
  17234. if (map != 0) {
  17235. sp_256_map_8(r, rt, t);
  17236. }
  17237. else {
  17238. XMEMCPY(r, rt, sizeof(sp_point));
  17239. }
  17240. }
  17241. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17242. if (t != NULL) {
  17243. XFREE(t, heap, DYNAMIC_TYPE_ECC);
  17244. }
  17245. #endif
  17246. sp_ecc_point_free(p, 0, heap);
  17247. sp_ecc_point_free(rt, 0, heap);
  17248. return err;
  17249. }
  17250. #ifdef FP_ECC
  17251. #ifndef FP_ENTRIES
  17252. #define FP_ENTRIES 16
  17253. #endif
  17254. typedef struct sp_cache_t {
  17255. sp_digit x[8];
  17256. sp_digit y[8];
  17257. sp_table_entry table[16];
  17258. uint32_t cnt;
  17259. int set;
  17260. } sp_cache_t;
  17261. static THREAD_LS_T sp_cache_t sp_cache[FP_ENTRIES];
  17262. static THREAD_LS_T int sp_cache_last = -1;
  17263. static THREAD_LS_T int sp_cache_inited = 0;
  17264. #ifndef HAVE_THREAD_LS
  17265. static volatile int initCacheMutex = 0;
  17266. static wolfSSL_Mutex sp_cache_lock;
  17267. #endif
  17268. static void sp_ecc_get_cache(const sp_point* g, sp_cache_t** cache)
  17269. {
  17270. int i, j;
  17271. uint32_t least;
  17272. if (sp_cache_inited == 0) {
  17273. for (i=0; i<FP_ENTRIES; i++) {
  17274. sp_cache[i].set = 0;
  17275. }
  17276. sp_cache_inited = 1;
  17277. }
  17278. /* Compare point with those in cache. */
  17279. for (i=0; i<FP_ENTRIES; i++) {
  17280. if (!sp_cache[i].set)
  17281. continue;
  17282. if (sp_256_cmp_equal_8(g->x, sp_cache[i].x) &
  17283. sp_256_cmp_equal_8(g->y, sp_cache[i].y)) {
  17284. sp_cache[i].cnt++;
  17285. break;
  17286. }
  17287. }
  17288. /* No match. */
  17289. if (i == FP_ENTRIES) {
  17290. /* Find empty entry. */
  17291. i = (sp_cache_last + 1) % FP_ENTRIES;
  17292. for (; i != sp_cache_last; i=(i+1)%FP_ENTRIES) {
  17293. if (!sp_cache[i].set) {
  17294. break;
  17295. }
  17296. }
  17297. /* Evict least used. */
  17298. if (i == sp_cache_last) {
  17299. least = sp_cache[0].cnt;
  17300. for (j=1; j<FP_ENTRIES; j++) {
  17301. if (sp_cache[j].cnt < least) {
  17302. i = j;
  17303. least = sp_cache[i].cnt;
  17304. }
  17305. }
  17306. }
  17307. XMEMCPY(sp_cache[i].x, g->x, sizeof(sp_cache[i].x));
  17308. XMEMCPY(sp_cache[i].y, g->y, sizeof(sp_cache[i].y));
  17309. sp_cache[i].set = 1;
  17310. sp_cache[i].cnt = 1;
  17311. }
  17312. *cache = &sp_cache[i];
  17313. sp_cache_last = i;
  17314. }
  17315. #endif /* FP_ECC */
  17316. /* Multiply the base point of P256 by the scalar and return the result.
  17317. * If map is true then convert result to affine coordinates.
  17318. *
  17319. * r Resulting point.
  17320. * g Point to multiply.
  17321. * k Scalar to multiply by.
  17322. * map Indicates whether to convert result to affine.
  17323. * heap Heap to use for allocation.
  17324. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17325. */
  17326. static int sp_256_ecc_mulmod_8(sp_point* r, const sp_point* g, const sp_digit* k,
  17327. int map, void* heap)
  17328. {
  17329. #ifndef FP_ECC
  17330. return sp_256_ecc_mulmod_fast_8(r, g, k, map, heap);
  17331. #else
  17332. sp_digit tmp[2 * 8 * 5];
  17333. sp_cache_t* cache;
  17334. int err = MP_OKAY;
  17335. #ifndef HAVE_THREAD_LS
  17336. if (initCacheMutex == 0) {
  17337. wc_InitMutex(&sp_cache_lock);
  17338. initCacheMutex = 1;
  17339. }
  17340. if (wc_LockMutex(&sp_cache_lock) != 0)
  17341. err = BAD_MUTEX_E;
  17342. #endif /* HAVE_THREAD_LS */
  17343. if (err == MP_OKAY) {
  17344. sp_ecc_get_cache(g, &cache);
  17345. if (cache->cnt == 2)
  17346. sp_256_gen_stripe_table_8(g, cache->table, tmp, heap);
  17347. #ifndef HAVE_THREAD_LS
  17348. wc_UnLockMutex(&sp_cache_lock);
  17349. #endif /* HAVE_THREAD_LS */
  17350. if (cache->cnt < 2) {
  17351. err = sp_256_ecc_mulmod_fast_8(r, g, k, map, heap);
  17352. }
  17353. else {
  17354. err = sp_256_ecc_mulmod_stripe_8(r, g, cache->table, k,
  17355. map, heap);
  17356. }
  17357. }
  17358. return err;
  17359. #endif
  17360. }
  17361. #else
  17362. #ifdef FP_ECC
  17363. /* Generate the pre-computed table of points for the base point.
  17364. *
  17365. * a The base point.
  17366. * table Place to store generated point data.
  17367. * tmp Temporary data.
  17368. * heap Heap to use for allocation.
  17369. */
  17370. static int sp_256_gen_stripe_table_8(const sp_point* a,
  17371. sp_table_entry* table, sp_digit* tmp, void* heap)
  17372. {
  17373. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  17374. sp_point td, s1d, s2d;
  17375. #endif
  17376. sp_point* t;
  17377. sp_point* s1 = NULL;
  17378. sp_point* s2 = NULL;
  17379. int i, j;
  17380. int err;
  17381. (void)heap;
  17382. err = sp_ecc_point_new(heap, td, t);
  17383. if (err == MP_OKAY) {
  17384. err = sp_ecc_point_new(heap, s1d, s1);
  17385. }
  17386. if (err == MP_OKAY) {
  17387. err = sp_ecc_point_new(heap, s2d, s2);
  17388. }
  17389. if (err == MP_OKAY) {
  17390. err = sp_256_mod_mul_norm_8(t->x, a->x, p256_mod);
  17391. }
  17392. if (err == MP_OKAY) {
  17393. err = sp_256_mod_mul_norm_8(t->y, a->y, p256_mod);
  17394. }
  17395. if (err == MP_OKAY) {
  17396. err = sp_256_mod_mul_norm_8(t->z, a->z, p256_mod);
  17397. }
  17398. if (err == MP_OKAY) {
  17399. t->infinity = 0;
  17400. sp_256_proj_to_affine_8(t, tmp);
  17401. XMEMCPY(s1->z, p256_norm_mod, sizeof(p256_norm_mod));
  17402. s1->infinity = 0;
  17403. XMEMCPY(s2->z, p256_norm_mod, sizeof(p256_norm_mod));
  17404. s2->infinity = 0;
  17405. /* table[0] = {0, 0, infinity} */
  17406. XMEMSET(&table[0], 0, sizeof(sp_table_entry));
  17407. /* table[1] = Affine version of 'a' in Montgomery form */
  17408. XMEMCPY(table[1].x, t->x, sizeof(table->x));
  17409. XMEMCPY(table[1].y, t->y, sizeof(table->y));
  17410. for (i=1; i<8; i++) {
  17411. sp_256_proj_point_dbl_n_8(t, t, 32, tmp);
  17412. sp_256_proj_to_affine_8(t, tmp);
  17413. XMEMCPY(table[1<<i].x, t->x, sizeof(table->x));
  17414. XMEMCPY(table[1<<i].y, t->y, sizeof(table->y));
  17415. }
  17416. for (i=1; i<8; i++) {
  17417. XMEMCPY(s1->x, table[1<<i].x, sizeof(table->x));
  17418. XMEMCPY(s1->y, table[1<<i].y, sizeof(table->y));
  17419. for (j=(1<<i)+1; j<(1<<(i+1)); j++) {
  17420. XMEMCPY(s2->x, table[j-(1<<i)].x, sizeof(table->x));
  17421. XMEMCPY(s2->y, table[j-(1<<i)].y, sizeof(table->y));
  17422. sp_256_proj_point_add_qz1_8(t, s1, s2, tmp);
  17423. sp_256_proj_to_affine_8(t, tmp);
  17424. XMEMCPY(table[j].x, t->x, sizeof(table->x));
  17425. XMEMCPY(table[j].y, t->y, sizeof(table->y));
  17426. }
  17427. }
  17428. }
  17429. sp_ecc_point_free(s2, 0, heap);
  17430. sp_ecc_point_free(s1, 0, heap);
  17431. sp_ecc_point_free( t, 0, heap);
  17432. return err;
  17433. }
  17434. #endif /* FP_ECC */
  17435. /* Multiply the point by the scalar and return the result.
  17436. * If map is true then convert result to affine coordinates.
  17437. *
  17438. * r Resulting point.
  17439. * k Scalar to multiply by.
  17440. * map Indicates whether to convert result to affine.
  17441. * heap Heap to use for allocation.
  17442. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17443. */
  17444. static int sp_256_ecc_mulmod_stripe_8(sp_point* r, const sp_point* g,
  17445. const sp_table_entry* table, const sp_digit* k, int map, void* heap)
  17446. {
  17447. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  17448. sp_point rtd;
  17449. sp_point pd;
  17450. sp_digit td[2 * 8 * 5];
  17451. #endif
  17452. sp_point* rt;
  17453. sp_point* p = NULL;
  17454. sp_digit* t;
  17455. int i, j;
  17456. int y, x;
  17457. int err;
  17458. (void)g;
  17459. (void)heap;
  17460. err = sp_ecc_point_new(heap, rtd, rt);
  17461. if (err == MP_OKAY) {
  17462. err = sp_ecc_point_new(heap, pd, p);
  17463. }
  17464. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17465. t = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 5, heap,
  17466. DYNAMIC_TYPE_ECC);
  17467. if (t == NULL) {
  17468. err = MEMORY_E;
  17469. }
  17470. #else
  17471. t = td;
  17472. #endif
  17473. if (err == MP_OKAY) {
  17474. XMEMCPY(p->z, p256_norm_mod, sizeof(p256_norm_mod));
  17475. XMEMCPY(rt->z, p256_norm_mod, sizeof(p256_norm_mod));
  17476. y = 0;
  17477. for (j=0,x=31; j<8; j++,x+=32) {
  17478. y |= ((k[x / 32] >> (x % 32)) & 1) << j;
  17479. }
  17480. XMEMCPY(rt->x, table[y].x, sizeof(table[y].x));
  17481. XMEMCPY(rt->y, table[y].y, sizeof(table[y].y));
  17482. rt->infinity = !y;
  17483. for (i=30; i>=0; i--) {
  17484. y = 0;
  17485. for (j=0,x=i; j<8; j++,x+=32) {
  17486. y |= ((k[x / 32] >> (x % 32)) & 1) << j;
  17487. }
  17488. sp_256_proj_point_dbl_8(rt, rt, t);
  17489. XMEMCPY(p->x, table[y].x, sizeof(table[y].x));
  17490. XMEMCPY(p->y, table[y].y, sizeof(table[y].y));
  17491. p->infinity = !y;
  17492. sp_256_proj_point_add_qz1_8(rt, rt, p, t);
  17493. }
  17494. if (map != 0) {
  17495. sp_256_map_8(r, rt, t);
  17496. }
  17497. else {
  17498. XMEMCPY(r, rt, sizeof(sp_point));
  17499. }
  17500. }
  17501. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17502. if (t != NULL) {
  17503. XFREE(t, heap, DYNAMIC_TYPE_ECC);
  17504. }
  17505. #endif
  17506. sp_ecc_point_free(p, 0, heap);
  17507. sp_ecc_point_free(rt, 0, heap);
  17508. return err;
  17509. }
  17510. #ifdef FP_ECC
  17511. #ifndef FP_ENTRIES
  17512. #define FP_ENTRIES 16
  17513. #endif
  17514. typedef struct sp_cache_t {
  17515. sp_digit x[8];
  17516. sp_digit y[8];
  17517. sp_table_entry table[256];
  17518. uint32_t cnt;
  17519. int set;
  17520. } sp_cache_t;
  17521. static THREAD_LS_T sp_cache_t sp_cache[FP_ENTRIES];
  17522. static THREAD_LS_T int sp_cache_last = -1;
  17523. static THREAD_LS_T int sp_cache_inited = 0;
  17524. #ifndef HAVE_THREAD_LS
  17525. static volatile int initCacheMutex = 0;
  17526. static wolfSSL_Mutex sp_cache_lock;
  17527. #endif
  17528. static void sp_ecc_get_cache(const sp_point* g, sp_cache_t** cache)
  17529. {
  17530. int i, j;
  17531. uint32_t least;
  17532. if (sp_cache_inited == 0) {
  17533. for (i=0; i<FP_ENTRIES; i++) {
  17534. sp_cache[i].set = 0;
  17535. }
  17536. sp_cache_inited = 1;
  17537. }
  17538. /* Compare point with those in cache. */
  17539. for (i=0; i<FP_ENTRIES; i++) {
  17540. if (!sp_cache[i].set)
  17541. continue;
  17542. if (sp_256_cmp_equal_8(g->x, sp_cache[i].x) &
  17543. sp_256_cmp_equal_8(g->y, sp_cache[i].y)) {
  17544. sp_cache[i].cnt++;
  17545. break;
  17546. }
  17547. }
  17548. /* No match. */
  17549. if (i == FP_ENTRIES) {
  17550. /* Find empty entry. */
  17551. i = (sp_cache_last + 1) % FP_ENTRIES;
  17552. for (; i != sp_cache_last; i=(i+1)%FP_ENTRIES) {
  17553. if (!sp_cache[i].set) {
  17554. break;
  17555. }
  17556. }
  17557. /* Evict least used. */
  17558. if (i == sp_cache_last) {
  17559. least = sp_cache[0].cnt;
  17560. for (j=1; j<FP_ENTRIES; j++) {
  17561. if (sp_cache[j].cnt < least) {
  17562. i = j;
  17563. least = sp_cache[i].cnt;
  17564. }
  17565. }
  17566. }
  17567. XMEMCPY(sp_cache[i].x, g->x, sizeof(sp_cache[i].x));
  17568. XMEMCPY(sp_cache[i].y, g->y, sizeof(sp_cache[i].y));
  17569. sp_cache[i].set = 1;
  17570. sp_cache[i].cnt = 1;
  17571. }
  17572. *cache = &sp_cache[i];
  17573. sp_cache_last = i;
  17574. }
  17575. #endif /* FP_ECC */
  17576. /* Multiply the base point of P256 by the scalar and return the result.
  17577. * If map is true then convert result to affine coordinates.
  17578. *
  17579. * r Resulting point.
  17580. * g Point to multiply.
  17581. * k Scalar to multiply by.
  17582. * map Indicates whether to convert result to affine.
  17583. * heap Heap to use for allocation.
  17584. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17585. */
  17586. static int sp_256_ecc_mulmod_8(sp_point* r, const sp_point* g, const sp_digit* k,
  17587. int map, void* heap)
  17588. {
  17589. #ifndef FP_ECC
  17590. return sp_256_ecc_mulmod_fast_8(r, g, k, map, heap);
  17591. #else
  17592. sp_digit tmp[2 * 8 * 5];
  17593. sp_cache_t* cache;
  17594. int err = MP_OKAY;
  17595. #ifndef HAVE_THREAD_LS
  17596. if (initCacheMutex == 0) {
  17597. wc_InitMutex(&sp_cache_lock);
  17598. initCacheMutex = 1;
  17599. }
  17600. if (wc_LockMutex(&sp_cache_lock) != 0)
  17601. err = BAD_MUTEX_E;
  17602. #endif /* HAVE_THREAD_LS */
  17603. if (err == MP_OKAY) {
  17604. sp_ecc_get_cache(g, &cache);
  17605. if (cache->cnt == 2)
  17606. sp_256_gen_stripe_table_8(g, cache->table, tmp, heap);
  17607. #ifndef HAVE_THREAD_LS
  17608. wc_UnLockMutex(&sp_cache_lock);
  17609. #endif /* HAVE_THREAD_LS */
  17610. if (cache->cnt < 2) {
  17611. err = sp_256_ecc_mulmod_fast_8(r, g, k, map, heap);
  17612. }
  17613. else {
  17614. err = sp_256_ecc_mulmod_stripe_8(r, g, cache->table, k,
  17615. map, heap);
  17616. }
  17617. }
  17618. return err;
  17619. #endif
  17620. }
  17621. #endif /* WOLFSSL_SP_SMALL */
  17622. /* Multiply the point by the scalar and return the result.
  17623. * If map is true then convert result to affine coordinates.
  17624. *
  17625. * km Scalar to multiply by.
  17626. * p Point to multiply.
  17627. * r Resulting point.
  17628. * map Indicates whether to convert result to affine.
  17629. * heap Heap to use for allocation.
  17630. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17631. */
  17632. int sp_ecc_mulmod_256(mp_int* km, ecc_point* gm, ecc_point* r, int map,
  17633. void* heap)
  17634. {
  17635. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  17636. sp_point p;
  17637. sp_digit k[8];
  17638. #else
  17639. sp_digit* k = NULL;
  17640. #endif
  17641. sp_point* point;
  17642. int err;
  17643. err = sp_ecc_point_new(heap, p, point);
  17644. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17645. if (err == MP_OKAY) {
  17646. k = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8, heap,
  17647. DYNAMIC_TYPE_ECC);
  17648. if (k == NULL)
  17649. err = MEMORY_E;
  17650. }
  17651. #endif
  17652. if (err == MP_OKAY) {
  17653. sp_256_from_mp(k, 8, km);
  17654. sp_256_point_from_ecc_point_8(point, gm);
  17655. err = sp_256_ecc_mulmod_8(point, point, k, map, heap);
  17656. }
  17657. if (err == MP_OKAY) {
  17658. err = sp_256_point_to_ecc_point_8(point, r);
  17659. }
  17660. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  17661. if (k != NULL) {
  17662. XFREE(k, heap, DYNAMIC_TYPE_ECC);
  17663. }
  17664. #endif
  17665. sp_ecc_point_free(point, 0, heap);
  17666. return err;
  17667. }
  17668. #ifdef WOLFSSL_SP_SMALL
  17669. static const sp_table_entry p256_table[16] = {
  17670. /* 0 */
  17671. { { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 },
  17672. { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 } },
  17673. /* 1 */
  17674. { { 0x18a9143c,0x79e730d4,0x5fedb601,0x75ba95fc,0x77622510,0x79fb732b,
  17675. 0xa53755c6,0x18905f76 },
  17676. { 0xce95560a,0xddf25357,0xba19e45c,0x8b4ab8e4,0xdd21f325,0xd2e88688,
  17677. 0x25885d85,0x8571ff18 } },
  17678. /* 2 */
  17679. { { 0x16a0d2bb,0x4f922fc5,0x1a623499,0x0d5cc16c,0x57c62c8b,0x9241cf3a,
  17680. 0xfd1b667f,0x2f5e6961 },
  17681. { 0xf5a01797,0x5c15c70b,0x60956192,0x3d20b44d,0x071fdb52,0x04911b37,
  17682. 0x8d6f0f7b,0xf648f916 } },
  17683. /* 3 */
  17684. { { 0xe137bbbc,0x9e566847,0x8a6a0bec,0xe434469e,0x79d73463,0xb1c42761,
  17685. 0x133d0015,0x5abe0285 },
  17686. { 0xc04c7dab,0x92aa837c,0x43260c07,0x573d9f4c,0x78e6cc37,0x0c931562,
  17687. 0x6b6f7383,0x94bb725b } },
  17688. /* 4 */
  17689. { { 0xbfe20925,0x62a8c244,0x8fdce867,0x91c19ac3,0xdd387063,0x5a96a5d5,
  17690. 0x21d324f6,0x61d587d4 },
  17691. { 0xa37173ea,0xe87673a2,0x53778b65,0x23848008,0x05bab43e,0x10f8441e,
  17692. 0x4621efbe,0xfa11fe12 } },
  17693. /* 5 */
  17694. { { 0x2cb19ffd,0x1c891f2b,0xb1923c23,0x01ba8d5b,0x8ac5ca8e,0xb6d03d67,
  17695. 0x1f13bedc,0x586eb04c },
  17696. { 0x27e8ed09,0x0c35c6e5,0x1819ede2,0x1e81a33c,0x56c652fa,0x278fd6c0,
  17697. 0x70864f11,0x19d5ac08 } },
  17698. /* 6 */
  17699. { { 0xd2b533d5,0x62577734,0xa1bdddc0,0x673b8af6,0xa79ec293,0x577e7c9a,
  17700. 0xc3b266b1,0xbb6de651 },
  17701. { 0xb65259b3,0xe7e9303a,0xd03a7480,0xd6a0afd3,0x9b3cfc27,0xc5ac83d1,
  17702. 0x5d18b99b,0x60b4619a } },
  17703. /* 7 */
  17704. { { 0x1ae5aa1c,0xbd6a38e1,0x49e73658,0xb8b7652b,0xee5f87ed,0x0b130014,
  17705. 0xaeebffcd,0x9d0f27b2 },
  17706. { 0x7a730a55,0xca924631,0xddbbc83a,0x9c955b2f,0xac019a71,0x07c1dfe0,
  17707. 0x356ec48d,0x244a566d } },
  17708. /* 8 */
  17709. { { 0xf4f8b16a,0x56f8410e,0xc47b266a,0x97241afe,0x6d9c87c1,0x0a406b8e,
  17710. 0xcd42ab1b,0x803f3e02 },
  17711. { 0x04dbec69,0x7f0309a8,0x3bbad05f,0xa83b85f7,0xad8e197f,0xc6097273,
  17712. 0x5067adc1,0xc097440e } },
  17713. /* 9 */
  17714. { { 0xc379ab34,0x846a56f2,0x841df8d1,0xa8ee068b,0x176c68ef,0x20314459,
  17715. 0x915f1f30,0xf1af32d5 },
  17716. { 0x5d75bd50,0x99c37531,0xf72f67bc,0x837cffba,0x48d7723f,0x0613a418,
  17717. 0xe2d41c8b,0x23d0f130 } },
  17718. /* 10 */
  17719. { { 0xd5be5a2b,0xed93e225,0x5934f3c6,0x6fe79983,0x22626ffc,0x43140926,
  17720. 0x7990216a,0x50bbb4d9 },
  17721. { 0xe57ec63e,0x378191c6,0x181dcdb2,0x65422c40,0x0236e0f6,0x41a8099b,
  17722. 0x01fe49c3,0x2b100118 } },
  17723. /* 11 */
  17724. { { 0x9b391593,0xfc68b5c5,0x598270fc,0xc385f5a2,0xd19adcbb,0x7144f3aa,
  17725. 0x83fbae0c,0xdd558999 },
  17726. { 0x74b82ff4,0x93b88b8e,0x71e734c9,0xd2e03c40,0x43c0322a,0x9a7a9eaf,
  17727. 0x149d6041,0xe6e4c551 } },
  17728. /* 12 */
  17729. { { 0x80ec21fe,0x5fe14bfe,0xc255be82,0xf6ce116a,0x2f4a5d67,0x98bc5a07,
  17730. 0xdb7e63af,0xfad27148 },
  17731. { 0x29ab05b3,0x90c0b6ac,0x4e251ae6,0x37a9a83c,0xc2aade7d,0x0a7dc875,
  17732. 0x9f0e1a84,0x77387de3 } },
  17733. /* 13 */
  17734. { { 0xa56c0dd7,0x1e9ecc49,0x46086c74,0xa5cffcd8,0xf505aece,0x8f7a1408,
  17735. 0xbef0c47e,0xb37b85c0 },
  17736. { 0xcc0e6a8f,0x3596b6e4,0x6b388f23,0xfd6d4bbf,0xc39cef4e,0xaba453fa,
  17737. 0xf9f628d5,0x9c135ac8 } },
  17738. /* 14 */
  17739. { { 0x95c8f8be,0x0a1c7294,0x3bf362bf,0x2961c480,0xdf63d4ac,0x9e418403,
  17740. 0x91ece900,0xc109f9cb },
  17741. { 0x58945705,0xc2d095d0,0xddeb85c0,0xb9083d96,0x7a40449b,0x84692b8d,
  17742. 0x2eee1ee1,0x9bc3344f } },
  17743. /* 15 */
  17744. { { 0x42913074,0x0d5ae356,0x48a542b1,0x55491b27,0xb310732a,0x469ca665,
  17745. 0x5f1a4cc1,0x29591d52 },
  17746. { 0xb84f983f,0xe76f5b6b,0x9f5f84e1,0xbe7eef41,0x80baa189,0x1200d496,
  17747. 0x18ef332c,0x6376551f } },
  17748. };
  17749. /* Multiply the base point of P256 by the scalar and return the result.
  17750. * If map is true then convert result to affine coordinates.
  17751. *
  17752. * r Resulting point.
  17753. * k Scalar to multiply by.
  17754. * map Indicates whether to convert result to affine.
  17755. * heap Heap to use for allocation.
  17756. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  17757. */
  17758. static int sp_256_ecc_mulmod_base_8(sp_point* r, const sp_digit* k,
  17759. int map, void* heap)
  17760. {
  17761. return sp_256_ecc_mulmod_stripe_8(r, &p256_base, p256_table,
  17762. k, map, heap);
  17763. }
  17764. #else
  17765. static const sp_table_entry p256_table[256] = {
  17766. /* 0 */
  17767. { { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 },
  17768. { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 } },
  17769. /* 1 */
  17770. { { 0x18a9143c,0x79e730d4,0x5fedb601,0x75ba95fc,0x77622510,0x79fb732b,
  17771. 0xa53755c6,0x18905f76 },
  17772. { 0xce95560a,0xddf25357,0xba19e45c,0x8b4ab8e4,0xdd21f325,0xd2e88688,
  17773. 0x25885d85,0x8571ff18 } },
  17774. /* 2 */
  17775. { { 0x4147519a,0x20288602,0x26b372f0,0xd0981eac,0xa785ebc8,0xa9d4a7ca,
  17776. 0xdbdf58e9,0xd953c50d },
  17777. { 0xfd590f8f,0x9d6361cc,0x44e6c917,0x72e9626b,0x22eb64cf,0x7fd96110,
  17778. 0x9eb288f3,0x863ebb7e } },
  17779. /* 3 */
  17780. { { 0x5cdb6485,0x7856b623,0x2f0a2f97,0x808f0ea2,0x4f7e300b,0x3e68d954,
  17781. 0xb5ff80a0,0x00076055 },
  17782. { 0x838d2010,0x7634eb9b,0x3243708a,0x54014fbb,0x842a6606,0xe0e47d39,
  17783. 0x34373ee0,0x83087761 } },
  17784. /* 4 */
  17785. { { 0x16a0d2bb,0x4f922fc5,0x1a623499,0x0d5cc16c,0x57c62c8b,0x9241cf3a,
  17786. 0xfd1b667f,0x2f5e6961 },
  17787. { 0xf5a01797,0x5c15c70b,0x60956192,0x3d20b44d,0x071fdb52,0x04911b37,
  17788. 0x8d6f0f7b,0xf648f916 } },
  17789. /* 5 */
  17790. { { 0xe137bbbc,0x9e566847,0x8a6a0bec,0xe434469e,0x79d73463,0xb1c42761,
  17791. 0x133d0015,0x5abe0285 },
  17792. { 0xc04c7dab,0x92aa837c,0x43260c07,0x573d9f4c,0x78e6cc37,0x0c931562,
  17793. 0x6b6f7383,0x94bb725b } },
  17794. /* 6 */
  17795. { { 0x720f141c,0xbbf9b48f,0x2df5bc74,0x6199b3cd,0x411045c4,0xdc3f6129,
  17796. 0x2f7dc4ef,0xcdd6bbcb },
  17797. { 0xeaf436fd,0xcca6700b,0xb99326be,0x6f647f6d,0x014f2522,0x0c0fa792,
  17798. 0x4bdae5f6,0xa361bebd } },
  17799. /* 7 */
  17800. { { 0x597c13c7,0x28aa2558,0x50b7c3e1,0xc38d635f,0xf3c09d1d,0x07039aec,
  17801. 0xc4b5292c,0xba12ca09 },
  17802. { 0x59f91dfd,0x9e408fa4,0xceea07fb,0x3af43b66,0x9d780b29,0x1eceb089,
  17803. 0x701fef4b,0x53ebb99d } },
  17804. /* 8 */
  17805. { { 0xb0e63d34,0x4fe7ee31,0xa9e54fab,0xf4600572,0xd5e7b5a4,0xc0493334,
  17806. 0x06d54831,0x8589fb92 },
  17807. { 0x6583553a,0xaa70f5cc,0xe25649e5,0x0879094a,0x10044652,0xcc904507,
  17808. 0x02541c4f,0xebb0696d } },
  17809. /* 9 */
  17810. { { 0xac1647c5,0x4616ca15,0xc4cf5799,0xb8127d47,0x764dfbac,0xdc666aa3,
  17811. 0xd1b27da3,0xeb2820cb },
  17812. { 0x6a87e008,0x9406f8d8,0x922378f3,0xd87dfa9d,0x80ccecb2,0x56ed2e42,
  17813. 0x55a7da1d,0x1f28289b } },
  17814. /* 10 */
  17815. { { 0x3b89da99,0xabbaa0c0,0xb8284022,0xa6f2d79e,0xb81c05e8,0x27847862,
  17816. 0x05e54d63,0x337a4b59 },
  17817. { 0x21f7794a,0x3c67500d,0x7d6d7f61,0x207005b7,0x04cfd6e8,0x0a5a3781,
  17818. 0xf4c2fbd6,0x0d65e0d5 } },
  17819. /* 11 */
  17820. { { 0xb5275d38,0xd9d09bbe,0x0be0a358,0x4268a745,0x973eb265,0xf0762ff4,
  17821. 0x52f4a232,0xc23da242 },
  17822. { 0x0b94520c,0x5da1b84f,0xb05bd78e,0x09666763,0x94d29ea1,0x3a4dcb86,
  17823. 0xc790cff1,0x19de3b8c } },
  17824. /* 12 */
  17825. { { 0x26c5fe04,0x183a716c,0x3bba1bdb,0x3b28de0b,0xa4cb712c,0x7432c586,
  17826. 0x91fccbfd,0xe34dcbd4 },
  17827. { 0xaaa58403,0xb408d46b,0x82e97a53,0x9a697486,0x36aaa8af,0x9e390127,
  17828. 0x7b4e0f7f,0xe7641f44 } },
  17829. /* 13 */
  17830. { { 0xdf64ba59,0x7d753941,0x0b0242fc,0xd33f10ec,0xa1581859,0x4f06dfc6,
  17831. 0x052a57bf,0x4a12df57 },
  17832. { 0x9439dbd0,0xbfa6338f,0xbde53e1f,0xd3c24bd4,0x21f1b314,0xfd5e4ffa,
  17833. 0xbb5bea46,0x6af5aa93 } },
  17834. /* 14 */
  17835. { { 0x10c91999,0xda10b699,0x2a580491,0x0a24b440,0xb8cc2090,0x3e0094b4,
  17836. 0x66a44013,0x5fe3475a },
  17837. { 0xf93e7b4b,0xb0f8cabd,0x7c23f91a,0x292b501a,0xcd1e6263,0x42e889ae,
  17838. 0xecfea916,0xb544e308 } },
  17839. /* 15 */
  17840. { { 0x16ddfdce,0x6478c6e9,0xf89179e6,0x2c329166,0x4d4e67e1,0x4e8d6e76,
  17841. 0xa6b0c20b,0xe0b6b2bd },
  17842. { 0xbb7efb57,0x0d312df2,0x790c4007,0x1aac0dde,0x679bc944,0xf90336ad,
  17843. 0x25a63774,0x71c023de } },
  17844. /* 16 */
  17845. { { 0xbfe20925,0x62a8c244,0x8fdce867,0x91c19ac3,0xdd387063,0x5a96a5d5,
  17846. 0x21d324f6,0x61d587d4 },
  17847. { 0xa37173ea,0xe87673a2,0x53778b65,0x23848008,0x05bab43e,0x10f8441e,
  17848. 0x4621efbe,0xfa11fe12 } },
  17849. /* 17 */
  17850. { { 0x2cb19ffd,0x1c891f2b,0xb1923c23,0x01ba8d5b,0x8ac5ca8e,0xb6d03d67,
  17851. 0x1f13bedc,0x586eb04c },
  17852. { 0x27e8ed09,0x0c35c6e5,0x1819ede2,0x1e81a33c,0x56c652fa,0x278fd6c0,
  17853. 0x70864f11,0x19d5ac08 } },
  17854. /* 18 */
  17855. { { 0x309a4e1f,0x1e99f581,0xe9270074,0xab7de71b,0xefd28d20,0x26a5ef0b,
  17856. 0x7f9c563f,0xe7c0073f },
  17857. { 0x0ef59f76,0x1f6d663a,0x20fcb050,0x669b3b54,0x7a6602d4,0xc08c1f7a,
  17858. 0xc65b3c0a,0xe08504fe } },
  17859. /* 19 */
  17860. { { 0xa031b3ca,0xf098f68d,0xe6da6d66,0x6d1cab9e,0x94f246e8,0x5bfd81fa,
  17861. 0x5b0996b4,0x78f01882 },
  17862. { 0x3a25787f,0xb7eefde4,0x1dccac9b,0x8016f80d,0xb35bfc36,0x0cea4877,
  17863. 0x7e94747a,0x43a773b8 } },
  17864. /* 20 */
  17865. { { 0xd2b533d5,0x62577734,0xa1bdddc0,0x673b8af6,0xa79ec293,0x577e7c9a,
  17866. 0xc3b266b1,0xbb6de651 },
  17867. { 0xb65259b3,0xe7e9303a,0xd03a7480,0xd6a0afd3,0x9b3cfc27,0xc5ac83d1,
  17868. 0x5d18b99b,0x60b4619a } },
  17869. /* 21 */
  17870. { { 0x1ae5aa1c,0xbd6a38e1,0x49e73658,0xb8b7652b,0xee5f87ed,0x0b130014,
  17871. 0xaeebffcd,0x9d0f27b2 },
  17872. { 0x7a730a55,0xca924631,0xddbbc83a,0x9c955b2f,0xac019a71,0x07c1dfe0,
  17873. 0x356ec48d,0x244a566d } },
  17874. /* 22 */
  17875. { { 0xeacf1f96,0x6db0394a,0x024c271c,0x9f2122a9,0x82cbd3b9,0x2626ac1b,
  17876. 0x3581ef69,0x45e58c87 },
  17877. { 0xa38f9dbc,0xd3ff479d,0xe888a040,0xa8aaf146,0x46e0bed7,0x945adfb2,
  17878. 0xc1e4b7a4,0xc040e21c } },
  17879. /* 23 */
  17880. { { 0x6f8117b6,0x847af000,0x73a35433,0x651969ff,0x1d9475eb,0x482b3576,
  17881. 0x682c6ec7,0x1cdf5c97 },
  17882. { 0x11f04839,0x7db775b4,0x48de1698,0x7dbeacf4,0xb70b3219,0xb2921dd1,
  17883. 0xa92dff3d,0x046755f8 } },
  17884. /* 24 */
  17885. { { 0xbce8ffcd,0xcc8ac5d2,0x2fe61a82,0x0d53c48b,0x7202d6c7,0xf6f16172,
  17886. 0x3b83a5f3,0x046e5e11 },
  17887. { 0xd8007f01,0xe7b8ff64,0x5af43183,0x7fb1ef12,0x35e1a03c,0x045c5ea6,
  17888. 0x303d005b,0x6e0106c3 } },
  17889. /* 25 */
  17890. { { 0x88dd73b1,0x48c73584,0x995ed0d9,0x7670708f,0xc56a2ab7,0x38385ea8,
  17891. 0xe901cf1f,0x442594ed },
  17892. { 0x12d4b65b,0xf8faa2c9,0x96c90c37,0x94c2343b,0x5e978d1f,0xd326e4a1,
  17893. 0x4c2ee68e,0xa796fa51 } },
  17894. /* 26 */
  17895. { { 0x823addd7,0x359fb604,0xe56693b3,0x9e2a6183,0x3cbf3c80,0xf885b78e,
  17896. 0xc69766e9,0xe4ad2da9 },
  17897. { 0x8e048a61,0x357f7f42,0xc092d9a0,0x082d198c,0xc03ed8ef,0xfc3a1af4,
  17898. 0xc37b5143,0xc5e94046 } },
  17899. /* 27 */
  17900. { { 0x2be75f9e,0x476a538c,0xcb123a78,0x6fd1a9e8,0xb109c04b,0xd85e4df0,
  17901. 0xdb464747,0x63283daf },
  17902. { 0xbaf2df15,0xce728cf7,0x0ad9a7f4,0xe592c455,0xe834bcc3,0xfab226ad,
  17903. 0x1981a938,0x68bd19ab } },
  17904. /* 28 */
  17905. { { 0x1887d659,0xc08ead51,0xb359305a,0x3374d5f4,0xcfe74fe3,0x96986981,
  17906. 0x3c6fdfd6,0x495292f5 },
  17907. { 0x1acec896,0x4a878c9e,0xec5b4484,0xd964b210,0x664d60a7,0x6696f7e2,
  17908. 0x26036837,0x0ec7530d } },
  17909. /* 29 */
  17910. { { 0xad2687bb,0x2da13a05,0xf32e21fa,0xa1f83b6a,0x1dd4607b,0x390f5ef5,
  17911. 0x64863f0b,0x0f6207a6 },
  17912. { 0x0f138233,0xbd67e3bb,0x272aa718,0xdd66b96c,0x26ec88ae,0x8ed00407,
  17913. 0x08ed6dcf,0xff0db072 } },
  17914. /* 30 */
  17915. { { 0x4c95d553,0x749fa101,0x5d680a8a,0xa44052fd,0xff3b566f,0x183b4317,
  17916. 0x88740ea3,0x313b513c },
  17917. { 0x08d11549,0xb402e2ac,0xb4dee21c,0x071ee10b,0x47f2320e,0x26b987dd,
  17918. 0x86f19f81,0x2d3abcf9 } },
  17919. /* 31 */
  17920. { { 0x815581a2,0x4c288501,0x632211af,0x9a0a6d56,0x0cab2e99,0x19ba7a0f,
  17921. 0xded98cdf,0xc036fa10 },
  17922. { 0xc1fbd009,0x29ae08ba,0x06d15816,0x0b68b190,0x9b9e0d8f,0xc2eb3277,
  17923. 0xb6d40194,0xa6b2a2c4 } },
  17924. /* 32 */
  17925. { { 0x6d3549cf,0xd433e50f,0xfacd665e,0x6f33696f,0xce11fcb4,0x695bfdac,
  17926. 0xaf7c9860,0x810ee252 },
  17927. { 0x7159bb2c,0x65450fe1,0x758b357b,0xf7dfbebe,0xd69fea72,0x2b057e74,
  17928. 0x92731745,0xd485717a } },
  17929. /* 33 */
  17930. { { 0xf0cb5a98,0x11741a8a,0x1f3110bf,0xd3da8f93,0xab382adf,0x1994e2cb,
  17931. 0x2f9a604e,0x6a6045a7 },
  17932. { 0xa2b2411d,0x170c0d3f,0x510e96e0,0xbe0eb83e,0x8865b3cc,0x3bcc9f73,
  17933. 0xf9e15790,0xd3e45cfa } },
  17934. /* 34 */
  17935. { { 0xe83f7669,0xce1f69bb,0x72877d6b,0x09f8ae82,0x3244278d,0x9548ae54,
  17936. 0xe3c2c19c,0x207755de },
  17937. { 0x6fef1945,0x87bd61d9,0xb12d28c3,0x18813cef,0x72df64aa,0x9fbcd1d6,
  17938. 0x7154b00d,0x48dc5ee5 } },
  17939. /* 35 */
  17940. { { 0xf7e5a199,0x123790bf,0x989ccbb7,0xe0efb8cf,0x0a519c79,0xc27a2bfe,
  17941. 0xdff6f445,0xf2fb0aed },
  17942. { 0xf0b5025f,0x41c09575,0x40fa9f22,0x550543d7,0x380bfbd0,0x8fa3c8ad,
  17943. 0xdb28d525,0xa13e9015 } },
  17944. /* 36 */
  17945. { { 0xa2b65cbc,0xf9f7a350,0x2a464226,0x0b04b972,0xe23f07a1,0x265ce241,
  17946. 0x1497526f,0x2bf0d6b0 },
  17947. { 0x4b216fb7,0xd3d4dd3f,0xfbdda26a,0xf7d7b867,0x6708505c,0xaeb7b83f,
  17948. 0x162fe89f,0x42a94a5a } },
  17949. /* 37 */
  17950. { { 0xeaadf191,0x5846ad0b,0x25a268d7,0x0f8a4890,0x494dc1f6,0xe8603050,
  17951. 0xc65ede3d,0x2c2dd969 },
  17952. { 0x93849c17,0x6d02171d,0x1da250dd,0x460488ba,0x3c3a5485,0x4810c706,
  17953. 0x42c56dbc,0xf437fa1f } },
  17954. /* 38 */
  17955. { { 0x4a0f7dab,0x6aa0d714,0x1776e9ac,0x0f049793,0xf5f39786,0x52c0a050,
  17956. 0x54707aa8,0xaaf45b33 },
  17957. { 0xc18d364a,0x85e37c33,0x3e497165,0xd40b9b06,0x15ec5444,0xf4171681,
  17958. 0xf4f272bc,0xcdf6310d } },
  17959. /* 39 */
  17960. { { 0x8ea8b7ef,0x7473c623,0x85bc2287,0x08e93518,0x2bda8e34,0x41956772,
  17961. 0xda9e2ff2,0xf0d008ba },
  17962. { 0x2414d3b1,0x2912671d,0xb019ea76,0xb3754985,0x453bcbdb,0x5c61b96d,
  17963. 0xca887b8b,0x5bd5c2f5 } },
  17964. /* 40 */
  17965. { { 0xf49a3154,0xef0f469e,0x6e2b2e9a,0x3e85a595,0xaa924a9c,0x45aaec1e,
  17966. 0xa09e4719,0xaa12dfc8 },
  17967. { 0x4df69f1d,0x26f27227,0xa2ff5e73,0xe0e4c82c,0xb7a9dd44,0xb9d8ce73,
  17968. 0xe48ca901,0x6c036e73 } },
  17969. /* 41 */
  17970. { { 0x0f6e3138,0x5cfae12a,0x25ad345a,0x6966ef00,0x45672bc5,0x8993c64b,
  17971. 0x96afbe24,0x292ff658 },
  17972. { 0x5e213402,0xd5250d44,0x4392c9fe,0xf6580e27,0xda1c72e8,0x097b397f,
  17973. 0x311b7276,0x644e0c90 } },
  17974. /* 42 */
  17975. { { 0xa47153f0,0xe1e421e1,0x920418c9,0xb86c3b79,0x705d7672,0x93bdce87,
  17976. 0xcab79a77,0xf25ae793 },
  17977. { 0x6d869d0c,0x1f3194a3,0x4986c264,0x9d55c882,0x096e945e,0x49fb5ea3,
  17978. 0x13db0a3e,0x39b8e653 } },
  17979. /* 43 */
  17980. { { 0xb6fd2e59,0x37754200,0x9255c98f,0x35e2c066,0x0e2a5739,0xd9dab21a,
  17981. 0x0f19db06,0x39122f2f },
  17982. { 0x03cad53c,0xcfbce1e0,0xe65c17e3,0x225b2c0f,0x9aa13877,0x72baf1d2,
  17983. 0xce80ff8d,0x8de80af8 } },
  17984. /* 44 */
  17985. { { 0x207bbb76,0xafbea8d9,0x21782758,0x921c7e7c,0x1c0436b1,0xdfa2b74b,
  17986. 0x2e368c04,0x87194906 },
  17987. { 0xa3993df5,0xb5f928bb,0xf3b3d26a,0x639d75b5,0x85b55050,0x011aa78a,
  17988. 0x5b74fde1,0xfc315e6a } },
  17989. /* 45 */
  17990. { { 0xe8d6ecfa,0x561fd41a,0x1aec7f86,0x5f8c44f6,0x4924741d,0x98452a7b,
  17991. 0xee389088,0xe6d4a7ad },
  17992. { 0x4593c75d,0x60552ed1,0xdd271162,0x70a70da4,0x7ba2c7db,0xd2aede93,
  17993. 0x9be2ae57,0x35dfaf9a } },
  17994. /* 46 */
  17995. { { 0xaa736636,0x6b956fcd,0xae2cab7e,0x09f51d97,0x0f349966,0xfb10bf41,
  17996. 0x1c830d2b,0x1da5c7d7 },
  17997. { 0x3cce6825,0x5c41e483,0xf9573c3b,0x15ad118f,0xf23036b8,0xa28552c7,
  17998. 0xdbf4b9d6,0x7077c0fd } },
  17999. /* 47 */
  18000. { { 0x46b9661c,0xbf63ff8d,0x0d2cfd71,0xa1dfd36b,0xa847f8f7,0x0373e140,
  18001. 0xe50efe44,0x53a8632e },
  18002. { 0x696d8051,0x0976ff68,0xc74f468a,0xdaec0c95,0x5e4e26bd,0x62994dc3,
  18003. 0x34e1fcc1,0x028ca76d } },
  18004. /* 48 */
  18005. { { 0xfc9877ee,0xd11d47dc,0x801d0002,0xc8b36210,0x54c260b6,0xd002c117,
  18006. 0x6962f046,0x04c17cd8 },
  18007. { 0xb0daddf5,0x6d9bd094,0x24ce55c0,0xbea23575,0x72da03b5,0x663356e6,
  18008. 0xfed97474,0xf7ba4de9 } },
  18009. /* 49 */
  18010. { { 0xebe1263f,0xd0dbfa34,0x71ae7ce6,0x55763735,0x82a6f523,0xd2440553,
  18011. 0x52131c41,0xe31f9600 },
  18012. { 0xea6b6ec6,0xd1bb9216,0x73c2fc44,0x37a1d12e,0x89d0a294,0xc10e7eac,
  18013. 0xce34d47b,0xaa3a6259 } },
  18014. /* 50 */
  18015. { { 0x36f3dcd3,0xfbcf9df5,0xd2bf7360,0x6ceded50,0xdf504f5b,0x491710fa,
  18016. 0x7e79daee,0x2398dd62 },
  18017. { 0x6d09569e,0xcf4705a3,0x5149f769,0xea0619bb,0x35f6034c,0xff9c0377,
  18018. 0x1c046210,0x5717f5b2 } },
  18019. /* 51 */
  18020. { { 0x21dd895e,0x9fe229c9,0x40c28451,0x8e518500,0x1d637ecd,0xfa13d239,
  18021. 0x0e3c28de,0x660a2c56 },
  18022. { 0xd67fcbd0,0x9cca88ae,0x0ea9f096,0xc8472478,0x72e92b4d,0x32b2f481,
  18023. 0x4f522453,0x624ee54c } },
  18024. /* 52 */
  18025. { { 0xd897eccc,0x09549ce4,0x3f9880aa,0x4d49d1d9,0x043a7c20,0x723c2423,
  18026. 0x92bdfbc0,0x4f392afb },
  18027. { 0x7de44fd9,0x6969f8fa,0x57b32156,0xb66cfbe4,0x368ebc3c,0xdb2fa803,
  18028. 0xccdb399c,0x8a3e7977 } },
  18029. /* 53 */
  18030. { { 0x06c4b125,0xdde1881f,0xf6e3ca8c,0xae34e300,0x5c7a13e9,0xef6999de,
  18031. 0x70c24404,0x3888d023 },
  18032. { 0x44f91081,0x76280356,0x5f015504,0x3d9fcf61,0x632cd36e,0x1827edc8,
  18033. 0x18102336,0xa5e62e47 } },
  18034. /* 54 */
  18035. { { 0x2facd6c8,0x1a825ee3,0x54bcbc66,0x699c6354,0x98df9931,0x0ce3edf7,
  18036. 0x466a5adc,0x2c4768e6 },
  18037. { 0x90a64bc9,0xb346ff8c,0xe4779f5c,0x630a6020,0xbc05e884,0xd949d064,
  18038. 0xf9e652a0,0x7b5e6441 } },
  18039. /* 55 */
  18040. { { 0x1d28444a,0x2169422c,0xbe136a39,0xe996c5d8,0xfb0c7fce,0x2387afe5,
  18041. 0x0c8d744a,0xb8af73cb },
  18042. { 0x338b86fd,0x5fde83aa,0xa58a5cff,0xfee3f158,0x20ac9433,0xc9ee8f6f,
  18043. 0x7f3f0895,0xa036395f } },
  18044. /* 56 */
  18045. { { 0xa10f7770,0x8c73c6bb,0xa12a0e24,0xa6f16d81,0x51bc2b9f,0x100df682,
  18046. 0x875fb533,0x4be36b01 },
  18047. { 0x9fb56dbb,0x9226086e,0x07e7a4f8,0x306fef8b,0x66d52f20,0xeeaccc05,
  18048. 0x1bdc00c0,0x8cbc9a87 } },
  18049. /* 57 */
  18050. { { 0xc0dac4ab,0xe131895c,0x712ff112,0xa874a440,0x6a1cee57,0x6332ae7c,
  18051. 0x0c0835f8,0x44e7553e },
  18052. { 0x7734002d,0x6d503fff,0x0b34425c,0x9d35cb8b,0x0e8738b5,0x95f70276,
  18053. 0x5eb8fc18,0x470a683a } },
  18054. /* 58 */
  18055. { { 0x90513482,0x81b761dc,0x01e9276a,0x0287202a,0x0ce73083,0xcda441ee,
  18056. 0xc63dc6ef,0x16410690 },
  18057. { 0x6d06a2ed,0xf5034a06,0x189b100b,0xdd4d7745,0xab8218c9,0xd914ae72,
  18058. 0x7abcbb4f,0xd73479fd } },
  18059. /* 59 */
  18060. { { 0x5ad4c6e5,0x7edefb16,0x5b06d04d,0x262cf08f,0x8575cb14,0x12ed5bb1,
  18061. 0x0771666b,0x816469e3 },
  18062. { 0x561e291e,0xd7ab9d79,0xc1de1661,0xeb9daf22,0x135e0513,0xf49827eb,
  18063. 0xf0dd3f9c,0x0a36dd23 } },
  18064. /* 60 */
  18065. { { 0x41d5533c,0x098d32c7,0x8684628f,0x7c5f5a9e,0xe349bd11,0x39a228ad,
  18066. 0xfdbab118,0xe331dfd6 },
  18067. { 0x6bcc6ed8,0x5100ab68,0xef7a260e,0x7160c3bd,0xbce850d7,0x9063d9a7,
  18068. 0x492e3389,0xd3b4782a } },
  18069. /* 61 */
  18070. { { 0xf3821f90,0xa149b6e8,0x66eb7aad,0x92edd9ed,0x1a013116,0x0bb66953,
  18071. 0x4c86a5bd,0x7281275a },
  18072. { 0xd3ff47e5,0x503858f7,0x61016441,0x5e1616bc,0x7dfd9bb1,0x62b0f11a,
  18073. 0xce145059,0x2c062e7e } },
  18074. /* 62 */
  18075. { { 0x0159ac2e,0xa76f996f,0xcbdb2713,0x281e7736,0x08e46047,0x2ad6d288,
  18076. 0x2c4e7ef1,0x282a35f9 },
  18077. { 0xc0ce5cd2,0x9c354b1e,0x1379c229,0xcf99efc9,0x3e82c11e,0x992caf38,
  18078. 0x554d2abd,0xc71cd513 } },
  18079. /* 63 */
  18080. { { 0x09b578f4,0x4885de9c,0xe3affa7a,0x1884e258,0x59182f1f,0x8f76b1b7,
  18081. 0xcf47f3a3,0xc50f6740 },
  18082. { 0x374b68ea,0xa9c4adf3,0x69965fe2,0xa406f323,0x85a53050,0x2f86a222,
  18083. 0x212958dc,0xb9ecb3a7 } },
  18084. /* 64 */
  18085. { { 0xf4f8b16a,0x56f8410e,0xc47b266a,0x97241afe,0x6d9c87c1,0x0a406b8e,
  18086. 0xcd42ab1b,0x803f3e02 },
  18087. { 0x04dbec69,0x7f0309a8,0x3bbad05f,0xa83b85f7,0xad8e197f,0xc6097273,
  18088. 0x5067adc1,0xc097440e } },
  18089. /* 65 */
  18090. { { 0xc379ab34,0x846a56f2,0x841df8d1,0xa8ee068b,0x176c68ef,0x20314459,
  18091. 0x915f1f30,0xf1af32d5 },
  18092. { 0x5d75bd50,0x99c37531,0xf72f67bc,0x837cffba,0x48d7723f,0x0613a418,
  18093. 0xe2d41c8b,0x23d0f130 } },
  18094. /* 66 */
  18095. { { 0xf41500d9,0x857ab6ed,0xfcbeada8,0x0d890ae5,0x89725951,0x52fe8648,
  18096. 0xc0a3fadd,0xb0288dd6 },
  18097. { 0x650bcb08,0x85320f30,0x695d6e16,0x71af6313,0xb989aa76,0x31f520a7,
  18098. 0xf408c8d2,0xffd3724f } },
  18099. /* 67 */
  18100. { { 0xb458e6cb,0x53968e64,0x317a5d28,0x992dad20,0x7aa75f56,0x3814ae0b,
  18101. 0xd78c26df,0xf5590f4a },
  18102. { 0xcf0ba55a,0x0fc24bd3,0x0c778bae,0x0fc4724a,0x683b674a,0x1ce9864f,
  18103. 0xf6f74a20,0x18d6da54 } },
  18104. /* 68 */
  18105. { { 0xd5be5a2b,0xed93e225,0x5934f3c6,0x6fe79983,0x22626ffc,0x43140926,
  18106. 0x7990216a,0x50bbb4d9 },
  18107. { 0xe57ec63e,0x378191c6,0x181dcdb2,0x65422c40,0x0236e0f6,0x41a8099b,
  18108. 0x01fe49c3,0x2b100118 } },
  18109. /* 69 */
  18110. { { 0x9b391593,0xfc68b5c5,0x598270fc,0xc385f5a2,0xd19adcbb,0x7144f3aa,
  18111. 0x83fbae0c,0xdd558999 },
  18112. { 0x74b82ff4,0x93b88b8e,0x71e734c9,0xd2e03c40,0x43c0322a,0x9a7a9eaf,
  18113. 0x149d6041,0xe6e4c551 } },
  18114. /* 70 */
  18115. { { 0x1e9af288,0x55f655bb,0xf7ada931,0x647e1a64,0xcb2820e5,0x43697e4b,
  18116. 0x07ed56ff,0x51e00db1 },
  18117. { 0x771c327e,0x43d169b8,0x4a96c2ad,0x29cdb20b,0x3deb4779,0xc07d51f5,
  18118. 0x49829177,0xe22f4241 } },
  18119. /* 71 */
  18120. { { 0x635f1abb,0xcd45e8f4,0x68538874,0x7edc0cb5,0xb5a8034d,0xc9472c1f,
  18121. 0x52dc48c9,0xf709373d },
  18122. { 0xa8af30d6,0x401966bb,0xf137b69c,0x95bf5f4a,0x9361c47e,0x3966162a,
  18123. 0xe7275b11,0xbd52d288 } },
  18124. /* 72 */
  18125. { { 0x9c5fa877,0xab155c7a,0x7d3a3d48,0x17dad672,0x73d189d8,0x43f43f9e,
  18126. 0xc8aa77a6,0xa0d0f8e4 },
  18127. { 0xcc94f92d,0x0bbeafd8,0x0c4ddb3a,0xd818c8be,0xb82eba14,0x22cc65f8,
  18128. 0x946d6a00,0xa56c78c7 } },
  18129. /* 73 */
  18130. { { 0x0dd09529,0x2962391b,0x3daddfcf,0x803e0ea6,0x5b5bf481,0x2c77351f,
  18131. 0x731a367a,0xd8befdf8 },
  18132. { 0xfc0157f4,0xab919d42,0xfec8e650,0xf51caed7,0x02d48b0a,0xcdf9cb40,
  18133. 0xce9f6478,0x854a68a5 } },
  18134. /* 74 */
  18135. { { 0x63506ea5,0xdc35f67b,0xa4fe0d66,0x9286c489,0xfe95cd4d,0x3f101d3b,
  18136. 0x98846a95,0x5cacea0b },
  18137. { 0x9ceac44d,0xa90df60c,0x354d1c3a,0x3db29af4,0xad5dbabe,0x08dd3de8,
  18138. 0x35e4efa9,0xe4982d12 } },
  18139. /* 75 */
  18140. { { 0xc34cd55e,0x23104a22,0x2680d132,0x58695bb3,0x1fa1d943,0xfb345afa,
  18141. 0x16b20499,0x8046b7f6 },
  18142. { 0x38e7d098,0xb533581e,0xf46f0b70,0xd7f61e8d,0x44cb78c4,0x30dea9ea,
  18143. 0x9082af55,0xeb17ca7b } },
  18144. /* 76 */
  18145. { { 0x76a145b9,0x1751b598,0xc1bc71ec,0xa5cf6b0f,0x392715bb,0xd3e03565,
  18146. 0xfab5e131,0x097b00ba },
  18147. { 0x565f69e1,0xaa66c8e9,0xb5be5199,0x77e8f75a,0xda4fd984,0x6033ba11,
  18148. 0xafdbcc9e,0xf95c747b } },
  18149. /* 77 */
  18150. { { 0xbebae45e,0x558f01d3,0xc4bc6955,0xa8ebe9f0,0xdbc64fc6,0xaeb705b1,
  18151. 0x566ed837,0x3512601e },
  18152. { 0xfa1161cd,0x9336f1e1,0x4c65ef87,0x328ab8d5,0x724f21e5,0x4757eee2,
  18153. 0x6068ab6b,0x0ef97123 } },
  18154. /* 78 */
  18155. { { 0x54ca4226,0x02598cf7,0xf8642c8e,0x5eede138,0x468e1790,0x48963f74,
  18156. 0x3b4fbc95,0xfc16d933 },
  18157. { 0xe7c800ca,0xbe96fb31,0x2678adaa,0x13806331,0x6ff3e8b5,0x3d624497,
  18158. 0xb95d7a17,0x14ca4af1 } },
  18159. /* 79 */
  18160. { { 0xbd2f81d5,0x7a4771ba,0x01f7d196,0x1a5f9d69,0xcad9c907,0xd898bef7,
  18161. 0xf59c231d,0x4057b063 },
  18162. { 0x89c05c0a,0xbffd82fe,0x1dc0df85,0xe4911c6f,0xa35a16db,0x3befccae,
  18163. 0xf1330b13,0x1c3b5d64 } },
  18164. /* 80 */
  18165. { { 0x80ec21fe,0x5fe14bfe,0xc255be82,0xf6ce116a,0x2f4a5d67,0x98bc5a07,
  18166. 0xdb7e63af,0xfad27148 },
  18167. { 0x29ab05b3,0x90c0b6ac,0x4e251ae6,0x37a9a83c,0xc2aade7d,0x0a7dc875,
  18168. 0x9f0e1a84,0x77387de3 } },
  18169. /* 81 */
  18170. { { 0xa56c0dd7,0x1e9ecc49,0x46086c74,0xa5cffcd8,0xf505aece,0x8f7a1408,
  18171. 0xbef0c47e,0xb37b85c0 },
  18172. { 0xcc0e6a8f,0x3596b6e4,0x6b388f23,0xfd6d4bbf,0xc39cef4e,0xaba453fa,
  18173. 0xf9f628d5,0x9c135ac8 } },
  18174. /* 82 */
  18175. { { 0x84e35743,0x32aa3202,0x85a3cdef,0x320d6ab1,0x1df19819,0xb821b176,
  18176. 0xc433851f,0x5721361f },
  18177. { 0x71fc9168,0x1f0db36a,0x5e5c403c,0x5f98ba73,0x37bcd8f5,0xf64ca87e,
  18178. 0xe6bb11bd,0xdcbac3c9 } },
  18179. /* 83 */
  18180. { { 0x4518cbe2,0xf01d9968,0x9c9eb04e,0xd242fc18,0xe47feebf,0x727663c7,
  18181. 0x2d626862,0xb8c1c89e },
  18182. { 0xc8e1d569,0x51a58bdd,0xb7d88cd0,0x563809c8,0xf11f31eb,0x26c27fd9,
  18183. 0x2f9422d4,0x5d23bbda } },
  18184. /* 84 */
  18185. { { 0x95c8f8be,0x0a1c7294,0x3bf362bf,0x2961c480,0xdf63d4ac,0x9e418403,
  18186. 0x91ece900,0xc109f9cb },
  18187. { 0x58945705,0xc2d095d0,0xddeb85c0,0xb9083d96,0x7a40449b,0x84692b8d,
  18188. 0x2eee1ee1,0x9bc3344f } },
  18189. /* 85 */
  18190. { { 0x42913074,0x0d5ae356,0x48a542b1,0x55491b27,0xb310732a,0x469ca665,
  18191. 0x5f1a4cc1,0x29591d52 },
  18192. { 0xb84f983f,0xe76f5b6b,0x9f5f84e1,0xbe7eef41,0x80baa189,0x1200d496,
  18193. 0x18ef332c,0x6376551f } },
  18194. /* 86 */
  18195. { { 0x562976cc,0xbda5f14e,0x0ef12c38,0x22bca3e6,0x6cca9852,0xbbfa3064,
  18196. 0x08e2987a,0xbdb79dc8 },
  18197. { 0xcb06a772,0xfd2cb5c9,0xfe536dce,0x38f475aa,0x7c2b5db8,0xc2a3e022,
  18198. 0xadd3c14a,0x8ee86001 } },
  18199. /* 87 */
  18200. { { 0xa4ade873,0xcbe96981,0xc4fba48c,0x7ee9aa4d,0x5a054ba5,0x2cee2899,
  18201. 0x6f77aa4b,0x92e51d7a },
  18202. { 0x7190a34d,0x948bafa8,0xf6bd1ed1,0xd698f75b,0x0caf1144,0xd00ee6e3,
  18203. 0x0a56aaaa,0x5182f86f } },
  18204. /* 88 */
  18205. { { 0x7a4cc99c,0xfba6212c,0x3e6d9ca1,0xff609b68,0x5ac98c5a,0x5dbb27cb,
  18206. 0x4073a6f2,0x91dcab5d },
  18207. { 0x5f575a70,0x01b6cc3d,0x6f8d87fa,0x0cb36139,0x89981736,0x165d4e8c,
  18208. 0x97974f2b,0x17a0cedb } },
  18209. /* 89 */
  18210. { { 0x076c8d3a,0x38861e2a,0x210f924b,0x701aad39,0x13a835d9,0x94d0eae4,
  18211. 0x7f4cdf41,0x2e8ce36c },
  18212. { 0x037a862b,0x91273dab,0x60e4c8fa,0x01ba9bb7,0x33baf2dd,0xf9645388,
  18213. 0x34f668f3,0xf4ccc6cb } },
  18214. /* 90 */
  18215. { { 0xf1f79687,0x44ef525c,0x92efa815,0x7c595495,0xa5c78d29,0xe1231741,
  18216. 0x9a0df3c9,0xac0db488 },
  18217. { 0xdf01747f,0x86bfc711,0xef17df13,0x592b9358,0x5ccb6bb5,0xe5880e4f,
  18218. 0x94c974a2,0x95a64a61 } },
  18219. /* 91 */
  18220. { { 0xc15a4c93,0x72c1efda,0x82585141,0x40269b73,0x16cb0bad,0x6a8dfb1c,
  18221. 0x29210677,0x231e54ba },
  18222. { 0x8ae6d2dc,0xa70df917,0x39112918,0x4d6aa63f,0x5e5b7223,0xf627726b,
  18223. 0xd8a731e1,0xab0be032 } },
  18224. /* 92 */
  18225. { { 0x8d131f2d,0x097ad0e9,0x3b04f101,0x637f09e3,0xd5e9a748,0x1ac86196,
  18226. 0x2cf6a679,0xf1bcc880 },
  18227. { 0xe8daacb4,0x25c69140,0x60f65009,0x3c4e4055,0x477937a6,0x591cc8fc,
  18228. 0x5aebb271,0x85169469 } },
  18229. /* 93 */
  18230. { { 0xf1dcf593,0xde35c143,0xb018be3b,0x78202b29,0x9bdd9d3d,0xe9cdadc2,
  18231. 0xdaad55d8,0x8f67d9d2 },
  18232. { 0x7481ea5f,0x84111656,0xe34c590c,0xe7d2dde9,0x05053fa8,0xffdd43f4,
  18233. 0xc0728b5d,0xf84572b9 } },
  18234. /* 94 */
  18235. { { 0x97af71c9,0x5e1a7a71,0x7a736565,0xa1449444,0x0e1d5063,0xa1b4ae07,
  18236. 0x616b2c19,0xedee2710 },
  18237. { 0x11734121,0xb2f034f5,0x4a25e9f0,0x1cac6e55,0xa40c2ecf,0x8dc148f3,
  18238. 0x44ebd7f4,0x9fd27e9b } },
  18239. /* 95 */
  18240. { { 0xf6e2cb16,0x3cc7658a,0xfe5919b6,0xe3eb7d2c,0x168d5583,0x5a8c5816,
  18241. 0x958ff387,0xa40c2fb6 },
  18242. { 0xfedcc158,0x8c9ec560,0x55f23056,0x7ad804c6,0x9a307e12,0xd9396704,
  18243. 0x7dc6decf,0x99bc9bb8 } },
  18244. /* 96 */
  18245. { { 0x927dafc6,0x84a9521d,0x5c09cd19,0x52c1fb69,0xf9366dde,0x9d9581a0,
  18246. 0xa16d7e64,0x9abe210b },
  18247. { 0x48915220,0x480af84a,0x4dd816c6,0xfa73176a,0x1681ca5a,0xc7d53987,
  18248. 0x87f344b0,0x7881c257 } },
  18249. /* 97 */
  18250. { { 0xe0bcf3ff,0x93399b51,0x127f74f6,0x0d02cbc5,0xdd01d968,0x8fb465a2,
  18251. 0xa30e8940,0x15e6e319 },
  18252. { 0x3e0e05f4,0x646d6e0d,0x43588404,0xfad7bddc,0xc4f850d3,0xbe61c7d1,
  18253. 0x191172ce,0x0e55facf } },
  18254. /* 98 */
  18255. { { 0xf8787564,0x7e9d9806,0x31e85ce6,0x1a331721,0xb819e8d6,0x6b0158ca,
  18256. 0x6fe96577,0xd73d0976 },
  18257. { 0x1eb7206e,0x42483425,0xc618bb42,0xa519290f,0x5e30a520,0x5dcbb859,
  18258. 0x8f15a50b,0x9250a374 } },
  18259. /* 99 */
  18260. { { 0xbe577410,0xcaff08f8,0x5077a8c6,0xfd408a03,0xec0a63a4,0xf1f63289,
  18261. 0xc1cc8c0b,0x77414082 },
  18262. { 0xeb0991cd,0x05a40fa6,0x49fdc296,0xc1ca0866,0xb324fd40,0x3a68a3c7,
  18263. 0x12eb20b9,0x8cb04f4d } },
  18264. /* 100 */
  18265. { { 0x6906171c,0xb1c2d055,0xb0240c3f,0x9073e9cd,0xd8906841,0xdb8e6b4f,
  18266. 0x47123b51,0xe4e429ef },
  18267. { 0x38ec36f4,0x0b8dd53c,0xff4b6a27,0xf9d2dc01,0x879a9a48,0x5d066e07,
  18268. 0x3c6e6552,0x37bca2ff } },
  18269. /* 101 */
  18270. { { 0xdf562470,0x4cd2e3c7,0xc0964ac9,0x44f272a2,0x80c793be,0x7c6d5df9,
  18271. 0x3002b22a,0x59913edc },
  18272. { 0x5750592a,0x7a139a83,0xe783de02,0x99e01d80,0xea05d64f,0xcf8c0375,
  18273. 0xb013e226,0x43786e4a } },
  18274. /* 102 */
  18275. { { 0x9e56b5a6,0xff32b0ed,0xd9fc68f9,0x0750d9a6,0x597846a7,0xec15e845,
  18276. 0xb7e79e7a,0x8638ca98 },
  18277. { 0x0afc24b2,0x2f5ae096,0x4dace8f2,0x05398eaf,0xaecba78f,0x3b765dd0,
  18278. 0x7b3aa6f0,0x1ecdd36a } },
  18279. /* 103 */
  18280. { { 0x6c5ff2f3,0x5d3acd62,0x2873a978,0xa2d516c0,0xd2110d54,0xad94c9fa,
  18281. 0xd459f32d,0xd85d0f85 },
  18282. { 0x10b11da3,0x9f700b8d,0xa78318c4,0xd2c22c30,0x9208decd,0x556988f4,
  18283. 0xb4ed3c62,0xa04f19c3 } },
  18284. /* 104 */
  18285. { { 0xed7f93bd,0x087924c8,0x392f51f6,0xcb64ac5d,0x821b71af,0x7cae330a,
  18286. 0x5c0950b0,0x92b2eeea },
  18287. { 0x85b6e235,0x85ac4c94,0x2936c0f0,0xab2ca4a9,0xe0508891,0x80faa6b3,
  18288. 0x5834276c,0x1ee78221 } },
  18289. /* 105 */
  18290. { { 0xe63e79f7,0xa60a2e00,0xf399d906,0xf590e7b2,0x6607c09d,0x9021054a,
  18291. 0x57a6e150,0xf3f2ced8 },
  18292. { 0xf10d9b55,0x200510f3,0xd8642648,0x9d2fcfac,0xe8bd0e7c,0xe5631aa7,
  18293. 0x3da3e210,0x0f56a454 } },
  18294. /* 106 */
  18295. { { 0x1043e0df,0x5b21bffa,0x9c007e6d,0x6c74b6cc,0xd4a8517a,0x1a656ec0,
  18296. 0x1969e263,0xbd8f1741 },
  18297. { 0xbeb7494a,0x8a9bbb86,0x45f3b838,0x1567d46f,0xa4e5a79a,0xdf7a12a7,
  18298. 0x30ccfa09,0x2d1a1c35 } },
  18299. /* 107 */
  18300. { { 0x506508da,0x192e3813,0xa1d795a7,0x336180c4,0x7a9944b3,0xcddb5949,
  18301. 0xb91fba46,0xa107a65e },
  18302. { 0x0f94d639,0xe6d1d1c5,0x8a58b7d7,0x8b4af375,0xbd37ca1c,0x1a7c5584,
  18303. 0xf87a9af2,0x183d760a } },
  18304. /* 108 */
  18305. { { 0x0dde59a4,0x29d69711,0x0e8bef87,0xf1ad8d07,0x4f2ebe78,0x229b4963,
  18306. 0xc269d754,0x1d44179d },
  18307. { 0x8390d30e,0xb32dc0cf,0x0de8110c,0x0a3b2753,0x2bc0339a,0x31af1dc5,
  18308. 0x9606d262,0x771f9cc2 } },
  18309. /* 109 */
  18310. { { 0x85040739,0x99993e77,0x8026a939,0x44539db9,0xf5f8fc26,0xcf40f6f2,
  18311. 0x0362718e,0x64427a31 },
  18312. { 0x85428aa8,0x4f4f2d87,0xebfb49a8,0x7b7adc3f,0xf23d01ac,0x201b2c6d,
  18313. 0x6ae90d6d,0x49d9b749 } },
  18314. /* 110 */
  18315. { { 0x435d1099,0xcc78d8bc,0x8e8d1a08,0x2adbcd4e,0x2cb68a41,0x02c2e2a0,
  18316. 0x3f605445,0x9037d81b },
  18317. { 0x074c7b61,0x7cdbac27,0x57bfd72e,0xfe2031ab,0x596d5352,0x61ccec96,
  18318. 0x7cc0639c,0x08c3de6a } },
  18319. /* 111 */
  18320. { { 0xf6d552ab,0x20fdd020,0x05cd81f1,0x56baff98,0x91351291,0x06fb7c3e,
  18321. 0x45796b2f,0xc6909442 },
  18322. { 0x41231bd1,0x17b3ae9c,0x5cc58205,0x1eac6e87,0xf9d6a122,0x208837ab,
  18323. 0xcafe3ac0,0x3fa3db02 } },
  18324. /* 112 */
  18325. { { 0x05058880,0xd75a3e65,0x643943f2,0x7da365ef,0xfab24925,0x4147861c,
  18326. 0xfdb808ff,0xc5c4bdb0 },
  18327. { 0xb272b56b,0x73513e34,0x11b9043a,0xc8327e95,0xf8844969,0xfd8ce37d,
  18328. 0x46c2b6b5,0x2d56db94 } },
  18329. /* 113 */
  18330. { { 0xff46ac6b,0x2461782f,0x07a2e425,0xd19f7926,0x09a48de1,0xfafea3c4,
  18331. 0xe503ba42,0x0f56bd9d },
  18332. { 0x345cda49,0x137d4ed1,0x816f299d,0x821158fc,0xaeb43402,0xe7c6a54a,
  18333. 0x1173b5f1,0x4003bb9d } },
  18334. /* 114 */
  18335. { { 0xa0803387,0x3b8e8189,0x39cbd404,0xece115f5,0xd2877f21,0x4297208d,
  18336. 0xa07f2f9e,0x53765522 },
  18337. { 0xa8a4182d,0xa4980a21,0x3219df79,0xa2bbd07a,0x1a19a2d4,0x674d0a2e,
  18338. 0x6c5d4549,0x7a056f58 } },
  18339. /* 115 */
  18340. { { 0x9d8a2a47,0x646b2558,0xc3df2773,0x5b582948,0xabf0d539,0x51ec000e,
  18341. 0x7a1a2675,0x77d482f1 },
  18342. { 0x87853948,0xb8a1bd95,0x6cfbffee,0xa6f817bd,0x80681e47,0xab6ec057,
  18343. 0x2b38b0e4,0x4115012b } },
  18344. /* 116 */
  18345. { { 0x6de28ced,0x3c73f0f4,0x9b13ec47,0x1d5da760,0x6e5c6392,0x61b8ce9e,
  18346. 0xfbea0946,0xcdf04572 },
  18347. { 0x6c53c3b0,0x1cb3c58b,0x447b843c,0x97fe3c10,0x2cb9780e,0xfb2b8ae1,
  18348. 0x97383109,0xee703dda } },
  18349. /* 117 */
  18350. { { 0xff57e43a,0x34515140,0xb1b811b8,0xd44660d3,0x8f42b986,0x2b3b5dff,
  18351. 0xa162ce21,0x2a0ad89d },
  18352. { 0x6bc277ba,0x64e4a694,0xc141c276,0xc788c954,0xcabf6274,0x141aa64c,
  18353. 0xac2b4659,0xd62d0b67 } },
  18354. /* 118 */
  18355. { { 0x2c054ac4,0x39c5d87b,0xf27df788,0x57005859,0xb18128d6,0xedf7cbf3,
  18356. 0x991c2426,0xb39a23f2 },
  18357. { 0xf0b16ae5,0x95284a15,0xa136f51b,0x0c6a05b1,0xf2700783,0x1d63c137,
  18358. 0xc0674cc5,0x04ed0092 } },
  18359. /* 119 */
  18360. { { 0x9ae90393,0x1f4185d1,0x4a3d64e6,0x3047b429,0x9854fc14,0xae0001a6,
  18361. 0x0177c387,0xa0a91fc1 },
  18362. { 0xae2c831e,0xff0a3f01,0x2b727e16,0xbb76ae82,0x5a3075b4,0x8f12c8a1,
  18363. 0x9ed20c41,0x084cf988 } },
  18364. /* 120 */
  18365. { { 0xfca6becf,0xd98509de,0x7dffb328,0x2fceae80,0x4778e8b9,0x5d8a15c4,
  18366. 0x73abf77e,0xd57955b2 },
  18367. { 0x31b5d4f1,0x210da79e,0x3cfa7a1c,0xaa52f04b,0xdc27c20b,0xd4d12089,
  18368. 0x02d141f1,0x8e14ea42 } },
  18369. /* 121 */
  18370. { { 0xf2897042,0xeed50345,0x43402c4a,0x8d05331f,0xc8bdfb21,0xc8d9c194,
  18371. 0x2aa4d158,0x597e1a37 },
  18372. { 0xcf0bd68c,0x0327ec1a,0xab024945,0x6d4be0dc,0xc9fe3e84,0x5b9c8d7a,
  18373. 0x199b4dea,0xca3f0236 } },
  18374. /* 122 */
  18375. { { 0x6170bd20,0x592a10b5,0x6d3f5de7,0x0ea897f1,0x44b2ade2,0xa3363ff1,
  18376. 0x309c07e4,0xbde7fd7e },
  18377. { 0xb8f5432c,0x516bb6d2,0xe043444b,0x210dc1cb,0xf8f95b5a,0x3db01e6f,
  18378. 0x0a7dd198,0xb623ad0e } },
  18379. /* 123 */
  18380. { { 0x60c7b65b,0xa75bd675,0x23a4a289,0xab8c5590,0xd7b26795,0xf8220fd0,
  18381. 0x58ec137b,0xd6aa2e46 },
  18382. { 0x5138bb85,0x10abc00b,0xd833a95c,0x8c31d121,0x1702a32e,0xb24ff00b,
  18383. 0x2dcc513a,0x111662e0 } },
  18384. /* 124 */
  18385. { { 0xefb42b87,0x78114015,0x1b6c4dff,0xbd9f5d70,0xa7d7c129,0x66ecccd7,
  18386. 0x94b750f8,0xdb3ee1cb },
  18387. { 0xf34837cf,0xb26f3db0,0xb9578d4f,0xe7eed18b,0x7c56657d,0x5d2cdf93,
  18388. 0x52206a59,0x886a6442 } },
  18389. /* 125 */
  18390. { { 0x65b569ea,0x3c234cfb,0xf72119c1,0x20011141,0xa15a619e,0x8badc85d,
  18391. 0x018a17bc,0xa70cf4eb },
  18392. { 0x8c4a6a65,0x224f97ae,0x0134378f,0x36e5cf27,0x4f7e0960,0xbe3a609e,
  18393. 0xd1747b77,0xaa4772ab } },
  18394. /* 126 */
  18395. { { 0x7aa60cc0,0x67676131,0x0368115f,0xc7916361,0xbbc1bb5a,0xded98bb4,
  18396. 0x30faf974,0x611a6ddc },
  18397. { 0xc15ee47a,0x30e78cbc,0x4e0d96a5,0x2e896282,0x3dd9ed88,0x36f35adf,
  18398. 0x16429c88,0x5cfffaf8 } },
  18399. /* 127 */
  18400. { { 0x9b7a99cd,0xc0d54cff,0x843c45a1,0x7bf3b99d,0x62c739e1,0x038a908f,
  18401. 0x7dc1994c,0x6e5a6b23 },
  18402. { 0x0ba5db77,0xef8b454e,0xacf60d63,0xb7b8807f,0x76608378,0xe591c0c6,
  18403. 0x242dabcc,0x481a238d } },
  18404. /* 128 */
  18405. { { 0x35d0b34a,0xe3417bc0,0x8327c0a7,0x440b386b,0xac0362d1,0x8fb7262d,
  18406. 0xe0cdf943,0x2c41114c },
  18407. { 0xad95a0b1,0x2ba5cef1,0x67d54362,0xc09b37a8,0x01e486c9,0x26d6cdd2,
  18408. 0x42ff9297,0x20477abf } },
  18409. /* 129 */
  18410. { { 0x18d65dbf,0x2f75173c,0x339edad8,0x77bf940e,0xdcf1001c,0x7022d26b,
  18411. 0xc77396b6,0xac66409a },
  18412. { 0xc6261cc3,0x8b0bb36f,0x190e7e90,0x213f7bc9,0xa45e6c10,0x6541ceba,
  18413. 0xcc122f85,0xce8e6975 } },
  18414. /* 130 */
  18415. { { 0xbc0a67d2,0x0f121b41,0x444d248a,0x62d4760a,0x659b4737,0x0e044f1d,
  18416. 0x250bb4a8,0x08fde365 },
  18417. { 0x848bf287,0xaceec3da,0xd3369d6e,0xc2a62182,0x92449482,0x3582dfdc,
  18418. 0x565d6cd7,0x2f7e2fd2 } },
  18419. /* 131 */
  18420. { { 0xc3770fa7,0xae4b92db,0x379043f9,0x095e8d5c,0x17761171,0x54f34e9d,
  18421. 0x907702ae,0xc65be92e },
  18422. { 0xf6fd0a40,0x2758a303,0xbcce784b,0xe7d822e3,0x4f9767bf,0x7ae4f585,
  18423. 0xd1193b3a,0x4bff8e47 } },
  18424. /* 132 */
  18425. { { 0x00ff1480,0xcd41d21f,0x0754db16,0x2ab8fb7d,0xbbe0f3ea,0xac81d2ef,
  18426. 0x5772967d,0x3e4e4ae6 },
  18427. { 0x3c5303e6,0x7e18f36d,0x92262397,0x3bd9994b,0x1324c3c0,0x9ed70e26,
  18428. 0x58ec6028,0x5388aefd } },
  18429. /* 133 */
  18430. { { 0x5e5d7713,0xad1317eb,0x75de49da,0x09b985ee,0xc74fb261,0x32f5bc4f,
  18431. 0x4f75be0e,0x5cf908d1 },
  18432. { 0x8e657b12,0x76043510,0xb96ed9e6,0xbfd421a5,0x8970ccc2,0x0e29f51f,
  18433. 0x60f00ce2,0xa698ba40 } },
  18434. /* 134 */
  18435. { { 0xef748fec,0x73db1686,0x7e9d2cf9,0xe6e755a2,0xce265eff,0x630b6544,
  18436. 0x7aebad8d,0xb142ef8a },
  18437. { 0x17d5770a,0xad31af9f,0x2cb3412f,0x66af3b67,0xdf3359de,0x6bd60d1b,
  18438. 0x58515075,0xd1896a96 } },
  18439. /* 135 */
  18440. { { 0x33c41c08,0xec5957ab,0x5468e2e1,0x87de94ac,0xac472f6c,0x18816b73,
  18441. 0x7981da39,0x267b0e0b },
  18442. { 0x8e62b988,0x6e554e5d,0x116d21e7,0xd8ddc755,0x3d2a6f99,0x4610faf0,
  18443. 0xa1119393,0xb54e287a } },
  18444. /* 136 */
  18445. { { 0x178a876b,0x0a0122b5,0x085104b4,0x51ff96ff,0x14f29f76,0x050b31ab,
  18446. 0x5f87d4e6,0x84abb28b },
  18447. { 0x8270790a,0xd5ed439f,0x85e3f46b,0x2d6cb59d,0x6c1e2212,0x75f55c1b,
  18448. 0x17655640,0xe5436f67 } },
  18449. /* 137 */
  18450. { { 0x2286e8d5,0x53f9025e,0x864453be,0x353c95b4,0xe408e3a0,0xd832f5bd,
  18451. 0x5b9ce99e,0x0404f68b },
  18452. { 0xa781e8e5,0xcad33bde,0x163c2f5b,0x3cdf5018,0x0119caa3,0x57576960,
  18453. 0x0ac1c701,0x3a4263df } },
  18454. /* 138 */
  18455. { { 0x9aeb596d,0xc2965ecc,0x023c92b4,0x01ea03e7,0x2e013961,0x4704b4b6,
  18456. 0x905ea367,0x0ca8fd3f },
  18457. { 0x551b2b61,0x92523a42,0x390fcd06,0x1eb7a89c,0x0392a63e,0xe7f1d2be,
  18458. 0x4ddb0c33,0x96dca264 } },
  18459. /* 139 */
  18460. { { 0x387510af,0x203bb43a,0xa9a36a01,0x846feaa8,0x2f950378,0xd23a5770,
  18461. 0x3aad59dc,0x4363e212 },
  18462. { 0x40246a47,0xca43a1c7,0xe55dd24d,0xb362b8d2,0x5d8faf96,0xf9b08604,
  18463. 0xd8bb98c4,0x840e115c } },
  18464. /* 140 */
  18465. { { 0x1023e8a7,0xf12205e2,0xd8dc7a0b,0xc808a8cd,0x163a5ddf,0xe292a272,
  18466. 0x30ded6d4,0x5e0d6abd },
  18467. { 0x7cfc0f64,0x07a721c2,0x0e55ed88,0x42eec01d,0x1d1f9db2,0x26a7bef9,
  18468. 0x2945a25a,0x7dea48f4 } },
  18469. /* 141 */
  18470. { { 0xe5060a81,0xabdf6f1c,0xf8f95615,0xe79f9c72,0x06ac268b,0xcfd36c54,
  18471. 0xebfd16d1,0xabc2a2be },
  18472. { 0xd3e2eac7,0x8ac66f91,0xd2dd0466,0x6f10ba63,0x0282d31b,0x6790e377,
  18473. 0x6c7eefc1,0x4ea35394 } },
  18474. /* 142 */
  18475. { { 0x5266309d,0xed8a2f8d,0x81945a3e,0x0a51c6c0,0x578c5dc1,0xcecaf45a,
  18476. 0x1c94ffc3,0x3a76e689 },
  18477. { 0x7d7b0d0f,0x9aace8a4,0x8f584a5f,0x963ace96,0x4e697fbe,0x51a30c72,
  18478. 0x465e6464,0x8212a10a } },
  18479. /* 143 */
  18480. { { 0xcfab8caa,0xef7c61c3,0x0e142390,0x18eb8e84,0x7e9733ca,0xcd1dff67,
  18481. 0x599cb164,0xaa7cab71 },
  18482. { 0xbc837bd1,0x02fc9273,0xc36af5d7,0xc06407d0,0xf423da49,0x17621292,
  18483. 0xfe0617c3,0x40e38073 } },
  18484. /* 144 */
  18485. { { 0xa7bf9b7c,0xf4f80824,0x3fbe30d0,0x365d2320,0x97cf9ce3,0xbfbe5320,
  18486. 0xb3055526,0xe3604700 },
  18487. { 0x6cc6c2c7,0x4dcb9911,0xba4cbee6,0x72683708,0x637ad9ec,0xdcded434,
  18488. 0xa3dee15f,0x6542d677 } },
  18489. /* 145 */
  18490. { { 0x7b6c377a,0x3f32b6d0,0x903448be,0x6cb03847,0x20da8af7,0xd6fdd3a8,
  18491. 0x09bb6f21,0xa6534aee },
  18492. { 0x1035facf,0x30a1780d,0x9dcb47e6,0x35e55a33,0xc447f393,0x6ea50fe1,
  18493. 0xdc9aef22,0xf3cb672f } },
  18494. /* 146 */
  18495. { { 0x3b55fd83,0xeb3719fe,0x875ddd10,0xe0d7a46c,0x05cea784,0x33ac9fa9,
  18496. 0xaae870e7,0x7cafaa2e },
  18497. { 0x1d53b338,0x9b814d04,0xef87e6c6,0xe0acc0a0,0x11672b0f,0xfb93d108,
  18498. 0xb9bd522e,0x0aab13c1 } },
  18499. /* 147 */
  18500. { { 0xd2681297,0xddcce278,0xb509546a,0xcb350eb1,0x7661aaf2,0x2dc43173,
  18501. 0x847012e9,0x4b91a602 },
  18502. { 0x72f8ddcf,0xdcff1095,0x9a911af4,0x08ebf61e,0xc372430e,0x48f4360a,
  18503. 0x72321cab,0x49534c53 } },
  18504. /* 148 */
  18505. { { 0xf07b7e9d,0x83df7d71,0x13cd516f,0xa478efa3,0x6c047ee3,0x78ef264b,
  18506. 0xd65ac5ee,0xcaf46c4f },
  18507. { 0x92aa8266,0xa04d0c77,0x913684bb,0xedf45466,0xae4b16b0,0x56e65168,
  18508. 0x04c6770f,0x14ce9e57 } },
  18509. /* 149 */
  18510. { { 0x965e8f91,0x99445e3e,0xcb0f2492,0xd3aca1ba,0x90c8a0a0,0xd31cc70f,
  18511. 0x3e4c9a71,0x1bb708a5 },
  18512. { 0x558bdd7a,0xd5ca9e69,0x018a26b1,0x734a0508,0x4c9cf1ec,0xb093aa71,
  18513. 0xda300102,0xf9d126f2 } },
  18514. /* 150 */
  18515. { { 0xaff9563e,0x749bca7a,0xb49914a0,0xdd077afe,0xbf5f1671,0xe27a0311,
  18516. 0x729ecc69,0x807afcb9 },
  18517. { 0xc9b08b77,0x7f8a9337,0x443c7e38,0x86c3a785,0x476fd8ba,0x85fafa59,
  18518. 0x6568cd8c,0x751adcd1 } },
  18519. /* 151 */
  18520. { { 0x10715c0d,0x8aea38b4,0x8f7697f7,0xd113ea71,0x93fbf06d,0x665eab14,
  18521. 0x2537743f,0x29ec4468 },
  18522. { 0xb50bebbc,0x3d94719c,0xe4505422,0x399ee5bf,0x8d2dedb1,0x90cd5b3a,
  18523. 0x92a4077d,0xff9370e3 } },
  18524. /* 152 */
  18525. { { 0xc6b75b65,0x59a2d69b,0x266651c5,0x4188f8d5,0x3de9d7d2,0x28a9f33e,
  18526. 0xa2a9d01a,0x9776478b },
  18527. { 0x929af2c7,0x8852622d,0x4e690923,0x334f5d6d,0xa89a51e9,0xce6cc7e5,
  18528. 0xac2f82fa,0x74a6313f } },
  18529. /* 153 */
  18530. { { 0xb75f079c,0xb2f4dfdd,0x18e36fbb,0x85b07c95,0xe7cd36dd,0x1b6cfcf0,
  18531. 0x0ff4863d,0xab75be15 },
  18532. { 0x173fc9b7,0x81b367c0,0xd2594fd0,0xb90a7420,0xc4091236,0x15fdbf03,
  18533. 0x0b4459f6,0x4ebeac2e } },
  18534. /* 154 */
  18535. { { 0x5c9f2c53,0xeb6c5fe7,0x8eae9411,0xd2522011,0xf95ac5d8,0xc8887633,
  18536. 0x2c1baffc,0xdf99887b },
  18537. { 0x850aaecb,0xbb78eed2,0x01d6a272,0x9d49181b,0xb1cdbcac,0x978dd511,
  18538. 0x779f4058,0x27b040a7 } },
  18539. /* 155 */
  18540. { { 0xf73b2eb2,0x90405db7,0x8e1b2118,0xe0df8508,0x5962327e,0x501b7152,
  18541. 0xe4cfa3f5,0xb393dd37 },
  18542. { 0x3fd75165,0xa1230e7b,0xbcd33554,0xd66344c2,0x0f7b5022,0x6c36f1be,
  18543. 0xd0463419,0x09588c12 } },
  18544. /* 156 */
  18545. { { 0x02601c3b,0xe086093f,0xcf5c335f,0xfb0252f8,0x894aff28,0x955cf280,
  18546. 0xdb9f648b,0x81c879a9 },
  18547. { 0xc6f56c51,0x040e687c,0x3f17618c,0xfed47169,0x9059353b,0x44f88a41,
  18548. 0x5fc11bc4,0xfa0d48f5 } },
  18549. /* 157 */
  18550. { { 0xe1608e4d,0xbc6e1c9d,0x3582822c,0x010dda11,0x157ec2d7,0xf6b7ddc1,
  18551. 0xb6a367d6,0x8ea0e156 },
  18552. { 0x2383b3b4,0xa354e02f,0x3f01f53c,0x69966b94,0x2de03ca5,0x4ff6632b,
  18553. 0xfa00b5ac,0x3f5ab924 } },
  18554. /* 158 */
  18555. { { 0x59739efb,0x337bb0d9,0xe7ebec0d,0xc751b0f4,0x411a67d1,0x2da52dd6,
  18556. 0x2b74256e,0x8bc76887 },
  18557. { 0x82d3d253,0xa5be3b72,0xf58d779f,0xa9f679a1,0xe16767bb,0xa1cac168,
  18558. 0x60fcf34f,0xb386f190 } },
  18559. /* 159 */
  18560. { { 0x2fedcfc2,0x31f3c135,0x62f8af0d,0x5396bf62,0xe57288c2,0x9a02b4ea,
  18561. 0x1b069c4d,0x4cb460f7 },
  18562. { 0x5b8095ea,0xae67b4d3,0x6fc07603,0x92bbf859,0xb614a165,0xe1475f66,
  18563. 0x95ef5223,0x52c0d508 } },
  18564. /* 160 */
  18565. { { 0x15339848,0x231c210e,0x70778c8d,0xe87a28e8,0x6956e170,0x9d1de661,
  18566. 0x2bb09c0b,0x4ac3c938 },
  18567. { 0x6998987d,0x19be0551,0xae09f4d6,0x8b2376c4,0x1a3f933d,0x1de0b765,
  18568. 0xe39705f4,0x380d94c7 } },
  18569. /* 161 */
  18570. { { 0x81542e75,0x01a355aa,0xee01b9b7,0x96c724a1,0x624d7087,0x6b3a2977,
  18571. 0xde2637af,0x2ce3e171 },
  18572. { 0xf5d5bc1a,0xcfefeb49,0x2777e2b5,0xa655607e,0x9513756c,0x4feaac2f,
  18573. 0x0b624e4d,0x2e6cd852 } },
  18574. /* 162 */
  18575. { { 0x8c31c31d,0x3685954b,0x5bf21a0c,0x68533d00,0x75c79ec9,0x0bd7626e,
  18576. 0x42c69d54,0xca177547 },
  18577. { 0xf6d2dbb2,0xcc6edaff,0x174a9d18,0xfd0d8cbd,0xaa4578e8,0x875e8793,
  18578. 0x9cab2ce6,0xa976a713 } },
  18579. /* 163 */
  18580. { { 0x93fb353d,0x0a651f1b,0x57fcfa72,0xd75cab8b,0x31b15281,0xaa88cfa7,
  18581. 0x0a1f4999,0x8720a717 },
  18582. { 0x693e1b90,0x8c3e8d37,0x16f6dfc3,0xd345dc0b,0xb52a8742,0x8ea8d00a,
  18583. 0xc769893c,0x9719ef29 } },
  18584. /* 164 */
  18585. { { 0x58e35909,0x820eed8d,0x33ddc116,0x9366d8dc,0x6e205026,0xd7f999d0,
  18586. 0xe15704c1,0xa5072976 },
  18587. { 0xc4e70b2e,0x002a37ea,0x6890aa8a,0x84dcf657,0x645b2a5c,0xcd71bf18,
  18588. 0xf7b77725,0x99389c9d } },
  18589. /* 165 */
  18590. { { 0x7ada7a4b,0x238c08f2,0xfd389366,0x3abe9d03,0x766f512c,0x6b672e89,
  18591. 0x202c82e4,0xa88806aa },
  18592. { 0xd380184e,0x6602044a,0x126a8b85,0xa8cb78c4,0xad844f17,0x79d670c0,
  18593. 0x4738dcfe,0x0043bffb } },
  18594. /* 166 */
  18595. { { 0x36d5192e,0x8d59b5dc,0x4590b2af,0xacf885d3,0x11601781,0x83566d0a,
  18596. 0xba6c4866,0x52f3ef01 },
  18597. { 0x0edcb64d,0x3986732a,0x8068379f,0x0a482c23,0x7040f309,0x16cbe5fa,
  18598. 0x9ef27e75,0x3296bd89 } },
  18599. /* 167 */
  18600. { { 0x454d81d7,0x476aba89,0x51eb9b3c,0x9eade7ef,0x81c57986,0x619a21cd,
  18601. 0xaee571e9,0x3b90febf },
  18602. { 0x5496f7cb,0x9393023e,0x7fb51bc4,0x55be41d8,0x99beb5ce,0x03f1dd48,
  18603. 0x9f810b18,0x6e88069d } },
  18604. /* 168 */
  18605. { { 0xb43ea1db,0xce37ab11,0x5259d292,0x0a7ff1a9,0x8f84f186,0x851b0221,
  18606. 0xdefaad13,0xa7222bea },
  18607. { 0x2b0a9144,0xa2ac78ec,0xf2fa59c5,0x5a024051,0x6147ce38,0x91d1eca5,
  18608. 0xbc2ac690,0xbe94d523 } },
  18609. /* 169 */
  18610. { { 0x0b226ce7,0x72f4945e,0x967e8b70,0xb8afd747,0x85a6c63e,0xedea46f1,
  18611. 0x9be8c766,0x7782defe },
  18612. { 0x3db38626,0x760d2aa4,0x76f67ad1,0x460ae787,0x54499cdb,0x341b86fc,
  18613. 0xa2892e4b,0x03838567 } },
  18614. /* 170 */
  18615. { { 0x79ec1a0f,0x2d8daefd,0xceb39c97,0x3bbcd6fd,0x58f61a95,0xf5575ffc,
  18616. 0xadf7b420,0xdbd986c4 },
  18617. { 0x15f39eb7,0x81aa8814,0xb98d976c,0x6ee2fcf5,0xcf2f717d,0x5465475d,
  18618. 0x6860bbd0,0x8e24d3c4 } },
  18619. /* 171 */
  18620. { { 0x9a587390,0x749d8e54,0x0cbec588,0x12bb194f,0xb25983c6,0x46e07da4,
  18621. 0x407bafc8,0x541a99c4 },
  18622. { 0x624c8842,0xdb241692,0xd86c05ff,0x6044c12a,0x4f7fcf62,0xc59d14b4,
  18623. 0xf57d35d1,0xc0092c49 } },
  18624. /* 172 */
  18625. { { 0xdf2e61ef,0xd3cc75c3,0x2e1b35ca,0x7e8841c8,0x909f29f4,0xc62d30d1,
  18626. 0x7286944d,0x75e40634 },
  18627. { 0xbbc237d0,0xe7d41fc5,0xec4f01c9,0xc9537bf0,0x282bd534,0x91c51a16,
  18628. 0xc7848586,0x5b7cb658 } },
  18629. /* 173 */
  18630. { { 0x8a28ead1,0x964a7084,0xfd3b47f6,0x802dc508,0x767e5b39,0x9ae4bfd1,
  18631. 0x8df097a1,0x7ae13eba },
  18632. { 0xeadd384e,0xfd216ef8,0xb6b2ff06,0x0361a2d9,0x4bcdb5f3,0x204b9878,
  18633. 0xe2a8e3fd,0x787d8074 } },
  18634. /* 174 */
  18635. { { 0x757fbb1c,0xc5e25d6b,0xca201deb,0xe47bddb2,0x6d2233ff,0x4a55e9a3,
  18636. 0x9ef28484,0x5c222819 },
  18637. { 0x88315250,0x773d4a85,0x827097c1,0x21b21a2b,0xdef5d33f,0xab7c4ea1,
  18638. 0xbaf0f2b0,0xe45d37ab } },
  18639. /* 175 */
  18640. { { 0x28511c8a,0xd2df1e34,0xbdca6cd3,0xebb229c8,0x627c39a7,0x578a71a7,
  18641. 0x84dfb9d3,0xed7bc122 },
  18642. { 0x93dea561,0xcf22a6df,0xd48f0ed1,0x5443f18d,0x5bad23e8,0xd8b86140,
  18643. 0x45ca6d27,0xaac97cc9 } },
  18644. /* 176 */
  18645. { { 0xa16bd00a,0xeb54ea74,0xf5c0bcc1,0xd839e9ad,0x1f9bfc06,0x092bb7f1,
  18646. 0x1163dc4e,0x318f97b3 },
  18647. { 0xc30d7138,0xecc0c5be,0xabc30220,0x44e8df23,0xb0223606,0x2bb7972f,
  18648. 0x9a84ff4d,0xfa41faa1 } },
  18649. /* 177 */
  18650. { { 0xa6642269,0x4402d974,0x9bb783bd,0xc81814ce,0x7941e60b,0x398d38e4,
  18651. 0x1d26e9e2,0x38bb6b2c },
  18652. { 0x6a577f87,0xc64e4a25,0xdc11fe1c,0x8b52d253,0x62280728,0xff336abf,
  18653. 0xce7601a5,0x94dd0905 } },
  18654. /* 178 */
  18655. { { 0xde93f92a,0x156cf7dc,0x89b5f315,0xa01333cb,0xc995e750,0x02404df9,
  18656. 0xd25c2ae9,0x92077867 },
  18657. { 0x0bf39d44,0xe2471e01,0x96bb53d7,0x5f2c9020,0x5c9c3d8f,0x4c44b7b3,
  18658. 0xd29beb51,0x81e8428b } },
  18659. /* 179 */
  18660. { { 0xc477199f,0x6dd9c2ba,0x6b5ecdd9,0x8cb8eeee,0xee40fd0e,0x8af7db3f,
  18661. 0xdbbfa4b1,0x1b94ab62 },
  18662. { 0xce47f143,0x44f0d8b3,0x63f46163,0x51e623fc,0xcc599383,0xf18f270f,
  18663. 0x055590ee,0x06a38e28 } },
  18664. /* 180 */
  18665. { { 0xb3355b49,0x2e5b0139,0xb4ebf99b,0x20e26560,0xd269f3dc,0xc08ffa6b,
  18666. 0x83d9d4f8,0xa7b36c20 },
  18667. { 0x1b3e8830,0x64d15c3a,0xa89f9c0b,0xd5fceae1,0xe2d16930,0xcfeee4a2,
  18668. 0xa2822a20,0xbe54c6b4 } },
  18669. /* 181 */
  18670. { { 0x8d91167c,0xd6cdb3df,0xe7a6625e,0x517c3f79,0x346ac7f4,0x7105648f,
  18671. 0xeae022bb,0xbf30a5ab },
  18672. { 0x93828a68,0x8e7785be,0x7f3ef036,0x5161c332,0x592146b2,0xe11b5feb,
  18673. 0x2732d13a,0xd1c820de } },
  18674. /* 182 */
  18675. { { 0x9038b363,0x043e1347,0x6b05e519,0x58c11f54,0x6026cad1,0x4fe57abe,
  18676. 0x68a18da3,0xb7d17bed },
  18677. { 0xe29c2559,0x44ca5891,0x5bfffd84,0x4f7a0376,0x74e46948,0x498de4af,
  18678. 0x6412cc64,0x3997fd5e } },
  18679. /* 183 */
  18680. { { 0x8bd61507,0xf2074682,0x34a64d2a,0x29e132d5,0x8a8a15e3,0xffeddfb0,
  18681. 0x3c6c13e8,0x0eeb8929 },
  18682. { 0xa7e259f8,0xe9b69a3e,0xd13e7e67,0xce1db7e6,0xad1fa685,0x277318f6,
  18683. 0xc922b6ef,0x228916f8 } },
  18684. /* 184 */
  18685. { { 0x0a12ab5b,0x959ae25b,0x957bc136,0xcc11171f,0xd16e2b0c,0x8058429e,
  18686. 0x6e93097e,0xec05ad1d },
  18687. { 0xac3f3708,0x157ba5be,0x30b59d77,0x31baf935,0x118234e5,0x47b55237,
  18688. 0x7ff11b37,0x7d314156 } },
  18689. /* 185 */
  18690. { { 0xf6dfefab,0x7bd9c05c,0xdcb37707,0xbe2f2268,0x3a38bb95,0xe53ead97,
  18691. 0x9bc1d7a3,0xe9ce66fc },
  18692. { 0x6f6a02a1,0x75aa1576,0x60e600ed,0x38c087df,0x68cdc1b9,0xf8947f34,
  18693. 0x72280651,0xd9650b01 } },
  18694. /* 186 */
  18695. { { 0x5a057e60,0x504b4c4a,0x8def25e4,0xcbccc3be,0x17c1ccbd,0xa6353208,
  18696. 0x804eb7a2,0x14d6699a },
  18697. { 0xdb1f411a,0x2c8a8415,0xf80d769c,0x09fbaf0b,0x1c2f77ad,0xb4deef90,
  18698. 0x0d43598a,0x6f4c6841 } },
  18699. /* 187 */
  18700. { { 0x96c24a96,0x8726df4e,0xfcbd99a3,0x534dbc85,0x8b2ae30a,0x3c466ef2,
  18701. 0x61189abb,0x4c4350fd },
  18702. { 0xf855b8da,0x2967f716,0x463c38a1,0x41a42394,0xeae93343,0xc37e1413,
  18703. 0x5a3118b5,0xa726d242 } },
  18704. /* 188 */
  18705. { { 0x948c1086,0xdae6b3ee,0xcbd3a2e1,0xf1de503d,0x03d022f3,0x3f35ed3f,
  18706. 0xcc6cf392,0x13639e82 },
  18707. { 0xcdafaa86,0x9ac938fb,0x2654a258,0xf45bc5fb,0x45051329,0x1963b26e,
  18708. 0xc1a335a3,0xca9365e1 } },
  18709. /* 189 */
  18710. { { 0x4c3b2d20,0x3615ac75,0x904e241b,0x742a5417,0xcc9d071d,0xb08521c4,
  18711. 0x970b72a5,0x9ce29c34 },
  18712. { 0x6d3e0ad6,0x8cc81f73,0xf2f8434c,0x8060da9e,0x6ce862d9,0x35ed1d1a,
  18713. 0xab42af98,0x48c4abd7 } },
  18714. /* 190 */
  18715. { { 0x40c7485a,0xd221b0cc,0xe5274dbf,0xead455bb,0x9263d2e8,0x493c7698,
  18716. 0xf67b33cb,0x78017c32 },
  18717. { 0x930cb5ee,0xb9d35769,0x0c408ed2,0xc0d14e94,0x272f1a4d,0xf8b7bf55,
  18718. 0xde5c1c04,0x53cd0454 } },
  18719. /* 191 */
  18720. { { 0x5d28ccac,0xbcd585fa,0x005b746e,0x5f823e56,0xcd0123aa,0x7c79f0a1,
  18721. 0xd3d7fa8f,0xeea465c1 },
  18722. { 0x0551803b,0x7810659f,0x7ce6af70,0x6c0b599f,0x29288e70,0x4195a770,
  18723. 0x7ae69193,0x1b6e42a4 } },
  18724. /* 192 */
  18725. { { 0xf67d04c3,0x2e80937c,0x89eeb811,0x1e312be2,0x92594d60,0x56b5d887,
  18726. 0x187fbd3d,0x0224da14 },
  18727. { 0x0c5fe36f,0x87abb863,0x4ef51f5f,0x580f3c60,0xb3b429ec,0x964fb1bf,
  18728. 0x42bfff33,0x60838ef0 } },
  18729. /* 193 */
  18730. { { 0x7e0bbe99,0x432cb2f2,0x04aa39ee,0x7bda44f3,0x9fa93903,0x5f497c7a,
  18731. 0x2d331643,0x636eb202 },
  18732. { 0x93ae00aa,0xfcfd0e61,0x31ae6d2f,0x875a00fe,0x9f93901c,0xf43658a2,
  18733. 0x39218bac,0x8844eeb6 } },
  18734. /* 194 */
  18735. { { 0x6b3bae58,0x114171d2,0x17e39f3e,0x7db3df71,0x81a8eada,0xcd37bc7f,
  18736. 0x51fb789e,0x27ba83dc },
  18737. { 0xfbf54de5,0xa7df439f,0xb5fe1a71,0x7277030b,0xdb297a48,0x42ee8e35,
  18738. 0x87f3a4ab,0xadb62d34 } },
  18739. /* 195 */
  18740. { { 0xa175df2a,0x9b1168a2,0x618c32e9,0x082aa04f,0x146b0916,0xc9e4f2e7,
  18741. 0x75e7c8b2,0xb990fd76 },
  18742. { 0x4df37313,0x0829d96b,0xd0b40789,0x1c205579,0x78087711,0x66c9ae4a,
  18743. 0x4d10d18d,0x81707ef9 } },
  18744. /* 196 */
  18745. { { 0x03d6ff96,0x97d7cab2,0x0d843360,0x5b851bfc,0xd042db4b,0x268823c4,
  18746. 0xd5a8aa5c,0x3792daea },
  18747. { 0x941afa0b,0x52818865,0x42d83671,0xf3e9e741,0x5be4e0a7,0x17c82527,
  18748. 0x94b001ba,0x5abd635e } },
  18749. /* 197 */
  18750. { { 0x0ac4927c,0x727fa84e,0xa7c8cf23,0xe3886035,0x4adca0df,0xa4bcd5ea,
  18751. 0x846ab610,0x5995bf21 },
  18752. { 0x829dfa33,0xe90f860b,0x958fc18b,0xcaafe2ae,0x78630366,0x9b3baf44,
  18753. 0xd483411e,0x44c32ca2 } },
  18754. /* 198 */
  18755. { { 0xe40ed80c,0xa74a97f1,0x31d2ca82,0x5f938cb1,0x7c2d6ad9,0x53f2124b,
  18756. 0x8082a54c,0x1f2162fb },
  18757. { 0x720b173e,0x7e467cc5,0x085f12f9,0x40e8a666,0x4c9d65dc,0x8cebc20e,
  18758. 0xc3e907c9,0x8f1d402b } },
  18759. /* 199 */
  18760. { { 0xfbc4058a,0x4f592f9c,0x292f5670,0xb15e14b6,0xbc1d8c57,0xc55cfe37,
  18761. 0x926edbf9,0xb1980f43 },
  18762. { 0x32c76b09,0x98c33e09,0x33b07f78,0x1df5279d,0x863bb461,0x6f08ead4,
  18763. 0x37448e45,0x2828ad9b } },
  18764. /* 200 */
  18765. { { 0xc4cf4ac5,0x696722c4,0xdde64afb,0xf5ac1a3f,0xe0890832,0x0551baa2,
  18766. 0x5a14b390,0x4973f127 },
  18767. { 0x322eac5d,0xe59d8335,0x0bd9b568,0x5e07eef5,0xa2588393,0xab36720f,
  18768. 0xdb168ac7,0x6dac8ed0 } },
  18769. /* 201 */
  18770. { { 0xeda835ef,0xf7b545ae,0x1d10ed51,0x4aa113d2,0x13741b09,0x035a65e0,
  18771. 0x20b9de4c,0x4b23ef59 },
  18772. { 0x3c4c7341,0xe82bb680,0x3f58bc37,0xd457706d,0xa51e3ee8,0x73527863,
  18773. 0xddf49a4e,0x4dd71534 } },
  18774. /* 202 */
  18775. { { 0x95476cd9,0xbf944672,0xe31a725b,0x648d072f,0xfc4b67e0,0x1441c8b8,
  18776. 0x2f4a4dbb,0xfd317000 },
  18777. { 0x8995d0e1,0x1cb43ff4,0x0ef729aa,0x76e695d1,0x41798982,0xe0d5f976,
  18778. 0x9569f365,0x14fac58c } },
  18779. /* 203 */
  18780. { { 0xf312ae18,0xad9a0065,0xfcc93fc9,0x51958dc0,0x8a7d2846,0xd9a14240,
  18781. 0x36abda50,0xed7c7651 },
  18782. { 0x25d4abbc,0x46270f1a,0xf1a113ea,0x9b5dd8f3,0x5b51952f,0xc609b075,
  18783. 0x4d2e9f53,0xfefcb7f7 } },
  18784. /* 204 */
  18785. { { 0xba119185,0xbd09497a,0xaac45ba4,0xd54e8c30,0xaa521179,0x492479de,
  18786. 0x87e0d80b,0x1801a57e },
  18787. { 0xfcafffb0,0x073d3f8d,0xae255240,0x6cf33c0b,0x5b5fdfbc,0x781d763b,
  18788. 0x1ead1064,0x9f8fc11e } },
  18789. /* 205 */
  18790. { { 0x5e69544c,0x1583a171,0xf04b7813,0x0eaf8567,0x278a4c32,0x1e22a8fd,
  18791. 0x3d3a69a9,0xa9d3809d },
  18792. { 0x59a2da3b,0x936c2c2c,0x1895c847,0x38ccbcf6,0x63d50869,0x5e65244e,
  18793. 0xe1178ef7,0x3006b9ae } },
  18794. /* 206 */
  18795. { { 0xc9eead28,0x0bb1f2b0,0x89f4dfbc,0x7eef635d,0xb2ce8939,0x074757fd,
  18796. 0x45f8f761,0x0ab85fd7 },
  18797. { 0x3e5b4549,0xecda7c93,0x97922f21,0x4be2bb5c,0xb43b8040,0x261a1274,
  18798. 0x11e942c2,0xb122d675 } },
  18799. /* 207 */
  18800. { { 0x66a5ae7a,0x3be607be,0x76adcbe3,0x01e703fa,0x4eb6e5c5,0xaf904301,
  18801. 0x097dbaec,0x9f599dc1 },
  18802. { 0x0ff250ed,0x6d75b718,0x349a20dc,0x8eb91574,0x10b227a3,0x425605a4,
  18803. 0x8a294b78,0x7d5528e0 } },
  18804. /* 208 */
  18805. { { 0x20c26def,0xf0f58f66,0x582b2d1e,0x025585ea,0x01ce3881,0xfbe7d79b,
  18806. 0x303f1730,0x28ccea01 },
  18807. { 0x79644ba5,0xd1dabcd1,0x06fff0b8,0x1fc643e8,0x66b3e17b,0xa60a76fc,
  18808. 0xa1d013bf,0xc18baf48 } },
  18809. /* 209 */
  18810. { { 0x5dc4216d,0x34e638c8,0x206142ac,0x00c01067,0x95f5064a,0xd453a171,
  18811. 0xb7a9596b,0x9def809d },
  18812. { 0x67ab8d2c,0x41e8642e,0x6237a2b6,0xb4240433,0x64c4218b,0x7d506a6d,
  18813. 0x68808ce5,0x0357f8b0 } },
  18814. /* 210 */
  18815. { { 0x4cd2cc88,0x8e9dbe64,0xf0b8f39d,0xcc61c28d,0xcd30a0c8,0x4a309874,
  18816. 0x1b489887,0xe4a01add },
  18817. { 0xf57cd8f9,0x2ed1eeac,0xbd594c48,0x1b767d3e,0x7bd2f787,0xa7295c71,
  18818. 0xce10cc30,0x466d7d79 } },
  18819. /* 211 */
  18820. { { 0x9dada2c7,0x47d31892,0x8f9aa27d,0x4fa0a6c3,0x820a59e1,0x90e4fd28,
  18821. 0x451ead1a,0xc672a522 },
  18822. { 0x5d86b655,0x30607cc8,0xf9ad4af1,0xf0235d3b,0x571172a6,0x99a08680,
  18823. 0xf2a67513,0x5e3d64fa } },
  18824. /* 212 */
  18825. { { 0x9b3b4416,0xaa6410c7,0xeab26d99,0xcd8fcf85,0xdb656a74,0x5ebff74a,
  18826. 0xeb8e42fc,0x6c8a7a95 },
  18827. { 0xb02a63bd,0x10c60ba7,0x8b8f0047,0x6b2f2303,0x312d90b0,0x8c6c3738,
  18828. 0xad82ca91,0x348ae422 } },
  18829. /* 213 */
  18830. { { 0x5ccda2fb,0x7f474663,0x8e0726d2,0x22accaa1,0x492b1f20,0x85adf782,
  18831. 0xd9ef2d2e,0xc1074de0 },
  18832. { 0xae9a65b3,0xfcf3ce44,0x05d7151b,0xfd71e4ac,0xce6a9788,0xd4711f50,
  18833. 0xc9e54ffc,0xfbadfbdb } },
  18834. /* 214 */
  18835. { { 0x20a99363,0x1713f1cd,0x6cf22775,0xb915658f,0x24d359b2,0x968175cd,
  18836. 0x83716fcd,0xb7f976b4 },
  18837. { 0x5d6dbf74,0x5758e24d,0x71c3af36,0x8d23bafd,0x0243dfe3,0x48f47760,
  18838. 0xcafcc805,0xf4d41b2e } },
  18839. /* 215 */
  18840. { { 0xfdabd48d,0x51f1cf28,0x32c078a4,0xce81be36,0x117146e9,0x6ace2974,
  18841. 0xe0160f10,0x180824ea },
  18842. { 0x66e58358,0x0387698b,0xce6ca358,0x63568752,0x5e41e6c5,0x82380e34,
  18843. 0x83cf6d25,0x67e5f639 } },
  18844. /* 216 */
  18845. { { 0xcf4899ef,0xf89ccb8d,0x9ebb44c0,0x949015f0,0xb2598ec9,0x546f9276,
  18846. 0x04c11fc6,0x9fef789a },
  18847. { 0x53d2a071,0x6d367ecf,0xa4519b09,0xb10e1a7f,0x611e2eef,0xca6b3fb0,
  18848. 0xa99c4e20,0xbc80c181 } },
  18849. /* 217 */
  18850. { { 0xe5eb82e6,0x972536f8,0xf56cb920,0x1a484fc7,0x50b5da5e,0xc78e2171,
  18851. 0x9f8cdf10,0x49270e62 },
  18852. { 0xea6b50ad,0x1a39b7bb,0xa2388ffc,0x9a0284c1,0x8107197b,0x5403eb17,
  18853. 0x61372f7f,0xd2ee52f9 } },
  18854. /* 218 */
  18855. { { 0x88e0362a,0xd37cd285,0x8fa5d94d,0x442fa8a7,0xa434a526,0xaff836e5,
  18856. 0xe5abb733,0xdfb478be },
  18857. { 0x673eede6,0xa91f1ce7,0x2b5b2f04,0xa5390ad4,0x5530da2f,0x5e66f7bf,
  18858. 0x08df473a,0xd9a140b4 } },
  18859. /* 219 */
  18860. { { 0x6e8ea498,0x0e0221b5,0x3563ee09,0x62347829,0x335d2ade,0xe06b8391,
  18861. 0x623f4b1a,0x760c058d },
  18862. { 0xc198aa79,0x0b89b58c,0xf07aba7f,0xf74890d2,0xfde2556a,0x4e204110,
  18863. 0x8f190409,0x7141982d } },
  18864. /* 220 */
  18865. { { 0x4d4b0f45,0x6f0a0e33,0x392a94e1,0xd9280b38,0xb3c61d5e,0x3af324c6,
  18866. 0x89d54e47,0x3af9d1ce },
  18867. { 0x20930371,0xfd8f7981,0x21c17097,0xeda2664c,0xdc42309b,0x0e9545dc,
  18868. 0x73957dd6,0xb1f815c3 } },
  18869. /* 221 */
  18870. { { 0x89fec44a,0x84faa78e,0x3caa4caf,0xc8c2ae47,0xc1b6a624,0x691c807d,
  18871. 0x1543f052,0xa41aed14 },
  18872. { 0x7d5ffe04,0x42435399,0x625b6e20,0x8bacb2df,0x87817775,0x85d660be,
  18873. 0x86fb60ef,0xd6e9c1dd } },
  18874. /* 222 */
  18875. { { 0xc6853264,0x3aa2e97e,0xe2304a0b,0x771533b7,0xb8eae9be,0x1b912bb7,
  18876. 0xae9bf8c2,0x9c9c6e10 },
  18877. { 0xe030b74c,0xa2309a59,0x6a631e90,0x4ed7494d,0xa49b79f2,0x89f44b23,
  18878. 0x40fa61b6,0x566bd596 } },
  18879. /* 223 */
  18880. { { 0xc18061f3,0x066c0118,0x7c83fc70,0x190b25d3,0x27273245,0xf05fc8e0,
  18881. 0xf525345e,0xcf2c7390 },
  18882. { 0x10eb30cf,0xa09bceb4,0x0d77703a,0xcfd2ebba,0x150ff255,0xe842c43a,
  18883. 0x8aa20979,0x02f51755 } },
  18884. /* 224 */
  18885. { { 0xaddb7d07,0x396ef794,0x24455500,0x0b4fc742,0xc78aa3ce,0xfaff8eac,
  18886. 0xe8d4d97d,0x14e9ada5 },
  18887. { 0x2f7079e2,0xdaa480a1,0xe4b0800e,0x45baa3cd,0x7838157d,0x01765e2d,
  18888. 0x8e9d9ae8,0xa0ad4fab } },
  18889. /* 225 */
  18890. { { 0x4a653618,0x0bfb7621,0x31eaaa5f,0x1872813c,0x44949d5e,0x1553e737,
  18891. 0x6e56ed1e,0xbcd530b8 },
  18892. { 0x32e9c47b,0x169be853,0xb50059ab,0xdc2776fe,0x192bfbb4,0xcdba9761,
  18893. 0x6979341d,0x909283cf } },
  18894. /* 226 */
  18895. { { 0x76e81a13,0x67b00324,0x62171239,0x9bee1a99,0xd32e19d6,0x08ed361b,
  18896. 0xace1549a,0x35eeb7c9 },
  18897. { 0x7e4e5bdc,0x1280ae5a,0xb6ceec6e,0x2dcd2cd3,0x6e266bc1,0x52e4224c,
  18898. 0x448ae864,0x9a8b2cf4 } },
  18899. /* 227 */
  18900. { { 0x09d03b59,0xf6471bf2,0xb65af2ab,0xc90e62a3,0xebd5eec9,0xff7ff168,
  18901. 0xd4491379,0x6bdb60f4 },
  18902. { 0x8a55bc30,0xdadafebc,0x10097fe0,0xc79ead16,0x4c1e3bdd,0x42e19741,
  18903. 0x94ba08a9,0x01ec3cfd } },
  18904. /* 228 */
  18905. { { 0xdc9485c2,0xba6277eb,0x22fb10c7,0x48cc9a79,0x70a28d8a,0x4f61d60f,
  18906. 0x475464f6,0xd1acb1c0 },
  18907. { 0x26f36612,0xd26902b1,0xe0618d8b,0x59c3a44e,0x308357ee,0x4df8a813,
  18908. 0x405626c2,0x7dcd079d } },
  18909. /* 229 */
  18910. { { 0xf05a4b48,0x5ce7d4d3,0x37230772,0xadcd2952,0x812a915a,0xd18f7971,
  18911. 0x377d19b8,0x0bf53589 },
  18912. { 0x6c68ea73,0x35ecd95a,0x823a584d,0xc7f3bbca,0xf473a723,0x9fb674c6,
  18913. 0xe16686fc,0xd28be4d9 } },
  18914. /* 230 */
  18915. { { 0x38fa8e4b,0x5d2b9906,0x893fd8fc,0x559f186e,0x436fb6fc,0x3a6de2aa,
  18916. 0x510f88ce,0xd76007aa },
  18917. { 0x523a4988,0x2d10aab6,0x74dd0273,0xb455cf44,0xa3407278,0x7f467082,
  18918. 0xb303bb01,0xf2b52f68 } },
  18919. /* 231 */
  18920. { { 0x9835b4ca,0x0d57eafa,0xbb669cbc,0x2d2232fc,0xc6643198,0x8eeeb680,
  18921. 0xcc5aed3a,0xd8dbe98e },
  18922. { 0xc5a02709,0xcba9be3f,0xf5ba1fa8,0x30be68e5,0xf10ea852,0xfebd43cd,
  18923. 0xee559705,0xe01593a3 } },
  18924. /* 232 */
  18925. { { 0xea75a0a6,0xd3e5af50,0x57858033,0x512226ac,0xd0176406,0x6fe6d50f,
  18926. 0xaeb8ef06,0xafec07b1 },
  18927. { 0x80bb0a31,0x7fb99567,0x37309aae,0x6f1af3cc,0x01abf389,0x9153a15a,
  18928. 0x6e2dbfdd,0xa71b9354 } },
  18929. /* 233 */
  18930. { { 0x18f593d2,0xbf8e12e0,0xa078122b,0xd1a90428,0x0ba4f2ad,0x150505db,
  18931. 0x628523d9,0x53a2005c },
  18932. { 0xe7f2b935,0x07c8b639,0xc182961a,0x2bff975a,0x7518ca2c,0x86bceea7,
  18933. 0x3d588e3d,0xbf47d19b } },
  18934. /* 234 */
  18935. { { 0xdd7665d5,0x672967a7,0x2f2f4de5,0x4e303057,0x80d4903f,0x144005ae,
  18936. 0x39c9a1b6,0x001c2c7f },
  18937. { 0x69efc6d6,0x143a8014,0x7bc7a724,0xc810bdaa,0xa78150a4,0x5f65670b,
  18938. 0x86ffb99b,0xfdadf8e7 } },
  18939. /* 235 */
  18940. { { 0xffc00785,0xfd38cb88,0x3b48eb67,0x77fa7591,0xbf368fbc,0x0454d055,
  18941. 0x5aa43c94,0x3a838e4d },
  18942. { 0x3e97bb9a,0x56166329,0x441d94d9,0x9eb93363,0x0adb2a83,0x515591a6,
  18943. 0x873e1da3,0x3cdb8257 } },
  18944. /* 236 */
  18945. { { 0x7de77eab,0x137140a9,0x41648109,0xf7e1c50d,0xceb1d0df,0x762dcad2,
  18946. 0xf1f57fba,0x5a60cc89 },
  18947. { 0x40d45673,0x80b36382,0x5913c655,0x1b82be19,0xdd64b741,0x057284b8,
  18948. 0xdbfd8fc0,0x922ff56f } },
  18949. /* 237 */
  18950. { { 0xc9a129a1,0x1b265dee,0xcc284e04,0xa5b1ce57,0xcebfbe3c,0x04380c46,
  18951. 0xf6c5cd62,0x72919a7d },
  18952. { 0x8fb90f9a,0x298f453a,0x88e4031b,0xd719c00b,0x796f1856,0xe32c0e77,
  18953. 0x3624089a,0x5e791780 } },
  18954. /* 238 */
  18955. { { 0x7f63cdfb,0x5c16ec55,0xf1cae4fd,0x8e6a3571,0x560597ca,0xfce26bea,
  18956. 0xe24c2fab,0x4e0a5371 },
  18957. { 0xa5765357,0x276a40d3,0x0d73a2b4,0x3c89af44,0x41d11a32,0xb8f370ae,
  18958. 0xd56604ee,0xf5ff7818 } },
  18959. /* 239 */
  18960. { { 0x1a09df21,0xfbf3e3fe,0xe66e8e47,0x26d5d28e,0x29c89015,0x2096bd0a,
  18961. 0x533f5e64,0xe41df0e9 },
  18962. { 0xb3ba9e3f,0x305fda40,0x2604d895,0xf2340ceb,0x7f0367c7,0x0866e192,
  18963. 0xac4f155f,0x8edd7d6e } },
  18964. /* 240 */
  18965. { { 0x0bfc8ff3,0xc9a1dc0e,0xe936f42f,0x14efd82b,0xcca381ef,0x67016f7c,
  18966. 0xed8aee96,0x1432c1ca },
  18967. { 0x70b23c26,0xec684829,0x0735b273,0xa64fe873,0xeaef0f5a,0xe389f6e5,
  18968. 0x5ac8d2c6,0xcaef480b } },
  18969. /* 241 */
  18970. { { 0x75315922,0x5245c978,0x3063cca5,0xd8295171,0xb64ef2cb,0xf3ce60d0,
  18971. 0x8efae236,0xd0ba177e },
  18972. { 0xb1b3af60,0x53a9ae8f,0x3d2da20e,0x1a796ae5,0xdf9eef28,0x01d63605,
  18973. 0x1c54ae16,0xf31c957c } },
  18974. /* 242 */
  18975. { { 0x49cc4597,0xc0f58d52,0xbae0a028,0xdc5015b0,0x734a814a,0xefc5fc55,
  18976. 0x96e17c3a,0x013404cb },
  18977. { 0xc9a824bf,0xb29e2585,0x001eaed7,0xd593185e,0x61ef68ac,0x8d6ee682,
  18978. 0x91933e6c,0x6f377c4b } },
  18979. /* 243 */
  18980. { { 0xa8333fd2,0x9f93bad1,0x5a2a95b8,0xa8930202,0xeaf75ace,0x211e5037,
  18981. 0xd2d09506,0x6dba3e4e },
  18982. { 0xd04399cd,0xa48ef98c,0xe6b73ade,0x1811c66e,0xc17ecaf3,0x72f60752,
  18983. 0x3becf4a7,0xf13cf342 } },
  18984. /* 244 */
  18985. { { 0xa919e2eb,0xceeb9ec0,0xf62c0f68,0x83a9a195,0x7aba2299,0xcfba3bb6,
  18986. 0x274bbad3,0xc83fa9a9 },
  18987. { 0x62fa1ce0,0x0d7d1b0b,0x3418efbf,0xe58b60f5,0x52706f04,0xbfa8ef9e,
  18988. 0x5d702683,0xb49d70f4 } },
  18989. /* 245 */
  18990. { { 0xfad5513b,0x914c7510,0xb1751e2d,0x05f32eec,0xd9fb9d59,0x6d850418,
  18991. 0x0c30f1cf,0x59cfadbb },
  18992. { 0x55cb7fd6,0xe167ac23,0x820426a3,0x249367b8,0x90a78864,0xeaeec58c,
  18993. 0x354a4b67,0x5babf362 } },
  18994. /* 246 */
  18995. { { 0xee424865,0x37c981d1,0xf2e5577f,0x8b002878,0xb9e0c058,0x702970f1,
  18996. 0x9026c8f0,0x6188c6a7 },
  18997. { 0xd0f244da,0x06f9a19b,0xfb080873,0x1ecced5c,0x9f213637,0x35470f9b,
  18998. 0xdf50b9d9,0x993fe475 } },
  18999. /* 247 */
  19000. { { 0x9b2c3609,0x68e31cdf,0x2c46d4ea,0x84eb19c0,0x9a775101,0x7ac9ec1a,
  19001. 0x4c80616b,0x81f76466 },
  19002. { 0x75fbe978,0x1d7c2a5a,0xf183b356,0x6743fed3,0x501dd2bf,0x838d1f04,
  19003. 0x5fe9060d,0x564a812a } },
  19004. /* 248 */
  19005. { { 0xfa817d1d,0x7a5a64f4,0xbea82e0f,0x55f96844,0xcd57f9aa,0xb5ff5a0f,
  19006. 0x00e51d6c,0x226bf3cf },
  19007. { 0x2f2833cf,0xd6d1a9f9,0x4f4f89a8,0x20a0a35a,0x8f3f7f77,0x11536c49,
  19008. 0xff257836,0x68779f47 } },
  19009. /* 249 */
  19010. { { 0x73043d08,0x79b0c1c1,0x1fc020fa,0xa5446774,0x9a6d26d0,0xd3767e28,
  19011. 0xeb092e0b,0x97bcb0d1 },
  19012. { 0xf32ed3c3,0x2ab6eaa8,0xb281bc48,0xc8a4f151,0xbfa178f3,0x4d1bf4f3,
  19013. 0x0a784655,0xa872ffe8 } },
  19014. /* 250 */
  19015. { { 0xa32b2086,0xb1ab7935,0x8160f486,0xe1eb710e,0x3b6ae6be,0x9bd0cd91,
  19016. 0xb732a36a,0x02812bfc },
  19017. { 0xcf605318,0xa63fd7ca,0xfdfd6d1d,0x646e5d50,0x2102d619,0xa1d68398,
  19018. 0xfe5396af,0x07391cc9 } },
  19019. /* 251 */
  19020. { { 0x8b80d02b,0xc50157f0,0x62877f7f,0x6b8333d1,0x78d542ae,0x7aca1af8,
  19021. 0x7e6d2a08,0x355d2adc },
  19022. { 0x287386e1,0xb41f335a,0xf8e43275,0xfd272a94,0xe79989ea,0x286ca2cd,
  19023. 0x7c2a3a79,0x3dc2b1e3 } },
  19024. /* 252 */
  19025. { { 0x04581352,0xd689d21c,0x376782be,0x0a00c825,0x9fed701f,0x203bd590,
  19026. 0x3ccd846b,0xc4786910 },
  19027. { 0x24c768ed,0x5dba7708,0x6841f657,0x72feea02,0x6accce0e,0x73313ed5,
  19028. 0xd5bb4d32,0xccc42968 } },
  19029. /* 253 */
  19030. { { 0x3d7620b9,0x94e50de1,0x5992a56a,0xd89a5c8a,0x675487c9,0xdc007640,
  19031. 0xaa4871cf,0xe147eb42 },
  19032. { 0xacf3ae46,0x274ab4ee,0x50350fbe,0xfd4936fb,0x48c840ea,0xdf2afe47,
  19033. 0x080e96e3,0x239ac047 } },
  19034. /* 254 */
  19035. { { 0x2bfee8d4,0x481d1f35,0xfa7b0fec,0xce80b5cf,0x2ce9af3c,0x105c4c9e,
  19036. 0xf5f7e59d,0xc55fa1a3 },
  19037. { 0x8257c227,0x3186f14e,0x342be00b,0xc5b1653f,0xaa904fb2,0x09afc998,
  19038. 0xd4f4b699,0x094cd99c } },
  19039. /* 255 */
  19040. { { 0xd703beba,0x8a981c84,0x32ceb291,0x8631d150,0xe3bd49ec,0xa445f2c9,
  19041. 0x42abad33,0xb90a30b6 },
  19042. { 0xb4a5abf9,0xb465404f,0x75db7603,0x004750c3,0xca35d89f,0x6f9a42cc,
  19043. 0x1b7924f7,0x019f8b9a } },
  19044. };
  19045. /* Multiply the base point of P256 by the scalar and return the result.
  19046. * If map is true then convert result to affine coordinates.
  19047. *
  19048. * r Resulting point.
  19049. * k Scalar to multiply by.
  19050. * map Indicates whether to convert result to affine.
  19051. * heap Heap to use for allocation.
  19052. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  19053. */
  19054. static int sp_256_ecc_mulmod_base_8(sp_point* r, const sp_digit* k,
  19055. int map, void* heap)
  19056. {
  19057. return sp_256_ecc_mulmod_stripe_8(r, &p256_base, p256_table,
  19058. k, map, heap);
  19059. }
  19060. #endif
  19061. /* Multiply the base point of P256 by the scalar and return the result.
  19062. * If map is true then convert result to affine coordinates.
  19063. *
  19064. * km Scalar to multiply by.
  19065. * r Resulting point.
  19066. * map Indicates whether to convert result to affine.
  19067. * heap Heap to use for allocation.
  19068. * returns MEMORY_E when memory allocation fails and MP_OKAY on success.
  19069. */
  19070. int sp_ecc_mulmod_base_256(mp_int* km, ecc_point* r, int map, void* heap)
  19071. {
  19072. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  19073. sp_point p;
  19074. sp_digit k[8];
  19075. #else
  19076. sp_digit* k = NULL;
  19077. #endif
  19078. sp_point* point;
  19079. int err;
  19080. err = sp_ecc_point_new(heap, p, point);
  19081. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19082. if (err == MP_OKAY) {
  19083. k = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8, heap,
  19084. DYNAMIC_TYPE_ECC);
  19085. if (k == NULL) {
  19086. err = MEMORY_E;
  19087. }
  19088. }
  19089. #endif
  19090. if (err == MP_OKAY) {
  19091. sp_256_from_mp(k, 8, km);
  19092. err = sp_256_ecc_mulmod_base_8(point, k, map, heap);
  19093. }
  19094. if (err == MP_OKAY) {
  19095. err = sp_256_point_to_ecc_point_8(point, r);
  19096. }
  19097. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19098. if (k != NULL) {
  19099. XFREE(k, heap, DYNAMIC_TYPE_ECC);
  19100. }
  19101. #endif
  19102. sp_ecc_point_free(point, 0, heap);
  19103. return err;
  19104. }
  19105. #if defined(WOLFSSL_VALIDATE_ECC_KEYGEN) || defined(HAVE_ECC_SIGN) || \
  19106. defined(HAVE_ECC_VERIFY)
  19107. /* Returns 1 if the number of zero.
  19108. * Implementation is constant time.
  19109. *
  19110. * a Number to check.
  19111. * returns 1 if the number is zero and 0 otherwise.
  19112. */
  19113. static int sp_256_iszero_8(const sp_digit* a)
  19114. {
  19115. return (a[0] | a[1] | a[2] | a[3] | a[4] | a[5] | a[6] | a[7]) == 0;
  19116. }
  19117. #endif /* WOLFSSL_VALIDATE_ECC_KEYGEN || HAVE_ECC_SIGN || HAVE_ECC_VERIFY */
  19118. /* Add 1 to a. (a = a + 1)
  19119. *
  19120. * a A single precision integer.
  19121. */
  19122. SP_NOINLINE static void sp_256_add_one_8(sp_digit* a)
  19123. {
  19124. __asm__ __volatile__ (
  19125. "mov r2, #1\n\t"
  19126. "ldr r1, [%[a], #0]\n\t"
  19127. "add r1, r2\n\t"
  19128. "mov r2, #0\n\t"
  19129. "str r1, [%[a], #0]\n\t"
  19130. "ldr r1, [%[a], #4]\n\t"
  19131. "adc r1, r2\n\t"
  19132. "str r1, [%[a], #4]\n\t"
  19133. "ldr r1, [%[a], #8]\n\t"
  19134. "adc r1, r2\n\t"
  19135. "str r1, [%[a], #8]\n\t"
  19136. "ldr r1, [%[a], #12]\n\t"
  19137. "adc r1, r2\n\t"
  19138. "str r1, [%[a], #12]\n\t"
  19139. "ldr r1, [%[a], #16]\n\t"
  19140. "adc r1, r2\n\t"
  19141. "str r1, [%[a], #16]\n\t"
  19142. "ldr r1, [%[a], #20]\n\t"
  19143. "adc r1, r2\n\t"
  19144. "str r1, [%[a], #20]\n\t"
  19145. "ldr r1, [%[a], #24]\n\t"
  19146. "adc r1, r2\n\t"
  19147. "str r1, [%[a], #24]\n\t"
  19148. "ldr r1, [%[a], #28]\n\t"
  19149. "adc r1, r2\n\t"
  19150. "str r1, [%[a], #28]\n\t"
  19151. :
  19152. : [a] "r" (a)
  19153. : "memory", "r1", "r2"
  19154. );
  19155. }
  19156. /* Read big endian unsigned byte array into r.
  19157. *
  19158. * r A single precision integer.
  19159. * size Maximum number of bytes to convert
  19160. * a Byte array.
  19161. * n Number of bytes in array to read.
  19162. */
  19163. static void sp_256_from_bin(sp_digit* r, int size, const byte* a, int n)
  19164. {
  19165. int i, j = 0;
  19166. word32 s = 0;
  19167. r[0] = 0;
  19168. for (i = n-1; i >= 0; i--) {
  19169. r[j] |= (((sp_digit)a[i]) << s);
  19170. if (s >= 24U) {
  19171. r[j] &= 0xffffffff;
  19172. s = 32U - s;
  19173. if (j + 1 >= size) {
  19174. break;
  19175. }
  19176. r[++j] = (sp_digit)a[i] >> s;
  19177. s = 8U - s;
  19178. }
  19179. else {
  19180. s += 8U;
  19181. }
  19182. }
  19183. for (j++; j < size; j++) {
  19184. r[j] = 0;
  19185. }
  19186. }
  19187. /* Generates a scalar that is in the range 1..order-1.
  19188. *
  19189. * rng Random number generator.
  19190. * k Scalar value.
  19191. * returns RNG failures, MEMORY_E when memory allocation fails and
  19192. * MP_OKAY on success.
  19193. */
  19194. static int sp_256_ecc_gen_k_8(WC_RNG* rng, sp_digit* k)
  19195. {
  19196. int err;
  19197. byte buf[32];
  19198. do {
  19199. err = wc_RNG_GenerateBlock(rng, buf, sizeof(buf));
  19200. if (err == 0) {
  19201. sp_256_from_bin(k, 8, buf, (int)sizeof(buf));
  19202. if (sp_256_cmp_8(k, p256_order2) < 0) {
  19203. sp_256_add_one_8(k);
  19204. break;
  19205. }
  19206. }
  19207. }
  19208. while (err == 0);
  19209. return err;
  19210. }
  19211. /* Makes a random EC key pair.
  19212. *
  19213. * rng Random number generator.
  19214. * priv Generated private value.
  19215. * pub Generated public point.
  19216. * heap Heap to use for allocation.
  19217. * returns ECC_INF_E when the point does not have the correct order, RNG
  19218. * failures, MEMORY_E when memory allocation fails and MP_OKAY on success.
  19219. */
  19220. int sp_ecc_make_key_256(WC_RNG* rng, mp_int* priv, ecc_point* pub, void* heap)
  19221. {
  19222. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  19223. sp_point p;
  19224. sp_digit k[8];
  19225. #ifdef WOLFSSL_VALIDATE_ECC_KEYGEN
  19226. sp_point inf;
  19227. #endif
  19228. #else
  19229. sp_digit* k = NULL;
  19230. #endif
  19231. sp_point* point;
  19232. #ifdef WOLFSSL_VALIDATE_ECC_KEYGEN
  19233. sp_point* infinity;
  19234. #endif
  19235. int err;
  19236. (void)heap;
  19237. err = sp_ecc_point_new(heap, p, point);
  19238. #ifdef WOLFSSL_VALIDATE_ECC_KEYGEN
  19239. if (err == MP_OKAY) {
  19240. err = sp_ecc_point_new(heap, inf, infinity);
  19241. }
  19242. #endif
  19243. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19244. if (err == MP_OKAY) {
  19245. k = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8, heap,
  19246. DYNAMIC_TYPE_ECC);
  19247. if (k == NULL) {
  19248. err = MEMORY_E;
  19249. }
  19250. }
  19251. #endif
  19252. if (err == MP_OKAY) {
  19253. err = sp_256_ecc_gen_k_8(rng, k);
  19254. }
  19255. if (err == MP_OKAY) {
  19256. err = sp_256_ecc_mulmod_base_8(point, k, 1, NULL);
  19257. }
  19258. #ifdef WOLFSSL_VALIDATE_ECC_KEYGEN
  19259. if (err == MP_OKAY) {
  19260. err = sp_256_ecc_mulmod_8(infinity, point, p256_order, 1, NULL);
  19261. }
  19262. if (err == MP_OKAY) {
  19263. if ((sp_256_iszero_8(point->x) == 0) || (sp_256_iszero_8(point->y) == 0)) {
  19264. err = ECC_INF_E;
  19265. }
  19266. }
  19267. #endif
  19268. if (err == MP_OKAY) {
  19269. err = sp_256_to_mp(k, priv);
  19270. }
  19271. if (err == MP_OKAY) {
  19272. err = sp_256_point_to_ecc_point_8(point, pub);
  19273. }
  19274. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19275. if (k != NULL) {
  19276. XFREE(k, heap, DYNAMIC_TYPE_ECC);
  19277. }
  19278. #endif
  19279. #ifdef WOLFSSL_VALIDATE_ECC_KEYGEN
  19280. sp_ecc_point_free(infinity, 1, heap);
  19281. #endif
  19282. sp_ecc_point_free(point, 1, heap);
  19283. return err;
  19284. }
  19285. #ifdef HAVE_ECC_DHE
  19286. /* Write r as big endian to byte array.
  19287. * Fixed length number of bytes written: 32
  19288. *
  19289. * r A single precision integer.
  19290. * a Byte array.
  19291. */
  19292. static void sp_256_to_bin(sp_digit* r, byte* a)
  19293. {
  19294. int i, j, s = 0, b;
  19295. j = 256 / 8 - 1;
  19296. a[j] = 0;
  19297. for (i=0; i<8 && j>=0; i++) {
  19298. b = 0;
  19299. /* lint allow cast of mismatch sp_digit and int */
  19300. a[j--] |= (byte)(r[i] << s); b += 8 - s; /*lint !e9033*/
  19301. if (j < 0) {
  19302. break;
  19303. }
  19304. while (b < 32) {
  19305. a[j--] = r[i] >> b; b += 8;
  19306. if (j < 0) {
  19307. break;
  19308. }
  19309. }
  19310. s = 8 - (b - 32);
  19311. if (j >= 0) {
  19312. a[j] = 0;
  19313. }
  19314. if (s != 0) {
  19315. j++;
  19316. }
  19317. }
  19318. }
  19319. /* Multiply the point by the scalar and serialize the X ordinate.
  19320. * The number is 0 padded to maximum size on output.
  19321. *
  19322. * priv Scalar to multiply the point by.
  19323. * pub Point to multiply.
  19324. * out Buffer to hold X ordinate.
  19325. * outLen On entry, size of the buffer in bytes.
  19326. * On exit, length of data in buffer in bytes.
  19327. * heap Heap to use for allocation.
  19328. * returns BUFFER_E if the buffer is to small for output size,
  19329. * MEMORY_E when memory allocation fails and MP_OKAY on success.
  19330. */
  19331. int sp_ecc_secret_gen_256(mp_int* priv, ecc_point* pub, byte* out,
  19332. word32* outLen, void* heap)
  19333. {
  19334. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  19335. sp_point p;
  19336. sp_digit k[8];
  19337. #else
  19338. sp_digit* k = NULL;
  19339. #endif
  19340. sp_point* point = NULL;
  19341. int err = MP_OKAY;
  19342. if (*outLen < 32U) {
  19343. err = BUFFER_E;
  19344. }
  19345. if (err == MP_OKAY) {
  19346. err = sp_ecc_point_new(heap, p, point);
  19347. }
  19348. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19349. if (err == MP_OKAY) {
  19350. k = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8, heap,
  19351. DYNAMIC_TYPE_ECC);
  19352. if (k == NULL)
  19353. err = MEMORY_E;
  19354. }
  19355. #endif
  19356. if (err == MP_OKAY) {
  19357. sp_256_from_mp(k, 8, priv);
  19358. sp_256_point_from_ecc_point_8(point, pub);
  19359. err = sp_256_ecc_mulmod_8(point, point, k, 1, heap);
  19360. }
  19361. if (err == MP_OKAY) {
  19362. sp_256_to_bin(point->x, out);
  19363. *outLen = 32;
  19364. }
  19365. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19366. if (k != NULL) {
  19367. XFREE(k, heap, DYNAMIC_TYPE_ECC);
  19368. }
  19369. #endif
  19370. sp_ecc_point_free(point, 0, heap);
  19371. return err;
  19372. }
  19373. #endif /* HAVE_ECC_DHE */
  19374. #if defined(HAVE_ECC_SIGN) || defined(HAVE_ECC_VERIFY)
  19375. #endif
  19376. #if defined(HAVE_ECC_SIGN) || defined(HAVE_ECC_VERIFY)
  19377. #ifdef WOLFSSL_SP_SMALL
  19378. /* Sub b from a into a. (a -= b)
  19379. *
  19380. * a A single precision integer.
  19381. * b A single precision integer.
  19382. */
  19383. SP_NOINLINE static sp_digit sp_256_sub_in_place_8(sp_digit* a,
  19384. const sp_digit* b)
  19385. {
  19386. sp_digit c = 0;
  19387. __asm__ __volatile__ (
  19388. "mov r7, %[a]\n\t"
  19389. "add r7, #32\n\t"
  19390. "\n1:\n\t"
  19391. "mov r5, #0\n\t"
  19392. "sub r5, %[c]\n\t"
  19393. "ldr r3, [%[a]]\n\t"
  19394. "ldr r4, [%[a], #4]\n\t"
  19395. "ldr r5, [%[b]]\n\t"
  19396. "ldr r6, [%[b], #4]\n\t"
  19397. "sbc r3, r5\n\t"
  19398. "sbc r4, r6\n\t"
  19399. "str r3, [%[a]]\n\t"
  19400. "str r4, [%[a], #4]\n\t"
  19401. "sbc %[c], %[c]\n\t"
  19402. "add %[a], #8\n\t"
  19403. "add %[b], #8\n\t"
  19404. "cmp %[a], r7\n\t"
  19405. "bne 1b\n\t"
  19406. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  19407. :
  19408. : "memory", "r3", "r4", "r5", "r6", "r7"
  19409. );
  19410. return c;
  19411. }
  19412. #else
  19413. /* Sub b from a into r. (r = a - b)
  19414. *
  19415. * r A single precision integer.
  19416. * a A single precision integer.
  19417. * b A single precision integer.
  19418. */
  19419. SP_NOINLINE static sp_digit sp_256_sub_in_place_8(sp_digit* a,
  19420. const sp_digit* b)
  19421. {
  19422. sp_digit c = 0;
  19423. __asm__ __volatile__ (
  19424. "ldr r3, [%[a], #0]\n\t"
  19425. "ldr r4, [%[a], #4]\n\t"
  19426. "ldr r5, [%[b], #0]\n\t"
  19427. "ldr r6, [%[b], #4]\n\t"
  19428. "sub r3, r5\n\t"
  19429. "sbc r4, r6\n\t"
  19430. "str r3, [%[a], #0]\n\t"
  19431. "str r4, [%[a], #4]\n\t"
  19432. "ldr r3, [%[a], #8]\n\t"
  19433. "ldr r4, [%[a], #12]\n\t"
  19434. "ldr r5, [%[b], #8]\n\t"
  19435. "ldr r6, [%[b], #12]\n\t"
  19436. "sbc r3, r5\n\t"
  19437. "sbc r4, r6\n\t"
  19438. "str r3, [%[a], #8]\n\t"
  19439. "str r4, [%[a], #12]\n\t"
  19440. "ldr r3, [%[a], #16]\n\t"
  19441. "ldr r4, [%[a], #20]\n\t"
  19442. "ldr r5, [%[b], #16]\n\t"
  19443. "ldr r6, [%[b], #20]\n\t"
  19444. "sbc r3, r5\n\t"
  19445. "sbc r4, r6\n\t"
  19446. "str r3, [%[a], #16]\n\t"
  19447. "str r4, [%[a], #20]\n\t"
  19448. "ldr r3, [%[a], #24]\n\t"
  19449. "ldr r4, [%[a], #28]\n\t"
  19450. "ldr r5, [%[b], #24]\n\t"
  19451. "ldr r6, [%[b], #28]\n\t"
  19452. "sbc r3, r5\n\t"
  19453. "sbc r4, r6\n\t"
  19454. "str r3, [%[a], #24]\n\t"
  19455. "str r4, [%[a], #28]\n\t"
  19456. "sbc %[c], %[c]\n\t"
  19457. : [c] "+r" (c), [a] "+r" (a), [b] "+r" (b)
  19458. :
  19459. : "memory", "r3", "r4", "r5", "r6"
  19460. );
  19461. return c;
  19462. }
  19463. #endif /* WOLFSSL_SP_SMALL */
  19464. /* Mul a by digit b into r. (r = a * b)
  19465. *
  19466. * r A single precision integer.
  19467. * a A single precision integer.
  19468. * b A single precision digit.
  19469. */
  19470. SP_NOINLINE static void sp_256_mul_d_8(sp_digit* r, const sp_digit* a,
  19471. sp_digit b)
  19472. {
  19473. __asm__ __volatile__ (
  19474. "mov r6, #32\n\t"
  19475. "add r6, %[a]\n\t"
  19476. "mov r8, %[r]\n\t"
  19477. "mov r9, r6\n\t"
  19478. "mov r3, #0\n\t"
  19479. "mov r4, #0\n\t"
  19480. "1:\n\t"
  19481. "mov %[r], #0\n\t"
  19482. "mov r5, #0\n\t"
  19483. "# A[] * B\n\t"
  19484. "ldr r6, [%[a]]\n\t"
  19485. "lsl r6, r6, #16\n\t"
  19486. "lsl r7, %[b], #16\n\t"
  19487. "lsr r6, r6, #16\n\t"
  19488. "lsr r7, r7, #16\n\t"
  19489. "mul r7, r6\n\t"
  19490. "add r3, r7\n\t"
  19491. "adc r4, %[r]\n\t"
  19492. "adc r5, %[r]\n\t"
  19493. "lsr r7, %[b], #16\n\t"
  19494. "mul r6, r7\n\t"
  19495. "lsr r7, r6, #16\n\t"
  19496. "lsl r6, r6, #16\n\t"
  19497. "add r3, r6\n\t"
  19498. "adc r4, r7\n\t"
  19499. "adc r5, %[r]\n\t"
  19500. "ldr r6, [%[a]]\n\t"
  19501. "lsr r6, r6, #16\n\t"
  19502. "lsr r7, %[b], #16\n\t"
  19503. "mul r7, r6\n\t"
  19504. "add r4, r7\n\t"
  19505. "adc r5, %[r]\n\t"
  19506. "lsl r7, %[b], #16\n\t"
  19507. "lsr r7, r7, #16\n\t"
  19508. "mul r6, r7\n\t"
  19509. "lsr r7, r6, #16\n\t"
  19510. "lsl r6, r6, #16\n\t"
  19511. "add r3, r6\n\t"
  19512. "adc r4, r7\n\t"
  19513. "adc r5, %[r]\n\t"
  19514. "# A[] * B - Done\n\t"
  19515. "mov %[r], r8\n\t"
  19516. "str r3, [%[r]]\n\t"
  19517. "mov r3, r4\n\t"
  19518. "mov r4, r5\n\t"
  19519. "add %[r], #4\n\t"
  19520. "add %[a], #4\n\t"
  19521. "mov r8, %[r]\n\t"
  19522. "cmp %[a], r9\n\t"
  19523. "blt 1b\n\t"
  19524. "str r3, [%[r]]\n\t"
  19525. : [r] "+r" (r), [a] "+r" (a)
  19526. : [b] "r" (b)
  19527. : "memory", "r3", "r4", "r5", "r6", "r7", "r8", "r9"
  19528. );
  19529. }
  19530. /* Divide the double width number (d1|d0) by the dividend. (d1|d0 / div)
  19531. *
  19532. * d1 The high order half of the number to divide.
  19533. * d0 The low order half of the number to divide.
  19534. * div The dividend.
  19535. * returns the result of the division.
  19536. *
  19537. * Note that this is an approximate div. It may give an answer 1 larger.
  19538. */
  19539. SP_NOINLINE static sp_digit div_256_word_8(sp_digit d1, sp_digit d0,
  19540. sp_digit div)
  19541. {
  19542. sp_digit r = 0;
  19543. __asm__ __volatile__ (
  19544. "lsr r5, %[div], #1\n\t"
  19545. "add r5, #1\n\t"
  19546. "mov r8, %[d0]\n\t"
  19547. "mov r9, %[d1]\n\t"
  19548. "# Do top 32\n\t"
  19549. "mov r6, r5\n\t"
  19550. "sub r6, %[d1]\n\t"
  19551. "sbc r6, r6\n\t"
  19552. "add %[r], %[r]\n\t"
  19553. "sub %[r], r6\n\t"
  19554. "and r6, r5\n\t"
  19555. "sub %[d1], r6\n\t"
  19556. "# Next 30 bits\n\t"
  19557. "mov r4, #29\n\t"
  19558. "1:\n\t"
  19559. "lsl %[d0], %[d0], #1\n\t"
  19560. "adc %[d1], %[d1]\n\t"
  19561. "mov r6, r5\n\t"
  19562. "sub r6, %[d1]\n\t"
  19563. "sbc r6, r6\n\t"
  19564. "add %[r], %[r]\n\t"
  19565. "sub %[r], r6\n\t"
  19566. "and r6, r5\n\t"
  19567. "sub %[d1], r6\n\t"
  19568. "sub r4, #1\n\t"
  19569. "bpl 1b\n\t"
  19570. "mov r7, #0\n\t"
  19571. "add %[r], %[r]\n\t"
  19572. "add %[r], #1\n\t"
  19573. "# r * div - Start\n\t"
  19574. "lsl %[d1], %[r], #16\n\t"
  19575. "lsl r4, %[div], #16\n\t"
  19576. "lsr %[d1], %[d1], #16\n\t"
  19577. "lsr r4, r4, #16\n\t"
  19578. "mul r4, %[d1]\n\t"
  19579. "lsr r6, %[div], #16\n\t"
  19580. "mul %[d1], r6\n\t"
  19581. "lsr r5, %[d1], #16\n\t"
  19582. "lsl %[d1], %[d1], #16\n\t"
  19583. "add r4, %[d1]\n\t"
  19584. "adc r5, r7\n\t"
  19585. "lsr %[d1], %[r], #16\n\t"
  19586. "mul r6, %[d1]\n\t"
  19587. "add r5, r6\n\t"
  19588. "lsl r6, %[div], #16\n\t"
  19589. "lsr r6, r6, #16\n\t"
  19590. "mul %[d1], r6\n\t"
  19591. "lsr r6, %[d1], #16\n\t"
  19592. "lsl %[d1], %[d1], #16\n\t"
  19593. "add r4, %[d1]\n\t"
  19594. "adc r5, r6\n\t"
  19595. "# r * div - Done\n\t"
  19596. "mov %[d1], r8\n\t"
  19597. "sub %[d1], r4\n\t"
  19598. "mov r4, %[d1]\n\t"
  19599. "mov %[d1], r9\n\t"
  19600. "sbc %[d1], r5\n\t"
  19601. "mov r5, %[d1]\n\t"
  19602. "add %[r], r5\n\t"
  19603. "# r * div - Start\n\t"
  19604. "lsl %[d1], %[r], #16\n\t"
  19605. "lsl r4, %[div], #16\n\t"
  19606. "lsr %[d1], %[d1], #16\n\t"
  19607. "lsr r4, r4, #16\n\t"
  19608. "mul r4, %[d1]\n\t"
  19609. "lsr r6, %[div], #16\n\t"
  19610. "mul %[d1], r6\n\t"
  19611. "lsr r5, %[d1], #16\n\t"
  19612. "lsl %[d1], %[d1], #16\n\t"
  19613. "add r4, %[d1]\n\t"
  19614. "adc r5, r7\n\t"
  19615. "lsr %[d1], %[r], #16\n\t"
  19616. "mul r6, %[d1]\n\t"
  19617. "add r5, r6\n\t"
  19618. "lsl r6, %[div], #16\n\t"
  19619. "lsr r6, r6, #16\n\t"
  19620. "mul %[d1], r6\n\t"
  19621. "lsr r6, %[d1], #16\n\t"
  19622. "lsl %[d1], %[d1], #16\n\t"
  19623. "add r4, %[d1]\n\t"
  19624. "adc r5, r6\n\t"
  19625. "# r * div - Done\n\t"
  19626. "mov %[d1], r8\n\t"
  19627. "mov r6, r9\n\t"
  19628. "sub r4, %[d1], r4\n\t"
  19629. "sbc r6, r5\n\t"
  19630. "mov r5, r6\n\t"
  19631. "add %[r], r5\n\t"
  19632. "# r * div - Start\n\t"
  19633. "lsl %[d1], %[r], #16\n\t"
  19634. "lsl r4, %[div], #16\n\t"
  19635. "lsr %[d1], %[d1], #16\n\t"
  19636. "lsr r4, r4, #16\n\t"
  19637. "mul r4, %[d1]\n\t"
  19638. "lsr r6, %[div], #16\n\t"
  19639. "mul %[d1], r6\n\t"
  19640. "lsr r5, %[d1], #16\n\t"
  19641. "lsl %[d1], %[d1], #16\n\t"
  19642. "add r4, %[d1]\n\t"
  19643. "adc r5, r7\n\t"
  19644. "lsr %[d1], %[r], #16\n\t"
  19645. "mul r6, %[d1]\n\t"
  19646. "add r5, r6\n\t"
  19647. "lsl r6, %[div], #16\n\t"
  19648. "lsr r6, r6, #16\n\t"
  19649. "mul %[d1], r6\n\t"
  19650. "lsr r6, %[d1], #16\n\t"
  19651. "lsl %[d1], %[d1], #16\n\t"
  19652. "add r4, %[d1]\n\t"
  19653. "adc r5, r6\n\t"
  19654. "# r * div - Done\n\t"
  19655. "mov %[d1], r8\n\t"
  19656. "mov r6, r9\n\t"
  19657. "sub r4, %[d1], r4\n\t"
  19658. "sbc r6, r5\n\t"
  19659. "mov r5, r6\n\t"
  19660. "add %[r], r5\n\t"
  19661. "mov r6, %[div]\n\t"
  19662. "sub r6, r4\n\t"
  19663. "sbc r6, r6\n\t"
  19664. "sub %[r], r6\n\t"
  19665. : [r] "+r" (r)
  19666. : [d1] "r" (d1), [d0] "r" (d0), [div] "r" (div)
  19667. : "r4", "r5", "r7", "r6", "r8", "r9"
  19668. );
  19669. return r;
  19670. }
  19671. /* AND m into each word of a and store in r.
  19672. *
  19673. * r A single precision integer.
  19674. * a A single precision integer.
  19675. * m Mask to AND against each digit.
  19676. */
  19677. static void sp_256_mask_8(sp_digit* r, const sp_digit* a, sp_digit m)
  19678. {
  19679. #ifdef WOLFSSL_SP_SMALL
  19680. int i;
  19681. for (i=0; i<8; i++) {
  19682. r[i] = a[i] & m;
  19683. }
  19684. #else
  19685. r[0] = a[0] & m;
  19686. r[1] = a[1] & m;
  19687. r[2] = a[2] & m;
  19688. r[3] = a[3] & m;
  19689. r[4] = a[4] & m;
  19690. r[5] = a[5] & m;
  19691. r[6] = a[6] & m;
  19692. r[7] = a[7] & m;
  19693. #endif
  19694. }
  19695. /* Divide d in a and put remainder into r (m*d + r = a)
  19696. * m is not calculated as it is not needed at this time.
  19697. *
  19698. * a Nmber to be divided.
  19699. * d Number to divide with.
  19700. * m Multiplier result.
  19701. * r Remainder from the division.
  19702. * returns MP_OKAY indicating success.
  19703. */
  19704. static WC_INLINE int sp_256_div_8(const sp_digit* a, const sp_digit* d, sp_digit* m,
  19705. sp_digit* r)
  19706. {
  19707. sp_digit t1[16], t2[9];
  19708. sp_digit div, r1;
  19709. int i;
  19710. (void)m;
  19711. div = d[7];
  19712. XMEMCPY(t1, a, sizeof(*t1) * 2 * 8);
  19713. for (i=7; i>=0; i--) {
  19714. r1 = div_256_word_8(t1[8 + i], t1[8 + i - 1], div);
  19715. sp_256_mul_d_8(t2, d, r1);
  19716. t1[8 + i] += sp_256_sub_in_place_8(&t1[i], t2);
  19717. t1[8 + i] -= t2[8];
  19718. sp_256_mask_8(t2, d, t1[8 + i]);
  19719. t1[8 + i] += sp_256_add_8(&t1[i], &t1[i], t2);
  19720. sp_256_mask_8(t2, d, t1[8 + i]);
  19721. t1[8 + i] += sp_256_add_8(&t1[i], &t1[i], t2);
  19722. }
  19723. r1 = sp_256_cmp_8(t1, d) >= 0;
  19724. sp_256_cond_sub_8(r, t1, d, (sp_digit)0 - r1);
  19725. return MP_OKAY;
  19726. }
  19727. /* Reduce a modulo m into r. (r = a mod m)
  19728. *
  19729. * r A single precision number that is the reduced result.
  19730. * a A single precision number that is to be reduced.
  19731. * m A single precision number that is the modulus to reduce with.
  19732. * returns MP_OKAY indicating success.
  19733. */
  19734. static WC_INLINE int sp_256_mod_8(sp_digit* r, const sp_digit* a, const sp_digit* m)
  19735. {
  19736. return sp_256_div_8(a, m, NULL, r);
  19737. }
  19738. #endif
  19739. #if defined(HAVE_ECC_SIGN) || defined(HAVE_ECC_VERIFY)
  19740. #ifdef WOLFSSL_SP_SMALL
  19741. /* Order-2 for the P256 curve. */
  19742. static const uint32_t p256_order_2[8] = {
  19743. 0xfc63254fU,0xf3b9cac2U,0xa7179e84U,0xbce6faadU,0xffffffffU,0xffffffffU,
  19744. 0x00000000U,0xffffffffU
  19745. };
  19746. #else
  19747. /* The low half of the order-2 of the P256 curve. */
  19748. static const uint32_t p256_order_low[4] = {
  19749. 0xfc63254fU,0xf3b9cac2U,0xa7179e84U,0xbce6faadU
  19750. };
  19751. #endif /* WOLFSSL_SP_SMALL */
  19752. /* Multiply two number mod the order of P256 curve. (r = a * b mod order)
  19753. *
  19754. * r Result of the multiplication.
  19755. * a First operand of the multiplication.
  19756. * b Second operand of the multiplication.
  19757. */
  19758. static void sp_256_mont_mul_order_8(sp_digit* r, const sp_digit* a, const sp_digit* b)
  19759. {
  19760. sp_256_mul_8(r, a, b);
  19761. sp_256_mont_reduce_order_8(r, p256_order, p256_mp_order);
  19762. }
  19763. /* Square number mod the order of P256 curve. (r = a * a mod order)
  19764. *
  19765. * r Result of the squaring.
  19766. * a Number to square.
  19767. */
  19768. static void sp_256_mont_sqr_order_8(sp_digit* r, const sp_digit* a)
  19769. {
  19770. sp_256_sqr_8(r, a);
  19771. sp_256_mont_reduce_order_8(r, p256_order, p256_mp_order);
  19772. }
  19773. #ifndef WOLFSSL_SP_SMALL
  19774. /* Square number mod the order of P256 curve a number of times.
  19775. * (r = a ^ n mod order)
  19776. *
  19777. * r Result of the squaring.
  19778. * a Number to square.
  19779. */
  19780. static void sp_256_mont_sqr_n_order_8(sp_digit* r, const sp_digit* a, int n)
  19781. {
  19782. int i;
  19783. sp_256_mont_sqr_order_8(r, a);
  19784. for (i=1; i<n; i++) {
  19785. sp_256_mont_sqr_order_8(r, r);
  19786. }
  19787. }
  19788. #endif /* !WOLFSSL_SP_SMALL */
  19789. /* Invert the number, in Montgomery form, modulo the order of the P256 curve.
  19790. * (r = 1 / a mod order)
  19791. *
  19792. * r Inverse result.
  19793. * a Number to invert.
  19794. * td Temporary data.
  19795. */
  19796. static void sp_256_mont_inv_order_8(sp_digit* r, const sp_digit* a,
  19797. sp_digit* td)
  19798. {
  19799. #ifdef WOLFSSL_SP_SMALL
  19800. sp_digit* t = td;
  19801. int i;
  19802. XMEMCPY(t, a, sizeof(sp_digit) * 8);
  19803. for (i=254; i>=0; i--) {
  19804. sp_256_mont_sqr_order_8(t, t);
  19805. if ((p256_order_2[i / 32] & ((sp_int_digit)1 << (i % 32))) != 0) {
  19806. sp_256_mont_mul_order_8(t, t, a);
  19807. }
  19808. }
  19809. XMEMCPY(r, t, sizeof(sp_digit) * 8U);
  19810. #else
  19811. sp_digit* t = td;
  19812. sp_digit* t2 = td + 2 * 8;
  19813. sp_digit* t3 = td + 4 * 8;
  19814. int i;
  19815. /* t = a^2 */
  19816. sp_256_mont_sqr_order_8(t, a);
  19817. /* t = a^3 = t * a */
  19818. sp_256_mont_mul_order_8(t, t, a);
  19819. /* t2= a^c = t ^ 2 ^ 2 */
  19820. sp_256_mont_sqr_n_order_8(t2, t, 2);
  19821. /* t3= a^f = t2 * t */
  19822. sp_256_mont_mul_order_8(t3, t2, t);
  19823. /* t2= a^f0 = t3 ^ 2 ^ 4 */
  19824. sp_256_mont_sqr_n_order_8(t2, t3, 4);
  19825. /* t = a^ff = t2 * t3 */
  19826. sp_256_mont_mul_order_8(t, t2, t3);
  19827. /* t3= a^ff00 = t ^ 2 ^ 8 */
  19828. sp_256_mont_sqr_n_order_8(t2, t, 8);
  19829. /* t = a^ffff = t2 * t */
  19830. sp_256_mont_mul_order_8(t, t2, t);
  19831. /* t2= a^ffff0000 = t ^ 2 ^ 16 */
  19832. sp_256_mont_sqr_n_order_8(t2, t, 16);
  19833. /* t = a^ffffffff = t2 * t */
  19834. sp_256_mont_mul_order_8(t, t2, t);
  19835. /* t2= a^ffffffff0000000000000000 = t ^ 2 ^ 64 */
  19836. sp_256_mont_sqr_n_order_8(t2, t, 64);
  19837. /* t2= a^ffffffff00000000ffffffff = t2 * t */
  19838. sp_256_mont_mul_order_8(t2, t2, t);
  19839. /* t2= a^ffffffff00000000ffffffff00000000 = t2 ^ 2 ^ 32 */
  19840. sp_256_mont_sqr_n_order_8(t2, t2, 32);
  19841. /* t2= a^ffffffff00000000ffffffffffffffff = t2 * t */
  19842. sp_256_mont_mul_order_8(t2, t2, t);
  19843. /* t2= a^ffffffff00000000ffffffffffffffffbce6 */
  19844. for (i=127; i>=112; i--) {
  19845. sp_256_mont_sqr_order_8(t2, t2);
  19846. if (((sp_digit)p256_order_low[i / 32] & ((sp_int_digit)1 << (i % 32))) != 0) {
  19847. sp_256_mont_mul_order_8(t2, t2, a);
  19848. }
  19849. }
  19850. /* t2= a^ffffffff00000000ffffffffffffffffbce6f */
  19851. sp_256_mont_sqr_n_order_8(t2, t2, 4);
  19852. sp_256_mont_mul_order_8(t2, t2, t3);
  19853. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84 */
  19854. for (i=107; i>=64; i--) {
  19855. sp_256_mont_sqr_order_8(t2, t2);
  19856. if (((sp_digit)p256_order_low[i / 32] & ((sp_int_digit)1 << (i % 32))) != 0) {
  19857. sp_256_mont_mul_order_8(t2, t2, a);
  19858. }
  19859. }
  19860. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84f */
  19861. sp_256_mont_sqr_n_order_8(t2, t2, 4);
  19862. sp_256_mont_mul_order_8(t2, t2, t3);
  19863. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84f3b9cac2 */
  19864. for (i=59; i>=32; i--) {
  19865. sp_256_mont_sqr_order_8(t2, t2);
  19866. if (((sp_digit)p256_order_low[i / 32] & ((sp_int_digit)1 << (i % 32))) != 0) {
  19867. sp_256_mont_mul_order_8(t2, t2, a);
  19868. }
  19869. }
  19870. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84f3b9cac2f */
  19871. sp_256_mont_sqr_n_order_8(t2, t2, 4);
  19872. sp_256_mont_mul_order_8(t2, t2, t3);
  19873. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84f3b9cac2fc63254 */
  19874. for (i=27; i>=0; i--) {
  19875. sp_256_mont_sqr_order_8(t2, t2);
  19876. if (((sp_digit)p256_order_low[i / 32] & ((sp_int_digit)1 << (i % 32))) != 0) {
  19877. sp_256_mont_mul_order_8(t2, t2, a);
  19878. }
  19879. }
  19880. /* t2= a^ffffffff00000000ffffffffffffffffbce6faada7179e84f3b9cac2fc632540 */
  19881. sp_256_mont_sqr_n_order_8(t2, t2, 4);
  19882. /* r = a^ffffffff00000000ffffffffffffffffbce6faada7179e84f3b9cac2fc63254f */
  19883. sp_256_mont_mul_order_8(r, t2, t3);
  19884. #endif /* WOLFSSL_SP_SMALL */
  19885. }
  19886. #endif /* HAVE_ECC_SIGN || HAVE_ECC_VERIFY */
  19887. #ifdef HAVE_ECC_SIGN
  19888. #ifndef SP_ECC_MAX_SIG_GEN
  19889. #define SP_ECC_MAX_SIG_GEN 64
  19890. #endif
  19891. /* Sign the hash using the private key.
  19892. * e = [hash, 256 bits] from binary
  19893. * r = (k.G)->x mod order
  19894. * s = (r * x + e) / k mod order
  19895. * The hash is truncated to the first 256 bits.
  19896. *
  19897. * hash Hash to sign.
  19898. * hashLen Length of the hash data.
  19899. * rng Random number generator.
  19900. * priv Private part of key - scalar.
  19901. * rm First part of result as an mp_int.
  19902. * sm Sirst part of result as an mp_int.
  19903. * heap Heap to use for allocation.
  19904. * returns RNG failures, MEMORY_E when memory allocation fails and
  19905. * MP_OKAY on success.
  19906. */
  19907. int sp_ecc_sign_256(const byte* hash, word32 hashLen, WC_RNG* rng, mp_int* priv,
  19908. mp_int* rm, mp_int* sm, mp_int* km, void* heap)
  19909. {
  19910. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19911. sp_digit* d = NULL;
  19912. #else
  19913. sp_digit ed[2*8];
  19914. sp_digit xd[2*8];
  19915. sp_digit kd[2*8];
  19916. sp_digit rd[2*8];
  19917. sp_digit td[3 * 2*8];
  19918. sp_point p;
  19919. #endif
  19920. sp_digit* e = NULL;
  19921. sp_digit* x = NULL;
  19922. sp_digit* k = NULL;
  19923. sp_digit* r = NULL;
  19924. sp_digit* tmp = NULL;
  19925. sp_point* point = NULL;
  19926. sp_digit carry;
  19927. sp_digit* s = NULL;
  19928. sp_digit* kInv = NULL;
  19929. int err;
  19930. int32_t c;
  19931. int i;
  19932. (void)heap;
  19933. err = sp_ecc_point_new(heap, p, point);
  19934. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19935. if (err == MP_OKAY) {
  19936. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 7 * 2 * 8, heap,
  19937. DYNAMIC_TYPE_ECC);
  19938. if (d == NULL) {
  19939. err = MEMORY_E;
  19940. }
  19941. }
  19942. #endif
  19943. if (err == MP_OKAY) {
  19944. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  19945. e = d + 0 * 8;
  19946. x = d + 2 * 8;
  19947. k = d + 4 * 8;
  19948. r = d + 6 * 8;
  19949. tmp = d + 8 * 8;
  19950. #else
  19951. e = ed;
  19952. x = xd;
  19953. k = kd;
  19954. r = rd;
  19955. tmp = td;
  19956. #endif
  19957. s = e;
  19958. kInv = k;
  19959. if (hashLen > 32U) {
  19960. hashLen = 32U;
  19961. }
  19962. sp_256_from_bin(e, 8, hash, (int)hashLen);
  19963. }
  19964. for (i = SP_ECC_MAX_SIG_GEN; err == MP_OKAY && i > 0; i--) {
  19965. sp_256_from_mp(x, 8, priv);
  19966. /* New random point. */
  19967. if (km == NULL || mp_iszero(km)) {
  19968. err = sp_256_ecc_gen_k_8(rng, k);
  19969. }
  19970. else {
  19971. sp_256_from_mp(k, 8, km);
  19972. mp_zero(km);
  19973. }
  19974. if (err == MP_OKAY) {
  19975. err = sp_256_ecc_mulmod_base_8(point, k, 1, NULL);
  19976. }
  19977. if (err == MP_OKAY) {
  19978. /* r = point->x mod order */
  19979. XMEMCPY(r, point->x, sizeof(sp_digit) * 8U);
  19980. sp_256_norm_8(r);
  19981. c = sp_256_cmp_8(r, p256_order);
  19982. sp_256_cond_sub_8(r, r, p256_order, 0L - (sp_digit)(c >= 0));
  19983. sp_256_norm_8(r);
  19984. /* Conv k to Montgomery form (mod order) */
  19985. sp_256_mul_8(k, k, p256_norm_order);
  19986. err = sp_256_mod_8(k, k, p256_order);
  19987. }
  19988. if (err == MP_OKAY) {
  19989. sp_256_norm_8(k);
  19990. /* kInv = 1/k mod order */
  19991. sp_256_mont_inv_order_8(kInv, k, tmp);
  19992. sp_256_norm_8(kInv);
  19993. /* s = r * x + e */
  19994. sp_256_mul_8(x, x, r);
  19995. err = sp_256_mod_8(x, x, p256_order);
  19996. }
  19997. if (err == MP_OKAY) {
  19998. sp_256_norm_8(x);
  19999. carry = sp_256_add_8(s, e, x);
  20000. sp_256_cond_sub_8(s, s, p256_order, 0 - carry);
  20001. sp_256_norm_8(s);
  20002. c = sp_256_cmp_8(s, p256_order);
  20003. sp_256_cond_sub_8(s, s, p256_order, 0L - (sp_digit)(c >= 0));
  20004. sp_256_norm_8(s);
  20005. /* s = s * k^-1 mod order */
  20006. sp_256_mont_mul_order_8(s, s, kInv);
  20007. sp_256_norm_8(s);
  20008. /* Check that signature is usable. */
  20009. if (sp_256_iszero_8(s) == 0) {
  20010. break;
  20011. }
  20012. }
  20013. }
  20014. if (i == 0) {
  20015. err = RNG_FAILURE_E;
  20016. }
  20017. if (err == MP_OKAY) {
  20018. err = sp_256_to_mp(r, rm);
  20019. }
  20020. if (err == MP_OKAY) {
  20021. err = sp_256_to_mp(s, sm);
  20022. }
  20023. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20024. if (d != NULL) {
  20025. XMEMSET(d, 0, sizeof(sp_digit) * 8 * 8);
  20026. XFREE(d, heap, DYNAMIC_TYPE_ECC);
  20027. }
  20028. #else
  20029. XMEMSET(e, 0, sizeof(sp_digit) * 2U * 8U);
  20030. XMEMSET(x, 0, sizeof(sp_digit) * 2U * 8U);
  20031. XMEMSET(k, 0, sizeof(sp_digit) * 2U * 8U);
  20032. XMEMSET(r, 0, sizeof(sp_digit) * 2U * 8U);
  20033. XMEMSET(r, 0, sizeof(sp_digit) * 2U * 8U);
  20034. XMEMSET(tmp, 0, sizeof(sp_digit) * 3U * 2U * 8U);
  20035. #endif
  20036. sp_ecc_point_free(point, 1, heap);
  20037. return err;
  20038. }
  20039. #endif /* HAVE_ECC_SIGN */
  20040. #ifdef HAVE_ECC_VERIFY
  20041. /* Verify the signature values with the hash and public key.
  20042. * e = Truncate(hash, 256)
  20043. * u1 = e/s mod order
  20044. * u2 = r/s mod order
  20045. * r == (u1.G + u2.Q)->x mod order
  20046. * Optimization: Leave point in projective form.
  20047. * (x, y, 1) == (x' / z'*z', y' / z'*z'*z', z' / z')
  20048. * (r + n*order).z'.z' mod prime == (u1.G + u2.Q)->x'
  20049. * The hash is truncated to the first 256 bits.
  20050. *
  20051. * hash Hash to sign.
  20052. * hashLen Length of the hash data.
  20053. * rng Random number generator.
  20054. * priv Private part of key - scalar.
  20055. * rm First part of result as an mp_int.
  20056. * sm Sirst part of result as an mp_int.
  20057. * heap Heap to use for allocation.
  20058. * returns RNG failures, MEMORY_E when memory allocation fails and
  20059. * MP_OKAY on success.
  20060. */
  20061. int sp_ecc_verify_256(const byte* hash, word32 hashLen, mp_int* pX,
  20062. mp_int* pY, mp_int* pZ, mp_int* r, mp_int* sm, int* res, void* heap)
  20063. {
  20064. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20065. sp_digit* d = NULL;
  20066. #else
  20067. sp_digit u1d[2*8];
  20068. sp_digit u2d[2*8];
  20069. sp_digit sd[2*8];
  20070. sp_digit tmpd[2*8 * 5];
  20071. sp_point p1d;
  20072. sp_point p2d;
  20073. #endif
  20074. sp_digit* u1 = NULL;
  20075. sp_digit* u2 = NULL;
  20076. sp_digit* s = NULL;
  20077. sp_digit* tmp = NULL;
  20078. sp_point* p1;
  20079. sp_point* p2 = NULL;
  20080. sp_digit carry;
  20081. int32_t c;
  20082. int err;
  20083. err = sp_ecc_point_new(heap, p1d, p1);
  20084. if (err == MP_OKAY) {
  20085. err = sp_ecc_point_new(heap, p2d, p2);
  20086. }
  20087. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20088. if (err == MP_OKAY) {
  20089. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 16 * 8, heap,
  20090. DYNAMIC_TYPE_ECC);
  20091. if (d == NULL) {
  20092. err = MEMORY_E;
  20093. }
  20094. }
  20095. #endif
  20096. if (err == MP_OKAY) {
  20097. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20098. u1 = d + 0 * 8;
  20099. u2 = d + 2 * 8;
  20100. s = d + 4 * 8;
  20101. tmp = d + 6 * 8;
  20102. #else
  20103. u1 = u1d;
  20104. u2 = u2d;
  20105. s = sd;
  20106. tmp = tmpd;
  20107. #endif
  20108. if (hashLen > 32U) {
  20109. hashLen = 32U;
  20110. }
  20111. sp_256_from_bin(u1, 8, hash, (int)hashLen);
  20112. sp_256_from_mp(u2, 8, r);
  20113. sp_256_from_mp(s, 8, sm);
  20114. sp_256_from_mp(p2->x, 8, pX);
  20115. sp_256_from_mp(p2->y, 8, pY);
  20116. sp_256_from_mp(p2->z, 8, pZ);
  20117. {
  20118. sp_256_mul_8(s, s, p256_norm_order);
  20119. }
  20120. err = sp_256_mod_8(s, s, p256_order);
  20121. }
  20122. if (err == MP_OKAY) {
  20123. sp_256_norm_8(s);
  20124. {
  20125. sp_256_mont_inv_order_8(s, s, tmp);
  20126. sp_256_mont_mul_order_8(u1, u1, s);
  20127. sp_256_mont_mul_order_8(u2, u2, s);
  20128. }
  20129. err = sp_256_ecc_mulmod_base_8(p1, u1, 0, heap);
  20130. }
  20131. if (err == MP_OKAY) {
  20132. err = sp_256_ecc_mulmod_8(p2, p2, u2, 0, heap);
  20133. }
  20134. if (err == MP_OKAY) {
  20135. {
  20136. sp_256_proj_point_add_8(p1, p1, p2, tmp);
  20137. if (sp_256_iszero_8(p1->z)) {
  20138. if (sp_256_iszero_8(p1->x) && sp_256_iszero_8(p1->y)) {
  20139. sp_256_proj_point_dbl_8(p1, p2, tmp);
  20140. }
  20141. else {
  20142. /* Y ordinate is not used from here - don't set. */
  20143. p1->x[0] = 0;
  20144. p1->x[1] = 0;
  20145. p1->x[2] = 0;
  20146. p1->x[3] = 0;
  20147. p1->x[4] = 0;
  20148. p1->x[5] = 0;
  20149. p1->x[6] = 0;
  20150. p1->x[7] = 0;
  20151. XMEMCPY(p1->z, p256_norm_mod, sizeof(p256_norm_mod));
  20152. }
  20153. }
  20154. }
  20155. /* (r + n*order).z'.z' mod prime == (u1.G + u2.Q)->x' */
  20156. /* Reload r and convert to Montgomery form. */
  20157. sp_256_from_mp(u2, 8, r);
  20158. err = sp_256_mod_mul_norm_8(u2, u2, p256_mod);
  20159. }
  20160. if (err == MP_OKAY) {
  20161. /* u1 = r.z'.z' mod prime */
  20162. sp_256_mont_sqr_8(p1->z, p1->z, p256_mod, p256_mp_mod);
  20163. sp_256_mont_mul_8(u1, u2, p1->z, p256_mod, p256_mp_mod);
  20164. *res = (int)(sp_256_cmp_8(p1->x, u1) == 0);
  20165. if (*res == 0) {
  20166. /* Reload r and add order. */
  20167. sp_256_from_mp(u2, 8, r);
  20168. carry = sp_256_add_8(u2, u2, p256_order);
  20169. /* Carry means result is greater than mod and is not valid. */
  20170. if (carry == 0) {
  20171. sp_256_norm_8(u2);
  20172. /* Compare with mod and if greater or equal then not valid. */
  20173. c = sp_256_cmp_8(u2, p256_mod);
  20174. if (c < 0) {
  20175. /* Convert to Montogomery form */
  20176. err = sp_256_mod_mul_norm_8(u2, u2, p256_mod);
  20177. if (err == MP_OKAY) {
  20178. /* u1 = (r + 1*order).z'.z' mod prime */
  20179. sp_256_mont_mul_8(u1, u2, p1->z, p256_mod,
  20180. p256_mp_mod);
  20181. *res = (int)(sp_256_cmp_8(p1->x, u1) == 0);
  20182. }
  20183. }
  20184. }
  20185. }
  20186. }
  20187. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20188. if (d != NULL)
  20189. XFREE(d, heap, DYNAMIC_TYPE_ECC);
  20190. #endif
  20191. sp_ecc_point_free(p1, 0, heap);
  20192. sp_ecc_point_free(p2, 0, heap);
  20193. return err;
  20194. }
  20195. #endif /* HAVE_ECC_VERIFY */
  20196. #ifdef HAVE_ECC_CHECK_KEY
  20197. /* Check that the x and y oridinates are a valid point on the curve.
  20198. *
  20199. * point EC point.
  20200. * heap Heap to use if dynamically allocating.
  20201. * returns MEMORY_E if dynamic memory allocation fails, MP_VAL if the point is
  20202. * not on the curve and MP_OKAY otherwise.
  20203. */
  20204. static int sp_256_ecc_is_point_8(sp_point* point, void* heap)
  20205. {
  20206. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20207. sp_digit* d;
  20208. #else
  20209. sp_digit t1d[2*8];
  20210. sp_digit t2d[2*8];
  20211. #endif
  20212. sp_digit* t1;
  20213. sp_digit* t2;
  20214. int err = MP_OKAY;
  20215. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20216. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8 * 4, heap, DYNAMIC_TYPE_ECC);
  20217. if (d == NULL) {
  20218. err = MEMORY_E;
  20219. }
  20220. #endif
  20221. if (err == MP_OKAY) {
  20222. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20223. t1 = d + 0 * 8;
  20224. t2 = d + 2 * 8;
  20225. #else
  20226. (void)heap;
  20227. t1 = t1d;
  20228. t2 = t2d;
  20229. #endif
  20230. sp_256_sqr_8(t1, point->y);
  20231. (void)sp_256_mod_8(t1, t1, p256_mod);
  20232. sp_256_sqr_8(t2, point->x);
  20233. (void)sp_256_mod_8(t2, t2, p256_mod);
  20234. sp_256_mul_8(t2, t2, point->x);
  20235. (void)sp_256_mod_8(t2, t2, p256_mod);
  20236. (void)sp_256_sub_8(t2, p256_mod, t2);
  20237. sp_256_mont_add_8(t1, t1, t2, p256_mod);
  20238. sp_256_mont_add_8(t1, t1, point->x, p256_mod);
  20239. sp_256_mont_add_8(t1, t1, point->x, p256_mod);
  20240. sp_256_mont_add_8(t1, t1, point->x, p256_mod);
  20241. if (sp_256_cmp_8(t1, p256_b) != 0) {
  20242. err = MP_VAL;
  20243. }
  20244. }
  20245. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20246. if (d != NULL) {
  20247. XFREE(d, heap, DYNAMIC_TYPE_ECC);
  20248. }
  20249. #endif
  20250. return err;
  20251. }
  20252. /* Check that the x and y oridinates are a valid point on the curve.
  20253. *
  20254. * pX X ordinate of EC point.
  20255. * pY Y ordinate of EC point.
  20256. * returns MEMORY_E if dynamic memory allocation fails, MP_VAL if the point is
  20257. * not on the curve and MP_OKAY otherwise.
  20258. */
  20259. int sp_ecc_is_point_256(mp_int* pX, mp_int* pY)
  20260. {
  20261. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  20262. sp_point pubd;
  20263. #endif
  20264. sp_point* pub;
  20265. byte one[1] = { 1 };
  20266. int err;
  20267. err = sp_ecc_point_new(NULL, pubd, pub);
  20268. if (err == MP_OKAY) {
  20269. sp_256_from_mp(pub->x, 8, pX);
  20270. sp_256_from_mp(pub->y, 8, pY);
  20271. sp_256_from_bin(pub->z, 8, one, (int)sizeof(one));
  20272. err = sp_256_ecc_is_point_8(pub, NULL);
  20273. }
  20274. sp_ecc_point_free(pub, 0, NULL);
  20275. return err;
  20276. }
  20277. /* Check that the private scalar generates the EC point (px, py), the point is
  20278. * on the curve and the point has the correct order.
  20279. *
  20280. * pX X ordinate of EC point.
  20281. * pY Y ordinate of EC point.
  20282. * privm Private scalar that generates EC point.
  20283. * returns MEMORY_E if dynamic memory allocation fails, MP_VAL if the point is
  20284. * not on the curve, ECC_INF_E if the point does not have the correct order,
  20285. * ECC_PRIV_KEY_E when the private scalar doesn't generate the EC point and
  20286. * MP_OKAY otherwise.
  20287. */
  20288. int sp_ecc_check_key_256(mp_int* pX, mp_int* pY, mp_int* privm, void* heap)
  20289. {
  20290. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  20291. sp_digit privd[8];
  20292. sp_point pubd;
  20293. sp_point pd;
  20294. #endif
  20295. sp_digit* priv = NULL;
  20296. sp_point* pub;
  20297. sp_point* p = NULL;
  20298. byte one[1] = { 1 };
  20299. int err;
  20300. err = sp_ecc_point_new(heap, pubd, pub);
  20301. if (err == MP_OKAY) {
  20302. err = sp_ecc_point_new(heap, pd, p);
  20303. }
  20304. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20305. if (err == MP_OKAY) {
  20306. priv = (sp_digit*)XMALLOC(sizeof(sp_digit) * 8, heap,
  20307. DYNAMIC_TYPE_ECC);
  20308. if (priv == NULL) {
  20309. err = MEMORY_E;
  20310. }
  20311. }
  20312. #endif
  20313. if (err == MP_OKAY) {
  20314. #if !(defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK))
  20315. priv = privd;
  20316. #endif
  20317. sp_256_from_mp(pub->x, 8, pX);
  20318. sp_256_from_mp(pub->y, 8, pY);
  20319. sp_256_from_bin(pub->z, 8, one, (int)sizeof(one));
  20320. sp_256_from_mp(priv, 8, privm);
  20321. /* Check point at infinitiy. */
  20322. if ((sp_256_iszero_8(pub->x) != 0) &&
  20323. (sp_256_iszero_8(pub->y) != 0)) {
  20324. err = ECC_INF_E;
  20325. }
  20326. }
  20327. if (err == MP_OKAY) {
  20328. /* Check range of X and Y */
  20329. if (sp_256_cmp_8(pub->x, p256_mod) >= 0 ||
  20330. sp_256_cmp_8(pub->y, p256_mod) >= 0) {
  20331. err = ECC_OUT_OF_RANGE_E;
  20332. }
  20333. }
  20334. if (err == MP_OKAY) {
  20335. /* Check point is on curve */
  20336. err = sp_256_ecc_is_point_8(pub, heap);
  20337. }
  20338. if (err == MP_OKAY) {
  20339. /* Point * order = infinity */
  20340. err = sp_256_ecc_mulmod_8(p, pub, p256_order, 1, heap);
  20341. }
  20342. if (err == MP_OKAY) {
  20343. /* Check result is infinity */
  20344. if ((sp_256_iszero_8(p->x) == 0) ||
  20345. (sp_256_iszero_8(p->y) == 0)) {
  20346. err = ECC_INF_E;
  20347. }
  20348. }
  20349. if (err == MP_OKAY) {
  20350. /* Base * private = point */
  20351. err = sp_256_ecc_mulmod_base_8(p, priv, 1, heap);
  20352. }
  20353. if (err == MP_OKAY) {
  20354. /* Check result is public key */
  20355. if (sp_256_cmp_8(p->x, pub->x) != 0 ||
  20356. sp_256_cmp_8(p->y, pub->y) != 0) {
  20357. err = ECC_PRIV_KEY_E;
  20358. }
  20359. }
  20360. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20361. if (priv != NULL) {
  20362. XFREE(priv, heap, DYNAMIC_TYPE_ECC);
  20363. }
  20364. #endif
  20365. sp_ecc_point_free(p, 0, heap);
  20366. sp_ecc_point_free(pub, 0, heap);
  20367. return err;
  20368. }
  20369. #endif
  20370. #ifdef WOLFSSL_PUBLIC_ECC_ADD_DBL
  20371. /* Add two projective EC points together.
  20372. * (pX, pY, pZ) + (qX, qY, qZ) = (rX, rY, rZ)
  20373. *
  20374. * pX First EC point's X ordinate.
  20375. * pY First EC point's Y ordinate.
  20376. * pZ First EC point's Z ordinate.
  20377. * qX Second EC point's X ordinate.
  20378. * qY Second EC point's Y ordinate.
  20379. * qZ Second EC point's Z ordinate.
  20380. * rX Resultant EC point's X ordinate.
  20381. * rY Resultant EC point's Y ordinate.
  20382. * rZ Resultant EC point's Z ordinate.
  20383. * returns MEMORY_E if dynamic memory allocation fails and MP_OKAY otherwise.
  20384. */
  20385. int sp_ecc_proj_add_point_256(mp_int* pX, mp_int* pY, mp_int* pZ,
  20386. mp_int* qX, mp_int* qY, mp_int* qZ,
  20387. mp_int* rX, mp_int* rY, mp_int* rZ)
  20388. {
  20389. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  20390. sp_digit tmpd[2 * 8 * 5];
  20391. sp_point pd;
  20392. sp_point qd;
  20393. #endif
  20394. sp_digit* tmp;
  20395. sp_point* p;
  20396. sp_point* q = NULL;
  20397. int err;
  20398. err = sp_ecc_point_new(NULL, pd, p);
  20399. if (err == MP_OKAY) {
  20400. err = sp_ecc_point_new(NULL, qd, q);
  20401. }
  20402. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20403. if (err == MP_OKAY) {
  20404. tmp = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 5, NULL,
  20405. DYNAMIC_TYPE_ECC);
  20406. if (tmp == NULL) {
  20407. err = MEMORY_E;
  20408. }
  20409. }
  20410. #else
  20411. tmp = tmpd;
  20412. #endif
  20413. if (err == MP_OKAY) {
  20414. sp_256_from_mp(p->x, 8, pX);
  20415. sp_256_from_mp(p->y, 8, pY);
  20416. sp_256_from_mp(p->z, 8, pZ);
  20417. sp_256_from_mp(q->x, 8, qX);
  20418. sp_256_from_mp(q->y, 8, qY);
  20419. sp_256_from_mp(q->z, 8, qZ);
  20420. sp_256_proj_point_add_8(p, p, q, tmp);
  20421. }
  20422. if (err == MP_OKAY) {
  20423. err = sp_256_to_mp(p->x, rX);
  20424. }
  20425. if (err == MP_OKAY) {
  20426. err = sp_256_to_mp(p->y, rY);
  20427. }
  20428. if (err == MP_OKAY) {
  20429. err = sp_256_to_mp(p->z, rZ);
  20430. }
  20431. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20432. if (tmp != NULL) {
  20433. XFREE(tmp, NULL, DYNAMIC_TYPE_ECC);
  20434. }
  20435. #endif
  20436. sp_ecc_point_free(q, 0, NULL);
  20437. sp_ecc_point_free(p, 0, NULL);
  20438. return err;
  20439. }
  20440. /* Double a projective EC point.
  20441. * (pX, pY, pZ) + (pX, pY, pZ) = (rX, rY, rZ)
  20442. *
  20443. * pX EC point's X ordinate.
  20444. * pY EC point's Y ordinate.
  20445. * pZ EC point's Z ordinate.
  20446. * rX Resultant EC point's X ordinate.
  20447. * rY Resultant EC point's Y ordinate.
  20448. * rZ Resultant EC point's Z ordinate.
  20449. * returns MEMORY_E if dynamic memory allocation fails and MP_OKAY otherwise.
  20450. */
  20451. int sp_ecc_proj_dbl_point_256(mp_int* pX, mp_int* pY, mp_int* pZ,
  20452. mp_int* rX, mp_int* rY, mp_int* rZ)
  20453. {
  20454. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  20455. sp_digit tmpd[2 * 8 * 2];
  20456. sp_point pd;
  20457. #endif
  20458. sp_digit* tmp;
  20459. sp_point* p;
  20460. int err;
  20461. err = sp_ecc_point_new(NULL, pd, p);
  20462. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20463. if (err == MP_OKAY) {
  20464. tmp = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 2, NULL,
  20465. DYNAMIC_TYPE_ECC);
  20466. if (tmp == NULL) {
  20467. err = MEMORY_E;
  20468. }
  20469. }
  20470. #else
  20471. tmp = tmpd;
  20472. #endif
  20473. if (err == MP_OKAY) {
  20474. sp_256_from_mp(p->x, 8, pX);
  20475. sp_256_from_mp(p->y, 8, pY);
  20476. sp_256_from_mp(p->z, 8, pZ);
  20477. sp_256_proj_point_dbl_8(p, p, tmp);
  20478. }
  20479. if (err == MP_OKAY) {
  20480. err = sp_256_to_mp(p->x, rX);
  20481. }
  20482. if (err == MP_OKAY) {
  20483. err = sp_256_to_mp(p->y, rY);
  20484. }
  20485. if (err == MP_OKAY) {
  20486. err = sp_256_to_mp(p->z, rZ);
  20487. }
  20488. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20489. if (tmp != NULL) {
  20490. XFREE(tmp, NULL, DYNAMIC_TYPE_ECC);
  20491. }
  20492. #endif
  20493. sp_ecc_point_free(p, 0, NULL);
  20494. return err;
  20495. }
  20496. /* Map a projective EC point to affine in place.
  20497. * pZ will be one.
  20498. *
  20499. * pX EC point's X ordinate.
  20500. * pY EC point's Y ordinate.
  20501. * pZ EC point's Z ordinate.
  20502. * returns MEMORY_E if dynamic memory allocation fails and MP_OKAY otherwise.
  20503. */
  20504. int sp_ecc_map_256(mp_int* pX, mp_int* pY, mp_int* pZ)
  20505. {
  20506. #if !defined(WOLFSSL_SP_SMALL) && !defined(WOLFSSL_SMALL_STACK)
  20507. sp_digit tmpd[2 * 8 * 4];
  20508. sp_point pd;
  20509. #endif
  20510. sp_digit* tmp;
  20511. sp_point* p;
  20512. int err;
  20513. err = sp_ecc_point_new(NULL, pd, p);
  20514. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20515. if (err == MP_OKAY) {
  20516. tmp = (sp_digit*)XMALLOC(sizeof(sp_digit) * 2 * 8 * 4, NULL,
  20517. DYNAMIC_TYPE_ECC);
  20518. if (tmp == NULL) {
  20519. err = MEMORY_E;
  20520. }
  20521. }
  20522. #else
  20523. tmp = tmpd;
  20524. #endif
  20525. if (err == MP_OKAY) {
  20526. sp_256_from_mp(p->x, 8, pX);
  20527. sp_256_from_mp(p->y, 8, pY);
  20528. sp_256_from_mp(p->z, 8, pZ);
  20529. sp_256_map_8(p, p, tmp);
  20530. }
  20531. if (err == MP_OKAY) {
  20532. err = sp_256_to_mp(p->x, pX);
  20533. }
  20534. if (err == MP_OKAY) {
  20535. err = sp_256_to_mp(p->y, pY);
  20536. }
  20537. if (err == MP_OKAY) {
  20538. err = sp_256_to_mp(p->z, pZ);
  20539. }
  20540. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20541. if (tmp != NULL) {
  20542. XFREE(tmp, NULL, DYNAMIC_TYPE_ECC);
  20543. }
  20544. #endif
  20545. sp_ecc_point_free(p, 0, NULL);
  20546. return err;
  20547. }
  20548. #endif /* WOLFSSL_PUBLIC_ECC_ADD_DBL */
  20549. #ifdef HAVE_COMP_KEY
  20550. /* Find the square root of a number mod the prime of the curve.
  20551. *
  20552. * y The number to operate on and the result.
  20553. * returns MEMORY_E if dynamic memory allocation fails and MP_OKAY otherwise.
  20554. */
  20555. static int sp_256_mont_sqrt_8(sp_digit* y)
  20556. {
  20557. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20558. sp_digit* d;
  20559. #else
  20560. sp_digit t1d[2 * 8];
  20561. sp_digit t2d[2 * 8];
  20562. #endif
  20563. sp_digit* t1;
  20564. sp_digit* t2;
  20565. int err = MP_OKAY;
  20566. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20567. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 4 * 8, NULL, DYNAMIC_TYPE_ECC);
  20568. if (d == NULL) {
  20569. err = MEMORY_E;
  20570. }
  20571. #endif
  20572. if (err == MP_OKAY) {
  20573. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20574. t1 = d + 0 * 8;
  20575. t2 = d + 2 * 8;
  20576. #else
  20577. t1 = t1d;
  20578. t2 = t2d;
  20579. #endif
  20580. {
  20581. /* t2 = y ^ 0x2 */
  20582. sp_256_mont_sqr_8(t2, y, p256_mod, p256_mp_mod);
  20583. /* t1 = y ^ 0x3 */
  20584. sp_256_mont_mul_8(t1, t2, y, p256_mod, p256_mp_mod);
  20585. /* t2 = y ^ 0xc */
  20586. sp_256_mont_sqr_n_8(t2, t1, 2, p256_mod, p256_mp_mod);
  20587. /* t1 = y ^ 0xf */
  20588. sp_256_mont_mul_8(t1, t1, t2, p256_mod, p256_mp_mod);
  20589. /* t2 = y ^ 0xf0 */
  20590. sp_256_mont_sqr_n_8(t2, t1, 4, p256_mod, p256_mp_mod);
  20591. /* t1 = y ^ 0xff */
  20592. sp_256_mont_mul_8(t1, t1, t2, p256_mod, p256_mp_mod);
  20593. /* t2 = y ^ 0xff00 */
  20594. sp_256_mont_sqr_n_8(t2, t1, 8, p256_mod, p256_mp_mod);
  20595. /* t1 = y ^ 0xffff */
  20596. sp_256_mont_mul_8(t1, t1, t2, p256_mod, p256_mp_mod);
  20597. /* t2 = y ^ 0xffff0000 */
  20598. sp_256_mont_sqr_n_8(t2, t1, 16, p256_mod, p256_mp_mod);
  20599. /* t1 = y ^ 0xffffffff */
  20600. sp_256_mont_mul_8(t1, t1, t2, p256_mod, p256_mp_mod);
  20601. /* t1 = y ^ 0xffffffff00000000 */
  20602. sp_256_mont_sqr_n_8(t1, t1, 32, p256_mod, p256_mp_mod);
  20603. /* t1 = y ^ 0xffffffff00000001 */
  20604. sp_256_mont_mul_8(t1, t1, y, p256_mod, p256_mp_mod);
  20605. /* t1 = y ^ 0xffffffff00000001000000000000000000000000 */
  20606. sp_256_mont_sqr_n_8(t1, t1, 96, p256_mod, p256_mp_mod);
  20607. /* t1 = y ^ 0xffffffff00000001000000000000000000000001 */
  20608. sp_256_mont_mul_8(t1, t1, y, p256_mod, p256_mp_mod);
  20609. sp_256_mont_sqr_n_8(y, t1, 94, p256_mod, p256_mp_mod);
  20610. }
  20611. }
  20612. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20613. if (d != NULL) {
  20614. XFREE(d, NULL, DYNAMIC_TYPE_ECC);
  20615. }
  20616. #endif
  20617. return err;
  20618. }
  20619. /* Uncompress the point given the X ordinate.
  20620. *
  20621. * xm X ordinate.
  20622. * odd Whether the Y ordinate is odd.
  20623. * ym Calculated Y ordinate.
  20624. * returns MEMORY_E if dynamic memory allocation fails and MP_OKAY otherwise.
  20625. */
  20626. int sp_ecc_uncompress_256(mp_int* xm, int odd, mp_int* ym)
  20627. {
  20628. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20629. sp_digit* d;
  20630. #else
  20631. sp_digit xd[2 * 8];
  20632. sp_digit yd[2 * 8];
  20633. #endif
  20634. sp_digit* x = NULL;
  20635. sp_digit* y = NULL;
  20636. int err = MP_OKAY;
  20637. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20638. d = (sp_digit*)XMALLOC(sizeof(sp_digit) * 4 * 8, NULL, DYNAMIC_TYPE_ECC);
  20639. if (d == NULL) {
  20640. err = MEMORY_E;
  20641. }
  20642. #endif
  20643. if (err == MP_OKAY) {
  20644. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20645. x = d + 0 * 8;
  20646. y = d + 2 * 8;
  20647. #else
  20648. x = xd;
  20649. y = yd;
  20650. #endif
  20651. sp_256_from_mp(x, 8, xm);
  20652. err = sp_256_mod_mul_norm_8(x, x, p256_mod);
  20653. }
  20654. if (err == MP_OKAY) {
  20655. /* y = x^3 */
  20656. {
  20657. sp_256_mont_sqr_8(y, x, p256_mod, p256_mp_mod);
  20658. sp_256_mont_mul_8(y, y, x, p256_mod, p256_mp_mod);
  20659. }
  20660. /* y = x^3 - 3x */
  20661. sp_256_mont_sub_8(y, y, x, p256_mod);
  20662. sp_256_mont_sub_8(y, y, x, p256_mod);
  20663. sp_256_mont_sub_8(y, y, x, p256_mod);
  20664. /* y = x^3 - 3x + b */
  20665. err = sp_256_mod_mul_norm_8(x, p256_b, p256_mod);
  20666. }
  20667. if (err == MP_OKAY) {
  20668. sp_256_mont_add_8(y, y, x, p256_mod);
  20669. /* y = sqrt(x^3 - 3x + b) */
  20670. err = sp_256_mont_sqrt_8(y);
  20671. }
  20672. if (err == MP_OKAY) {
  20673. XMEMSET(y + 8, 0, 8U * sizeof(sp_digit));
  20674. sp_256_mont_reduce_8(y, p256_mod, p256_mp_mod);
  20675. if ((((word32)y[0] ^ (word32)odd) & 1U) != 0U) {
  20676. sp_256_mont_sub_8(y, p256_mod, y, p256_mod);
  20677. }
  20678. err = sp_256_to_mp(y, ym);
  20679. }
  20680. #if defined(WOLFSSL_SP_SMALL) || defined(WOLFSSL_SMALL_STACK)
  20681. if (d != NULL) {
  20682. XFREE(d, NULL, DYNAMIC_TYPE_ECC);
  20683. }
  20684. #endif
  20685. return err;
  20686. }
  20687. #endif
  20688. #endif /* !WOLFSSL_SP_NO_256 */
  20689. #endif /* WOLFSSL_HAVE_SP_ECC */
  20690. #endif /* WOLFSSL_SP_ARM_THUMB_ASM */
  20691. #endif /* WOLFSSL_HAVE_SP_RSA || WOLFSSL_HAVE_SP_DH || WOLFSSL_HAVE_SP_ECC */