12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309231023112312231323142315231623172318231923202321232223232324232523262327232823292330233123322333233423352336233723382339234023412342234323442345234623472348234923502351235223532354235523562357235823592360236123622363236423652366236723682369237023712372237323742375237623772378237923802381238223832384238523862387238823892390239123922393239423952396239723982399240024012402240324042405240624072408240924102411241224132414241524162417241824192420242124222423242424252426242724282429243024312432243324342435243624372438243924402441244224432444244524462447244824492450245124522453245424552456245724582459246024612462246324642465246624672468246924702471247224732474247524762477247824792480248124822483248424852486248724882489249024912492249324942495249624972498249925002501250225032504250525062507250825092510251125122513251425152516251725182519252025212522252325242525252625272528252925302531253225332534253525362537253825392540254125422543254425452546254725482549255025512552255325542555255625572558255925602561256225632564256525662567256825692570257125722573257425752576257725782579258025812582258325842585258625872588258925902591259225932594259525962597259825992600260126022603260426052606260726082609261026112612261326142615261626172618261926202621262226232624262526262627262826292630263126322633263426352636263726382639264026412642264326442645264626472648264926502651265226532654265526562657265826592660266126622663266426652666266726682669267026712672267326742675267626772678267926802681268226832684268526862687268826892690269126922693269426952696269726982699270027012702270327042705270627072708270927102711271227132714271527162717271827192720272127222723272427252726272727282729273027312732273327342735273627372738273927402741274227432744274527462747274827492750275127522753275427552756275727582759276027612762276327642765276627672768276927702771277227732774277527762777277827792780278127822783278427852786278727882789279027912792279327942795279627972798279928002801280228032804280528062807280828092810281128122813281428152816281728182819282028212822282328242825282628272828282928302831283228332834283528362837283828392840284128422843284428452846284728482849285028512852285328542855285628572858285928602861286228632864286528662867286828692870287128722873287428752876287728782879288028812882288328842885288628872888288928902891289228932894289528962897289828992900290129022903290429052906290729082909291029112912291329142915291629172918291929202921292229232924292529262927292829292930293129322933293429352936293729382939294029412942294329442945294629472948294929502951295229532954295529562957295829592960296129622963296429652966296729682969297029712972297329742975297629772978297929802981298229832984298529862987298829892990299129922993299429952996299729982999300030013002300330043005300630073008300930103011301230133014301530163017301830193020302130223023302430253026302730283029303030313032303330343035303630373038303930403041304230433044304530463047304830493050305130523053305430553056305730583059306030613062306330643065306630673068306930703071307230733074307530763077307830793080308130823083308430853086308730883089309030913092309330943095309630973098309931003101310231033104310531063107310831093110311131123113311431153116311731183119312031213122312331243125312631273128312931303131313231333134313531363137313831393140314131423143314431453146314731483149315031513152315331543155315631573158315931603161316231633164316531663167316831693170317131723173317431753176317731783179318031813182318331843185318631873188318931903191319231933194319531963197319831993200320132023203320432053206320732083209321032113212321332143215321632173218321932203221322232233224322532263227322832293230323132323233323432353236323732383239324032413242324332443245324632473248324932503251325232533254325532563257325832593260326132623263326432653266326732683269327032713272327332743275327632773278327932803281328232833284328532863287328832893290329132923293329432953296329732983299330033013302330333043305330633073308330933103311331233133314331533163317331833193320332133223323332433253326332733283329333033313332333333343335333633373338333933403341334233433344334533463347334833493350335133523353335433553356335733583359336033613362336333643365336633673368336933703371337233733374337533763377337833793380338133823383338433853386338733883389339033913392339333943395339633973398339934003401340234033404340534063407340834093410341134123413341434153416341734183419342034213422342334243425342634273428342934303431343234333434343534363437343834393440344134423443344434453446344734483449345034513452345334543455345634573458345934603461346234633464346534663467346834693470347134723473347434753476347734783479348034813482348334843485348634873488348934903491349234933494349534963497349834993500350135023503350435053506350735083509351035113512351335143515351635173518351935203521352235233524352535263527352835293530353135323533353435353536353735383539354035413542354335443545354635473548354935503551355235533554355535563557355835593560356135623563356435653566356735683569357035713572357335743575357635773578357935803581358235833584358535863587358835893590359135923593359435953596359735983599360036013602360336043605360636073608360936103611361236133614361536163617361836193620362136223623362436253626362736283629363036313632363336343635363636373638363936403641364236433644364536463647364836493650365136523653365436553656365736583659366036613662366336643665366636673668366936703671367236733674367536763677367836793680368136823683368436853686368736883689369036913692369336943695369636973698369937003701370237033704370537063707370837093710371137123713371437153716371737183719372037213722372337243725372637273728372937303731373237333734373537363737373837393740374137423743374437453746374737483749375037513752375337543755375637573758375937603761376237633764376537663767376837693770377137723773377437753776377737783779378037813782378337843785378637873788378937903791379237933794379537963797379837993800380138023803380438053806380738083809381038113812381338143815381638173818381938203821382238233824382538263827382838293830383138323833383438353836383738383839384038413842384338443845384638473848384938503851385238533854385538563857385838593860386138623863386438653866386738683869387038713872387338743875387638773878387938803881388238833884388538863887388838893890389138923893389438953896389738983899390039013902390339043905390639073908390939103911391239133914391539163917391839193920392139223923392439253926392739283929393039313932393339343935393639373938393939403941394239433944394539463947394839493950395139523953395439553956395739583959396039613962396339643965396639673968396939703971397239733974397539763977397839793980398139823983398439853986398739883989399039913992399339943995399639973998399940004001400240034004400540064007400840094010401140124013401440154016401740184019402040214022402340244025402640274028402940304031403240334034403540364037403840394040404140424043404440454046404740484049405040514052405340544055405640574058405940604061406240634064406540664067406840694070407140724073407440754076407740784079408040814082408340844085408640874088408940904091409240934094409540964097409840994100410141024103410441054106410741084109411041114112411341144115411641174118411941204121412241234124412541264127412841294130413141324133413441354136413741384139414041414142414341444145414641474148414941504151415241534154415541564157415841594160416141624163416441654166416741684169417041714172417341744175417641774178417941804181418241834184418541864187418841894190419141924193419441954196419741984199420042014202420342044205420642074208420942104211421242134214421542164217421842194220422142224223422442254226422742284229423042314232423342344235423642374238423942404241424242434244424542464247424842494250425142524253425442554256425742584259426042614262426342644265426642674268426942704271427242734274427542764277427842794280428142824283428442854286428742884289429042914292429342944295429642974298429943004301430243034304430543064307430843094310431143124313431443154316431743184319432043214322432343244325432643274328432943304331433243334334433543364337433843394340434143424343434443454346434743484349435043514352435343544355435643574358435943604361436243634364436543664367436843694370437143724373437443754376437743784379438043814382438343844385438643874388438943904391439243934394439543964397439843994400440144024403440444054406440744084409441044114412441344144415441644174418441944204421442244234424442544264427442844294430443144324433443444354436443744384439444044414442444344444445444644474448444944504451445244534454445544564457445844594460446144624463446444654466446744684469447044714472447344744475447644774478447944804481448244834484448544864487448844894490449144924493449444954496449744984499450045014502450345044505450645074508450945104511451245134514451545164517451845194520452145224523452445254526452745284529453045314532453345344535453645374538453945404541454245434544454545464547454845494550455145524553455445554556455745584559456045614562456345644565456645674568456945704571457245734574457545764577457845794580458145824583458445854586458745884589459045914592459345944595459645974598459946004601460246034604460546064607460846094610461146124613461446154616461746184619462046214622462346244625462646274628462946304631463246334634463546364637463846394640464146424643464446454646464746484649465046514652465346544655465646574658465946604661466246634664466546664667466846694670467146724673467446754676467746784679468046814682468346844685468646874688468946904691469246934694469546964697469846994700470147024703470447054706470747084709471047114712471347144715471647174718471947204721472247234724472547264727472847294730473147324733473447354736473747384739474047414742474347444745474647474748474947504751475247534754475547564757475847594760476147624763476447654766476747684769477047714772477347744775477647774778477947804781478247834784478547864787478847894790479147924793479447954796479747984799480048014802480348044805480648074808480948104811481248134814481548164817481848194820482148224823482448254826482748284829483048314832483348344835483648374838483948404841484248434844484548464847484848494850485148524853485448554856485748584859486048614862486348644865486648674868486948704871487248734874487548764877487848794880488148824883488448854886488748884889489048914892489348944895489648974898489949004901490249034904490549064907490849094910491149124913491449154916491749184919492049214922492349244925492649274928492949304931493249334934493549364937493849394940494149424943494449454946494749484949495049514952495349544955495649574958495949604961496249634964496549664967496849694970497149724973497449754976497749784979498049814982498349844985498649874988498949904991499249934994499549964997499849995000500150025003500450055006500750085009501050115012501350145015501650175018501950205021502250235024502550265027502850295030503150325033503450355036503750385039504050415042504350445045504650475048504950505051505250535054505550565057505850595060506150625063506450655066506750685069507050715072507350745075507650775078507950805081508250835084508550865087508850895090509150925093509450955096509750985099510051015102510351045105510651075108510951105111511251135114511551165117511851195120512151225123512451255126512751285129513051315132513351345135513651375138513951405141514251435144514551465147514851495150515151525153515451555156515751585159516051615162516351645165516651675168516951705171517251735174517551765177517851795180518151825183518451855186518751885189519051915192519351945195519651975198519952005201520252035204520552065207520852095210521152125213521452155216521752185219522052215222522352245225522652275228522952305231523252335234523552365237523852395240524152425243524452455246524752485249525052515252525352545255525652575258525952605261526252635264526552665267526852695270527152725273527452755276527752785279528052815282528352845285528652875288528952905291529252935294529552965297529852995300530153025303530453055306530753085309531053115312531353145315531653175318531953205321532253235324532553265327532853295330533153325333533453355336533753385339534053415342534353445345534653475348534953505351535253535354535553565357535853595360536153625363536453655366536753685369537053715372537353745375537653775378537953805381538253835384538553865387538853895390539153925393539453955396539753985399540054015402540354045405540654075408540954105411541254135414541554165417541854195420542154225423542454255426542754285429543054315432543354345435543654375438543954405441544254435444544554465447544854495450545154525453545454555456545754585459546054615462546354645465546654675468546954705471547254735474547554765477547854795480548154825483548454855486548754885489549054915492549354945495549654975498549955005501550255035504550555065507550855095510551155125513551455155516551755185519552055215522552355245525552655275528552955305531553255335534553555365537553855395540554155425543554455455546554755485549555055515552555355545555555655575558555955605561556255635564556555665567556855695570557155725573557455755576557755785579558055815582558355845585558655875588558955905591559255935594559555965597559855995600560156025603560456055606560756085609561056115612561356145615561656175618561956205621562256235624562556265627562856295630563156325633563456355636563756385639564056415642564356445645564656475648564956505651565256535654565556565657565856595660566156625663566456655666566756685669567056715672567356745675567656775678567956805681568256835684568556865687568856895690569156925693569456955696569756985699570057015702570357045705570657075708570957105711571257135714571557165717571857195720572157225723572457255726572757285729573057315732573357345735573657375738573957405741574257435744574557465747574857495750575157525753575457555756575757585759576057615762576357645765576657675768576957705771577257735774577557765777577857795780578157825783578457855786578757885789579057915792579357945795579657975798579958005801580258035804580558065807580858095810581158125813581458155816581758185819582058215822582358245825582658275828582958305831583258335834583558365837583858395840584158425843584458455846584758485849585058515852585358545855585658575858585958605861586258635864586558665867586858695870587158725873587458755876587758785879588058815882588358845885588658875888588958905891589258935894589558965897589858995900590159025903590459055906590759085909591059115912591359145915591659175918591959205921592259235924592559265927592859295930593159325933593459355936593759385939594059415942594359445945594659475948594959505951595259535954595559565957595859595960596159625963596459655966596759685969597059715972597359745975597659775978597959805981598259835984598559865987598859895990599159925993599459955996599759985999600060016002600360046005600660076008600960106011601260136014601560166017601860196020602160226023602460256026602760286029603060316032603360346035603660376038603960406041604260436044604560466047604860496050605160526053605460556056605760586059606060616062606360646065606660676068606960706071607260736074607560766077607860796080608160826083608460856086608760886089609060916092609360946095609660976098609961006101610261036104610561066107610861096110611161126113611461156116611761186119612061216122612361246125612661276128612961306131613261336134613561366137613861396140614161426143614461456146614761486149615061516152615361546155615661576158615961606161616261636164616561666167616861696170617161726173617461756176617761786179618061816182618361846185618661876188618961906191619261936194619561966197619861996200620162026203620462056206620762086209621062116212621362146215621662176218621962206221622262236224622562266227622862296230623162326233623462356236623762386239624062416242624362446245624662476248624962506251625262536254625562566257625862596260626162626263626462656266626762686269627062716272627362746275627662776278627962806281628262836284628562866287628862896290629162926293629462956296629762986299630063016302630363046305630663076308630963106311631263136314631563166317631863196320632163226323632463256326632763286329633063316332633363346335633663376338633963406341634263436344634563466347634863496350635163526353635463556356635763586359636063616362636363646365636663676368636963706371637263736374637563766377637863796380638163826383638463856386638763886389639063916392639363946395639663976398639964006401640264036404640564066407640864096410641164126413641464156416641764186419642064216422642364246425642664276428642964306431643264336434643564366437643864396440644164426443644464456446644764486449645064516452645364546455645664576458645964606461646264636464646564666467646864696470647164726473647464756476647764786479648064816482648364846485648664876488648964906491649264936494649564966497649864996500650165026503650465056506650765086509651065116512651365146515651665176518651965206521652265236524652565266527652865296530653165326533653465356536653765386539654065416542654365446545654665476548654965506551655265536554655565566557655865596560656165626563656465656566656765686569657065716572657365746575657665776578657965806581658265836584658565866587658865896590659165926593659465956596659765986599660066016602660366046605660666076608660966106611661266136614661566166617661866196620662166226623662466256626662766286629663066316632663366346635663666376638663966406641664266436644664566466647664866496650665166526653665466556656665766586659666066616662666366646665666666676668666966706671667266736674667566766677667866796680668166826683668466856686668766886689669066916692669366946695669666976698669967006701670267036704670567066707670867096710671167126713671467156716671767186719672067216722672367246725672667276728672967306731673267336734673567366737673867396740674167426743674467456746674767486749675067516752675367546755675667576758675967606761676267636764676567666767676867696770677167726773677467756776677767786779678067816782678367846785678667876788678967906791679267936794679567966797679867996800680168026803680468056806680768086809681068116812681368146815681668176818681968206821682268236824682568266827682868296830683168326833683468356836683768386839684068416842684368446845684668476848684968506851685268536854685568566857685868596860686168626863686468656866686768686869687068716872687368746875687668776878687968806881688268836884688568866887688868896890689168926893689468956896689768986899690069016902690369046905690669076908690969106911691269136914691569166917691869196920692169226923692469256926692769286929693069316932693369346935693669376938693969406941694269436944694569466947694869496950695169526953695469556956695769586959696069616962696369646965696669676968696969706971697269736974697569766977697869796980698169826983698469856986698769886989699069916992699369946995699669976998699970007001700270037004700570067007700870097010701170127013701470157016701770187019702070217022702370247025702670277028702970307031703270337034703570367037703870397040704170427043704470457046704770487049705070517052705370547055705670577058705970607061706270637064706570667067706870697070707170727073707470757076707770787079708070817082708370847085708670877088708970907091709270937094709570967097709870997100710171027103710471057106710771087109711071117112711371147115711671177118711971207121712271237124712571267127712871297130713171327133713471357136713771387139714071417142714371447145714671477148714971507151715271537154715571567157715871597160716171627163716471657166716771687169717071717172717371747175717671777178717971807181718271837184718571867187718871897190719171927193719471957196719771987199720072017202720372047205720672077208720972107211721272137214721572167217721872197220722172227223722472257226722772287229723072317232723372347235723672377238723972407241724272437244724572467247724872497250725172527253725472557256725772587259726072617262726372647265726672677268726972707271727272737274727572767277727872797280728172827283728472857286728772887289729072917292729372947295729672977298729973007301730273037304730573067307730873097310731173127313731473157316731773187319732073217322732373247325732673277328732973307331733273337334733573367337733873397340734173427343734473457346734773487349735073517352735373547355735673577358735973607361736273637364736573667367736873697370737173727373737473757376737773787379738073817382738373847385738673877388738973907391739273937394739573967397739873997400740174027403740474057406740774087409741074117412741374147415741674177418741974207421742274237424742574267427742874297430743174327433743474357436743774387439744074417442744374447445744674477448744974507451745274537454745574567457745874597460746174627463746474657466746774687469747074717472747374747475747674777478747974807481748274837484748574867487748874897490749174927493749474957496749774987499750075017502750375047505750675077508750975107511751275137514751575167517751875197520752175227523752475257526752775287529753075317532753375347535753675377538753975407541754275437544754575467547754875497550755175527553755475557556755775587559756075617562756375647565756675677568756975707571757275737574757575767577757875797580758175827583758475857586758775887589759075917592759375947595759675977598759976007601760276037604760576067607760876097610761176127613761476157616761776187619762076217622762376247625762676277628762976307631763276337634763576367637763876397640764176427643764476457646764776487649765076517652765376547655765676577658765976607661766276637664766576667667766876697670767176727673767476757676767776787679768076817682768376847685768676877688768976907691769276937694769576967697769876997700770177027703770477057706770777087709771077117712771377147715771677177718771977207721772277237724772577267727772877297730773177327733773477357736773777387739774077417742774377447745774677477748774977507751775277537754775577567757775877597760776177627763776477657766776777687769777077717772777377747775777677777778777977807781778277837784778577867787778877897790779177927793779477957796779777987799780078017802780378047805780678077808780978107811781278137814781578167817781878197820782178227823782478257826782778287829783078317832783378347835783678377838783978407841784278437844784578467847784878497850785178527853785478557856785778587859786078617862786378647865786678677868786978707871787278737874787578767877787878797880788178827883788478857886788778887889789078917892789378947895789678977898789979007901790279037904790579067907790879097910791179127913791479157916791779187919792079217922792379247925792679277928792979307931793279337934793579367937793879397940794179427943794479457946794779487949795079517952795379547955795679577958795979607961796279637964796579667967796879697970797179727973797479757976797779787979798079817982798379847985798679877988798979907991799279937994799579967997799879998000800180028003800480058006800780088009801080118012801380148015801680178018801980208021802280238024802580268027802880298030803180328033803480358036803780388039804080418042804380448045804680478048804980508051805280538054805580568057805880598060806180628063806480658066806780688069807080718072807380748075807680778078807980808081808280838084808580868087808880898090809180928093809480958096809780988099810081018102810381048105810681078108810981108111811281138114811581168117811881198120812181228123812481258126812781288129813081318132813381348135813681378138813981408141814281438144814581468147814881498150815181528153815481558156815781588159816081618162816381648165816681678168816981708171817281738174817581768177817881798180818181828183818481858186818781888189819081918192819381948195819681978198819982008201820282038204820582068207820882098210821182128213821482158216821782188219822082218222822382248225822682278228822982308231823282338234823582368237823882398240824182428243824482458246824782488249825082518252825382548255825682578258825982608261826282638264826582668267826882698270827182728273827482758276827782788279828082818282828382848285828682878288828982908291829282938294829582968297829882998300830183028303830483058306830783088309831083118312831383148315831683178318831983208321832283238324832583268327832883298330833183328333833483358336833783388339834083418342834383448345834683478348834983508351835283538354835583568357835883598360836183628363836483658366836783688369837083718372837383748375837683778378837983808381838283838384838583868387838883898390839183928393839483958396839783988399840084018402840384048405840684078408840984108411841284138414841584168417841884198420842184228423842484258426842784288429843084318432843384348435843684378438843984408441844284438444844584468447844884498450845184528453845484558456845784588459846084618462846384648465846684678468846984708471847284738474847584768477847884798480848184828483848484858486848784888489849084918492849384948495849684978498849985008501850285038504850585068507850885098510851185128513851485158516851785188519852085218522852385248525852685278528852985308531853285338534853585368537853885398540854185428543854485458546854785488549855085518552855385548555855685578558855985608561856285638564856585668567856885698570857185728573857485758576857785788579858085818582858385848585858685878588858985908591859285938594859585968597859885998600860186028603860486058606860786088609861086118612861386148615861686178618861986208621862286238624862586268627862886298630863186328633863486358636863786388639864086418642864386448645864686478648864986508651865286538654865586568657865886598660866186628663866486658666866786688669867086718672867386748675867686778678867986808681868286838684868586868687868886898690869186928693869486958696869786988699870087018702870387048705870687078708870987108711871287138714871587168717871887198720872187228723872487258726872787288729873087318732873387348735873687378738873987408741874287438744874587468747874887498750875187528753875487558756875787588759876087618762876387648765876687678768876987708771877287738774877587768777877887798780878187828783878487858786878787888789879087918792879387948795879687978798879988008801880288038804880588068807880888098810881188128813881488158816881788188819882088218822882388248825882688278828882988308831883288338834883588368837883888398840884188428843884488458846884788488849885088518852885388548855885688578858885988608861886288638864886588668867886888698870887188728873887488758876887788788879888088818882888388848885888688878888888988908891889288938894889588968897889888998900890189028903890489058906890789088909891089118912891389148915891689178918891989208921892289238924892589268927892889298930893189328933893489358936893789388939894089418942894389448945894689478948894989508951895289538954895589568957895889598960896189628963896489658966896789688969897089718972897389748975897689778978897989808981898289838984898589868987898889898990899189928993899489958996899789988999900090019002900390049005900690079008900990109011901290139014901590169017901890199020 |
- /*
- * ARM translation
- *
- * Copyright (c) 2003 Fabrice Bellard
- * Copyright (c) 2005-2007 CodeSourcery
- * Copyright (c) 2007 OpenedHand, Ltd.
- *
- * This library is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This library is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with this library; if not, see <http://www.gnu.org/licenses/>.
- */
- #include "qemu/osdep.h"
- #include "cpu.h"
- #include "internals.h"
- #include "disas/disas.h"
- #include "exec/exec-all.h"
- #include "tcg/tcg-op.h"
- #include "tcg/tcg-op-gvec.h"
- #include "qemu/log.h"
- #include "qemu/bitops.h"
- #include "arm_ldst.h"
- #include "hw/semihosting/semihost.h"
- #include "exec/helper-proto.h"
- #include "exec/helper-gen.h"
- #include "trace-tcg.h"
- #include "exec/log.h"
- #define ENABLE_ARCH_4T arm_dc_feature(s, ARM_FEATURE_V4T)
- #define ENABLE_ARCH_5 arm_dc_feature(s, ARM_FEATURE_V5)
- /* currently all emulated v5 cores are also v5TE, so don't bother */
- #define ENABLE_ARCH_5TE arm_dc_feature(s, ARM_FEATURE_V5)
- #define ENABLE_ARCH_5J dc_isar_feature(aa32_jazelle, s)
- #define ENABLE_ARCH_6 arm_dc_feature(s, ARM_FEATURE_V6)
- #define ENABLE_ARCH_6K arm_dc_feature(s, ARM_FEATURE_V6K)
- #define ENABLE_ARCH_6T2 arm_dc_feature(s, ARM_FEATURE_THUMB2)
- #define ENABLE_ARCH_7 arm_dc_feature(s, ARM_FEATURE_V7)
- #define ENABLE_ARCH_8 arm_dc_feature(s, ARM_FEATURE_V8)
- #define ARCH(x) do { if (!ENABLE_ARCH_##x) goto illegal_op; } while(0)
- #include "translate.h"
- #if defined(CONFIG_USER_ONLY)
- #define IS_USER(s) 1
- #else
- #define IS_USER(s) (s->user)
- #endif
- /* We reuse the same 64-bit temporaries for efficiency. */
- static TCGv_i64 cpu_V0, cpu_V1, cpu_M0;
- static TCGv_i32 cpu_R[16];
- TCGv_i32 cpu_CF, cpu_NF, cpu_VF, cpu_ZF;
- TCGv_i64 cpu_exclusive_addr;
- TCGv_i64 cpu_exclusive_val;
- #include "exec/gen-icount.h"
- static const char * const regnames[] =
- { "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
- "r8", "r9", "r10", "r11", "r12", "r13", "r14", "pc" };
- /* Function prototypes for gen_ functions calling Neon helpers. */
- typedef void NeonGenThreeOpEnvFn(TCGv_i32, TCGv_env, TCGv_i32,
- TCGv_i32, TCGv_i32);
- /* Function prototypes for gen_ functions for fix point conversions */
- typedef void VFPGenFixPointFn(TCGv_i32, TCGv_i32, TCGv_i32, TCGv_ptr);
- /* initialize TCG globals. */
- void arm_translate_init(void)
- {
- int i;
- for (i = 0; i < 16; i++) {
- cpu_R[i] = tcg_global_mem_new_i32(cpu_env,
- offsetof(CPUARMState, regs[i]),
- regnames[i]);
- }
- cpu_CF = tcg_global_mem_new_i32(cpu_env, offsetof(CPUARMState, CF), "CF");
- cpu_NF = tcg_global_mem_new_i32(cpu_env, offsetof(CPUARMState, NF), "NF");
- cpu_VF = tcg_global_mem_new_i32(cpu_env, offsetof(CPUARMState, VF), "VF");
- cpu_ZF = tcg_global_mem_new_i32(cpu_env, offsetof(CPUARMState, ZF), "ZF");
- cpu_exclusive_addr = tcg_global_mem_new_i64(cpu_env,
- offsetof(CPUARMState, exclusive_addr), "exclusive_addr");
- cpu_exclusive_val = tcg_global_mem_new_i64(cpu_env,
- offsetof(CPUARMState, exclusive_val), "exclusive_val");
- a64_translate_init();
- }
- /* Flags for the disas_set_da_iss info argument:
- * lower bits hold the Rt register number, higher bits are flags.
- */
- typedef enum ISSInfo {
- ISSNone = 0,
- ISSRegMask = 0x1f,
- ISSInvalid = (1 << 5),
- ISSIsAcqRel = (1 << 6),
- ISSIsWrite = (1 << 7),
- ISSIs16Bit = (1 << 8),
- } ISSInfo;
- /* Save the syndrome information for a Data Abort */
- static void disas_set_da_iss(DisasContext *s, MemOp memop, ISSInfo issinfo)
- {
- uint32_t syn;
- int sas = memop & MO_SIZE;
- bool sse = memop & MO_SIGN;
- bool is_acqrel = issinfo & ISSIsAcqRel;
- bool is_write = issinfo & ISSIsWrite;
- bool is_16bit = issinfo & ISSIs16Bit;
- int srt = issinfo & ISSRegMask;
- if (issinfo & ISSInvalid) {
- /* Some callsites want to conditionally provide ISS info,
- * eg "only if this was not a writeback"
- */
- return;
- }
- if (srt == 15) {
- /* For AArch32, insns where the src/dest is R15 never generate
- * ISS information. Catching that here saves checking at all
- * the call sites.
- */
- return;
- }
- syn = syn_data_abort_with_iss(0, sas, sse, srt, 0, is_acqrel,
- 0, 0, 0, is_write, 0, is_16bit);
- disas_set_insn_syndrome(s, syn);
- }
- static inline int get_a32_user_mem_index(DisasContext *s)
- {
- /* Return the core mmu_idx to use for A32/T32 "unprivileged load/store"
- * insns:
- * if PL2, UNPREDICTABLE (we choose to implement as if PL0)
- * otherwise, access as if at PL0.
- */
- switch (s->mmu_idx) {
- case ARMMMUIdx_E2: /* this one is UNPREDICTABLE */
- case ARMMMUIdx_E10_0:
- case ARMMMUIdx_E10_1:
- case ARMMMUIdx_E10_1_PAN:
- return arm_to_core_mmu_idx(ARMMMUIdx_E10_0);
- case ARMMMUIdx_SE3:
- case ARMMMUIdx_SE10_0:
- case ARMMMUIdx_SE10_1:
- case ARMMMUIdx_SE10_1_PAN:
- return arm_to_core_mmu_idx(ARMMMUIdx_SE10_0);
- case ARMMMUIdx_MUser:
- case ARMMMUIdx_MPriv:
- return arm_to_core_mmu_idx(ARMMMUIdx_MUser);
- case ARMMMUIdx_MUserNegPri:
- case ARMMMUIdx_MPrivNegPri:
- return arm_to_core_mmu_idx(ARMMMUIdx_MUserNegPri);
- case ARMMMUIdx_MSUser:
- case ARMMMUIdx_MSPriv:
- return arm_to_core_mmu_idx(ARMMMUIdx_MSUser);
- case ARMMMUIdx_MSUserNegPri:
- case ARMMMUIdx_MSPrivNegPri:
- return arm_to_core_mmu_idx(ARMMMUIdx_MSUserNegPri);
- default:
- g_assert_not_reached();
- }
- }
- static inline TCGv_i32 load_cpu_offset(int offset)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_ld_i32(tmp, cpu_env, offset);
- return tmp;
- }
- #define load_cpu_field(name) load_cpu_offset(offsetof(CPUARMState, name))
- static inline void store_cpu_offset(TCGv_i32 var, int offset)
- {
- tcg_gen_st_i32(var, cpu_env, offset);
- tcg_temp_free_i32(var);
- }
- #define store_cpu_field(var, name) \
- store_cpu_offset(var, offsetof(CPUARMState, name))
- /* The architectural value of PC. */
- static uint32_t read_pc(DisasContext *s)
- {
- return s->pc_curr + (s->thumb ? 4 : 8);
- }
- /* Set a variable to the value of a CPU register. */
- static void load_reg_var(DisasContext *s, TCGv_i32 var, int reg)
- {
- if (reg == 15) {
- tcg_gen_movi_i32(var, read_pc(s));
- } else {
- tcg_gen_mov_i32(var, cpu_R[reg]);
- }
- }
- /* Create a new temporary and set it to the value of a CPU register. */
- static inline TCGv_i32 load_reg(DisasContext *s, int reg)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- load_reg_var(s, tmp, reg);
- return tmp;
- }
- /*
- * Create a new temp, REG + OFS, except PC is ALIGN(PC, 4).
- * This is used for load/store for which use of PC implies (literal),
- * or ADD that implies ADR.
- */
- static TCGv_i32 add_reg_for_lit(DisasContext *s, int reg, int ofs)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- if (reg == 15) {
- tcg_gen_movi_i32(tmp, (read_pc(s) & ~3) + ofs);
- } else {
- tcg_gen_addi_i32(tmp, cpu_R[reg], ofs);
- }
- return tmp;
- }
- /* Set a CPU register. The source must be a temporary and will be
- marked as dead. */
- static void store_reg(DisasContext *s, int reg, TCGv_i32 var)
- {
- if (reg == 15) {
- /* In Thumb mode, we must ignore bit 0.
- * In ARM mode, for ARMv4 and ARMv5, it is UNPREDICTABLE if bits [1:0]
- * are not 0b00, but for ARMv6 and above, we must ignore bits [1:0].
- * We choose to ignore [1:0] in ARM mode for all architecture versions.
- */
- tcg_gen_andi_i32(var, var, s->thumb ? ~1 : ~3);
- s->base.is_jmp = DISAS_JUMP;
- }
- tcg_gen_mov_i32(cpu_R[reg], var);
- tcg_temp_free_i32(var);
- }
- /*
- * Variant of store_reg which applies v8M stack-limit checks before updating
- * SP. If the check fails this will result in an exception being taken.
- * We disable the stack checks for CONFIG_USER_ONLY because we have
- * no idea what the stack limits should be in that case.
- * If stack checking is not being done this just acts like store_reg().
- */
- static void store_sp_checked(DisasContext *s, TCGv_i32 var)
- {
- #ifndef CONFIG_USER_ONLY
- if (s->v8m_stackcheck) {
- gen_helper_v8m_stackcheck(cpu_env, var);
- }
- #endif
- store_reg(s, 13, var);
- }
- /* Value extensions. */
- #define gen_uxtb(var) tcg_gen_ext8u_i32(var, var)
- #define gen_uxth(var) tcg_gen_ext16u_i32(var, var)
- #define gen_sxtb(var) tcg_gen_ext8s_i32(var, var)
- #define gen_sxth(var) tcg_gen_ext16s_i32(var, var)
- #define gen_sxtb16(var) gen_helper_sxtb16(var, var)
- #define gen_uxtb16(var) gen_helper_uxtb16(var, var)
- static inline void gen_set_cpsr(TCGv_i32 var, uint32_t mask)
- {
- TCGv_i32 tmp_mask = tcg_const_i32(mask);
- gen_helper_cpsr_write(cpu_env, var, tmp_mask);
- tcg_temp_free_i32(tmp_mask);
- }
- /* Set NZCV flags from the high 4 bits of var. */
- #define gen_set_nzcv(var) gen_set_cpsr(var, CPSR_NZCV)
- static void gen_exception_internal(int excp)
- {
- TCGv_i32 tcg_excp = tcg_const_i32(excp);
- assert(excp_is_internal(excp));
- gen_helper_exception_internal(cpu_env, tcg_excp);
- tcg_temp_free_i32(tcg_excp);
- }
- static void gen_step_complete_exception(DisasContext *s)
- {
- /* We just completed step of an insn. Move from Active-not-pending
- * to Active-pending, and then also take the swstep exception.
- * This corresponds to making the (IMPDEF) choice to prioritize
- * swstep exceptions over asynchronous exceptions taken to an exception
- * level where debug is disabled. This choice has the advantage that
- * we do not need to maintain internal state corresponding to the
- * ISV/EX syndrome bits between completion of the step and generation
- * of the exception, and our syndrome information is always correct.
- */
- gen_ss_advance(s);
- gen_swstep_exception(s, 1, s->is_ldex);
- s->base.is_jmp = DISAS_NORETURN;
- }
- static void gen_singlestep_exception(DisasContext *s)
- {
- /* Generate the right kind of exception for singlestep, which is
- * either the architectural singlestep or EXCP_DEBUG for QEMU's
- * gdb singlestepping.
- */
- if (s->ss_active) {
- gen_step_complete_exception(s);
- } else {
- gen_exception_internal(EXCP_DEBUG);
- }
- }
- static inline bool is_singlestepping(DisasContext *s)
- {
- /* Return true if we are singlestepping either because of
- * architectural singlestep or QEMU gdbstub singlestep. This does
- * not include the command line '-singlestep' mode which is rather
- * misnamed as it only means "one instruction per TB" and doesn't
- * affect the code we generate.
- */
- return s->base.singlestep_enabled || s->ss_active;
- }
- static void gen_smul_dual(TCGv_i32 a, TCGv_i32 b)
- {
- TCGv_i32 tmp1 = tcg_temp_new_i32();
- TCGv_i32 tmp2 = tcg_temp_new_i32();
- tcg_gen_ext16s_i32(tmp1, a);
- tcg_gen_ext16s_i32(tmp2, b);
- tcg_gen_mul_i32(tmp1, tmp1, tmp2);
- tcg_temp_free_i32(tmp2);
- tcg_gen_sari_i32(a, a, 16);
- tcg_gen_sari_i32(b, b, 16);
- tcg_gen_mul_i32(b, b, a);
- tcg_gen_mov_i32(a, tmp1);
- tcg_temp_free_i32(tmp1);
- }
- /* Byteswap each halfword. */
- static void gen_rev16(TCGv_i32 dest, TCGv_i32 var)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- TCGv_i32 mask = tcg_const_i32(0x00ff00ff);
- tcg_gen_shri_i32(tmp, var, 8);
- tcg_gen_and_i32(tmp, tmp, mask);
- tcg_gen_and_i32(var, var, mask);
- tcg_gen_shli_i32(var, var, 8);
- tcg_gen_or_i32(dest, var, tmp);
- tcg_temp_free_i32(mask);
- tcg_temp_free_i32(tmp);
- }
- /* Byteswap low halfword and sign extend. */
- static void gen_revsh(TCGv_i32 dest, TCGv_i32 var)
- {
- tcg_gen_ext16u_i32(var, var);
- tcg_gen_bswap16_i32(var, var);
- tcg_gen_ext16s_i32(dest, var);
- }
- /* Swap low and high halfwords. */
- static void gen_swap_half(TCGv_i32 dest, TCGv_i32 var)
- {
- tcg_gen_rotri_i32(dest, var, 16);
- }
- /* Dual 16-bit add. Result placed in t0 and t1 is marked as dead.
- tmp = (t0 ^ t1) & 0x8000;
- t0 &= ~0x8000;
- t1 &= ~0x8000;
- t0 = (t0 + t1) ^ tmp;
- */
- static void gen_add16(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_xor_i32(tmp, t0, t1);
- tcg_gen_andi_i32(tmp, tmp, 0x8000);
- tcg_gen_andi_i32(t0, t0, ~0x8000);
- tcg_gen_andi_i32(t1, t1, ~0x8000);
- tcg_gen_add_i32(t0, t0, t1);
- tcg_gen_xor_i32(dest, t0, tmp);
- tcg_temp_free_i32(tmp);
- }
- /* Set N and Z flags from var. */
- static inline void gen_logic_CC(TCGv_i32 var)
- {
- tcg_gen_mov_i32(cpu_NF, var);
- tcg_gen_mov_i32(cpu_ZF, var);
- }
- /* dest = T0 + T1 + CF. */
- static void gen_add_carry(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- tcg_gen_add_i32(dest, t0, t1);
- tcg_gen_add_i32(dest, dest, cpu_CF);
- }
- /* dest = T0 - T1 + CF - 1. */
- static void gen_sub_carry(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- tcg_gen_sub_i32(dest, t0, t1);
- tcg_gen_add_i32(dest, dest, cpu_CF);
- tcg_gen_subi_i32(dest, dest, 1);
- }
- /* dest = T0 + T1. Compute C, N, V and Z flags */
- static void gen_add_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_movi_i32(tmp, 0);
- tcg_gen_add2_i32(cpu_NF, cpu_CF, t0, tmp, t1, tmp);
- tcg_gen_mov_i32(cpu_ZF, cpu_NF);
- tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
- tcg_gen_xor_i32(tmp, t0, t1);
- tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
- tcg_temp_free_i32(tmp);
- tcg_gen_mov_i32(dest, cpu_NF);
- }
- /* dest = T0 + T1 + CF. Compute C, N, V and Z flags */
- static void gen_adc_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- if (TCG_TARGET_HAS_add2_i32) {
- tcg_gen_movi_i32(tmp, 0);
- tcg_gen_add2_i32(cpu_NF, cpu_CF, t0, tmp, cpu_CF, tmp);
- tcg_gen_add2_i32(cpu_NF, cpu_CF, cpu_NF, cpu_CF, t1, tmp);
- } else {
- TCGv_i64 q0 = tcg_temp_new_i64();
- TCGv_i64 q1 = tcg_temp_new_i64();
- tcg_gen_extu_i32_i64(q0, t0);
- tcg_gen_extu_i32_i64(q1, t1);
- tcg_gen_add_i64(q0, q0, q1);
- tcg_gen_extu_i32_i64(q1, cpu_CF);
- tcg_gen_add_i64(q0, q0, q1);
- tcg_gen_extr_i64_i32(cpu_NF, cpu_CF, q0);
- tcg_temp_free_i64(q0);
- tcg_temp_free_i64(q1);
- }
- tcg_gen_mov_i32(cpu_ZF, cpu_NF);
- tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
- tcg_gen_xor_i32(tmp, t0, t1);
- tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
- tcg_temp_free_i32(tmp);
- tcg_gen_mov_i32(dest, cpu_NF);
- }
- /* dest = T0 - T1. Compute C, N, V and Z flags */
- static void gen_sub_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp;
- tcg_gen_sub_i32(cpu_NF, t0, t1);
- tcg_gen_mov_i32(cpu_ZF, cpu_NF);
- tcg_gen_setcond_i32(TCG_COND_GEU, cpu_CF, t0, t1);
- tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
- tmp = tcg_temp_new_i32();
- tcg_gen_xor_i32(tmp, t0, t1);
- tcg_gen_and_i32(cpu_VF, cpu_VF, tmp);
- tcg_temp_free_i32(tmp);
- tcg_gen_mov_i32(dest, cpu_NF);
- }
- /* dest = T0 + ~T1 + CF. Compute C, N, V and Z flags */
- static void gen_sbc_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_not_i32(tmp, t1);
- gen_adc_CC(dest, t0, tmp);
- tcg_temp_free_i32(tmp);
- }
- #define GEN_SHIFT(name) \
- static void gen_##name(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1) \
- { \
- TCGv_i32 tmp1, tmp2, tmp3; \
- tmp1 = tcg_temp_new_i32(); \
- tcg_gen_andi_i32(tmp1, t1, 0xff); \
- tmp2 = tcg_const_i32(0); \
- tmp3 = tcg_const_i32(0x1f); \
- tcg_gen_movcond_i32(TCG_COND_GTU, tmp2, tmp1, tmp3, tmp2, t0); \
- tcg_temp_free_i32(tmp3); \
- tcg_gen_andi_i32(tmp1, tmp1, 0x1f); \
- tcg_gen_##name##_i32(dest, tmp2, tmp1); \
- tcg_temp_free_i32(tmp2); \
- tcg_temp_free_i32(tmp1); \
- }
- GEN_SHIFT(shl)
- GEN_SHIFT(shr)
- #undef GEN_SHIFT
- static void gen_sar(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
- {
- TCGv_i32 tmp1, tmp2;
- tmp1 = tcg_temp_new_i32();
- tcg_gen_andi_i32(tmp1, t1, 0xff);
- tmp2 = tcg_const_i32(0x1f);
- tcg_gen_movcond_i32(TCG_COND_GTU, tmp1, tmp1, tmp2, tmp2, tmp1);
- tcg_temp_free_i32(tmp2);
- tcg_gen_sar_i32(dest, t0, tmp1);
- tcg_temp_free_i32(tmp1);
- }
- static void shifter_out_im(TCGv_i32 var, int shift)
- {
- tcg_gen_extract_i32(cpu_CF, var, shift, 1);
- }
- /* Shift by immediate. Includes special handling for shift == 0. */
- static inline void gen_arm_shift_im(TCGv_i32 var, int shiftop,
- int shift, int flags)
- {
- switch (shiftop) {
- case 0: /* LSL */
- if (shift != 0) {
- if (flags)
- shifter_out_im(var, 32 - shift);
- tcg_gen_shli_i32(var, var, shift);
- }
- break;
- case 1: /* LSR */
- if (shift == 0) {
- if (flags) {
- tcg_gen_shri_i32(cpu_CF, var, 31);
- }
- tcg_gen_movi_i32(var, 0);
- } else {
- if (flags)
- shifter_out_im(var, shift - 1);
- tcg_gen_shri_i32(var, var, shift);
- }
- break;
- case 2: /* ASR */
- if (shift == 0)
- shift = 32;
- if (flags)
- shifter_out_im(var, shift - 1);
- if (shift == 32)
- shift = 31;
- tcg_gen_sari_i32(var, var, shift);
- break;
- case 3: /* ROR/RRX */
- if (shift != 0) {
- if (flags)
- shifter_out_im(var, shift - 1);
- tcg_gen_rotri_i32(var, var, shift); break;
- } else {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_shli_i32(tmp, cpu_CF, 31);
- if (flags)
- shifter_out_im(var, 0);
- tcg_gen_shri_i32(var, var, 1);
- tcg_gen_or_i32(var, var, tmp);
- tcg_temp_free_i32(tmp);
- }
- }
- };
- static inline void gen_arm_shift_reg(TCGv_i32 var, int shiftop,
- TCGv_i32 shift, int flags)
- {
- if (flags) {
- switch (shiftop) {
- case 0: gen_helper_shl_cc(var, cpu_env, var, shift); break;
- case 1: gen_helper_shr_cc(var, cpu_env, var, shift); break;
- case 2: gen_helper_sar_cc(var, cpu_env, var, shift); break;
- case 3: gen_helper_ror_cc(var, cpu_env, var, shift); break;
- }
- } else {
- switch (shiftop) {
- case 0:
- gen_shl(var, var, shift);
- break;
- case 1:
- gen_shr(var, var, shift);
- break;
- case 2:
- gen_sar(var, var, shift);
- break;
- case 3: tcg_gen_andi_i32(shift, shift, 0x1f);
- tcg_gen_rotr_i32(var, var, shift); break;
- }
- }
- tcg_temp_free_i32(shift);
- }
- /*
- * Generate a conditional based on ARM condition code cc.
- * This is common between ARM and Aarch64 targets.
- */
- void arm_test_cc(DisasCompare *cmp, int cc)
- {
- TCGv_i32 value;
- TCGCond cond;
- bool global = true;
- switch (cc) {
- case 0: /* eq: Z */
- case 1: /* ne: !Z */
- cond = TCG_COND_EQ;
- value = cpu_ZF;
- break;
- case 2: /* cs: C */
- case 3: /* cc: !C */
- cond = TCG_COND_NE;
- value = cpu_CF;
- break;
- case 4: /* mi: N */
- case 5: /* pl: !N */
- cond = TCG_COND_LT;
- value = cpu_NF;
- break;
- case 6: /* vs: V */
- case 7: /* vc: !V */
- cond = TCG_COND_LT;
- value = cpu_VF;
- break;
- case 8: /* hi: C && !Z */
- case 9: /* ls: !C || Z -> !(C && !Z) */
- cond = TCG_COND_NE;
- value = tcg_temp_new_i32();
- global = false;
- /* CF is 1 for C, so -CF is an all-bits-set mask for C;
- ZF is non-zero for !Z; so AND the two subexpressions. */
- tcg_gen_neg_i32(value, cpu_CF);
- tcg_gen_and_i32(value, value, cpu_ZF);
- break;
- case 10: /* ge: N == V -> N ^ V == 0 */
- case 11: /* lt: N != V -> N ^ V != 0 */
- /* Since we're only interested in the sign bit, == 0 is >= 0. */
- cond = TCG_COND_GE;
- value = tcg_temp_new_i32();
- global = false;
- tcg_gen_xor_i32(value, cpu_VF, cpu_NF);
- break;
- case 12: /* gt: !Z && N == V */
- case 13: /* le: Z || N != V */
- cond = TCG_COND_NE;
- value = tcg_temp_new_i32();
- global = false;
- /* (N == V) is equal to the sign bit of ~(NF ^ VF). Propagate
- * the sign bit then AND with ZF to yield the result. */
- tcg_gen_xor_i32(value, cpu_VF, cpu_NF);
- tcg_gen_sari_i32(value, value, 31);
- tcg_gen_andc_i32(value, cpu_ZF, value);
- break;
- case 14: /* always */
- case 15: /* always */
- /* Use the ALWAYS condition, which will fold early.
- * It doesn't matter what we use for the value. */
- cond = TCG_COND_ALWAYS;
- value = cpu_ZF;
- goto no_invert;
- default:
- fprintf(stderr, "Bad condition code 0x%x\n", cc);
- abort();
- }
- if (cc & 1) {
- cond = tcg_invert_cond(cond);
- }
- no_invert:
- cmp->cond = cond;
- cmp->value = value;
- cmp->value_global = global;
- }
- void arm_free_cc(DisasCompare *cmp)
- {
- if (!cmp->value_global) {
- tcg_temp_free_i32(cmp->value);
- }
- }
- void arm_jump_cc(DisasCompare *cmp, TCGLabel *label)
- {
- tcg_gen_brcondi_i32(cmp->cond, cmp->value, 0, label);
- }
- void arm_gen_test_cc(int cc, TCGLabel *label)
- {
- DisasCompare cmp;
- arm_test_cc(&cmp, cc);
- arm_jump_cc(&cmp, label);
- arm_free_cc(&cmp);
- }
- static inline void gen_set_condexec(DisasContext *s)
- {
- if (s->condexec_mask) {
- uint32_t val = (s->condexec_cond << 4) | (s->condexec_mask >> 1);
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_movi_i32(tmp, val);
- store_cpu_field(tmp, condexec_bits);
- }
- }
- static inline void gen_set_pc_im(DisasContext *s, target_ulong val)
- {
- tcg_gen_movi_i32(cpu_R[15], val);
- }
- /* Set PC and Thumb state from var. var is marked as dead. */
- static inline void gen_bx(DisasContext *s, TCGv_i32 var)
- {
- s->base.is_jmp = DISAS_JUMP;
- tcg_gen_andi_i32(cpu_R[15], var, ~1);
- tcg_gen_andi_i32(var, var, 1);
- store_cpu_field(var, thumb);
- }
- /*
- * Set PC and Thumb state from var. var is marked as dead.
- * For M-profile CPUs, include logic to detect exception-return
- * branches and handle them. This is needed for Thumb POP/LDM to PC, LDR to PC,
- * and BX reg, and no others, and happens only for code in Handler mode.
- * The Security Extension also requires us to check for the FNC_RETURN
- * which signals a function return from non-secure state; this can happen
- * in both Handler and Thread mode.
- * To avoid having to do multiple comparisons in inline generated code,
- * we make the check we do here loose, so it will match for EXC_RETURN
- * in Thread mode. For system emulation do_v7m_exception_exit() checks
- * for these spurious cases and returns without doing anything (giving
- * the same behaviour as for a branch to a non-magic address).
- *
- * In linux-user mode it is unclear what the right behaviour for an
- * attempted FNC_RETURN should be, because in real hardware this will go
- * directly to Secure code (ie not the Linux kernel) which will then treat
- * the error in any way it chooses. For QEMU we opt to make the FNC_RETURN
- * attempt behave the way it would on a CPU without the security extension,
- * which is to say "like a normal branch". That means we can simply treat
- * all branches as normal with no magic address behaviour.
- */
- static inline void gen_bx_excret(DisasContext *s, TCGv_i32 var)
- {
- /* Generate the same code here as for a simple bx, but flag via
- * s->base.is_jmp that we need to do the rest of the work later.
- */
- gen_bx(s, var);
- #ifndef CONFIG_USER_ONLY
- if (arm_dc_feature(s, ARM_FEATURE_M_SECURITY) ||
- (s->v7m_handler_mode && arm_dc_feature(s, ARM_FEATURE_M))) {
- s->base.is_jmp = DISAS_BX_EXCRET;
- }
- #endif
- }
- static inline void gen_bx_excret_final_code(DisasContext *s)
- {
- /* Generate the code to finish possible exception return and end the TB */
- TCGLabel *excret_label = gen_new_label();
- uint32_t min_magic;
- if (arm_dc_feature(s, ARM_FEATURE_M_SECURITY)) {
- /* Covers FNC_RETURN and EXC_RETURN magic */
- min_magic = FNC_RETURN_MIN_MAGIC;
- } else {
- /* EXC_RETURN magic only */
- min_magic = EXC_RETURN_MIN_MAGIC;
- }
- /* Is the new PC value in the magic range indicating exception return? */
- tcg_gen_brcondi_i32(TCG_COND_GEU, cpu_R[15], min_magic, excret_label);
- /* No: end the TB as we would for a DISAS_JMP */
- if (is_singlestepping(s)) {
- gen_singlestep_exception(s);
- } else {
- tcg_gen_exit_tb(NULL, 0);
- }
- gen_set_label(excret_label);
- /* Yes: this is an exception return.
- * At this point in runtime env->regs[15] and env->thumb will hold
- * the exception-return magic number, which do_v7m_exception_exit()
- * will read. Nothing else will be able to see those values because
- * the cpu-exec main loop guarantees that we will always go straight
- * from raising the exception to the exception-handling code.
- *
- * gen_ss_advance(s) does nothing on M profile currently but
- * calling it is conceptually the right thing as we have executed
- * this instruction (compare SWI, HVC, SMC handling).
- */
- gen_ss_advance(s);
- gen_exception_internal(EXCP_EXCEPTION_EXIT);
- }
- static inline void gen_bxns(DisasContext *s, int rm)
- {
- TCGv_i32 var = load_reg(s, rm);
- /* The bxns helper may raise an EXCEPTION_EXIT exception, so in theory
- * we need to sync state before calling it, but:
- * - we don't need to do gen_set_pc_im() because the bxns helper will
- * always set the PC itself
- * - we don't need to do gen_set_condexec() because BXNS is UNPREDICTABLE
- * unless it's outside an IT block or the last insn in an IT block,
- * so we know that condexec == 0 (already set at the top of the TB)
- * is correct in the non-UNPREDICTABLE cases, and we can choose
- * "zeroes the IT bits" as our UNPREDICTABLE behaviour otherwise.
- */
- gen_helper_v7m_bxns(cpu_env, var);
- tcg_temp_free_i32(var);
- s->base.is_jmp = DISAS_EXIT;
- }
- static inline void gen_blxns(DisasContext *s, int rm)
- {
- TCGv_i32 var = load_reg(s, rm);
- /* We don't need to sync condexec state, for the same reason as bxns.
- * We do however need to set the PC, because the blxns helper reads it.
- * The blxns helper may throw an exception.
- */
- gen_set_pc_im(s, s->base.pc_next);
- gen_helper_v7m_blxns(cpu_env, var);
- tcg_temp_free_i32(var);
- s->base.is_jmp = DISAS_EXIT;
- }
- /* Variant of store_reg which uses branch&exchange logic when storing
- to r15 in ARM architecture v7 and above. The source must be a temporary
- and will be marked as dead. */
- static inline void store_reg_bx(DisasContext *s, int reg, TCGv_i32 var)
- {
- if (reg == 15 && ENABLE_ARCH_7) {
- gen_bx(s, var);
- } else {
- store_reg(s, reg, var);
- }
- }
- /* Variant of store_reg which uses branch&exchange logic when storing
- * to r15 in ARM architecture v5T and above. This is used for storing
- * the results of a LDR/LDM/POP into r15, and corresponds to the cases
- * in the ARM ARM which use the LoadWritePC() pseudocode function. */
- static inline void store_reg_from_load(DisasContext *s, int reg, TCGv_i32 var)
- {
- if (reg == 15 && ENABLE_ARCH_5) {
- gen_bx_excret(s, var);
- } else {
- store_reg(s, reg, var);
- }
- }
- #ifdef CONFIG_USER_ONLY
- #define IS_USER_ONLY 1
- #else
- #define IS_USER_ONLY 0
- #endif
- /* Abstractions of "generate code to do a guest load/store for
- * AArch32", where a vaddr is always 32 bits (and is zero
- * extended if we're a 64 bit core) and data is also
- * 32 bits unless specifically doing a 64 bit access.
- * These functions work like tcg_gen_qemu_{ld,st}* except
- * that the address argument is TCGv_i32 rather than TCGv.
- */
- static inline TCGv gen_aa32_addr(DisasContext *s, TCGv_i32 a32, MemOp op)
- {
- TCGv addr = tcg_temp_new();
- tcg_gen_extu_i32_tl(addr, a32);
- /* Not needed for user-mode BE32, where we use MO_BE instead. */
- if (!IS_USER_ONLY && s->sctlr_b && (op & MO_SIZE) < MO_32) {
- tcg_gen_xori_tl(addr, addr, 4 - (1 << (op & MO_SIZE)));
- }
- return addr;
- }
- static void gen_aa32_ld_i32(DisasContext *s, TCGv_i32 val, TCGv_i32 a32,
- int index, MemOp opc)
- {
- TCGv addr;
- if (arm_dc_feature(s, ARM_FEATURE_M) &&
- !arm_dc_feature(s, ARM_FEATURE_M_MAIN)) {
- opc |= MO_ALIGN;
- }
- addr = gen_aa32_addr(s, a32, opc);
- tcg_gen_qemu_ld_i32(val, addr, index, opc);
- tcg_temp_free(addr);
- }
- static void gen_aa32_st_i32(DisasContext *s, TCGv_i32 val, TCGv_i32 a32,
- int index, MemOp opc)
- {
- TCGv addr;
- if (arm_dc_feature(s, ARM_FEATURE_M) &&
- !arm_dc_feature(s, ARM_FEATURE_M_MAIN)) {
- opc |= MO_ALIGN;
- }
- addr = gen_aa32_addr(s, a32, opc);
- tcg_gen_qemu_st_i32(val, addr, index, opc);
- tcg_temp_free(addr);
- }
- #define DO_GEN_LD(SUFF, OPC) \
- static inline void gen_aa32_ld##SUFF(DisasContext *s, TCGv_i32 val, \
- TCGv_i32 a32, int index) \
- { \
- gen_aa32_ld_i32(s, val, a32, index, OPC | s->be_data); \
- }
- #define DO_GEN_ST(SUFF, OPC) \
- static inline void gen_aa32_st##SUFF(DisasContext *s, TCGv_i32 val, \
- TCGv_i32 a32, int index) \
- { \
- gen_aa32_st_i32(s, val, a32, index, OPC | s->be_data); \
- }
- static inline void gen_aa32_frob64(DisasContext *s, TCGv_i64 val)
- {
- /* Not needed for user-mode BE32, where we use MO_BE instead. */
- if (!IS_USER_ONLY && s->sctlr_b) {
- tcg_gen_rotri_i64(val, val, 32);
- }
- }
- static void gen_aa32_ld_i64(DisasContext *s, TCGv_i64 val, TCGv_i32 a32,
- int index, MemOp opc)
- {
- TCGv addr = gen_aa32_addr(s, a32, opc);
- tcg_gen_qemu_ld_i64(val, addr, index, opc);
- gen_aa32_frob64(s, val);
- tcg_temp_free(addr);
- }
- static inline void gen_aa32_ld64(DisasContext *s, TCGv_i64 val,
- TCGv_i32 a32, int index)
- {
- gen_aa32_ld_i64(s, val, a32, index, MO_Q | s->be_data);
- }
- static void gen_aa32_st_i64(DisasContext *s, TCGv_i64 val, TCGv_i32 a32,
- int index, MemOp opc)
- {
- TCGv addr = gen_aa32_addr(s, a32, opc);
- /* Not needed for user-mode BE32, where we use MO_BE instead. */
- if (!IS_USER_ONLY && s->sctlr_b) {
- TCGv_i64 tmp = tcg_temp_new_i64();
- tcg_gen_rotri_i64(tmp, val, 32);
- tcg_gen_qemu_st_i64(tmp, addr, index, opc);
- tcg_temp_free_i64(tmp);
- } else {
- tcg_gen_qemu_st_i64(val, addr, index, opc);
- }
- tcg_temp_free(addr);
- }
- static inline void gen_aa32_st64(DisasContext *s, TCGv_i64 val,
- TCGv_i32 a32, int index)
- {
- gen_aa32_st_i64(s, val, a32, index, MO_Q | s->be_data);
- }
- DO_GEN_LD(8u, MO_UB)
- DO_GEN_LD(16u, MO_UW)
- DO_GEN_LD(32u, MO_UL)
- DO_GEN_ST(8, MO_UB)
- DO_GEN_ST(16, MO_UW)
- DO_GEN_ST(32, MO_UL)
- static inline void gen_hvc(DisasContext *s, int imm16)
- {
- /* The pre HVC helper handles cases when HVC gets trapped
- * as an undefined insn by runtime configuration (ie before
- * the insn really executes).
- */
- gen_set_pc_im(s, s->pc_curr);
- gen_helper_pre_hvc(cpu_env);
- /* Otherwise we will treat this as a real exception which
- * happens after execution of the insn. (The distinction matters
- * for the PC value reported to the exception handler and also
- * for single stepping.)
- */
- s->svc_imm = imm16;
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_HVC;
- }
- static inline void gen_smc(DisasContext *s)
- {
- /* As with HVC, we may take an exception either before or after
- * the insn executes.
- */
- TCGv_i32 tmp;
- gen_set_pc_im(s, s->pc_curr);
- tmp = tcg_const_i32(syn_aa32_smc());
- gen_helper_pre_smc(cpu_env, tmp);
- tcg_temp_free_i32(tmp);
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_SMC;
- }
- static void gen_exception_internal_insn(DisasContext *s, uint32_t pc, int excp)
- {
- gen_set_condexec(s);
- gen_set_pc_im(s, pc);
- gen_exception_internal(excp);
- s->base.is_jmp = DISAS_NORETURN;
- }
- static void gen_exception_insn(DisasContext *s, uint32_t pc, int excp,
- int syn, uint32_t target_el)
- {
- gen_set_condexec(s);
- gen_set_pc_im(s, pc);
- gen_exception(excp, syn, target_el);
- s->base.is_jmp = DISAS_NORETURN;
- }
- static void gen_exception_bkpt_insn(DisasContext *s, uint32_t syn)
- {
- TCGv_i32 tcg_syn;
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- tcg_syn = tcg_const_i32(syn);
- gen_helper_exception_bkpt_insn(cpu_env, tcg_syn);
- tcg_temp_free_i32(tcg_syn);
- s->base.is_jmp = DISAS_NORETURN;
- }
- static void unallocated_encoding(DisasContext *s)
- {
- /* Unallocated and reserved encodings are uncategorized */
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF, syn_uncategorized(),
- default_exception_el(s));
- }
- /* Force a TB lookup after an instruction that changes the CPU state. */
- static inline void gen_lookup_tb(DisasContext *s)
- {
- tcg_gen_movi_i32(cpu_R[15], s->base.pc_next);
- s->base.is_jmp = DISAS_EXIT;
- }
- static inline void gen_hlt(DisasContext *s, int imm)
- {
- /* HLT. This has two purposes.
- * Architecturally, it is an external halting debug instruction.
- * Since QEMU doesn't implement external debug, we treat this as
- * it is required for halting debug disabled: it will UNDEF.
- * Secondly, "HLT 0x3C" is a T32 semihosting trap instruction,
- * and "HLT 0xF000" is an A32 semihosting syscall. These traps
- * must trigger semihosting even for ARMv7 and earlier, where
- * HLT was an undefined encoding.
- * In system mode, we don't allow userspace access to
- * semihosting, to provide some semblance of security
- * (and for consistency with our 32-bit semihosting).
- */
- if (semihosting_enabled() &&
- #ifndef CONFIG_USER_ONLY
- s->current_el != 0 &&
- #endif
- (imm == (s->thumb ? 0x3c : 0xf000))) {
- gen_exception_internal_insn(s, s->pc_curr, EXCP_SEMIHOST);
- return;
- }
- unallocated_encoding(s);
- }
- static TCGv_ptr get_fpstatus_ptr(int neon)
- {
- TCGv_ptr statusptr = tcg_temp_new_ptr();
- int offset;
- if (neon) {
- offset = offsetof(CPUARMState, vfp.standard_fp_status);
- } else {
- offset = offsetof(CPUARMState, vfp.fp_status);
- }
- tcg_gen_addi_ptr(statusptr, cpu_env, offset);
- return statusptr;
- }
- static inline long vfp_reg_offset(bool dp, unsigned reg)
- {
- if (dp) {
- return offsetof(CPUARMState, vfp.zregs[reg >> 1].d[reg & 1]);
- } else {
- long ofs = offsetof(CPUARMState, vfp.zregs[reg >> 2].d[(reg >> 1) & 1]);
- if (reg & 1) {
- ofs += offsetof(CPU_DoubleU, l.upper);
- } else {
- ofs += offsetof(CPU_DoubleU, l.lower);
- }
- return ofs;
- }
- }
- /* Return the offset of a 32-bit piece of a NEON register.
- zero is the least significant end of the register. */
- static inline long
- neon_reg_offset (int reg, int n)
- {
- int sreg;
- sreg = reg * 2 + n;
- return vfp_reg_offset(0, sreg);
- }
- static TCGv_i32 neon_load_reg(int reg, int pass)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_ld_i32(tmp, cpu_env, neon_reg_offset(reg, pass));
- return tmp;
- }
- static void neon_store_reg(int reg, int pass, TCGv_i32 var)
- {
- tcg_gen_st_i32(var, cpu_env, neon_reg_offset(reg, pass));
- tcg_temp_free_i32(var);
- }
- static inline void neon_load_reg64(TCGv_i64 var, int reg)
- {
- tcg_gen_ld_i64(var, cpu_env, vfp_reg_offset(1, reg));
- }
- static inline void neon_store_reg64(TCGv_i64 var, int reg)
- {
- tcg_gen_st_i64(var, cpu_env, vfp_reg_offset(1, reg));
- }
- static inline void neon_load_reg32(TCGv_i32 var, int reg)
- {
- tcg_gen_ld_i32(var, cpu_env, vfp_reg_offset(false, reg));
- }
- static inline void neon_store_reg32(TCGv_i32 var, int reg)
- {
- tcg_gen_st_i32(var, cpu_env, vfp_reg_offset(false, reg));
- }
- static TCGv_ptr vfp_reg_ptr(bool dp, int reg)
- {
- TCGv_ptr ret = tcg_temp_new_ptr();
- tcg_gen_addi_ptr(ret, cpu_env, vfp_reg_offset(dp, reg));
- return ret;
- }
- #define ARM_CP_RW_BIT (1 << 20)
- /* Include the VFP and Neon decoders */
- #include "translate-vfp.c.inc"
- #include "translate-neon.c.inc"
- static inline void iwmmxt_load_reg(TCGv_i64 var, int reg)
- {
- tcg_gen_ld_i64(var, cpu_env, offsetof(CPUARMState, iwmmxt.regs[reg]));
- }
- static inline void iwmmxt_store_reg(TCGv_i64 var, int reg)
- {
- tcg_gen_st_i64(var, cpu_env, offsetof(CPUARMState, iwmmxt.regs[reg]));
- }
- static inline TCGv_i32 iwmmxt_load_creg(int reg)
- {
- TCGv_i32 var = tcg_temp_new_i32();
- tcg_gen_ld_i32(var, cpu_env, offsetof(CPUARMState, iwmmxt.cregs[reg]));
- return var;
- }
- static inline void iwmmxt_store_creg(int reg, TCGv_i32 var)
- {
- tcg_gen_st_i32(var, cpu_env, offsetof(CPUARMState, iwmmxt.cregs[reg]));
- tcg_temp_free_i32(var);
- }
- static inline void gen_op_iwmmxt_movq_wRn_M0(int rn)
- {
- iwmmxt_store_reg(cpu_M0, rn);
- }
- static inline void gen_op_iwmmxt_movq_M0_wRn(int rn)
- {
- iwmmxt_load_reg(cpu_M0, rn);
- }
- static inline void gen_op_iwmmxt_orq_M0_wRn(int rn)
- {
- iwmmxt_load_reg(cpu_V1, rn);
- tcg_gen_or_i64(cpu_M0, cpu_M0, cpu_V1);
- }
- static inline void gen_op_iwmmxt_andq_M0_wRn(int rn)
- {
- iwmmxt_load_reg(cpu_V1, rn);
- tcg_gen_and_i64(cpu_M0, cpu_M0, cpu_V1);
- }
- static inline void gen_op_iwmmxt_xorq_M0_wRn(int rn)
- {
- iwmmxt_load_reg(cpu_V1, rn);
- tcg_gen_xor_i64(cpu_M0, cpu_M0, cpu_V1);
- }
- #define IWMMXT_OP(name) \
- static inline void gen_op_iwmmxt_##name##_M0_wRn(int rn) \
- { \
- iwmmxt_load_reg(cpu_V1, rn); \
- gen_helper_iwmmxt_##name(cpu_M0, cpu_M0, cpu_V1); \
- }
- #define IWMMXT_OP_ENV(name) \
- static inline void gen_op_iwmmxt_##name##_M0_wRn(int rn) \
- { \
- iwmmxt_load_reg(cpu_V1, rn); \
- gen_helper_iwmmxt_##name(cpu_M0, cpu_env, cpu_M0, cpu_V1); \
- }
- #define IWMMXT_OP_ENV_SIZE(name) \
- IWMMXT_OP_ENV(name##b) \
- IWMMXT_OP_ENV(name##w) \
- IWMMXT_OP_ENV(name##l)
- #define IWMMXT_OP_ENV1(name) \
- static inline void gen_op_iwmmxt_##name##_M0(void) \
- { \
- gen_helper_iwmmxt_##name(cpu_M0, cpu_env, cpu_M0); \
- }
- IWMMXT_OP(maddsq)
- IWMMXT_OP(madduq)
- IWMMXT_OP(sadb)
- IWMMXT_OP(sadw)
- IWMMXT_OP(mulslw)
- IWMMXT_OP(mulshw)
- IWMMXT_OP(mululw)
- IWMMXT_OP(muluhw)
- IWMMXT_OP(macsw)
- IWMMXT_OP(macuw)
- IWMMXT_OP_ENV_SIZE(unpackl)
- IWMMXT_OP_ENV_SIZE(unpackh)
- IWMMXT_OP_ENV1(unpacklub)
- IWMMXT_OP_ENV1(unpackluw)
- IWMMXT_OP_ENV1(unpacklul)
- IWMMXT_OP_ENV1(unpackhub)
- IWMMXT_OP_ENV1(unpackhuw)
- IWMMXT_OP_ENV1(unpackhul)
- IWMMXT_OP_ENV1(unpacklsb)
- IWMMXT_OP_ENV1(unpacklsw)
- IWMMXT_OP_ENV1(unpacklsl)
- IWMMXT_OP_ENV1(unpackhsb)
- IWMMXT_OP_ENV1(unpackhsw)
- IWMMXT_OP_ENV1(unpackhsl)
- IWMMXT_OP_ENV_SIZE(cmpeq)
- IWMMXT_OP_ENV_SIZE(cmpgtu)
- IWMMXT_OP_ENV_SIZE(cmpgts)
- IWMMXT_OP_ENV_SIZE(mins)
- IWMMXT_OP_ENV_SIZE(minu)
- IWMMXT_OP_ENV_SIZE(maxs)
- IWMMXT_OP_ENV_SIZE(maxu)
- IWMMXT_OP_ENV_SIZE(subn)
- IWMMXT_OP_ENV_SIZE(addn)
- IWMMXT_OP_ENV_SIZE(subu)
- IWMMXT_OP_ENV_SIZE(addu)
- IWMMXT_OP_ENV_SIZE(subs)
- IWMMXT_OP_ENV_SIZE(adds)
- IWMMXT_OP_ENV(avgb0)
- IWMMXT_OP_ENV(avgb1)
- IWMMXT_OP_ENV(avgw0)
- IWMMXT_OP_ENV(avgw1)
- IWMMXT_OP_ENV(packuw)
- IWMMXT_OP_ENV(packul)
- IWMMXT_OP_ENV(packuq)
- IWMMXT_OP_ENV(packsw)
- IWMMXT_OP_ENV(packsl)
- IWMMXT_OP_ENV(packsq)
- static void gen_op_iwmmxt_set_mup(void)
- {
- TCGv_i32 tmp;
- tmp = load_cpu_field(iwmmxt.cregs[ARM_IWMMXT_wCon]);
- tcg_gen_ori_i32(tmp, tmp, 2);
- store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCon]);
- }
- static void gen_op_iwmmxt_set_cup(void)
- {
- TCGv_i32 tmp;
- tmp = load_cpu_field(iwmmxt.cregs[ARM_IWMMXT_wCon]);
- tcg_gen_ori_i32(tmp, tmp, 1);
- store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCon]);
- }
- static void gen_op_iwmmxt_setpsr_nz(void)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- gen_helper_iwmmxt_setpsr_nz(tmp, cpu_M0);
- store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCASF]);
- }
- static inline void gen_op_iwmmxt_addl_M0_wRn(int rn)
- {
- iwmmxt_load_reg(cpu_V1, rn);
- tcg_gen_ext32u_i64(cpu_V1, cpu_V1);
- tcg_gen_add_i64(cpu_M0, cpu_M0, cpu_V1);
- }
- static inline int gen_iwmmxt_address(DisasContext *s, uint32_t insn,
- TCGv_i32 dest)
- {
- int rd;
- uint32_t offset;
- TCGv_i32 tmp;
- rd = (insn >> 16) & 0xf;
- tmp = load_reg(s, rd);
- offset = (insn & 0xff) << ((insn >> 7) & 2);
- if (insn & (1 << 24)) {
- /* Pre indexed */
- if (insn & (1 << 23))
- tcg_gen_addi_i32(tmp, tmp, offset);
- else
- tcg_gen_addi_i32(tmp, tmp, -offset);
- tcg_gen_mov_i32(dest, tmp);
- if (insn & (1 << 21))
- store_reg(s, rd, tmp);
- else
- tcg_temp_free_i32(tmp);
- } else if (insn & (1 << 21)) {
- /* Post indexed */
- tcg_gen_mov_i32(dest, tmp);
- if (insn & (1 << 23))
- tcg_gen_addi_i32(tmp, tmp, offset);
- else
- tcg_gen_addi_i32(tmp, tmp, -offset);
- store_reg(s, rd, tmp);
- } else if (!(insn & (1 << 23)))
- return 1;
- return 0;
- }
- static inline int gen_iwmmxt_shift(uint32_t insn, uint32_t mask, TCGv_i32 dest)
- {
- int rd = (insn >> 0) & 0xf;
- TCGv_i32 tmp;
- if (insn & (1 << 8)) {
- if (rd < ARM_IWMMXT_wCGR0 || rd > ARM_IWMMXT_wCGR3) {
- return 1;
- } else {
- tmp = iwmmxt_load_creg(rd);
- }
- } else {
- tmp = tcg_temp_new_i32();
- iwmmxt_load_reg(cpu_V0, rd);
- tcg_gen_extrl_i64_i32(tmp, cpu_V0);
- }
- tcg_gen_andi_i32(tmp, tmp, mask);
- tcg_gen_mov_i32(dest, tmp);
- tcg_temp_free_i32(tmp);
- return 0;
- }
- /* Disassemble an iwMMXt instruction. Returns nonzero if an error occurred
- (ie. an undefined instruction). */
- static int disas_iwmmxt_insn(DisasContext *s, uint32_t insn)
- {
- int rd, wrd;
- int rdhi, rdlo, rd0, rd1, i;
- TCGv_i32 addr;
- TCGv_i32 tmp, tmp2, tmp3;
- if ((insn & 0x0e000e00) == 0x0c000000) {
- if ((insn & 0x0fe00ff0) == 0x0c400000) {
- wrd = insn & 0xf;
- rdlo = (insn >> 12) & 0xf;
- rdhi = (insn >> 16) & 0xf;
- if (insn & ARM_CP_RW_BIT) { /* TMRRC */
- iwmmxt_load_reg(cpu_V0, wrd);
- tcg_gen_extrl_i64_i32(cpu_R[rdlo], cpu_V0);
- tcg_gen_extrh_i64_i32(cpu_R[rdhi], cpu_V0);
- } else { /* TMCRR */
- tcg_gen_concat_i32_i64(cpu_V0, cpu_R[rdlo], cpu_R[rdhi]);
- iwmmxt_store_reg(cpu_V0, wrd);
- gen_op_iwmmxt_set_mup();
- }
- return 0;
- }
- wrd = (insn >> 12) & 0xf;
- addr = tcg_temp_new_i32();
- if (gen_iwmmxt_address(s, insn, addr)) {
- tcg_temp_free_i32(addr);
- return 1;
- }
- if (insn & ARM_CP_RW_BIT) {
- if ((insn >> 28) == 0xf) { /* WLDRW wCx */
- tmp = tcg_temp_new_i32();
- gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
- iwmmxt_store_creg(wrd, tmp);
- } else {
- i = 1;
- if (insn & (1 << 8)) {
- if (insn & (1 << 22)) { /* WLDRD */
- gen_aa32_ld64(s, cpu_M0, addr, get_mem_index(s));
- i = 0;
- } else { /* WLDRW wRd */
- tmp = tcg_temp_new_i32();
- gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
- }
- } else {
- tmp = tcg_temp_new_i32();
- if (insn & (1 << 22)) { /* WLDRH */
- gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
- } else { /* WLDRB */
- gen_aa32_ld8u(s, tmp, addr, get_mem_index(s));
- }
- }
- if (i) {
- tcg_gen_extu_i32_i64(cpu_M0, tmp);
- tcg_temp_free_i32(tmp);
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- }
- } else {
- if ((insn >> 28) == 0xf) { /* WSTRW wCx */
- tmp = iwmmxt_load_creg(wrd);
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
- } else {
- gen_op_iwmmxt_movq_M0_wRn(wrd);
- tmp = tcg_temp_new_i32();
- if (insn & (1 << 8)) {
- if (insn & (1 << 22)) { /* WSTRD */
- gen_aa32_st64(s, cpu_M0, addr, get_mem_index(s));
- } else { /* WSTRW wRd */
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
- }
- } else {
- if (insn & (1 << 22)) { /* WSTRH */
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- gen_aa32_st16(s, tmp, addr, get_mem_index(s));
- } else { /* WSTRB */
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- gen_aa32_st8(s, tmp, addr, get_mem_index(s));
- }
- }
- }
- tcg_temp_free_i32(tmp);
- }
- tcg_temp_free_i32(addr);
- return 0;
- }
- if ((insn & 0x0f000000) != 0x0e000000)
- return 1;
- switch (((insn >> 12) & 0xf00) | ((insn >> 4) & 0xff)) {
- case 0x000: /* WOR */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 0) & 0xf;
- rd1 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- gen_op_iwmmxt_orq_M0_wRn(rd1);
- gen_op_iwmmxt_setpsr_nz();
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x011: /* TMCR */
- if (insn & 0xf)
- return 1;
- rd = (insn >> 12) & 0xf;
- wrd = (insn >> 16) & 0xf;
- switch (wrd) {
- case ARM_IWMMXT_wCID:
- case ARM_IWMMXT_wCASF:
- break;
- case ARM_IWMMXT_wCon:
- gen_op_iwmmxt_set_cup();
- /* Fall through. */
- case ARM_IWMMXT_wCSSF:
- tmp = iwmmxt_load_creg(wrd);
- tmp2 = load_reg(s, rd);
- tcg_gen_andc_i32(tmp, tmp, tmp2);
- tcg_temp_free_i32(tmp2);
- iwmmxt_store_creg(wrd, tmp);
- break;
- case ARM_IWMMXT_wCGR0:
- case ARM_IWMMXT_wCGR1:
- case ARM_IWMMXT_wCGR2:
- case ARM_IWMMXT_wCGR3:
- gen_op_iwmmxt_set_cup();
- tmp = load_reg(s, rd);
- iwmmxt_store_creg(wrd, tmp);
- break;
- default:
- return 1;
- }
- break;
- case 0x100: /* WXOR */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 0) & 0xf;
- rd1 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- gen_op_iwmmxt_xorq_M0_wRn(rd1);
- gen_op_iwmmxt_setpsr_nz();
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x111: /* TMRC */
- if (insn & 0xf)
- return 1;
- rd = (insn >> 12) & 0xf;
- wrd = (insn >> 16) & 0xf;
- tmp = iwmmxt_load_creg(wrd);
- store_reg(s, rd, tmp);
- break;
- case 0x300: /* WANDN */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 0) & 0xf;
- rd1 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tcg_gen_neg_i64(cpu_M0, cpu_M0);
- gen_op_iwmmxt_andq_M0_wRn(rd1);
- gen_op_iwmmxt_setpsr_nz();
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x200: /* WAND */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 0) & 0xf;
- rd1 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- gen_op_iwmmxt_andq_M0_wRn(rd1);
- gen_op_iwmmxt_setpsr_nz();
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x810: case 0xa10: /* WMADD */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 0) & 0xf;
- rd1 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- if (insn & (1 << 21))
- gen_op_iwmmxt_maddsq_M0_wRn(rd1);
- else
- gen_op_iwmmxt_madduq_M0_wRn(rd1);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x10e: case 0x50e: case 0x90e: case 0xd0e: /* WUNPCKIL */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- gen_op_iwmmxt_unpacklb_M0_wRn(rd1);
- break;
- case 1:
- gen_op_iwmmxt_unpacklw_M0_wRn(rd1);
- break;
- case 2:
- gen_op_iwmmxt_unpackll_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x10c: case 0x50c: case 0x90c: case 0xd0c: /* WUNPCKIH */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- gen_op_iwmmxt_unpackhb_M0_wRn(rd1);
- break;
- case 1:
- gen_op_iwmmxt_unpackhw_M0_wRn(rd1);
- break;
- case 2:
- gen_op_iwmmxt_unpackhl_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x012: case 0x112: case 0x412: case 0x512: /* WSAD */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- if (insn & (1 << 22))
- gen_op_iwmmxt_sadw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_sadb_M0_wRn(rd1);
- if (!(insn & (1 << 20)))
- gen_op_iwmmxt_addl_M0_wRn(wrd);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x010: case 0x110: case 0x210: case 0x310: /* WMUL */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- if (insn & (1 << 21)) {
- if (insn & (1 << 20))
- gen_op_iwmmxt_mulshw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_mulslw_M0_wRn(rd1);
- } else {
- if (insn & (1 << 20))
- gen_op_iwmmxt_muluhw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_mululw_M0_wRn(rd1);
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x410: case 0x510: case 0x610: case 0x710: /* WMAC */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- if (insn & (1 << 21))
- gen_op_iwmmxt_macsw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_macuw_M0_wRn(rd1);
- if (!(insn & (1 << 20))) {
- iwmmxt_load_reg(cpu_V1, wrd);
- tcg_gen_add_i64(cpu_M0, cpu_M0, cpu_V1);
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x006: case 0x406: case 0x806: case 0xc06: /* WCMPEQ */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- gen_op_iwmmxt_cmpeqb_M0_wRn(rd1);
- break;
- case 1:
- gen_op_iwmmxt_cmpeqw_M0_wRn(rd1);
- break;
- case 2:
- gen_op_iwmmxt_cmpeql_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x800: case 0x900: case 0xc00: case 0xd00: /* WAVG2 */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- if (insn & (1 << 22)) {
- if (insn & (1 << 20))
- gen_op_iwmmxt_avgw1_M0_wRn(rd1);
- else
- gen_op_iwmmxt_avgw0_M0_wRn(rd1);
- } else {
- if (insn & (1 << 20))
- gen_op_iwmmxt_avgb1_M0_wRn(rd1);
- else
- gen_op_iwmmxt_avgb0_M0_wRn(rd1);
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x802: case 0x902: case 0xa02: case 0xb02: /* WALIGNR */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = iwmmxt_load_creg(ARM_IWMMXT_wCGR0 + ((insn >> 20) & 3));
- tcg_gen_andi_i32(tmp, tmp, 7);
- iwmmxt_load_reg(cpu_V1, rd1);
- gen_helper_iwmmxt_align(cpu_M0, cpu_M0, cpu_V1, tmp);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x601: case 0x605: case 0x609: case 0x60d: /* TINSR */
- if (((insn >> 6) & 3) == 3)
- return 1;
- rd = (insn >> 12) & 0xf;
- wrd = (insn >> 16) & 0xf;
- tmp = load_reg(s, rd);
- gen_op_iwmmxt_movq_M0_wRn(wrd);
- switch ((insn >> 6) & 3) {
- case 0:
- tmp2 = tcg_const_i32(0xff);
- tmp3 = tcg_const_i32((insn & 7) << 3);
- break;
- case 1:
- tmp2 = tcg_const_i32(0xffff);
- tmp3 = tcg_const_i32((insn & 3) << 4);
- break;
- case 2:
- tmp2 = tcg_const_i32(0xffffffff);
- tmp3 = tcg_const_i32((insn & 1) << 5);
- break;
- default:
- tmp2 = NULL;
- tmp3 = NULL;
- }
- gen_helper_iwmmxt_insr(cpu_M0, cpu_M0, tmp, tmp2, tmp3);
- tcg_temp_free_i32(tmp3);
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x107: case 0x507: case 0x907: case 0xd07: /* TEXTRM */
- rd = (insn >> 12) & 0xf;
- wrd = (insn >> 16) & 0xf;
- if (rd == 15 || ((insn >> 22) & 3) == 3)
- return 1;
- gen_op_iwmmxt_movq_M0_wRn(wrd);
- tmp = tcg_temp_new_i32();
- switch ((insn >> 22) & 3) {
- case 0:
- tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 7) << 3);
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- if (insn & 8) {
- tcg_gen_ext8s_i32(tmp, tmp);
- } else {
- tcg_gen_andi_i32(tmp, tmp, 0xff);
- }
- break;
- case 1:
- tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 3) << 4);
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- if (insn & 8) {
- tcg_gen_ext16s_i32(tmp, tmp);
- } else {
- tcg_gen_andi_i32(tmp, tmp, 0xffff);
- }
- break;
- case 2:
- tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 1) << 5);
- tcg_gen_extrl_i64_i32(tmp, cpu_M0);
- break;
- }
- store_reg(s, rd, tmp);
- break;
- case 0x117: case 0x517: case 0x917: case 0xd17: /* TEXTRC */
- if ((insn & 0x000ff008) != 0x0003f000 || ((insn >> 22) & 3) == 3)
- return 1;
- tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
- switch ((insn >> 22) & 3) {
- case 0:
- tcg_gen_shri_i32(tmp, tmp, ((insn & 7) << 2) + 0);
- break;
- case 1:
- tcg_gen_shri_i32(tmp, tmp, ((insn & 3) << 3) + 4);
- break;
- case 2:
- tcg_gen_shri_i32(tmp, tmp, ((insn & 1) << 4) + 12);
- break;
- }
- tcg_gen_shli_i32(tmp, tmp, 28);
- gen_set_nzcv(tmp);
- tcg_temp_free_i32(tmp);
- break;
- case 0x401: case 0x405: case 0x409: case 0x40d: /* TBCST */
- if (((insn >> 6) & 3) == 3)
- return 1;
- rd = (insn >> 12) & 0xf;
- wrd = (insn >> 16) & 0xf;
- tmp = load_reg(s, rd);
- switch ((insn >> 6) & 3) {
- case 0:
- gen_helper_iwmmxt_bcstb(cpu_M0, tmp);
- break;
- case 1:
- gen_helper_iwmmxt_bcstw(cpu_M0, tmp);
- break;
- case 2:
- gen_helper_iwmmxt_bcstl(cpu_M0, tmp);
- break;
- }
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x113: case 0x513: case 0x913: case 0xd13: /* TANDC */
- if ((insn & 0x000ff00f) != 0x0003f000 || ((insn >> 22) & 3) == 3)
- return 1;
- tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
- tmp2 = tcg_temp_new_i32();
- tcg_gen_mov_i32(tmp2, tmp);
- switch ((insn >> 22) & 3) {
- case 0:
- for (i = 0; i < 7; i ++) {
- tcg_gen_shli_i32(tmp2, tmp2, 4);
- tcg_gen_and_i32(tmp, tmp, tmp2);
- }
- break;
- case 1:
- for (i = 0; i < 3; i ++) {
- tcg_gen_shli_i32(tmp2, tmp2, 8);
- tcg_gen_and_i32(tmp, tmp, tmp2);
- }
- break;
- case 2:
- tcg_gen_shli_i32(tmp2, tmp2, 16);
- tcg_gen_and_i32(tmp, tmp, tmp2);
- break;
- }
- gen_set_nzcv(tmp);
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- break;
- case 0x01c: case 0x41c: case 0x81c: case 0xc1c: /* WACC */
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- gen_helper_iwmmxt_addcb(cpu_M0, cpu_M0);
- break;
- case 1:
- gen_helper_iwmmxt_addcw(cpu_M0, cpu_M0);
- break;
- case 2:
- gen_helper_iwmmxt_addcl(cpu_M0, cpu_M0);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x115: case 0x515: case 0x915: case 0xd15: /* TORC */
- if ((insn & 0x000ff00f) != 0x0003f000 || ((insn >> 22) & 3) == 3)
- return 1;
- tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
- tmp2 = tcg_temp_new_i32();
- tcg_gen_mov_i32(tmp2, tmp);
- switch ((insn >> 22) & 3) {
- case 0:
- for (i = 0; i < 7; i ++) {
- tcg_gen_shli_i32(tmp2, tmp2, 4);
- tcg_gen_or_i32(tmp, tmp, tmp2);
- }
- break;
- case 1:
- for (i = 0; i < 3; i ++) {
- tcg_gen_shli_i32(tmp2, tmp2, 8);
- tcg_gen_or_i32(tmp, tmp, tmp2);
- }
- break;
- case 2:
- tcg_gen_shli_i32(tmp2, tmp2, 16);
- tcg_gen_or_i32(tmp, tmp, tmp2);
- break;
- }
- gen_set_nzcv(tmp);
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- break;
- case 0x103: case 0x503: case 0x903: case 0xd03: /* TMOVMSK */
- rd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- if ((insn & 0xf) != 0 || ((insn >> 22) & 3) == 3)
- return 1;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_temp_new_i32();
- switch ((insn >> 22) & 3) {
- case 0:
- gen_helper_iwmmxt_msbb(tmp, cpu_M0);
- break;
- case 1:
- gen_helper_iwmmxt_msbw(tmp, cpu_M0);
- break;
- case 2:
- gen_helper_iwmmxt_msbl(tmp, cpu_M0);
- break;
- }
- store_reg(s, rd, tmp);
- break;
- case 0x106: case 0x306: case 0x506: case 0x706: /* WCMPGT */
- case 0x906: case 0xb06: case 0xd06: case 0xf06:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- if (insn & (1 << 21))
- gen_op_iwmmxt_cmpgtsb_M0_wRn(rd1);
- else
- gen_op_iwmmxt_cmpgtub_M0_wRn(rd1);
- break;
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_cmpgtsw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_cmpgtuw_M0_wRn(rd1);
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_cmpgtsl_M0_wRn(rd1);
- else
- gen_op_iwmmxt_cmpgtul_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x00e: case 0x20e: case 0x40e: case 0x60e: /* WUNPCKEL */
- case 0x80e: case 0xa0e: case 0xc0e: case 0xe0e:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpacklsb_M0();
- else
- gen_op_iwmmxt_unpacklub_M0();
- break;
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpacklsw_M0();
- else
- gen_op_iwmmxt_unpackluw_M0();
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpacklsl_M0();
- else
- gen_op_iwmmxt_unpacklul_M0();
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x00c: case 0x20c: case 0x40c: case 0x60c: /* WUNPCKEH */
- case 0x80c: case 0xa0c: case 0xc0c: case 0xe0c:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpackhsb_M0();
- else
- gen_op_iwmmxt_unpackhub_M0();
- break;
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpackhsw_M0();
- else
- gen_op_iwmmxt_unpackhuw_M0();
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_unpackhsl_M0();
- else
- gen_op_iwmmxt_unpackhul_M0();
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x204: case 0x604: case 0xa04: case 0xe04: /* WSRL */
- case 0x214: case 0x614: case 0xa14: case 0xe14:
- if (((insn >> 22) & 3) == 0)
- return 1;
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_temp_new_i32();
- if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- switch ((insn >> 22) & 3) {
- case 1:
- gen_helper_iwmmxt_srlw(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 2:
- gen_helper_iwmmxt_srll(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 3:
- gen_helper_iwmmxt_srlq(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- }
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x004: case 0x404: case 0x804: case 0xc04: /* WSRA */
- case 0x014: case 0x414: case 0x814: case 0xc14:
- if (((insn >> 22) & 3) == 0)
- return 1;
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_temp_new_i32();
- if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- switch ((insn >> 22) & 3) {
- case 1:
- gen_helper_iwmmxt_sraw(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 2:
- gen_helper_iwmmxt_sral(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 3:
- gen_helper_iwmmxt_sraq(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- }
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x104: case 0x504: case 0x904: case 0xd04: /* WSLL */
- case 0x114: case 0x514: case 0x914: case 0xd14:
- if (((insn >> 22) & 3) == 0)
- return 1;
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_temp_new_i32();
- if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- switch ((insn >> 22) & 3) {
- case 1:
- gen_helper_iwmmxt_sllw(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 2:
- gen_helper_iwmmxt_slll(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 3:
- gen_helper_iwmmxt_sllq(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- }
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x304: case 0x704: case 0xb04: case 0xf04: /* WROR */
- case 0x314: case 0x714: case 0xb14: case 0xf14:
- if (((insn >> 22) & 3) == 0)
- return 1;
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_temp_new_i32();
- switch ((insn >> 22) & 3) {
- case 1:
- if (gen_iwmmxt_shift(insn, 0xf, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- gen_helper_iwmmxt_rorw(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 2:
- if (gen_iwmmxt_shift(insn, 0x1f, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- gen_helper_iwmmxt_rorl(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- case 3:
- if (gen_iwmmxt_shift(insn, 0x3f, tmp)) {
- tcg_temp_free_i32(tmp);
- return 1;
- }
- gen_helper_iwmmxt_rorq(cpu_M0, cpu_env, cpu_M0, tmp);
- break;
- }
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x116: case 0x316: case 0x516: case 0x716: /* WMIN */
- case 0x916: case 0xb16: case 0xd16: case 0xf16:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- if (insn & (1 << 21))
- gen_op_iwmmxt_minsb_M0_wRn(rd1);
- else
- gen_op_iwmmxt_minub_M0_wRn(rd1);
- break;
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_minsw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_minuw_M0_wRn(rd1);
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_minsl_M0_wRn(rd1);
- else
- gen_op_iwmmxt_minul_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x016: case 0x216: case 0x416: case 0x616: /* WMAX */
- case 0x816: case 0xa16: case 0xc16: case 0xe16:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 0:
- if (insn & (1 << 21))
- gen_op_iwmmxt_maxsb_M0_wRn(rd1);
- else
- gen_op_iwmmxt_maxub_M0_wRn(rd1);
- break;
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_maxsw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_maxuw_M0_wRn(rd1);
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_maxsl_M0_wRn(rd1);
- else
- gen_op_iwmmxt_maxul_M0_wRn(rd1);
- break;
- case 3:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x002: case 0x102: case 0x202: case 0x302: /* WALIGNI */
- case 0x402: case 0x502: case 0x602: case 0x702:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_const_i32((insn >> 20) & 3);
- iwmmxt_load_reg(cpu_V1, rd1);
- gen_helper_iwmmxt_align(cpu_M0, cpu_M0, cpu_V1, tmp);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- case 0x01a: case 0x11a: case 0x21a: case 0x31a: /* WSUB */
- case 0x41a: case 0x51a: case 0x61a: case 0x71a:
- case 0x81a: case 0x91a: case 0xa1a: case 0xb1a:
- case 0xc1a: case 0xd1a: case 0xe1a: case 0xf1a:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 20) & 0xf) {
- case 0x0:
- gen_op_iwmmxt_subnb_M0_wRn(rd1);
- break;
- case 0x1:
- gen_op_iwmmxt_subub_M0_wRn(rd1);
- break;
- case 0x3:
- gen_op_iwmmxt_subsb_M0_wRn(rd1);
- break;
- case 0x4:
- gen_op_iwmmxt_subnw_M0_wRn(rd1);
- break;
- case 0x5:
- gen_op_iwmmxt_subuw_M0_wRn(rd1);
- break;
- case 0x7:
- gen_op_iwmmxt_subsw_M0_wRn(rd1);
- break;
- case 0x8:
- gen_op_iwmmxt_subnl_M0_wRn(rd1);
- break;
- case 0x9:
- gen_op_iwmmxt_subul_M0_wRn(rd1);
- break;
- case 0xb:
- gen_op_iwmmxt_subsl_M0_wRn(rd1);
- break;
- default:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x01e: case 0x11e: case 0x21e: case 0x31e: /* WSHUFH */
- case 0x41e: case 0x51e: case 0x61e: case 0x71e:
- case 0x81e: case 0x91e: case 0xa1e: case 0xb1e:
- case 0xc1e: case 0xd1e: case 0xe1e: case 0xf1e:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- tmp = tcg_const_i32(((insn >> 16) & 0xf0) | (insn & 0x0f));
- gen_helper_iwmmxt_shufh(cpu_M0, cpu_env, cpu_M0, tmp);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x018: case 0x118: case 0x218: case 0x318: /* WADD */
- case 0x418: case 0x518: case 0x618: case 0x718:
- case 0x818: case 0x918: case 0xa18: case 0xb18:
- case 0xc18: case 0xd18: case 0xe18: case 0xf18:
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 20) & 0xf) {
- case 0x0:
- gen_op_iwmmxt_addnb_M0_wRn(rd1);
- break;
- case 0x1:
- gen_op_iwmmxt_addub_M0_wRn(rd1);
- break;
- case 0x3:
- gen_op_iwmmxt_addsb_M0_wRn(rd1);
- break;
- case 0x4:
- gen_op_iwmmxt_addnw_M0_wRn(rd1);
- break;
- case 0x5:
- gen_op_iwmmxt_adduw_M0_wRn(rd1);
- break;
- case 0x7:
- gen_op_iwmmxt_addsw_M0_wRn(rd1);
- break;
- case 0x8:
- gen_op_iwmmxt_addnl_M0_wRn(rd1);
- break;
- case 0x9:
- gen_op_iwmmxt_addul_M0_wRn(rd1);
- break;
- case 0xb:
- gen_op_iwmmxt_addsl_M0_wRn(rd1);
- break;
- default:
- return 1;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x008: case 0x108: case 0x208: case 0x308: /* WPACK */
- case 0x408: case 0x508: case 0x608: case 0x708:
- case 0x808: case 0x908: case 0xa08: case 0xb08:
- case 0xc08: case 0xd08: case 0xe08: case 0xf08:
- if (!(insn & (1 << 20)) || ((insn >> 22) & 3) == 0)
- return 1;
- wrd = (insn >> 12) & 0xf;
- rd0 = (insn >> 16) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- gen_op_iwmmxt_movq_M0_wRn(rd0);
- switch ((insn >> 22) & 3) {
- case 1:
- if (insn & (1 << 21))
- gen_op_iwmmxt_packsw_M0_wRn(rd1);
- else
- gen_op_iwmmxt_packuw_M0_wRn(rd1);
- break;
- case 2:
- if (insn & (1 << 21))
- gen_op_iwmmxt_packsl_M0_wRn(rd1);
- else
- gen_op_iwmmxt_packul_M0_wRn(rd1);
- break;
- case 3:
- if (insn & (1 << 21))
- gen_op_iwmmxt_packsq_M0_wRn(rd1);
- else
- gen_op_iwmmxt_packuq_M0_wRn(rd1);
- break;
- }
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- gen_op_iwmmxt_set_cup();
- break;
- case 0x201: case 0x203: case 0x205: case 0x207:
- case 0x209: case 0x20b: case 0x20d: case 0x20f:
- case 0x211: case 0x213: case 0x215: case 0x217:
- case 0x219: case 0x21b: case 0x21d: case 0x21f:
- wrd = (insn >> 5) & 0xf;
- rd0 = (insn >> 12) & 0xf;
- rd1 = (insn >> 0) & 0xf;
- if (rd0 == 0xf || rd1 == 0xf)
- return 1;
- gen_op_iwmmxt_movq_M0_wRn(wrd);
- tmp = load_reg(s, rd0);
- tmp2 = load_reg(s, rd1);
- switch ((insn >> 16) & 0xf) {
- case 0x0: /* TMIA */
- gen_helper_iwmmxt_muladdsl(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- case 0x8: /* TMIAPH */
- gen_helper_iwmmxt_muladdsw(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- case 0xc: case 0xd: case 0xe: case 0xf: /* TMIAxy */
- if (insn & (1 << 16))
- tcg_gen_shri_i32(tmp, tmp, 16);
- if (insn & (1 << 17))
- tcg_gen_shri_i32(tmp2, tmp2, 16);
- gen_helper_iwmmxt_muladdswl(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- default:
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- return 1;
- }
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(wrd);
- gen_op_iwmmxt_set_mup();
- break;
- default:
- return 1;
- }
- return 0;
- }
- /* Disassemble an XScale DSP instruction. Returns nonzero if an error occurred
- (ie. an undefined instruction). */
- static int disas_dsp_insn(DisasContext *s, uint32_t insn)
- {
- int acc, rd0, rd1, rdhi, rdlo;
- TCGv_i32 tmp, tmp2;
- if ((insn & 0x0ff00f10) == 0x0e200010) {
- /* Multiply with Internal Accumulate Format */
- rd0 = (insn >> 12) & 0xf;
- rd1 = insn & 0xf;
- acc = (insn >> 5) & 7;
- if (acc != 0)
- return 1;
- tmp = load_reg(s, rd0);
- tmp2 = load_reg(s, rd1);
- switch ((insn >> 16) & 0xf) {
- case 0x0: /* MIA */
- gen_helper_iwmmxt_muladdsl(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- case 0x8: /* MIAPH */
- gen_helper_iwmmxt_muladdsw(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- case 0xc: /* MIABB */
- case 0xd: /* MIABT */
- case 0xe: /* MIATB */
- case 0xf: /* MIATT */
- if (insn & (1 << 16))
- tcg_gen_shri_i32(tmp, tmp, 16);
- if (insn & (1 << 17))
- tcg_gen_shri_i32(tmp2, tmp2, 16);
- gen_helper_iwmmxt_muladdswl(cpu_M0, cpu_M0, tmp, tmp2);
- break;
- default:
- return 1;
- }
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- gen_op_iwmmxt_movq_wRn_M0(acc);
- return 0;
- }
- if ((insn & 0x0fe00ff8) == 0x0c400000) {
- /* Internal Accumulator Access Format */
- rdhi = (insn >> 16) & 0xf;
- rdlo = (insn >> 12) & 0xf;
- acc = insn & 7;
- if (acc != 0)
- return 1;
- if (insn & ARM_CP_RW_BIT) { /* MRA */
- iwmmxt_load_reg(cpu_V0, acc);
- tcg_gen_extrl_i64_i32(cpu_R[rdlo], cpu_V0);
- tcg_gen_extrh_i64_i32(cpu_R[rdhi], cpu_V0);
- tcg_gen_andi_i32(cpu_R[rdhi], cpu_R[rdhi], (1 << (40 - 32)) - 1);
- } else { /* MAR */
- tcg_gen_concat_i32_i64(cpu_V0, cpu_R[rdlo], cpu_R[rdhi]);
- iwmmxt_store_reg(cpu_V0, acc);
- }
- return 0;
- }
- return 1;
- }
- #define VFP_REG_SHR(x, n) (((n) > 0) ? (x) >> (n) : (x) << -(n))
- #define VFP_DREG(reg, insn, bigbit, smallbit) do { \
- if (dc_isar_feature(aa32_simd_r32, s)) { \
- reg = (((insn) >> (bigbit)) & 0x0f) \
- | (((insn) >> ((smallbit) - 4)) & 0x10); \
- } else { \
- if (insn & (1 << (smallbit))) \
- return 1; \
- reg = ((insn) >> (bigbit)) & 0x0f; \
- }} while (0)
- #define VFP_DREG_D(reg, insn) VFP_DREG(reg, insn, 12, 22)
- #define VFP_DREG_N(reg, insn) VFP_DREG(reg, insn, 16, 7)
- #define VFP_DREG_M(reg, insn) VFP_DREG(reg, insn, 0, 5)
- static inline bool use_goto_tb(DisasContext *s, target_ulong dest)
- {
- #ifndef CONFIG_USER_ONLY
- return (s->base.tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) ||
- ((s->base.pc_next - 1) & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
- #else
- return true;
- #endif
- }
- static void gen_goto_ptr(void)
- {
- tcg_gen_lookup_and_goto_ptr();
- }
- /* This will end the TB but doesn't guarantee we'll return to
- * cpu_loop_exec. Any live exit_requests will be processed as we
- * enter the next TB.
- */
- static void gen_goto_tb(DisasContext *s, int n, target_ulong dest)
- {
- if (use_goto_tb(s, dest)) {
- tcg_gen_goto_tb(n);
- gen_set_pc_im(s, dest);
- tcg_gen_exit_tb(s->base.tb, n);
- } else {
- gen_set_pc_im(s, dest);
- gen_goto_ptr();
- }
- s->base.is_jmp = DISAS_NORETURN;
- }
- static inline void gen_jmp (DisasContext *s, uint32_t dest)
- {
- if (unlikely(is_singlestepping(s))) {
- /* An indirect jump so that we still trigger the debug exception. */
- gen_set_pc_im(s, dest);
- s->base.is_jmp = DISAS_JUMP;
- } else {
- gen_goto_tb(s, 0, dest);
- }
- }
- static inline void gen_mulxy(TCGv_i32 t0, TCGv_i32 t1, int x, int y)
- {
- if (x)
- tcg_gen_sari_i32(t0, t0, 16);
- else
- gen_sxth(t0);
- if (y)
- tcg_gen_sari_i32(t1, t1, 16);
- else
- gen_sxth(t1);
- tcg_gen_mul_i32(t0, t0, t1);
- }
- /* Return the mask of PSR bits set by a MSR instruction. */
- static uint32_t msr_mask(DisasContext *s, int flags, int spsr)
- {
- uint32_t mask = 0;
- if (flags & (1 << 0)) {
- mask |= 0xff;
- }
- if (flags & (1 << 1)) {
- mask |= 0xff00;
- }
- if (flags & (1 << 2)) {
- mask |= 0xff0000;
- }
- if (flags & (1 << 3)) {
- mask |= 0xff000000;
- }
- /* Mask out undefined and reserved bits. */
- mask &= aarch32_cpsr_valid_mask(s->features, s->isar);
- /* Mask out execution state. */
- if (!spsr) {
- mask &= ~CPSR_EXEC;
- }
- /* Mask out privileged bits. */
- if (IS_USER(s)) {
- mask &= CPSR_USER;
- }
- return mask;
- }
- /* Returns nonzero if access to the PSR is not permitted. Marks t0 as dead. */
- static int gen_set_psr(DisasContext *s, uint32_t mask, int spsr, TCGv_i32 t0)
- {
- TCGv_i32 tmp;
- if (spsr) {
- /* ??? This is also undefined in system mode. */
- if (IS_USER(s))
- return 1;
- tmp = load_cpu_field(spsr);
- tcg_gen_andi_i32(tmp, tmp, ~mask);
- tcg_gen_andi_i32(t0, t0, mask);
- tcg_gen_or_i32(tmp, tmp, t0);
- store_cpu_field(tmp, spsr);
- } else {
- gen_set_cpsr(t0, mask);
- }
- tcg_temp_free_i32(t0);
- gen_lookup_tb(s);
- return 0;
- }
- /* Returns nonzero if access to the PSR is not permitted. */
- static int gen_set_psr_im(DisasContext *s, uint32_t mask, int spsr, uint32_t val)
- {
- TCGv_i32 tmp;
- tmp = tcg_temp_new_i32();
- tcg_gen_movi_i32(tmp, val);
- return gen_set_psr(s, mask, spsr, tmp);
- }
- static bool msr_banked_access_decode(DisasContext *s, int r, int sysm, int rn,
- int *tgtmode, int *regno)
- {
- /* Decode the r and sysm fields of MSR/MRS banked accesses into
- * the target mode and register number, and identify the various
- * unpredictable cases.
- * MSR (banked) and MRS (banked) are CONSTRAINED UNPREDICTABLE if:
- * + executed in user mode
- * + using R15 as the src/dest register
- * + accessing an unimplemented register
- * + accessing a register that's inaccessible at current PL/security state*
- * + accessing a register that you could access with a different insn
- * We choose to UNDEF in all these cases.
- * Since we don't know which of the various AArch32 modes we are in
- * we have to defer some checks to runtime.
- * Accesses to Monitor mode registers from Secure EL1 (which implies
- * that EL3 is AArch64) must trap to EL3.
- *
- * If the access checks fail this function will emit code to take
- * an exception and return false. Otherwise it will return true,
- * and set *tgtmode and *regno appropriately.
- */
- int exc_target = default_exception_el(s);
- /* These instructions are present only in ARMv8, or in ARMv7 with the
- * Virtualization Extensions.
- */
- if (!arm_dc_feature(s, ARM_FEATURE_V8) &&
- !arm_dc_feature(s, ARM_FEATURE_EL2)) {
- goto undef;
- }
- if (IS_USER(s) || rn == 15) {
- goto undef;
- }
- /* The table in the v8 ARM ARM section F5.2.3 describes the encoding
- * of registers into (r, sysm).
- */
- if (r) {
- /* SPSRs for other modes */
- switch (sysm) {
- case 0xe: /* SPSR_fiq */
- *tgtmode = ARM_CPU_MODE_FIQ;
- break;
- case 0x10: /* SPSR_irq */
- *tgtmode = ARM_CPU_MODE_IRQ;
- break;
- case 0x12: /* SPSR_svc */
- *tgtmode = ARM_CPU_MODE_SVC;
- break;
- case 0x14: /* SPSR_abt */
- *tgtmode = ARM_CPU_MODE_ABT;
- break;
- case 0x16: /* SPSR_und */
- *tgtmode = ARM_CPU_MODE_UND;
- break;
- case 0x1c: /* SPSR_mon */
- *tgtmode = ARM_CPU_MODE_MON;
- break;
- case 0x1e: /* SPSR_hyp */
- *tgtmode = ARM_CPU_MODE_HYP;
- break;
- default: /* unallocated */
- goto undef;
- }
- /* We arbitrarily assign SPSR a register number of 16. */
- *regno = 16;
- } else {
- /* general purpose registers for other modes */
- switch (sysm) {
- case 0x0 ... 0x6: /* 0b00xxx : r8_usr ... r14_usr */
- *tgtmode = ARM_CPU_MODE_USR;
- *regno = sysm + 8;
- break;
- case 0x8 ... 0xe: /* 0b01xxx : r8_fiq ... r14_fiq */
- *tgtmode = ARM_CPU_MODE_FIQ;
- *regno = sysm;
- break;
- case 0x10 ... 0x11: /* 0b1000x : r14_irq, r13_irq */
- *tgtmode = ARM_CPU_MODE_IRQ;
- *regno = sysm & 1 ? 13 : 14;
- break;
- case 0x12 ... 0x13: /* 0b1001x : r14_svc, r13_svc */
- *tgtmode = ARM_CPU_MODE_SVC;
- *regno = sysm & 1 ? 13 : 14;
- break;
- case 0x14 ... 0x15: /* 0b1010x : r14_abt, r13_abt */
- *tgtmode = ARM_CPU_MODE_ABT;
- *regno = sysm & 1 ? 13 : 14;
- break;
- case 0x16 ... 0x17: /* 0b1011x : r14_und, r13_und */
- *tgtmode = ARM_CPU_MODE_UND;
- *regno = sysm & 1 ? 13 : 14;
- break;
- case 0x1c ... 0x1d: /* 0b1110x : r14_mon, r13_mon */
- *tgtmode = ARM_CPU_MODE_MON;
- *regno = sysm & 1 ? 13 : 14;
- break;
- case 0x1e ... 0x1f: /* 0b1111x : elr_hyp, r13_hyp */
- *tgtmode = ARM_CPU_MODE_HYP;
- /* Arbitrarily pick 17 for ELR_Hyp (which is not a banked LR!) */
- *regno = sysm & 1 ? 13 : 17;
- break;
- default: /* unallocated */
- goto undef;
- }
- }
- /* Catch the 'accessing inaccessible register' cases we can detect
- * at translate time.
- */
- switch (*tgtmode) {
- case ARM_CPU_MODE_MON:
- if (!arm_dc_feature(s, ARM_FEATURE_EL3) || s->ns) {
- goto undef;
- }
- if (s->current_el == 1) {
- /* If we're in Secure EL1 (which implies that EL3 is AArch64)
- * then accesses to Mon registers trap to EL3
- */
- exc_target = 3;
- goto undef;
- }
- break;
- case ARM_CPU_MODE_HYP:
- /*
- * SPSR_hyp and r13_hyp can only be accessed from Monitor mode
- * (and so we can forbid accesses from EL2 or below). elr_hyp
- * can be accessed also from Hyp mode, so forbid accesses from
- * EL0 or EL1.
- */
- if (!arm_dc_feature(s, ARM_FEATURE_EL2) || s->current_el < 2 ||
- (s->current_el < 3 && *regno != 17)) {
- goto undef;
- }
- break;
- default:
- break;
- }
- return true;
- undef:
- /* If we get here then some access check did not pass */
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
- syn_uncategorized(), exc_target);
- return false;
- }
- static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
- {
- TCGv_i32 tcg_reg, tcg_tgtmode, tcg_regno;
- int tgtmode = 0, regno = 0;
- if (!msr_banked_access_decode(s, r, sysm, rn, &tgtmode, ®no)) {
- return;
- }
- /* Sync state because msr_banked() can raise exceptions */
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- tcg_reg = load_reg(s, rn);
- tcg_tgtmode = tcg_const_i32(tgtmode);
- tcg_regno = tcg_const_i32(regno);
- gen_helper_msr_banked(cpu_env, tcg_reg, tcg_tgtmode, tcg_regno);
- tcg_temp_free_i32(tcg_tgtmode);
- tcg_temp_free_i32(tcg_regno);
- tcg_temp_free_i32(tcg_reg);
- s->base.is_jmp = DISAS_UPDATE_EXIT;
- }
- static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
- {
- TCGv_i32 tcg_reg, tcg_tgtmode, tcg_regno;
- int tgtmode = 0, regno = 0;
- if (!msr_banked_access_decode(s, r, sysm, rn, &tgtmode, ®no)) {
- return;
- }
- /* Sync state because mrs_banked() can raise exceptions */
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- tcg_reg = tcg_temp_new_i32();
- tcg_tgtmode = tcg_const_i32(tgtmode);
- tcg_regno = tcg_const_i32(regno);
- gen_helper_mrs_banked(tcg_reg, cpu_env, tcg_tgtmode, tcg_regno);
- tcg_temp_free_i32(tcg_tgtmode);
- tcg_temp_free_i32(tcg_regno);
- store_reg(s, rn, tcg_reg);
- s->base.is_jmp = DISAS_UPDATE_EXIT;
- }
- /* Store value to PC as for an exception return (ie don't
- * mask bits). The subsequent call to gen_helper_cpsr_write_eret()
- * will do the masking based on the new value of the Thumb bit.
- */
- static void store_pc_exc_ret(DisasContext *s, TCGv_i32 pc)
- {
- tcg_gen_mov_i32(cpu_R[15], pc);
- tcg_temp_free_i32(pc);
- }
- /* Generate a v6 exception return. Marks both values as dead. */
- static void gen_rfe(DisasContext *s, TCGv_i32 pc, TCGv_i32 cpsr)
- {
- store_pc_exc_ret(s, pc);
- /* The cpsr_write_eret helper will mask the low bits of PC
- * appropriately depending on the new Thumb bit, so it must
- * be called after storing the new PC.
- */
- if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
- gen_io_start();
- }
- gen_helper_cpsr_write_eret(cpu_env, cpsr);
- tcg_temp_free_i32(cpsr);
- /* Must exit loop to check un-masked IRQs */
- s->base.is_jmp = DISAS_EXIT;
- }
- /* Generate an old-style exception return. Marks pc as dead. */
- static void gen_exception_return(DisasContext *s, TCGv_i32 pc)
- {
- gen_rfe(s, pc, load_cpu_field(spsr));
- }
- static void gen_gvec_fn3_qc(uint32_t rd_ofs, uint32_t rn_ofs, uint32_t rm_ofs,
- uint32_t opr_sz, uint32_t max_sz,
- gen_helper_gvec_3_ptr *fn)
- {
- TCGv_ptr qc_ptr = tcg_temp_new_ptr();
- tcg_gen_addi_ptr(qc_ptr, cpu_env, offsetof(CPUARMState, vfp.qc));
- tcg_gen_gvec_3_ptr(rd_ofs, rn_ofs, rm_ofs, qc_ptr,
- opr_sz, max_sz, 0, fn);
- tcg_temp_free_ptr(qc_ptr);
- }
- void gen_gvec_sqrdmlah_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static gen_helper_gvec_3_ptr * const fns[2] = {
- gen_helper_gvec_qrdmlah_s16, gen_helper_gvec_qrdmlah_s32
- };
- tcg_debug_assert(vece >= 1 && vece <= 2);
- gen_gvec_fn3_qc(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, fns[vece - 1]);
- }
- void gen_gvec_sqrdmlsh_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static gen_helper_gvec_3_ptr * const fns[2] = {
- gen_helper_gvec_qrdmlsh_s16, gen_helper_gvec_qrdmlsh_s32
- };
- tcg_debug_assert(vece >= 1 && vece <= 2);
- gen_gvec_fn3_qc(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, fns[vece - 1]);
- }
- #define GEN_CMP0(NAME, COND) \
- static void gen_##NAME##0_i32(TCGv_i32 d, TCGv_i32 a) \
- { \
- tcg_gen_setcondi_i32(COND, d, a, 0); \
- tcg_gen_neg_i32(d, d); \
- } \
- static void gen_##NAME##0_i64(TCGv_i64 d, TCGv_i64 a) \
- { \
- tcg_gen_setcondi_i64(COND, d, a, 0); \
- tcg_gen_neg_i64(d, d); \
- } \
- static void gen_##NAME##0_vec(unsigned vece, TCGv_vec d, TCGv_vec a) \
- { \
- TCGv_vec zero = tcg_const_zeros_vec_matching(d); \
- tcg_gen_cmp_vec(COND, vece, d, a, zero); \
- tcg_temp_free_vec(zero); \
- } \
- void gen_gvec_##NAME##0(unsigned vece, uint32_t d, uint32_t m, \
- uint32_t opr_sz, uint32_t max_sz) \
- { \
- const GVecGen2 op[4] = { \
- { .fno = gen_helper_gvec_##NAME##0_b, \
- .fniv = gen_##NAME##0_vec, \
- .opt_opc = vecop_list_cmp, \
- .vece = MO_8 }, \
- { .fno = gen_helper_gvec_##NAME##0_h, \
- .fniv = gen_##NAME##0_vec, \
- .opt_opc = vecop_list_cmp, \
- .vece = MO_16 }, \
- { .fni4 = gen_##NAME##0_i32, \
- .fniv = gen_##NAME##0_vec, \
- .opt_opc = vecop_list_cmp, \
- .vece = MO_32 }, \
- { .fni8 = gen_##NAME##0_i64, \
- .fniv = gen_##NAME##0_vec, \
- .opt_opc = vecop_list_cmp, \
- .prefer_i64 = TCG_TARGET_REG_BITS == 64, \
- .vece = MO_64 }, \
- }; \
- tcg_gen_gvec_2(d, m, opr_sz, max_sz, &op[vece]); \
- }
- static const TCGOpcode vecop_list_cmp[] = {
- INDEX_op_cmp_vec, 0
- };
- GEN_CMP0(ceq, TCG_COND_EQ)
- GEN_CMP0(cle, TCG_COND_LE)
- GEN_CMP0(cge, TCG_COND_GE)
- GEN_CMP0(clt, TCG_COND_LT)
- GEN_CMP0(cgt, TCG_COND_GT)
- #undef GEN_CMP0
- static void gen_ssra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_vec_sar8i_i64(a, a, shift);
- tcg_gen_vec_add8_i64(d, d, a);
- }
- static void gen_ssra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_vec_sar16i_i64(a, a, shift);
- tcg_gen_vec_add16_i64(d, d, a);
- }
- static void gen_ssra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
- {
- tcg_gen_sari_i32(a, a, shift);
- tcg_gen_add_i32(d, d, a);
- }
- static void gen_ssra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_sari_i64(a, a, shift);
- tcg_gen_add_i64(d, d, a);
- }
- static void gen_ssra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- tcg_gen_sari_vec(vece, a, a, sh);
- tcg_gen_add_vec(vece, d, d, a);
- }
- void gen_gvec_ssra(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sari_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_ssra8_i64,
- .fniv = gen_ssra_vec,
- .fno = gen_helper_gvec_ssra_b,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni8 = gen_ssra16_i64,
- .fniv = gen_ssra_vec,
- .fno = gen_helper_gvec_ssra_h,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_ssra32_i32,
- .fniv = gen_ssra_vec,
- .fno = gen_helper_gvec_ssra_s,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_ssra64_i64,
- .fniv = gen_ssra_vec,
- .fno = gen_helper_gvec_ssra_b,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize]. */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- /*
- * Shifts larger than the element size are architecturally valid.
- * Signed results in all sign bits.
- */
- shift = MIN(shift, (8 << vece) - 1);
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- static void gen_usra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_vec_shr8i_i64(a, a, shift);
- tcg_gen_vec_add8_i64(d, d, a);
- }
- static void gen_usra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_vec_shr16i_i64(a, a, shift);
- tcg_gen_vec_add16_i64(d, d, a);
- }
- static void gen_usra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
- {
- tcg_gen_shri_i32(a, a, shift);
- tcg_gen_add_i32(d, d, a);
- }
- static void gen_usra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_shri_i64(a, a, shift);
- tcg_gen_add_i64(d, d, a);
- }
- static void gen_usra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- tcg_gen_shri_vec(vece, a, a, sh);
- tcg_gen_add_vec(vece, d, d, a);
- }
- void gen_gvec_usra(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_shri_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_usra8_i64,
- .fniv = gen_usra_vec,
- .fno = gen_helper_gvec_usra_b,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8, },
- { .fni8 = gen_usra16_i64,
- .fniv = gen_usra_vec,
- .fno = gen_helper_gvec_usra_h,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16, },
- { .fni4 = gen_usra32_i32,
- .fniv = gen_usra_vec,
- .fno = gen_helper_gvec_usra_s,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32, },
- { .fni8 = gen_usra64_i64,
- .fniv = gen_usra_vec,
- .fno = gen_helper_gvec_usra_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_64, },
- };
- /* tszimm encoding produces immediates in the range [1..esize]. */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- /*
- * Shifts larger than the element size are architecturally valid.
- * Unsigned results in all zeros as input to accumulate: nop.
- */
- if (shift < (8 << vece)) {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- } else {
- /* Nop, but we do need to clear the tail. */
- tcg_gen_gvec_mov(vece, rd_ofs, rd_ofs, opr_sz, max_sz);
- }
- }
- /*
- * Shift one less than the requested amount, and the low bit is
- * the rounding bit. For the 8 and 16-bit operations, because we
- * mask the low bit, we can perform a normal integer shift instead
- * of a vector shift.
- */
- static void gen_srshr8_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, sh - 1);
- tcg_gen_andi_i64(t, t, dup_const(MO_8, 1));
- tcg_gen_vec_sar8i_i64(d, a, sh);
- tcg_gen_vec_add8_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srshr16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, sh - 1);
- tcg_gen_andi_i64(t, t, dup_const(MO_16, 1));
- tcg_gen_vec_sar16i_i64(d, a, sh);
- tcg_gen_vec_add16_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srshr32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- tcg_gen_extract_i32(t, a, sh - 1, 1);
- tcg_gen_sari_i32(d, a, sh);
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_srshr64_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_extract_i64(t, a, sh - 1, 1);
- tcg_gen_sari_i64(d, a, sh);
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srshr_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- TCGv_vec ones = tcg_temp_new_vec_matching(d);
- tcg_gen_shri_vec(vece, t, a, sh - 1);
- tcg_gen_dupi_vec(vece, ones, 1);
- tcg_gen_and_vec(vece, t, t, ones);
- tcg_gen_sari_vec(vece, d, a, sh);
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- tcg_temp_free_vec(ones);
- }
- void gen_gvec_srshr(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_shri_vec, INDEX_op_sari_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_srshr8_i64,
- .fniv = gen_srshr_vec,
- .fno = gen_helper_gvec_srshr_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni8 = gen_srshr16_i64,
- .fniv = gen_srshr_vec,
- .fno = gen_helper_gvec_srshr_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_srshr32_i32,
- .fniv = gen_srshr_vec,
- .fno = gen_helper_gvec_srshr_s,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_srshr64_i64,
- .fniv = gen_srshr_vec,
- .fno = gen_helper_gvec_srshr_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize] */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- if (shift == (8 << vece)) {
- /*
- * Shifts larger than the element size are architecturally valid.
- * Signed results in all sign bits. With rounding, this produces
- * (-1 + 1) >> 1 == 0, or (0 + 1) >> 1 == 0.
- * I.e. always zero.
- */
- tcg_gen_gvec_dup_imm(vece, rd_ofs, opr_sz, max_sz, 0);
- } else {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- }
- static void gen_srsra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- gen_srshr8_i64(t, a, sh);
- tcg_gen_vec_add8_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srsra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- gen_srshr16_i64(t, a, sh);
- tcg_gen_vec_add16_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srsra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- gen_srshr32_i32(t, a, sh);
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_srsra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- gen_srshr64_i64(t, a, sh);
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_srsra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- gen_srshr_vec(vece, t, a, sh);
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_srsra(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_shri_vec, INDEX_op_sari_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_srsra8_i64,
- .fniv = gen_srsra_vec,
- .fno = gen_helper_gvec_srsra_b,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_8 },
- { .fni8 = gen_srsra16_i64,
- .fniv = gen_srsra_vec,
- .fno = gen_helper_gvec_srsra_h,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_16 },
- { .fni4 = gen_srsra32_i32,
- .fniv = gen_srsra_vec,
- .fno = gen_helper_gvec_srsra_s,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_32 },
- { .fni8 = gen_srsra64_i64,
- .fniv = gen_srsra_vec,
- .fno = gen_helper_gvec_srsra_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize] */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- /*
- * Shifts larger than the element size are architecturally valid.
- * Signed results in all sign bits. With rounding, this produces
- * (-1 + 1) >> 1 == 0, or (0 + 1) >> 1 == 0.
- * I.e. always zero. With accumulation, this leaves D unchanged.
- */
- if (shift == (8 << vece)) {
- /* Nop, but we do need to clear the tail. */
- tcg_gen_gvec_mov(vece, rd_ofs, rd_ofs, opr_sz, max_sz);
- } else {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- }
- static void gen_urshr8_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, sh - 1);
- tcg_gen_andi_i64(t, t, dup_const(MO_8, 1));
- tcg_gen_vec_shr8i_i64(d, a, sh);
- tcg_gen_vec_add8_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_urshr16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, sh - 1);
- tcg_gen_andi_i64(t, t, dup_const(MO_16, 1));
- tcg_gen_vec_shr16i_i64(d, a, sh);
- tcg_gen_vec_add16_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_urshr32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- tcg_gen_extract_i32(t, a, sh - 1, 1);
- tcg_gen_shri_i32(d, a, sh);
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_urshr64_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_extract_i64(t, a, sh - 1, 1);
- tcg_gen_shri_i64(d, a, sh);
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_urshr_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t shift)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- TCGv_vec ones = tcg_temp_new_vec_matching(d);
- tcg_gen_shri_vec(vece, t, a, shift - 1);
- tcg_gen_dupi_vec(vece, ones, 1);
- tcg_gen_and_vec(vece, t, t, ones);
- tcg_gen_shri_vec(vece, d, a, shift);
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- tcg_temp_free_vec(ones);
- }
- void gen_gvec_urshr(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_shri_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_urshr8_i64,
- .fniv = gen_urshr_vec,
- .fno = gen_helper_gvec_urshr_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni8 = gen_urshr16_i64,
- .fniv = gen_urshr_vec,
- .fno = gen_helper_gvec_urshr_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_urshr32_i32,
- .fniv = gen_urshr_vec,
- .fno = gen_helper_gvec_urshr_s,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_urshr64_i64,
- .fniv = gen_urshr_vec,
- .fno = gen_helper_gvec_urshr_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize] */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- if (shift == (8 << vece)) {
- /*
- * Shifts larger than the element size are architecturally valid.
- * Unsigned results in zero. With rounding, this produces a
- * copy of the most significant bit.
- */
- tcg_gen_gvec_shri(vece, rd_ofs, rm_ofs, shift - 1, opr_sz, max_sz);
- } else {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- }
- static void gen_ursra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- if (sh == 8) {
- tcg_gen_vec_shr8i_i64(t, a, 7);
- } else {
- gen_urshr8_i64(t, a, sh);
- }
- tcg_gen_vec_add8_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_ursra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- if (sh == 16) {
- tcg_gen_vec_shr16i_i64(t, a, 15);
- } else {
- gen_urshr16_i64(t, a, sh);
- }
- tcg_gen_vec_add16_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_ursra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- if (sh == 32) {
- tcg_gen_shri_i32(t, a, 31);
- } else {
- gen_urshr32_i32(t, a, sh);
- }
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_ursra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- if (sh == 64) {
- tcg_gen_shri_i64(t, a, 63);
- } else {
- gen_urshr64_i64(t, a, sh);
- }
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_ursra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- if (sh == (8 << vece)) {
- tcg_gen_shri_vec(vece, t, a, sh - 1);
- } else {
- gen_urshr_vec(vece, t, a, sh);
- }
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_ursra(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_shri_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen2i ops[4] = {
- { .fni8 = gen_ursra8_i64,
- .fniv = gen_ursra_vec,
- .fno = gen_helper_gvec_ursra_b,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_8 },
- { .fni8 = gen_ursra16_i64,
- .fniv = gen_ursra_vec,
- .fno = gen_helper_gvec_ursra_h,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_16 },
- { .fni4 = gen_ursra32_i32,
- .fniv = gen_ursra_vec,
- .fno = gen_helper_gvec_ursra_s,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_32 },
- { .fni8 = gen_ursra64_i64,
- .fniv = gen_ursra_vec,
- .fno = gen_helper_gvec_ursra_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize] */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- static void gen_shr8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- uint64_t mask = dup_const(MO_8, 0xff >> shift);
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, shift);
- tcg_gen_andi_i64(t, t, mask);
- tcg_gen_andi_i64(d, d, ~mask);
- tcg_gen_or_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_shr16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- uint64_t mask = dup_const(MO_16, 0xffff >> shift);
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shri_i64(t, a, shift);
- tcg_gen_andi_i64(t, t, mask);
- tcg_gen_andi_i64(d, d, ~mask);
- tcg_gen_or_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_shr32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
- {
- tcg_gen_shri_i32(a, a, shift);
- tcg_gen_deposit_i32(d, d, a, 0, 32 - shift);
- }
- static void gen_shr64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_shri_i64(a, a, shift);
- tcg_gen_deposit_i64(d, d, a, 0, 64 - shift);
- }
- static void gen_shr_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- TCGv_vec m = tcg_temp_new_vec_matching(d);
- tcg_gen_dupi_vec(vece, m, MAKE_64BIT_MASK((8 << vece) - sh, sh));
- tcg_gen_shri_vec(vece, t, a, sh);
- tcg_gen_and_vec(vece, d, d, m);
- tcg_gen_or_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- tcg_temp_free_vec(m);
- }
- void gen_gvec_sri(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = { INDEX_op_shri_vec, 0 };
- const GVecGen2i ops[4] = {
- { .fni8 = gen_shr8_ins_i64,
- .fniv = gen_shr_ins_vec,
- .fno = gen_helper_gvec_sri_b,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni8 = gen_shr16_ins_i64,
- .fniv = gen_shr_ins_vec,
- .fno = gen_helper_gvec_sri_h,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_shr32_ins_i32,
- .fniv = gen_shr_ins_vec,
- .fno = gen_helper_gvec_sri_s,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_shr64_ins_i64,
- .fniv = gen_shr_ins_vec,
- .fno = gen_helper_gvec_sri_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [1..esize]. */
- tcg_debug_assert(shift > 0);
- tcg_debug_assert(shift <= (8 << vece));
- /* Shift of esize leaves destination unchanged. */
- if (shift < (8 << vece)) {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- } else {
- /* Nop, but we do need to clear the tail. */
- tcg_gen_gvec_mov(vece, rd_ofs, rd_ofs, opr_sz, max_sz);
- }
- }
- static void gen_shl8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- uint64_t mask = dup_const(MO_8, 0xff << shift);
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shli_i64(t, a, shift);
- tcg_gen_andi_i64(t, t, mask);
- tcg_gen_andi_i64(d, d, ~mask);
- tcg_gen_or_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_shl16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- uint64_t mask = dup_const(MO_16, 0xffff << shift);
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_shli_i64(t, a, shift);
- tcg_gen_andi_i64(t, t, mask);
- tcg_gen_andi_i64(d, d, ~mask);
- tcg_gen_or_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_shl32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
- {
- tcg_gen_deposit_i32(d, d, a, shift, 32 - shift);
- }
- static void gen_shl64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
- {
- tcg_gen_deposit_i64(d, d, a, shift, 64 - shift);
- }
- static void gen_shl_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- TCGv_vec m = tcg_temp_new_vec_matching(d);
- tcg_gen_shli_vec(vece, t, a, sh);
- tcg_gen_dupi_vec(vece, m, MAKE_64BIT_MASK(0, sh));
- tcg_gen_and_vec(vece, d, d, m);
- tcg_gen_or_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- tcg_temp_free_vec(m);
- }
- void gen_gvec_sli(unsigned vece, uint32_t rd_ofs, uint32_t rm_ofs,
- int64_t shift, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = { INDEX_op_shli_vec, 0 };
- const GVecGen2i ops[4] = {
- { .fni8 = gen_shl8_ins_i64,
- .fniv = gen_shl_ins_vec,
- .fno = gen_helper_gvec_sli_b,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni8 = gen_shl16_ins_i64,
- .fniv = gen_shl_ins_vec,
- .fno = gen_helper_gvec_sli_h,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_shl32_ins_i32,
- .fniv = gen_shl_ins_vec,
- .fno = gen_helper_gvec_sli_s,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_shl64_ins_i64,
- .fniv = gen_shl_ins_vec,
- .fno = gen_helper_gvec_sli_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- /* tszimm encoding produces immediates in the range [0..esize-1]. */
- tcg_debug_assert(shift >= 0);
- tcg_debug_assert(shift < (8 << vece));
- if (shift == 0) {
- tcg_gen_gvec_mov(vece, rd_ofs, rm_ofs, opr_sz, max_sz);
- } else {
- tcg_gen_gvec_2i(rd_ofs, rm_ofs, opr_sz, max_sz, shift, &ops[vece]);
- }
- }
- static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- gen_helper_neon_mul_u8(a, a, b);
- gen_helper_neon_add_u8(d, d, a);
- }
- static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- gen_helper_neon_mul_u8(a, a, b);
- gen_helper_neon_sub_u8(d, d, a);
- }
- static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- gen_helper_neon_mul_u16(a, a, b);
- gen_helper_neon_add_u16(d, d, a);
- }
- static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- gen_helper_neon_mul_u16(a, a, b);
- gen_helper_neon_sub_u16(d, d, a);
- }
- static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- tcg_gen_mul_i32(a, a, b);
- tcg_gen_add_i32(d, d, a);
- }
- static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- tcg_gen_mul_i32(a, a, b);
- tcg_gen_sub_i32(d, d, a);
- }
- static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- tcg_gen_mul_i64(a, a, b);
- tcg_gen_add_i64(d, d, a);
- }
- static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- tcg_gen_mul_i64(a, a, b);
- tcg_gen_sub_i64(d, d, a);
- }
- static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- tcg_gen_mul_vec(vece, a, a, b);
- tcg_gen_add_vec(vece, d, d, a);
- }
- static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- tcg_gen_mul_vec(vece, a, a, b);
- tcg_gen_sub_vec(vece, d, d, a);
- }
- /* Note that while NEON does not support VMLA and VMLS as 64-bit ops,
- * these tables are shared with AArch64 which does support them.
- */
- void gen_gvec_mla(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_mul_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fni4 = gen_mla8_i32,
- .fniv = gen_mla_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni4 = gen_mla16_i32,
- .fniv = gen_mla_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_mla32_i32,
- .fniv = gen_mla_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_mla64_i64,
- .fniv = gen_mla_vec,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- void gen_gvec_mls(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_mul_vec, INDEX_op_sub_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fni4 = gen_mls8_i32,
- .fniv = gen_mls_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni4 = gen_mls16_i32,
- .fniv = gen_mls_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_mls32_i32,
- .fniv = gen_mls_vec,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_mls64_i64,
- .fniv = gen_mls_vec,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .load_dest = true,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- /* CMTST : test is "if (X & Y != 0)". */
- static void gen_cmtst_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- tcg_gen_and_i32(d, a, b);
- tcg_gen_setcondi_i32(TCG_COND_NE, d, d, 0);
- tcg_gen_neg_i32(d, d);
- }
- void gen_cmtst_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- tcg_gen_and_i64(d, a, b);
- tcg_gen_setcondi_i64(TCG_COND_NE, d, d, 0);
- tcg_gen_neg_i64(d, d);
- }
- static void gen_cmtst_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- tcg_gen_and_vec(vece, d, a, b);
- tcg_gen_dupi_vec(vece, a, 0);
- tcg_gen_cmp_vec(TCG_COND_NE, vece, d, d, a);
- }
- void gen_gvec_cmtst(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = { INDEX_op_cmp_vec, 0 };
- static const GVecGen3 ops[4] = {
- { .fni4 = gen_helper_neon_tst_u8,
- .fniv = gen_cmtst_vec,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fni4 = gen_helper_neon_tst_u16,
- .fniv = gen_cmtst_vec,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_cmtst_i32,
- .fniv = gen_cmtst_vec,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_cmtst_i64,
- .fniv = gen_cmtst_vec,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- void gen_ushl_i32(TCGv_i32 dst, TCGv_i32 src, TCGv_i32 shift)
- {
- TCGv_i32 lval = tcg_temp_new_i32();
- TCGv_i32 rval = tcg_temp_new_i32();
- TCGv_i32 lsh = tcg_temp_new_i32();
- TCGv_i32 rsh = tcg_temp_new_i32();
- TCGv_i32 zero = tcg_const_i32(0);
- TCGv_i32 max = tcg_const_i32(32);
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_ext8s_i32(lsh, shift);
- tcg_gen_neg_i32(rsh, lsh);
- tcg_gen_shl_i32(lval, src, lsh);
- tcg_gen_shr_i32(rval, src, rsh);
- tcg_gen_movcond_i32(TCG_COND_LTU, dst, lsh, max, lval, zero);
- tcg_gen_movcond_i32(TCG_COND_LTU, dst, rsh, max, rval, dst);
- tcg_temp_free_i32(lval);
- tcg_temp_free_i32(rval);
- tcg_temp_free_i32(lsh);
- tcg_temp_free_i32(rsh);
- tcg_temp_free_i32(zero);
- tcg_temp_free_i32(max);
- }
- void gen_ushl_i64(TCGv_i64 dst, TCGv_i64 src, TCGv_i64 shift)
- {
- TCGv_i64 lval = tcg_temp_new_i64();
- TCGv_i64 rval = tcg_temp_new_i64();
- TCGv_i64 lsh = tcg_temp_new_i64();
- TCGv_i64 rsh = tcg_temp_new_i64();
- TCGv_i64 zero = tcg_const_i64(0);
- TCGv_i64 max = tcg_const_i64(64);
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_ext8s_i64(lsh, shift);
- tcg_gen_neg_i64(rsh, lsh);
- tcg_gen_shl_i64(lval, src, lsh);
- tcg_gen_shr_i64(rval, src, rsh);
- tcg_gen_movcond_i64(TCG_COND_LTU, dst, lsh, max, lval, zero);
- tcg_gen_movcond_i64(TCG_COND_LTU, dst, rsh, max, rval, dst);
- tcg_temp_free_i64(lval);
- tcg_temp_free_i64(rval);
- tcg_temp_free_i64(lsh);
- tcg_temp_free_i64(rsh);
- tcg_temp_free_i64(zero);
- tcg_temp_free_i64(max);
- }
- static void gen_ushl_vec(unsigned vece, TCGv_vec dst,
- TCGv_vec src, TCGv_vec shift)
- {
- TCGv_vec lval = tcg_temp_new_vec_matching(dst);
- TCGv_vec rval = tcg_temp_new_vec_matching(dst);
- TCGv_vec lsh = tcg_temp_new_vec_matching(dst);
- TCGv_vec rsh = tcg_temp_new_vec_matching(dst);
- TCGv_vec msk, max;
- tcg_gen_neg_vec(vece, rsh, shift);
- if (vece == MO_8) {
- tcg_gen_mov_vec(lsh, shift);
- } else {
- msk = tcg_temp_new_vec_matching(dst);
- tcg_gen_dupi_vec(vece, msk, 0xff);
- tcg_gen_and_vec(vece, lsh, shift, msk);
- tcg_gen_and_vec(vece, rsh, rsh, msk);
- tcg_temp_free_vec(msk);
- }
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_shlv_vec(vece, lval, src, lsh);
- tcg_gen_shrv_vec(vece, rval, src, rsh);
- max = tcg_temp_new_vec_matching(dst);
- tcg_gen_dupi_vec(vece, max, 8 << vece);
- /*
- * The choice of LT (signed) and GEU (unsigned) are biased toward
- * the instructions of the x86_64 host. For MO_8, the whole byte
- * is significant so we must use an unsigned compare; otherwise we
- * have already masked to a byte and so a signed compare works.
- * Other tcg hosts have a full set of comparisons and do not care.
- */
- if (vece == MO_8) {
- tcg_gen_cmp_vec(TCG_COND_GEU, vece, lsh, lsh, max);
- tcg_gen_cmp_vec(TCG_COND_GEU, vece, rsh, rsh, max);
- tcg_gen_andc_vec(vece, lval, lval, lsh);
- tcg_gen_andc_vec(vece, rval, rval, rsh);
- } else {
- tcg_gen_cmp_vec(TCG_COND_LT, vece, lsh, lsh, max);
- tcg_gen_cmp_vec(TCG_COND_LT, vece, rsh, rsh, max);
- tcg_gen_and_vec(vece, lval, lval, lsh);
- tcg_gen_and_vec(vece, rval, rval, rsh);
- }
- tcg_gen_or_vec(vece, dst, lval, rval);
- tcg_temp_free_vec(max);
- tcg_temp_free_vec(lval);
- tcg_temp_free_vec(rval);
- tcg_temp_free_vec(lsh);
- tcg_temp_free_vec(rsh);
- }
- void gen_gvec_ushl(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_neg_vec, INDEX_op_shlv_vec,
- INDEX_op_shrv_vec, INDEX_op_cmp_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_ushl_vec,
- .fno = gen_helper_gvec_ushl_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fniv = gen_ushl_vec,
- .fno = gen_helper_gvec_ushl_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_ushl_i32,
- .fniv = gen_ushl_vec,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_ushl_i64,
- .fniv = gen_ushl_vec,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- void gen_sshl_i32(TCGv_i32 dst, TCGv_i32 src, TCGv_i32 shift)
- {
- TCGv_i32 lval = tcg_temp_new_i32();
- TCGv_i32 rval = tcg_temp_new_i32();
- TCGv_i32 lsh = tcg_temp_new_i32();
- TCGv_i32 rsh = tcg_temp_new_i32();
- TCGv_i32 zero = tcg_const_i32(0);
- TCGv_i32 max = tcg_const_i32(31);
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_ext8s_i32(lsh, shift);
- tcg_gen_neg_i32(rsh, lsh);
- tcg_gen_shl_i32(lval, src, lsh);
- tcg_gen_umin_i32(rsh, rsh, max);
- tcg_gen_sar_i32(rval, src, rsh);
- tcg_gen_movcond_i32(TCG_COND_LEU, lval, lsh, max, lval, zero);
- tcg_gen_movcond_i32(TCG_COND_LT, dst, lsh, zero, rval, lval);
- tcg_temp_free_i32(lval);
- tcg_temp_free_i32(rval);
- tcg_temp_free_i32(lsh);
- tcg_temp_free_i32(rsh);
- tcg_temp_free_i32(zero);
- tcg_temp_free_i32(max);
- }
- void gen_sshl_i64(TCGv_i64 dst, TCGv_i64 src, TCGv_i64 shift)
- {
- TCGv_i64 lval = tcg_temp_new_i64();
- TCGv_i64 rval = tcg_temp_new_i64();
- TCGv_i64 lsh = tcg_temp_new_i64();
- TCGv_i64 rsh = tcg_temp_new_i64();
- TCGv_i64 zero = tcg_const_i64(0);
- TCGv_i64 max = tcg_const_i64(63);
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_ext8s_i64(lsh, shift);
- tcg_gen_neg_i64(rsh, lsh);
- tcg_gen_shl_i64(lval, src, lsh);
- tcg_gen_umin_i64(rsh, rsh, max);
- tcg_gen_sar_i64(rval, src, rsh);
- tcg_gen_movcond_i64(TCG_COND_LEU, lval, lsh, max, lval, zero);
- tcg_gen_movcond_i64(TCG_COND_LT, dst, lsh, zero, rval, lval);
- tcg_temp_free_i64(lval);
- tcg_temp_free_i64(rval);
- tcg_temp_free_i64(lsh);
- tcg_temp_free_i64(rsh);
- tcg_temp_free_i64(zero);
- tcg_temp_free_i64(max);
- }
- static void gen_sshl_vec(unsigned vece, TCGv_vec dst,
- TCGv_vec src, TCGv_vec shift)
- {
- TCGv_vec lval = tcg_temp_new_vec_matching(dst);
- TCGv_vec rval = tcg_temp_new_vec_matching(dst);
- TCGv_vec lsh = tcg_temp_new_vec_matching(dst);
- TCGv_vec rsh = tcg_temp_new_vec_matching(dst);
- TCGv_vec tmp = tcg_temp_new_vec_matching(dst);
- /*
- * Rely on the TCG guarantee that out of range shifts produce
- * unspecified results, not undefined behaviour (i.e. no trap).
- * Discard out-of-range results after the fact.
- */
- tcg_gen_neg_vec(vece, rsh, shift);
- if (vece == MO_8) {
- tcg_gen_mov_vec(lsh, shift);
- } else {
- tcg_gen_dupi_vec(vece, tmp, 0xff);
- tcg_gen_and_vec(vece, lsh, shift, tmp);
- tcg_gen_and_vec(vece, rsh, rsh, tmp);
- }
- /* Bound rsh so out of bound right shift gets -1. */
- tcg_gen_dupi_vec(vece, tmp, (8 << vece) - 1);
- tcg_gen_umin_vec(vece, rsh, rsh, tmp);
- tcg_gen_cmp_vec(TCG_COND_GT, vece, tmp, lsh, tmp);
- tcg_gen_shlv_vec(vece, lval, src, lsh);
- tcg_gen_sarv_vec(vece, rval, src, rsh);
- /* Select in-bound left shift. */
- tcg_gen_andc_vec(vece, lval, lval, tmp);
- /* Select between left and right shift. */
- if (vece == MO_8) {
- tcg_gen_dupi_vec(vece, tmp, 0);
- tcg_gen_cmpsel_vec(TCG_COND_LT, vece, dst, lsh, tmp, rval, lval);
- } else {
- tcg_gen_dupi_vec(vece, tmp, 0x80);
- tcg_gen_cmpsel_vec(TCG_COND_LT, vece, dst, lsh, tmp, lval, rval);
- }
- tcg_temp_free_vec(lval);
- tcg_temp_free_vec(rval);
- tcg_temp_free_vec(lsh);
- tcg_temp_free_vec(rsh);
- tcg_temp_free_vec(tmp);
- }
- void gen_gvec_sshl(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_neg_vec, INDEX_op_umin_vec, INDEX_op_shlv_vec,
- INDEX_op_sarv_vec, INDEX_op_cmp_vec, INDEX_op_cmpsel_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_sshl_vec,
- .fno = gen_helper_gvec_sshl_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fniv = gen_sshl_vec,
- .fno = gen_helper_gvec_sshl_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_sshl_i32,
- .fniv = gen_sshl_vec,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_sshl_i64,
- .fniv = gen_sshl_vec,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_uqadd_vec(unsigned vece, TCGv_vec t, TCGv_vec sat,
- TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec x = tcg_temp_new_vec_matching(t);
- tcg_gen_add_vec(vece, x, a, b);
- tcg_gen_usadd_vec(vece, t, a, b);
- tcg_gen_cmp_vec(TCG_COND_NE, vece, x, x, t);
- tcg_gen_or_vec(vece, sat, sat, x);
- tcg_temp_free_vec(x);
- }
- void gen_gvec_uqadd_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_usadd_vec, INDEX_op_cmp_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen4 ops[4] = {
- { .fniv = gen_uqadd_vec,
- .fno = gen_helper_gvec_uqadd_b,
- .write_aofs = true,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fniv = gen_uqadd_vec,
- .fno = gen_helper_gvec_uqadd_h,
- .write_aofs = true,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fniv = gen_uqadd_vec,
- .fno = gen_helper_gvec_uqadd_s,
- .write_aofs = true,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fniv = gen_uqadd_vec,
- .fno = gen_helper_gvec_uqadd_d,
- .write_aofs = true,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
- rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_sqadd_vec(unsigned vece, TCGv_vec t, TCGv_vec sat,
- TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec x = tcg_temp_new_vec_matching(t);
- tcg_gen_add_vec(vece, x, a, b);
- tcg_gen_ssadd_vec(vece, t, a, b);
- tcg_gen_cmp_vec(TCG_COND_NE, vece, x, x, t);
- tcg_gen_or_vec(vece, sat, sat, x);
- tcg_temp_free_vec(x);
- }
- void gen_gvec_sqadd_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_ssadd_vec, INDEX_op_cmp_vec, INDEX_op_add_vec, 0
- };
- static const GVecGen4 ops[4] = {
- { .fniv = gen_sqadd_vec,
- .fno = gen_helper_gvec_sqadd_b,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_8 },
- { .fniv = gen_sqadd_vec,
- .fno = gen_helper_gvec_sqadd_h,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_16 },
- { .fniv = gen_sqadd_vec,
- .fno = gen_helper_gvec_sqadd_s,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_32 },
- { .fniv = gen_sqadd_vec,
- .fno = gen_helper_gvec_sqadd_d,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_64 },
- };
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
- rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_uqsub_vec(unsigned vece, TCGv_vec t, TCGv_vec sat,
- TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec x = tcg_temp_new_vec_matching(t);
- tcg_gen_sub_vec(vece, x, a, b);
- tcg_gen_ussub_vec(vece, t, a, b);
- tcg_gen_cmp_vec(TCG_COND_NE, vece, x, x, t);
- tcg_gen_or_vec(vece, sat, sat, x);
- tcg_temp_free_vec(x);
- }
- void gen_gvec_uqsub_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_ussub_vec, INDEX_op_cmp_vec, INDEX_op_sub_vec, 0
- };
- static const GVecGen4 ops[4] = {
- { .fniv = gen_uqsub_vec,
- .fno = gen_helper_gvec_uqsub_b,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_8 },
- { .fniv = gen_uqsub_vec,
- .fno = gen_helper_gvec_uqsub_h,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_16 },
- { .fniv = gen_uqsub_vec,
- .fno = gen_helper_gvec_uqsub_s,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_32 },
- { .fniv = gen_uqsub_vec,
- .fno = gen_helper_gvec_uqsub_d,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_64 },
- };
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
- rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_sqsub_vec(unsigned vece, TCGv_vec t, TCGv_vec sat,
- TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec x = tcg_temp_new_vec_matching(t);
- tcg_gen_sub_vec(vece, x, a, b);
- tcg_gen_sssub_vec(vece, t, a, b);
- tcg_gen_cmp_vec(TCG_COND_NE, vece, x, x, t);
- tcg_gen_or_vec(vece, sat, sat, x);
- tcg_temp_free_vec(x);
- }
- void gen_gvec_sqsub_qc(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sssub_vec, INDEX_op_cmp_vec, INDEX_op_sub_vec, 0
- };
- static const GVecGen4 ops[4] = {
- { .fniv = gen_sqsub_vec,
- .fno = gen_helper_gvec_sqsub_b,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_8 },
- { .fniv = gen_sqsub_vec,
- .fno = gen_helper_gvec_sqsub_h,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_16 },
- { .fniv = gen_sqsub_vec,
- .fno = gen_helper_gvec_sqsub_s,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_32 },
- { .fniv = gen_sqsub_vec,
- .fno = gen_helper_gvec_sqsub_d,
- .opt_opc = vecop_list,
- .write_aofs = true,
- .vece = MO_64 },
- };
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
- rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_sabd_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- tcg_gen_sub_i32(t, a, b);
- tcg_gen_sub_i32(d, b, a);
- tcg_gen_movcond_i32(TCG_COND_LT, d, a, b, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_sabd_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_sub_i64(t, a, b);
- tcg_gen_sub_i64(d, b, a);
- tcg_gen_movcond_i64(TCG_COND_LT, d, a, b, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_sabd_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- tcg_gen_smin_vec(vece, t, a, b);
- tcg_gen_smax_vec(vece, d, a, b);
- tcg_gen_sub_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_sabd(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sub_vec, INDEX_op_smin_vec, INDEX_op_smax_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_sabd_vec,
- .fno = gen_helper_gvec_sabd_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fniv = gen_sabd_vec,
- .fno = gen_helper_gvec_sabd_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_sabd_i32,
- .fniv = gen_sabd_vec,
- .fno = gen_helper_gvec_sabd_s,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_sabd_i64,
- .fniv = gen_sabd_vec,
- .fno = gen_helper_gvec_sabd_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_uabd_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- tcg_gen_sub_i32(t, a, b);
- tcg_gen_sub_i32(d, b, a);
- tcg_gen_movcond_i32(TCG_COND_LTU, d, a, b, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_uabd_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- tcg_gen_sub_i64(t, a, b);
- tcg_gen_sub_i64(d, b, a);
- tcg_gen_movcond_i64(TCG_COND_LTU, d, a, b, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_uabd_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- tcg_gen_umin_vec(vece, t, a, b);
- tcg_gen_umax_vec(vece, d, a, b);
- tcg_gen_sub_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_uabd(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sub_vec, INDEX_op_umin_vec, INDEX_op_umax_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_uabd_vec,
- .fno = gen_helper_gvec_uabd_b,
- .opt_opc = vecop_list,
- .vece = MO_8 },
- { .fniv = gen_uabd_vec,
- .fno = gen_helper_gvec_uabd_h,
- .opt_opc = vecop_list,
- .vece = MO_16 },
- { .fni4 = gen_uabd_i32,
- .fniv = gen_uabd_vec,
- .fno = gen_helper_gvec_uabd_s,
- .opt_opc = vecop_list,
- .vece = MO_32 },
- { .fni8 = gen_uabd_i64,
- .fniv = gen_uabd_vec,
- .fno = gen_helper_gvec_uabd_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_saba_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- gen_sabd_i32(t, a, b);
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_saba_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- gen_sabd_i64(t, a, b);
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_saba_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- gen_sabd_vec(vece, t, a, b);
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_saba(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sub_vec, INDEX_op_add_vec,
- INDEX_op_smin_vec, INDEX_op_smax_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_saba_vec,
- .fno = gen_helper_gvec_saba_b,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_8 },
- { .fniv = gen_saba_vec,
- .fno = gen_helper_gvec_saba_h,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_16 },
- { .fni4 = gen_saba_i32,
- .fniv = gen_saba_vec,
- .fno = gen_helper_gvec_saba_s,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_32 },
- { .fni8 = gen_saba_i64,
- .fniv = gen_saba_vec,
- .fno = gen_helper_gvec_saba_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static void gen_uaba_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
- {
- TCGv_i32 t = tcg_temp_new_i32();
- gen_uabd_i32(t, a, b);
- tcg_gen_add_i32(d, d, t);
- tcg_temp_free_i32(t);
- }
- static void gen_uaba_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
- {
- TCGv_i64 t = tcg_temp_new_i64();
- gen_uabd_i64(t, a, b);
- tcg_gen_add_i64(d, d, t);
- tcg_temp_free_i64(t);
- }
- static void gen_uaba_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
- {
- TCGv_vec t = tcg_temp_new_vec_matching(d);
- gen_uabd_vec(vece, t, a, b);
- tcg_gen_add_vec(vece, d, d, t);
- tcg_temp_free_vec(t);
- }
- void gen_gvec_uaba(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
- uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz)
- {
- static const TCGOpcode vecop_list[] = {
- INDEX_op_sub_vec, INDEX_op_add_vec,
- INDEX_op_umin_vec, INDEX_op_umax_vec, 0
- };
- static const GVecGen3 ops[4] = {
- { .fniv = gen_uaba_vec,
- .fno = gen_helper_gvec_uaba_b,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_8 },
- { .fniv = gen_uaba_vec,
- .fno = gen_helper_gvec_uaba_h,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_16 },
- { .fni4 = gen_uaba_i32,
- .fniv = gen_uaba_vec,
- .fno = gen_helper_gvec_uaba_s,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_32 },
- { .fni8 = gen_uaba_i64,
- .fniv = gen_uaba_vec,
- .fno = gen_helper_gvec_uaba_d,
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
- .opt_opc = vecop_list,
- .load_dest = true,
- .vece = MO_64 },
- };
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, &ops[vece]);
- }
- static int disas_coproc_insn(DisasContext *s, uint32_t insn)
- {
- int cpnum, is64, crn, crm, opc1, opc2, isread, rt, rt2;
- const ARMCPRegInfo *ri;
- cpnum = (insn >> 8) & 0xf;
- /* First check for coprocessor space used for XScale/iwMMXt insns */
- if (arm_dc_feature(s, ARM_FEATURE_XSCALE) && (cpnum < 2)) {
- if (extract32(s->c15_cpar, cpnum, 1) == 0) {
- return 1;
- }
- if (arm_dc_feature(s, ARM_FEATURE_IWMMXT)) {
- return disas_iwmmxt_insn(s, insn);
- } else if (arm_dc_feature(s, ARM_FEATURE_XSCALE)) {
- return disas_dsp_insn(s, insn);
- }
- return 1;
- }
- /* Otherwise treat as a generic register access */
- is64 = (insn & (1 << 25)) == 0;
- if (!is64 && ((insn & (1 << 4)) == 0)) {
- /* cdp */
- return 1;
- }
- crm = insn & 0xf;
- if (is64) {
- crn = 0;
- opc1 = (insn >> 4) & 0xf;
- opc2 = 0;
- rt2 = (insn >> 16) & 0xf;
- } else {
- crn = (insn >> 16) & 0xf;
- opc1 = (insn >> 21) & 7;
- opc2 = (insn >> 5) & 7;
- rt2 = 0;
- }
- isread = (insn >> 20) & 1;
- rt = (insn >> 12) & 0xf;
- ri = get_arm_cp_reginfo(s->cp_regs,
- ENCODE_CP_REG(cpnum, is64, s->ns, crn, crm, opc1, opc2));
- if (ri) {
- bool need_exit_tb;
- /* Check access permissions */
- if (!cp_access_ok(s->current_el, ri, isread)) {
- return 1;
- }
- if (s->hstr_active || ri->accessfn ||
- (arm_dc_feature(s, ARM_FEATURE_XSCALE) && cpnum < 14)) {
- /* Emit code to perform further access permissions checks at
- * runtime; this may result in an exception.
- * Note that on XScale all cp0..c13 registers do an access check
- * call in order to handle c15_cpar.
- */
- TCGv_ptr tmpptr;
- TCGv_i32 tcg_syn, tcg_isread;
- uint32_t syndrome;
- /* Note that since we are an implementation which takes an
- * exception on a trapped conditional instruction only if the
- * instruction passes its condition code check, we can take
- * advantage of the clause in the ARM ARM that allows us to set
- * the COND field in the instruction to 0xE in all cases.
- * We could fish the actual condition out of the insn (ARM)
- * or the condexec bits (Thumb) but it isn't necessary.
- */
- switch (cpnum) {
- case 14:
- if (is64) {
- syndrome = syn_cp14_rrt_trap(1, 0xe, opc1, crm, rt, rt2,
- isread, false);
- } else {
- syndrome = syn_cp14_rt_trap(1, 0xe, opc1, opc2, crn, crm,
- rt, isread, false);
- }
- break;
- case 15:
- if (is64) {
- syndrome = syn_cp15_rrt_trap(1, 0xe, opc1, crm, rt, rt2,
- isread, false);
- } else {
- syndrome = syn_cp15_rt_trap(1, 0xe, opc1, opc2, crn, crm,
- rt, isread, false);
- }
- break;
- default:
- /* ARMv8 defines that only coprocessors 14 and 15 exist,
- * so this can only happen if this is an ARMv7 or earlier CPU,
- * in which case the syndrome information won't actually be
- * guest visible.
- */
- assert(!arm_dc_feature(s, ARM_FEATURE_V8));
- syndrome = syn_uncategorized();
- break;
- }
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- tmpptr = tcg_const_ptr(ri);
- tcg_syn = tcg_const_i32(syndrome);
- tcg_isread = tcg_const_i32(isread);
- gen_helper_access_check_cp_reg(cpu_env, tmpptr, tcg_syn,
- tcg_isread);
- tcg_temp_free_ptr(tmpptr);
- tcg_temp_free_i32(tcg_syn);
- tcg_temp_free_i32(tcg_isread);
- } else if (ri->type & ARM_CP_RAISES_EXC) {
- /*
- * The readfn or writefn might raise an exception;
- * synchronize the CPU state in case it does.
- */
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- }
- /* Handle special cases first */
- switch (ri->type & ~(ARM_CP_FLAG_MASK & ~ARM_CP_SPECIAL)) {
- case ARM_CP_NOP:
- return 0;
- case ARM_CP_WFI:
- if (isread) {
- return 1;
- }
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_WFI;
- return 0;
- default:
- break;
- }
- if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
- gen_io_start();
- }
- if (isread) {
- /* Read */
- if (is64) {
- TCGv_i64 tmp64;
- TCGv_i32 tmp;
- if (ri->type & ARM_CP_CONST) {
- tmp64 = tcg_const_i64(ri->resetvalue);
- } else if (ri->readfn) {
- TCGv_ptr tmpptr;
- tmp64 = tcg_temp_new_i64();
- tmpptr = tcg_const_ptr(ri);
- gen_helper_get_cp_reg64(tmp64, cpu_env, tmpptr);
- tcg_temp_free_ptr(tmpptr);
- } else {
- tmp64 = tcg_temp_new_i64();
- tcg_gen_ld_i64(tmp64, cpu_env, ri->fieldoffset);
- }
- tmp = tcg_temp_new_i32();
- tcg_gen_extrl_i64_i32(tmp, tmp64);
- store_reg(s, rt, tmp);
- tmp = tcg_temp_new_i32();
- tcg_gen_extrh_i64_i32(tmp, tmp64);
- tcg_temp_free_i64(tmp64);
- store_reg(s, rt2, tmp);
- } else {
- TCGv_i32 tmp;
- if (ri->type & ARM_CP_CONST) {
- tmp = tcg_const_i32(ri->resetvalue);
- } else if (ri->readfn) {
- TCGv_ptr tmpptr;
- tmp = tcg_temp_new_i32();
- tmpptr = tcg_const_ptr(ri);
- gen_helper_get_cp_reg(tmp, cpu_env, tmpptr);
- tcg_temp_free_ptr(tmpptr);
- } else {
- tmp = load_cpu_offset(ri->fieldoffset);
- }
- if (rt == 15) {
- /* Destination register of r15 for 32 bit loads sets
- * the condition codes from the high 4 bits of the value
- */
- gen_set_nzcv(tmp);
- tcg_temp_free_i32(tmp);
- } else {
- store_reg(s, rt, tmp);
- }
- }
- } else {
- /* Write */
- if (ri->type & ARM_CP_CONST) {
- /* If not forbidden by access permissions, treat as WI */
- return 0;
- }
- if (is64) {
- TCGv_i32 tmplo, tmphi;
- TCGv_i64 tmp64 = tcg_temp_new_i64();
- tmplo = load_reg(s, rt);
- tmphi = load_reg(s, rt2);
- tcg_gen_concat_i32_i64(tmp64, tmplo, tmphi);
- tcg_temp_free_i32(tmplo);
- tcg_temp_free_i32(tmphi);
- if (ri->writefn) {
- TCGv_ptr tmpptr = tcg_const_ptr(ri);
- gen_helper_set_cp_reg64(cpu_env, tmpptr, tmp64);
- tcg_temp_free_ptr(tmpptr);
- } else {
- tcg_gen_st_i64(tmp64, cpu_env, ri->fieldoffset);
- }
- tcg_temp_free_i64(tmp64);
- } else {
- if (ri->writefn) {
- TCGv_i32 tmp;
- TCGv_ptr tmpptr;
- tmp = load_reg(s, rt);
- tmpptr = tcg_const_ptr(ri);
- gen_helper_set_cp_reg(cpu_env, tmpptr, tmp);
- tcg_temp_free_ptr(tmpptr);
- tcg_temp_free_i32(tmp);
- } else {
- TCGv_i32 tmp = load_reg(s, rt);
- store_cpu_offset(tmp, ri->fieldoffset);
- }
- }
- }
- /* I/O operations must end the TB here (whether read or write) */
- need_exit_tb = ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) &&
- (ri->type & ARM_CP_IO));
- if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
- /*
- * A write to any coprocessor register that ends a TB
- * must rebuild the hflags for the next TB.
- */
- TCGv_i32 tcg_el = tcg_const_i32(s->current_el);
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- gen_helper_rebuild_hflags_m32(cpu_env, tcg_el);
- } else {
- if (ri->type & ARM_CP_NEWEL) {
- gen_helper_rebuild_hflags_a32_newel(cpu_env);
- } else {
- gen_helper_rebuild_hflags_a32(cpu_env, tcg_el);
- }
- }
- tcg_temp_free_i32(tcg_el);
- /*
- * We default to ending the TB on a coprocessor register write,
- * but allow this to be suppressed by the register definition
- * (usually only necessary to work around guest bugs).
- */
- need_exit_tb = true;
- }
- if (need_exit_tb) {
- gen_lookup_tb(s);
- }
- return 0;
- }
- /* Unknown register; this might be a guest error or a QEMU
- * unimplemented feature.
- */
- if (is64) {
- qemu_log_mask(LOG_UNIMP, "%s access to unsupported AArch32 "
- "64 bit system register cp:%d opc1: %d crm:%d "
- "(%s)\n",
- isread ? "read" : "write", cpnum, opc1, crm,
- s->ns ? "non-secure" : "secure");
- } else {
- qemu_log_mask(LOG_UNIMP, "%s access to unsupported AArch32 "
- "system register cp:%d opc1:%d crn:%d crm:%d opc2:%d "
- "(%s)\n",
- isread ? "read" : "write", cpnum, opc1, crn, crm, opc2,
- s->ns ? "non-secure" : "secure");
- }
- return 1;
- }
- /* Store a 64-bit value to a register pair. Clobbers val. */
- static void gen_storeq_reg(DisasContext *s, int rlow, int rhigh, TCGv_i64 val)
- {
- TCGv_i32 tmp;
- tmp = tcg_temp_new_i32();
- tcg_gen_extrl_i64_i32(tmp, val);
- store_reg(s, rlow, tmp);
- tmp = tcg_temp_new_i32();
- tcg_gen_extrh_i64_i32(tmp, val);
- store_reg(s, rhigh, tmp);
- }
- /* load and add a 64-bit value from a register pair. */
- static void gen_addq(DisasContext *s, TCGv_i64 val, int rlow, int rhigh)
- {
- TCGv_i64 tmp;
- TCGv_i32 tmpl;
- TCGv_i32 tmph;
- /* Load 64-bit value rd:rn. */
- tmpl = load_reg(s, rlow);
- tmph = load_reg(s, rhigh);
- tmp = tcg_temp_new_i64();
- tcg_gen_concat_i32_i64(tmp, tmpl, tmph);
- tcg_temp_free_i32(tmpl);
- tcg_temp_free_i32(tmph);
- tcg_gen_add_i64(val, val, tmp);
- tcg_temp_free_i64(tmp);
- }
- /* Set N and Z flags from hi|lo. */
- static void gen_logicq_cc(TCGv_i32 lo, TCGv_i32 hi)
- {
- tcg_gen_mov_i32(cpu_NF, hi);
- tcg_gen_or_i32(cpu_ZF, lo, hi);
- }
- /* Load/Store exclusive instructions are implemented by remembering
- the value/address loaded, and seeing if these are the same
- when the store is performed. This should be sufficient to implement
- the architecturally mandated semantics, and avoids having to monitor
- regular stores. The compare vs the remembered value is done during
- the cmpxchg operation, but we must compare the addresses manually. */
- static void gen_load_exclusive(DisasContext *s, int rt, int rt2,
- TCGv_i32 addr, int size)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- MemOp opc = size | MO_ALIGN | s->be_data;
- s->is_ldex = true;
- if (size == 3) {
- TCGv_i32 tmp2 = tcg_temp_new_i32();
- TCGv_i64 t64 = tcg_temp_new_i64();
- /* For AArch32, architecturally the 32-bit word at the lowest
- * address is always Rt and the one at addr+4 is Rt2, even if
- * the CPU is big-endian. That means we don't want to do a
- * gen_aa32_ld_i64(), which invokes gen_aa32_frob64() as if
- * for an architecturally 64-bit access, but instead do a
- * 64-bit access using MO_BE if appropriate and then split
- * the two halves.
- * This only makes a difference for BE32 user-mode, where
- * frob64() must not flip the two halves of the 64-bit data
- * but this code must treat BE32 user-mode like BE32 system.
- */
- TCGv taddr = gen_aa32_addr(s, addr, opc);
- tcg_gen_qemu_ld_i64(t64, taddr, get_mem_index(s), opc);
- tcg_temp_free(taddr);
- tcg_gen_mov_i64(cpu_exclusive_val, t64);
- if (s->be_data == MO_BE) {
- tcg_gen_extr_i64_i32(tmp2, tmp, t64);
- } else {
- tcg_gen_extr_i64_i32(tmp, tmp2, t64);
- }
- tcg_temp_free_i64(t64);
- store_reg(s, rt2, tmp2);
- } else {
- gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s), opc);
- tcg_gen_extu_i32_i64(cpu_exclusive_val, tmp);
- }
- store_reg(s, rt, tmp);
- tcg_gen_extu_i32_i64(cpu_exclusive_addr, addr);
- }
- static void gen_clrex(DisasContext *s)
- {
- tcg_gen_movi_i64(cpu_exclusive_addr, -1);
- }
- static void gen_store_exclusive(DisasContext *s, int rd, int rt, int rt2,
- TCGv_i32 addr, int size)
- {
- TCGv_i32 t0, t1, t2;
- TCGv_i64 extaddr;
- TCGv taddr;
- TCGLabel *done_label;
- TCGLabel *fail_label;
- MemOp opc = size | MO_ALIGN | s->be_data;
- /* if (env->exclusive_addr == addr && env->exclusive_val == [addr]) {
- [addr] = {Rt};
- {Rd} = 0;
- } else {
- {Rd} = 1;
- } */
- fail_label = gen_new_label();
- done_label = gen_new_label();
- extaddr = tcg_temp_new_i64();
- tcg_gen_extu_i32_i64(extaddr, addr);
- tcg_gen_brcond_i64(TCG_COND_NE, extaddr, cpu_exclusive_addr, fail_label);
- tcg_temp_free_i64(extaddr);
- taddr = gen_aa32_addr(s, addr, opc);
- t0 = tcg_temp_new_i32();
- t1 = load_reg(s, rt);
- if (size == 3) {
- TCGv_i64 o64 = tcg_temp_new_i64();
- TCGv_i64 n64 = tcg_temp_new_i64();
- t2 = load_reg(s, rt2);
- /* For AArch32, architecturally the 32-bit word at the lowest
- * address is always Rt and the one at addr+4 is Rt2, even if
- * the CPU is big-endian. Since we're going to treat this as a
- * single 64-bit BE store, we need to put the two halves in the
- * opposite order for BE to LE, so that they end up in the right
- * places.
- * We don't want gen_aa32_frob64() because that does the wrong
- * thing for BE32 usermode.
- */
- if (s->be_data == MO_BE) {
- tcg_gen_concat_i32_i64(n64, t2, t1);
- } else {
- tcg_gen_concat_i32_i64(n64, t1, t2);
- }
- tcg_temp_free_i32(t2);
- tcg_gen_atomic_cmpxchg_i64(o64, taddr, cpu_exclusive_val, n64,
- get_mem_index(s), opc);
- tcg_temp_free_i64(n64);
- tcg_gen_setcond_i64(TCG_COND_NE, o64, o64, cpu_exclusive_val);
- tcg_gen_extrl_i64_i32(t0, o64);
- tcg_temp_free_i64(o64);
- } else {
- t2 = tcg_temp_new_i32();
- tcg_gen_extrl_i64_i32(t2, cpu_exclusive_val);
- tcg_gen_atomic_cmpxchg_i32(t0, taddr, t2, t1, get_mem_index(s), opc);
- tcg_gen_setcond_i32(TCG_COND_NE, t0, t0, t2);
- tcg_temp_free_i32(t2);
- }
- tcg_temp_free_i32(t1);
- tcg_temp_free(taddr);
- tcg_gen_mov_i32(cpu_R[rd], t0);
- tcg_temp_free_i32(t0);
- tcg_gen_br(done_label);
- gen_set_label(fail_label);
- tcg_gen_movi_i32(cpu_R[rd], 1);
- gen_set_label(done_label);
- tcg_gen_movi_i64(cpu_exclusive_addr, -1);
- }
- /* gen_srs:
- * @env: CPUARMState
- * @s: DisasContext
- * @mode: mode field from insn (which stack to store to)
- * @amode: addressing mode (DA/IA/DB/IB), encoded as per P,U bits in ARM insn
- * @writeback: true if writeback bit set
- *
- * Generate code for the SRS (Store Return State) insn.
- */
- static void gen_srs(DisasContext *s,
- uint32_t mode, uint32_t amode, bool writeback)
- {
- int32_t offset;
- TCGv_i32 addr, tmp;
- bool undef = false;
- /* SRS is:
- * - trapped to EL3 if EL3 is AArch64 and we are at Secure EL1
- * and specified mode is monitor mode
- * - UNDEFINED in Hyp mode
- * - UNPREDICTABLE in User or System mode
- * - UNPREDICTABLE if the specified mode is:
- * -- not implemented
- * -- not a valid mode number
- * -- a mode that's at a higher exception level
- * -- Monitor, if we are Non-secure
- * For the UNPREDICTABLE cases we choose to UNDEF.
- */
- if (s->current_el == 1 && !s->ns && mode == ARM_CPU_MODE_MON) {
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF, syn_uncategorized(), 3);
- return;
- }
- if (s->current_el == 0 || s->current_el == 2) {
- undef = true;
- }
- switch (mode) {
- case ARM_CPU_MODE_USR:
- case ARM_CPU_MODE_FIQ:
- case ARM_CPU_MODE_IRQ:
- case ARM_CPU_MODE_SVC:
- case ARM_CPU_MODE_ABT:
- case ARM_CPU_MODE_UND:
- case ARM_CPU_MODE_SYS:
- break;
- case ARM_CPU_MODE_HYP:
- if (s->current_el == 1 || !arm_dc_feature(s, ARM_FEATURE_EL2)) {
- undef = true;
- }
- break;
- case ARM_CPU_MODE_MON:
- /* No need to check specifically for "are we non-secure" because
- * we've already made EL0 UNDEF and handled the trap for S-EL1;
- * so if this isn't EL3 then we must be non-secure.
- */
- if (s->current_el != 3) {
- undef = true;
- }
- break;
- default:
- undef = true;
- }
- if (undef) {
- unallocated_encoding(s);
- return;
- }
- addr = tcg_temp_new_i32();
- tmp = tcg_const_i32(mode);
- /* get_r13_banked() will raise an exception if called from System mode */
- gen_set_condexec(s);
- gen_set_pc_im(s, s->pc_curr);
- gen_helper_get_r13_banked(addr, cpu_env, tmp);
- tcg_temp_free_i32(tmp);
- switch (amode) {
- case 0: /* DA */
- offset = -4;
- break;
- case 1: /* IA */
- offset = 0;
- break;
- case 2: /* DB */
- offset = -8;
- break;
- case 3: /* IB */
- offset = 4;
- break;
- default:
- abort();
- }
- tcg_gen_addi_i32(addr, addr, offset);
- tmp = load_reg(s, 14);
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
- tcg_temp_free_i32(tmp);
- tmp = load_cpu_field(spsr);
- tcg_gen_addi_i32(addr, addr, 4);
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
- tcg_temp_free_i32(tmp);
- if (writeback) {
- switch (amode) {
- case 0:
- offset = -8;
- break;
- case 1:
- offset = 4;
- break;
- case 2:
- offset = -4;
- break;
- case 3:
- offset = 0;
- break;
- default:
- abort();
- }
- tcg_gen_addi_i32(addr, addr, offset);
- tmp = tcg_const_i32(mode);
- gen_helper_set_r13_banked(cpu_env, tmp, addr);
- tcg_temp_free_i32(tmp);
- }
- tcg_temp_free_i32(addr);
- s->base.is_jmp = DISAS_UPDATE_EXIT;
- }
- /* Generate a label used for skipping this instruction */
- static void arm_gen_condlabel(DisasContext *s)
- {
- if (!s->condjmp) {
- s->condlabel = gen_new_label();
- s->condjmp = 1;
- }
- }
- /* Skip this instruction if the ARM condition is false */
- static void arm_skip_unless(DisasContext *s, uint32_t cond)
- {
- arm_gen_condlabel(s);
- arm_gen_test_cc(cond ^ 1, s->condlabel);
- }
- /*
- * Constant expanders for the decoders.
- */
- static int negate(DisasContext *s, int x)
- {
- return -x;
- }
- static int plus_2(DisasContext *s, int x)
- {
- return x + 2;
- }
- static int times_2(DisasContext *s, int x)
- {
- return x * 2;
- }
- static int times_4(DisasContext *s, int x)
- {
- return x * 4;
- }
- /* Return only the rotation part of T32ExpandImm. */
- static int t32_expandimm_rot(DisasContext *s, int x)
- {
- return x & 0xc00 ? extract32(x, 7, 5) : 0;
- }
- /* Return the unrotated immediate from T32ExpandImm. */
- static int t32_expandimm_imm(DisasContext *s, int x)
- {
- int imm = extract32(x, 0, 8);
- switch (extract32(x, 8, 4)) {
- case 0: /* XY */
- /* Nothing to do. */
- break;
- case 1: /* 00XY00XY */
- imm *= 0x00010001;
- break;
- case 2: /* XY00XY00 */
- imm *= 0x01000100;
- break;
- case 3: /* XYXYXYXY */
- imm *= 0x01010101;
- break;
- default:
- /* Rotated constant. */
- imm |= 0x80;
- break;
- }
- return imm;
- }
- static int t32_branch24(DisasContext *s, int x)
- {
- /* Convert J1:J2 at x[22:21] to I2:I1, which involves I=J^~S. */
- x ^= !(x < 0) * (3 << 21);
- /* Append the final zero. */
- return x << 1;
- }
- static int t16_setflags(DisasContext *s)
- {
- return s->condexec_mask == 0;
- }
- static int t16_push_list(DisasContext *s, int x)
- {
- return (x & 0xff) | (x & 0x100) << (14 - 8);
- }
- static int t16_pop_list(DisasContext *s, int x)
- {
- return (x & 0xff) | (x & 0x100) << (15 - 8);
- }
- /*
- * Include the generated decoders.
- */
- #include "decode-a32.c.inc"
- #include "decode-a32-uncond.c.inc"
- #include "decode-t32.c.inc"
- #include "decode-t16.c.inc"
- /* Helpers to swap operands for reverse-subtract. */
- static void gen_rsb(TCGv_i32 dst, TCGv_i32 a, TCGv_i32 b)
- {
- tcg_gen_sub_i32(dst, b, a);
- }
- static void gen_rsb_CC(TCGv_i32 dst, TCGv_i32 a, TCGv_i32 b)
- {
- gen_sub_CC(dst, b, a);
- }
- static void gen_rsc(TCGv_i32 dest, TCGv_i32 a, TCGv_i32 b)
- {
- gen_sub_carry(dest, b, a);
- }
- static void gen_rsc_CC(TCGv_i32 dest, TCGv_i32 a, TCGv_i32 b)
- {
- gen_sbc_CC(dest, b, a);
- }
- /*
- * Helpers for the data processing routines.
- *
- * After the computation store the results back.
- * This may be suppressed altogether (STREG_NONE), require a runtime
- * check against the stack limits (STREG_SP_CHECK), or generate an
- * exception return. Oh, or store into a register.
- *
- * Always return true, indicating success for a trans_* function.
- */
- typedef enum {
- STREG_NONE,
- STREG_NORMAL,
- STREG_SP_CHECK,
- STREG_EXC_RET,
- } StoreRegKind;
- static bool store_reg_kind(DisasContext *s, int rd,
- TCGv_i32 val, StoreRegKind kind)
- {
- switch (kind) {
- case STREG_NONE:
- tcg_temp_free_i32(val);
- return true;
- case STREG_NORMAL:
- /* See ALUWritePC: Interworking only from a32 mode. */
- if (s->thumb) {
- store_reg(s, rd, val);
- } else {
- store_reg_bx(s, rd, val);
- }
- return true;
- case STREG_SP_CHECK:
- store_sp_checked(s, val);
- return true;
- case STREG_EXC_RET:
- gen_exception_return(s, val);
- return true;
- }
- g_assert_not_reached();
- }
- /*
- * Data Processing (register)
- *
- * Operate, with set flags, one register source,
- * one immediate shifted register source, and a destination.
- */
- static bool op_s_rrr_shi(DisasContext *s, arg_s_rrr_shi *a,
- void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp1, tmp2;
- tmp2 = load_reg(s, a->rm);
- gen_arm_shift_im(tmp2, a->shty, a->shim, logic_cc);
- tmp1 = load_reg(s, a->rn);
- gen(tmp1, tmp1, tmp2);
- tcg_temp_free_i32(tmp2);
- if (logic_cc) {
- gen_logic_CC(tmp1);
- }
- return store_reg_kind(s, a->rd, tmp1, kind);
- }
- static bool op_s_rxr_shi(DisasContext *s, arg_s_rrr_shi *a,
- void (*gen)(TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp;
- tmp = load_reg(s, a->rm);
- gen_arm_shift_im(tmp, a->shty, a->shim, logic_cc);
- gen(tmp, tmp);
- if (logic_cc) {
- gen_logic_CC(tmp);
- }
- return store_reg_kind(s, a->rd, tmp, kind);
- }
- /*
- * Data-processing (register-shifted register)
- *
- * Operate, with set flags, one register source,
- * one register shifted register source, and a destination.
- */
- static bool op_s_rrr_shr(DisasContext *s, arg_s_rrr_shr *a,
- void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp1, tmp2;
- tmp1 = load_reg(s, a->rs);
- tmp2 = load_reg(s, a->rm);
- gen_arm_shift_reg(tmp2, a->shty, tmp1, logic_cc);
- tmp1 = load_reg(s, a->rn);
- gen(tmp1, tmp1, tmp2);
- tcg_temp_free_i32(tmp2);
- if (logic_cc) {
- gen_logic_CC(tmp1);
- }
- return store_reg_kind(s, a->rd, tmp1, kind);
- }
- static bool op_s_rxr_shr(DisasContext *s, arg_s_rrr_shr *a,
- void (*gen)(TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp1, tmp2;
- tmp1 = load_reg(s, a->rs);
- tmp2 = load_reg(s, a->rm);
- gen_arm_shift_reg(tmp2, a->shty, tmp1, logic_cc);
- gen(tmp2, tmp2);
- if (logic_cc) {
- gen_logic_CC(tmp2);
- }
- return store_reg_kind(s, a->rd, tmp2, kind);
- }
- /*
- * Data-processing (immediate)
- *
- * Operate, with set flags, one register source,
- * one rotated immediate, and a destination.
- *
- * Note that logic_cc && a->rot setting CF based on the msb of the
- * immediate is the reason why we must pass in the unrotated form
- * of the immediate.
- */
- static bool op_s_rri_rot(DisasContext *s, arg_s_rri_rot *a,
- void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp1, tmp2;
- uint32_t imm;
- imm = ror32(a->imm, a->rot);
- if (logic_cc && a->rot) {
- tcg_gen_movi_i32(cpu_CF, imm >> 31);
- }
- tmp2 = tcg_const_i32(imm);
- tmp1 = load_reg(s, a->rn);
- gen(tmp1, tmp1, tmp2);
- tcg_temp_free_i32(tmp2);
- if (logic_cc) {
- gen_logic_CC(tmp1);
- }
- return store_reg_kind(s, a->rd, tmp1, kind);
- }
- static bool op_s_rxi_rot(DisasContext *s, arg_s_rri_rot *a,
- void (*gen)(TCGv_i32, TCGv_i32),
- int logic_cc, StoreRegKind kind)
- {
- TCGv_i32 tmp;
- uint32_t imm;
- imm = ror32(a->imm, a->rot);
- if (logic_cc && a->rot) {
- tcg_gen_movi_i32(cpu_CF, imm >> 31);
- }
- tmp = tcg_const_i32(imm);
- gen(tmp, tmp);
- if (logic_cc) {
- gen_logic_CC(tmp);
- }
- return store_reg_kind(s, a->rd, tmp, kind);
- }
- #define DO_ANY3(NAME, OP, L, K) \
- static bool trans_##NAME##_rrri(DisasContext *s, arg_s_rrr_shi *a) \
- { StoreRegKind k = (K); return op_s_rrr_shi(s, a, OP, L, k); } \
- static bool trans_##NAME##_rrrr(DisasContext *s, arg_s_rrr_shr *a) \
- { StoreRegKind k = (K); return op_s_rrr_shr(s, a, OP, L, k); } \
- static bool trans_##NAME##_rri(DisasContext *s, arg_s_rri_rot *a) \
- { StoreRegKind k = (K); return op_s_rri_rot(s, a, OP, L, k); }
- #define DO_ANY2(NAME, OP, L, K) \
- static bool trans_##NAME##_rxri(DisasContext *s, arg_s_rrr_shi *a) \
- { StoreRegKind k = (K); return op_s_rxr_shi(s, a, OP, L, k); } \
- static bool trans_##NAME##_rxrr(DisasContext *s, arg_s_rrr_shr *a) \
- { StoreRegKind k = (K); return op_s_rxr_shr(s, a, OP, L, k); } \
- static bool trans_##NAME##_rxi(DisasContext *s, arg_s_rri_rot *a) \
- { StoreRegKind k = (K); return op_s_rxi_rot(s, a, OP, L, k); }
- #define DO_CMP2(NAME, OP, L) \
- static bool trans_##NAME##_xrri(DisasContext *s, arg_s_rrr_shi *a) \
- { return op_s_rrr_shi(s, a, OP, L, STREG_NONE); } \
- static bool trans_##NAME##_xrrr(DisasContext *s, arg_s_rrr_shr *a) \
- { return op_s_rrr_shr(s, a, OP, L, STREG_NONE); } \
- static bool trans_##NAME##_xri(DisasContext *s, arg_s_rri_rot *a) \
- { return op_s_rri_rot(s, a, OP, L, STREG_NONE); }
- DO_ANY3(AND, tcg_gen_and_i32, a->s, STREG_NORMAL)
- DO_ANY3(EOR, tcg_gen_xor_i32, a->s, STREG_NORMAL)
- DO_ANY3(ORR, tcg_gen_or_i32, a->s, STREG_NORMAL)
- DO_ANY3(BIC, tcg_gen_andc_i32, a->s, STREG_NORMAL)
- DO_ANY3(RSB, a->s ? gen_rsb_CC : gen_rsb, false, STREG_NORMAL)
- DO_ANY3(ADC, a->s ? gen_adc_CC : gen_add_carry, false, STREG_NORMAL)
- DO_ANY3(SBC, a->s ? gen_sbc_CC : gen_sub_carry, false, STREG_NORMAL)
- DO_ANY3(RSC, a->s ? gen_rsc_CC : gen_rsc, false, STREG_NORMAL)
- DO_CMP2(TST, tcg_gen_and_i32, true)
- DO_CMP2(TEQ, tcg_gen_xor_i32, true)
- DO_CMP2(CMN, gen_add_CC, false)
- DO_CMP2(CMP, gen_sub_CC, false)
- DO_ANY3(ADD, a->s ? gen_add_CC : tcg_gen_add_i32, false,
- a->rd == 13 && a->rn == 13 ? STREG_SP_CHECK : STREG_NORMAL)
- /*
- * Note for the computation of StoreRegKind we return out of the
- * middle of the functions that are expanded by DO_ANY3, and that
- * we modify a->s via that parameter before it is used by OP.
- */
- DO_ANY3(SUB, a->s ? gen_sub_CC : tcg_gen_sub_i32, false,
- ({
- StoreRegKind ret = STREG_NORMAL;
- if (a->rd == 15 && a->s) {
- /*
- * See ALUExceptionReturn:
- * In User mode, UNPREDICTABLE; we choose UNDEF.
- * In Hyp mode, UNDEFINED.
- */
- if (IS_USER(s) || s->current_el == 2) {
- unallocated_encoding(s);
- return true;
- }
- /* There is no writeback of nzcv to PSTATE. */
- a->s = 0;
- ret = STREG_EXC_RET;
- } else if (a->rd == 13 && a->rn == 13) {
- ret = STREG_SP_CHECK;
- }
- ret;
- }))
- DO_ANY2(MOV, tcg_gen_mov_i32, a->s,
- ({
- StoreRegKind ret = STREG_NORMAL;
- if (a->rd == 15 && a->s) {
- /*
- * See ALUExceptionReturn:
- * In User mode, UNPREDICTABLE; we choose UNDEF.
- * In Hyp mode, UNDEFINED.
- */
- if (IS_USER(s) || s->current_el == 2) {
- unallocated_encoding(s);
- return true;
- }
- /* There is no writeback of nzcv to PSTATE. */
- a->s = 0;
- ret = STREG_EXC_RET;
- } else if (a->rd == 13) {
- ret = STREG_SP_CHECK;
- }
- ret;
- }))
- DO_ANY2(MVN, tcg_gen_not_i32, a->s, STREG_NORMAL)
- /*
- * ORN is only available with T32, so there is no register-shifted-register
- * form of the insn. Using the DO_ANY3 macro would create an unused function.
- */
- static bool trans_ORN_rrri(DisasContext *s, arg_s_rrr_shi *a)
- {
- return op_s_rrr_shi(s, a, tcg_gen_orc_i32, a->s, STREG_NORMAL);
- }
- static bool trans_ORN_rri(DisasContext *s, arg_s_rri_rot *a)
- {
- return op_s_rri_rot(s, a, tcg_gen_orc_i32, a->s, STREG_NORMAL);
- }
- #undef DO_ANY3
- #undef DO_ANY2
- #undef DO_CMP2
- static bool trans_ADR(DisasContext *s, arg_ri *a)
- {
- store_reg_bx(s, a->rd, add_reg_for_lit(s, 15, a->imm));
- return true;
- }
- static bool trans_MOVW(DisasContext *s, arg_MOVW *a)
- {
- TCGv_i32 tmp;
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- tmp = tcg_const_i32(a->imm);
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_MOVT(DisasContext *s, arg_MOVW *a)
- {
- TCGv_i32 tmp;
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- tmp = load_reg(s, a->rd);
- tcg_gen_ext16u_i32(tmp, tmp);
- tcg_gen_ori_i32(tmp, tmp, a->imm << 16);
- store_reg(s, a->rd, tmp);
- return true;
- }
- /*
- * Multiply and multiply accumulate
- */
- static bool op_mla(DisasContext *s, arg_s_rrrr *a, bool add)
- {
- TCGv_i32 t1, t2;
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- tcg_gen_mul_i32(t1, t1, t2);
- tcg_temp_free_i32(t2);
- if (add) {
- t2 = load_reg(s, a->ra);
- tcg_gen_add_i32(t1, t1, t2);
- tcg_temp_free_i32(t2);
- }
- if (a->s) {
- gen_logic_CC(t1);
- }
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool trans_MUL(DisasContext *s, arg_MUL *a)
- {
- return op_mla(s, a, false);
- }
- static bool trans_MLA(DisasContext *s, arg_MLA *a)
- {
- return op_mla(s, a, true);
- }
- static bool trans_MLS(DisasContext *s, arg_MLS *a)
- {
- TCGv_i32 t1, t2;
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- tcg_gen_mul_i32(t1, t1, t2);
- tcg_temp_free_i32(t2);
- t2 = load_reg(s, a->ra);
- tcg_gen_sub_i32(t1, t2, t1);
- tcg_temp_free_i32(t2);
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool op_mlal(DisasContext *s, arg_s_rrrr *a, bool uns, bool add)
- {
- TCGv_i32 t0, t1, t2, t3;
- t0 = load_reg(s, a->rm);
- t1 = load_reg(s, a->rn);
- if (uns) {
- tcg_gen_mulu2_i32(t0, t1, t0, t1);
- } else {
- tcg_gen_muls2_i32(t0, t1, t0, t1);
- }
- if (add) {
- t2 = load_reg(s, a->ra);
- t3 = load_reg(s, a->rd);
- tcg_gen_add2_i32(t0, t1, t0, t1, t2, t3);
- tcg_temp_free_i32(t2);
- tcg_temp_free_i32(t3);
- }
- if (a->s) {
- gen_logicq_cc(t0, t1);
- }
- store_reg(s, a->ra, t0);
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool trans_UMULL(DisasContext *s, arg_UMULL *a)
- {
- return op_mlal(s, a, true, false);
- }
- static bool trans_SMULL(DisasContext *s, arg_SMULL *a)
- {
- return op_mlal(s, a, false, false);
- }
- static bool trans_UMLAL(DisasContext *s, arg_UMLAL *a)
- {
- return op_mlal(s, a, true, true);
- }
- static bool trans_SMLAL(DisasContext *s, arg_SMLAL *a)
- {
- return op_mlal(s, a, false, true);
- }
- static bool trans_UMAAL(DisasContext *s, arg_UMAAL *a)
- {
- TCGv_i32 t0, t1, t2, zero;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- t0 = load_reg(s, a->rm);
- t1 = load_reg(s, a->rn);
- tcg_gen_mulu2_i32(t0, t1, t0, t1);
- zero = tcg_const_i32(0);
- t2 = load_reg(s, a->ra);
- tcg_gen_add2_i32(t0, t1, t0, t1, t2, zero);
- tcg_temp_free_i32(t2);
- t2 = load_reg(s, a->rd);
- tcg_gen_add2_i32(t0, t1, t0, t1, t2, zero);
- tcg_temp_free_i32(t2);
- tcg_temp_free_i32(zero);
- store_reg(s, a->ra, t0);
- store_reg(s, a->rd, t1);
- return true;
- }
- /*
- * Saturating addition and subtraction
- */
- static bool op_qaddsub(DisasContext *s, arg_rrr *a, bool add, bool doub)
- {
- TCGv_i32 t0, t1;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_5TE) {
- return false;
- }
- t0 = load_reg(s, a->rm);
- t1 = load_reg(s, a->rn);
- if (doub) {
- gen_helper_add_saturate(t1, cpu_env, t1, t1);
- }
- if (add) {
- gen_helper_add_saturate(t0, cpu_env, t0, t1);
- } else {
- gen_helper_sub_saturate(t0, cpu_env, t0, t1);
- }
- tcg_temp_free_i32(t1);
- store_reg(s, a->rd, t0);
- return true;
- }
- #define DO_QADDSUB(NAME, ADD, DOUB) \
- static bool trans_##NAME(DisasContext *s, arg_rrr *a) \
- { \
- return op_qaddsub(s, a, ADD, DOUB); \
- }
- DO_QADDSUB(QADD, true, false)
- DO_QADDSUB(QSUB, false, false)
- DO_QADDSUB(QDADD, true, true)
- DO_QADDSUB(QDSUB, false, true)
- #undef DO_QADDSUB
- /*
- * Halfword multiply and multiply accumulate
- */
- static bool op_smlaxxx(DisasContext *s, arg_rrrr *a,
- int add_long, bool nt, bool mt)
- {
- TCGv_i32 t0, t1, tl, th;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_5TE) {
- return false;
- }
- t0 = load_reg(s, a->rn);
- t1 = load_reg(s, a->rm);
- gen_mulxy(t0, t1, nt, mt);
- tcg_temp_free_i32(t1);
- switch (add_long) {
- case 0:
- store_reg(s, a->rd, t0);
- break;
- case 1:
- t1 = load_reg(s, a->ra);
- gen_helper_add_setq(t0, cpu_env, t0, t1);
- tcg_temp_free_i32(t1);
- store_reg(s, a->rd, t0);
- break;
- case 2:
- tl = load_reg(s, a->ra);
- th = load_reg(s, a->rd);
- /* Sign-extend the 32-bit product to 64 bits. */
- t1 = tcg_temp_new_i32();
- tcg_gen_sari_i32(t1, t0, 31);
- tcg_gen_add2_i32(tl, th, tl, th, t0, t1);
- tcg_temp_free_i32(t0);
- tcg_temp_free_i32(t1);
- store_reg(s, a->ra, tl);
- store_reg(s, a->rd, th);
- break;
- default:
- g_assert_not_reached();
- }
- return true;
- }
- #define DO_SMLAX(NAME, add, nt, mt) \
- static bool trans_##NAME(DisasContext *s, arg_rrrr *a) \
- { \
- return op_smlaxxx(s, a, add, nt, mt); \
- }
- DO_SMLAX(SMULBB, 0, 0, 0)
- DO_SMLAX(SMULBT, 0, 0, 1)
- DO_SMLAX(SMULTB, 0, 1, 0)
- DO_SMLAX(SMULTT, 0, 1, 1)
- DO_SMLAX(SMLABB, 1, 0, 0)
- DO_SMLAX(SMLABT, 1, 0, 1)
- DO_SMLAX(SMLATB, 1, 1, 0)
- DO_SMLAX(SMLATT, 1, 1, 1)
- DO_SMLAX(SMLALBB, 2, 0, 0)
- DO_SMLAX(SMLALBT, 2, 0, 1)
- DO_SMLAX(SMLALTB, 2, 1, 0)
- DO_SMLAX(SMLALTT, 2, 1, 1)
- #undef DO_SMLAX
- static bool op_smlawx(DisasContext *s, arg_rrrr *a, bool add, bool mt)
- {
- TCGv_i32 t0, t1;
- if (!ENABLE_ARCH_5TE) {
- return false;
- }
- t0 = load_reg(s, a->rn);
- t1 = load_reg(s, a->rm);
- /*
- * Since the nominal result is product<47:16>, shift the 16-bit
- * input up by 16 bits, so that the result is at product<63:32>.
- */
- if (mt) {
- tcg_gen_andi_i32(t1, t1, 0xffff0000);
- } else {
- tcg_gen_shli_i32(t1, t1, 16);
- }
- tcg_gen_muls2_i32(t0, t1, t0, t1);
- tcg_temp_free_i32(t0);
- if (add) {
- t0 = load_reg(s, a->ra);
- gen_helper_add_setq(t1, cpu_env, t1, t0);
- tcg_temp_free_i32(t0);
- }
- store_reg(s, a->rd, t1);
- return true;
- }
- #define DO_SMLAWX(NAME, add, mt) \
- static bool trans_##NAME(DisasContext *s, arg_rrrr *a) \
- { \
- return op_smlawx(s, a, add, mt); \
- }
- DO_SMLAWX(SMULWB, 0, 0)
- DO_SMLAWX(SMULWT, 0, 1)
- DO_SMLAWX(SMLAWB, 1, 0)
- DO_SMLAWX(SMLAWT, 1, 1)
- #undef DO_SMLAWX
- /*
- * MSR (immediate) and hints
- */
- static bool trans_YIELD(DisasContext *s, arg_YIELD *a)
- {
- /*
- * When running single-threaded TCG code, use the helper to ensure that
- * the next round-robin scheduled vCPU gets a crack. When running in
- * MTTCG we don't generate jumps to the helper as it won't affect the
- * scheduling of other vCPUs.
- */
- if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) {
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_YIELD;
- }
- return true;
- }
- static bool trans_WFE(DisasContext *s, arg_WFE *a)
- {
- /*
- * When running single-threaded TCG code, use the helper to ensure that
- * the next round-robin scheduled vCPU gets a crack. In MTTCG mode we
- * just skip this instruction. Currently the SEV/SEVL instructions,
- * which are *one* of many ways to wake the CPU from WFE, are not
- * implemented so we can't sleep like WFI does.
- */
- if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) {
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_WFE;
- }
- return true;
- }
- static bool trans_WFI(DisasContext *s, arg_WFI *a)
- {
- /* For WFI, halt the vCPU until an IRQ. */
- gen_set_pc_im(s, s->base.pc_next);
- s->base.is_jmp = DISAS_WFI;
- return true;
- }
- static bool trans_NOP(DisasContext *s, arg_NOP *a)
- {
- return true;
- }
- static bool trans_MSR_imm(DisasContext *s, arg_MSR_imm *a)
- {
- uint32_t val = ror32(a->imm, a->rot * 2);
- uint32_t mask = msr_mask(s, a->mask, a->r);
- if (gen_set_psr_im(s, mask, a->r, val)) {
- unallocated_encoding(s);
- }
- return true;
- }
- /*
- * Cyclic Redundancy Check
- */
- static bool op_crc32(DisasContext *s, arg_rrr *a, bool c, MemOp sz)
- {
- TCGv_i32 t1, t2, t3;
- if (!dc_isar_feature(aa32_crc32, s)) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- switch (sz) {
- case MO_8:
- gen_uxtb(t2);
- break;
- case MO_16:
- gen_uxth(t2);
- break;
- case MO_32:
- break;
- default:
- g_assert_not_reached();
- }
- t3 = tcg_const_i32(1 << sz);
- if (c) {
- gen_helper_crc32c(t1, t1, t2, t3);
- } else {
- gen_helper_crc32(t1, t1, t2, t3);
- }
- tcg_temp_free_i32(t2);
- tcg_temp_free_i32(t3);
- store_reg(s, a->rd, t1);
- return true;
- }
- #define DO_CRC32(NAME, c, sz) \
- static bool trans_##NAME(DisasContext *s, arg_rrr *a) \
- { return op_crc32(s, a, c, sz); }
- DO_CRC32(CRC32B, false, MO_8)
- DO_CRC32(CRC32H, false, MO_16)
- DO_CRC32(CRC32W, false, MO_32)
- DO_CRC32(CRC32CB, true, MO_8)
- DO_CRC32(CRC32CH, true, MO_16)
- DO_CRC32(CRC32CW, true, MO_32)
- #undef DO_CRC32
- /*
- * Miscellaneous instructions
- */
- static bool trans_MRS_bank(DisasContext *s, arg_MRS_bank *a)
- {
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- gen_mrs_banked(s, a->r, a->sysm, a->rd);
- return true;
- }
- static bool trans_MSR_bank(DisasContext *s, arg_MSR_bank *a)
- {
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- gen_msr_banked(s, a->r, a->sysm, a->rn);
- return true;
- }
- static bool trans_MRS_reg(DisasContext *s, arg_MRS_reg *a)
- {
- TCGv_i32 tmp;
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (a->r) {
- if (IS_USER(s)) {
- unallocated_encoding(s);
- return true;
- }
- tmp = load_cpu_field(spsr);
- } else {
- tmp = tcg_temp_new_i32();
- gen_helper_cpsr_read(tmp, cpu_env);
- }
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_MSR_reg(DisasContext *s, arg_MSR_reg *a)
- {
- TCGv_i32 tmp;
- uint32_t mask = msr_mask(s, a->mask, a->r);
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- tmp = load_reg(s, a->rn);
- if (gen_set_psr(s, mask, a->r, tmp)) {
- unallocated_encoding(s);
- }
- return true;
- }
- static bool trans_MRS_v7m(DisasContext *s, arg_MRS_v7m *a)
- {
- TCGv_i32 tmp;
- if (!arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- tmp = tcg_const_i32(a->sysm);
- gen_helper_v7m_mrs(tmp, cpu_env, tmp);
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_MSR_v7m(DisasContext *s, arg_MSR_v7m *a)
- {
- TCGv_i32 addr, reg;
- if (!arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- addr = tcg_const_i32((a->mask << 10) | a->sysm);
- reg = load_reg(s, a->rn);
- gen_helper_v7m_msr(cpu_env, addr, reg);
- tcg_temp_free_i32(addr);
- tcg_temp_free_i32(reg);
- /* If we wrote to CONTROL, the EL might have changed */
- gen_helper_rebuild_hflags_m32_newel(cpu_env);
- gen_lookup_tb(s);
- return true;
- }
- static bool trans_BX(DisasContext *s, arg_BX *a)
- {
- if (!ENABLE_ARCH_4T) {
- return false;
- }
- gen_bx_excret(s, load_reg(s, a->rm));
- return true;
- }
- static bool trans_BXJ(DisasContext *s, arg_BXJ *a)
- {
- if (!ENABLE_ARCH_5J || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- /* Trivial implementation equivalent to bx. */
- gen_bx(s, load_reg(s, a->rm));
- return true;
- }
- static bool trans_BLX_r(DisasContext *s, arg_BLX_r *a)
- {
- TCGv_i32 tmp;
- if (!ENABLE_ARCH_5) {
- return false;
- }
- tmp = load_reg(s, a->rm);
- tcg_gen_movi_i32(cpu_R[14], s->base.pc_next | s->thumb);
- gen_bx(s, tmp);
- return true;
- }
- /*
- * BXNS/BLXNS: only exist for v8M with the security extensions,
- * and always UNDEF if NonSecure. We don't implement these in
- * the user-only mode either (in theory you can use them from
- * Secure User mode but they are too tied in to system emulation).
- */
- static bool trans_BXNS(DisasContext *s, arg_BXNS *a)
- {
- if (!s->v8m_secure || IS_USER_ONLY) {
- unallocated_encoding(s);
- } else {
- gen_bxns(s, a->rm);
- }
- return true;
- }
- static bool trans_BLXNS(DisasContext *s, arg_BLXNS *a)
- {
- if (!s->v8m_secure || IS_USER_ONLY) {
- unallocated_encoding(s);
- } else {
- gen_blxns(s, a->rm);
- }
- return true;
- }
- static bool trans_CLZ(DisasContext *s, arg_CLZ *a)
- {
- TCGv_i32 tmp;
- if (!ENABLE_ARCH_5) {
- return false;
- }
- tmp = load_reg(s, a->rm);
- tcg_gen_clzi_i32(tmp, tmp, 32);
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_ERET(DisasContext *s, arg_ERET *a)
- {
- TCGv_i32 tmp;
- if (!arm_dc_feature(s, ARM_FEATURE_V7VE)) {
- return false;
- }
- if (IS_USER(s)) {
- unallocated_encoding(s);
- return true;
- }
- if (s->current_el == 2) {
- /* ERET from Hyp uses ELR_Hyp, not LR */
- tmp = load_cpu_field(elr_el[2]);
- } else {
- tmp = load_reg(s, 14);
- }
- gen_exception_return(s, tmp);
- return true;
- }
- static bool trans_HLT(DisasContext *s, arg_HLT *a)
- {
- gen_hlt(s, a->imm);
- return true;
- }
- static bool trans_BKPT(DisasContext *s, arg_BKPT *a)
- {
- if (!ENABLE_ARCH_5) {
- return false;
- }
- if (arm_dc_feature(s, ARM_FEATURE_M) &&
- semihosting_enabled() &&
- #ifndef CONFIG_USER_ONLY
- !IS_USER(s) &&
- #endif
- (a->imm == 0xab)) {
- gen_exception_internal_insn(s, s->pc_curr, EXCP_SEMIHOST);
- } else {
- gen_exception_bkpt_insn(s, syn_aa32_bkpt(a->imm, false));
- }
- return true;
- }
- static bool trans_HVC(DisasContext *s, arg_HVC *a)
- {
- if (!ENABLE_ARCH_7 || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (IS_USER(s)) {
- unallocated_encoding(s);
- } else {
- gen_hvc(s, a->imm);
- }
- return true;
- }
- static bool trans_SMC(DisasContext *s, arg_SMC *a)
- {
- if (!ENABLE_ARCH_6K || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (IS_USER(s)) {
- unallocated_encoding(s);
- } else {
- gen_smc(s);
- }
- return true;
- }
- static bool trans_SG(DisasContext *s, arg_SG *a)
- {
- if (!arm_dc_feature(s, ARM_FEATURE_M) ||
- !arm_dc_feature(s, ARM_FEATURE_V8)) {
- return false;
- }
- /*
- * SG (v8M only)
- * The bulk of the behaviour for this instruction is implemented
- * in v7m_handle_execute_nsc(), which deals with the insn when
- * it is executed by a CPU in non-secure state from memory
- * which is Secure & NonSecure-Callable.
- * Here we only need to handle the remaining cases:
- * * in NS memory (including the "security extension not
- * implemented" case) : NOP
- * * in S memory but CPU already secure (clear IT bits)
- * We know that the attribute for the memory this insn is
- * in must match the current CPU state, because otherwise
- * get_phys_addr_pmsav8 would have generated an exception.
- */
- if (s->v8m_secure) {
- /* Like the IT insn, we don't need to generate any code */
- s->condexec_cond = 0;
- s->condexec_mask = 0;
- }
- return true;
- }
- static bool trans_TT(DisasContext *s, arg_TT *a)
- {
- TCGv_i32 addr, tmp;
- if (!arm_dc_feature(s, ARM_FEATURE_M) ||
- !arm_dc_feature(s, ARM_FEATURE_V8)) {
- return false;
- }
- if (a->rd == 13 || a->rd == 15 || a->rn == 15) {
- /* We UNDEF for these UNPREDICTABLE cases */
- unallocated_encoding(s);
- return true;
- }
- if (a->A && !s->v8m_secure) {
- /* This case is UNDEFINED. */
- unallocated_encoding(s);
- return true;
- }
- addr = load_reg(s, a->rn);
- tmp = tcg_const_i32((a->A << 1) | a->T);
- gen_helper_v7m_tt(tmp, cpu_env, addr, tmp);
- tcg_temp_free_i32(addr);
- store_reg(s, a->rd, tmp);
- return true;
- }
- /*
- * Load/store register index
- */
- static ISSInfo make_issinfo(DisasContext *s, int rd, bool p, bool w)
- {
- ISSInfo ret;
- /* ISS not valid if writeback */
- if (p && !w) {
- ret = rd;
- if (s->base.pc_next - s->pc_curr == 2) {
- ret |= ISSIs16Bit;
- }
- } else {
- ret = ISSInvalid;
- }
- return ret;
- }
- static TCGv_i32 op_addr_rr_pre(DisasContext *s, arg_ldst_rr *a)
- {
- TCGv_i32 addr = load_reg(s, a->rn);
- if (s->v8m_stackcheck && a->rn == 13 && a->w) {
- gen_helper_v8m_stackcheck(cpu_env, addr);
- }
- if (a->p) {
- TCGv_i32 ofs = load_reg(s, a->rm);
- gen_arm_shift_im(ofs, a->shtype, a->shimm, 0);
- if (a->u) {
- tcg_gen_add_i32(addr, addr, ofs);
- } else {
- tcg_gen_sub_i32(addr, addr, ofs);
- }
- tcg_temp_free_i32(ofs);
- }
- return addr;
- }
- static void op_addr_rr_post(DisasContext *s, arg_ldst_rr *a,
- TCGv_i32 addr, int address_offset)
- {
- if (!a->p) {
- TCGv_i32 ofs = load_reg(s, a->rm);
- gen_arm_shift_im(ofs, a->shtype, a->shimm, 0);
- if (a->u) {
- tcg_gen_add_i32(addr, addr, ofs);
- } else {
- tcg_gen_sub_i32(addr, addr, ofs);
- }
- tcg_temp_free_i32(ofs);
- } else if (!a->w) {
- tcg_temp_free_i32(addr);
- return;
- }
- tcg_gen_addi_i32(addr, addr, address_offset);
- store_reg(s, a->rn, addr);
- }
- static bool op_load_rr(DisasContext *s, arg_ldst_rr *a,
- MemOp mop, int mem_idx)
- {
- ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w);
- TCGv_i32 addr, tmp;
- addr = op_addr_rr_pre(s, a);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, mop | s->be_data);
- disas_set_da_iss(s, mop, issinfo);
- /*
- * Perform base writeback before the loaded value to
- * ensure correct behavior with overlapping index registers.
- */
- op_addr_rr_post(s, a, addr, 0);
- store_reg_from_load(s, a->rt, tmp);
- return true;
- }
- static bool op_store_rr(DisasContext *s, arg_ldst_rr *a,
- MemOp mop, int mem_idx)
- {
- ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w) | ISSIsWrite;
- TCGv_i32 addr, tmp;
- addr = op_addr_rr_pre(s, a);
- tmp = load_reg(s, a->rt);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, mop | s->be_data);
- disas_set_da_iss(s, mop, issinfo);
- tcg_temp_free_i32(tmp);
- op_addr_rr_post(s, a, addr, 0);
- return true;
- }
- static bool trans_LDRD_rr(DisasContext *s, arg_ldst_rr *a)
- {
- int mem_idx = get_mem_index(s);
- TCGv_i32 addr, tmp;
- if (!ENABLE_ARCH_5TE) {
- return false;
- }
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- addr = op_addr_rr_pre(s, a);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- store_reg(s, a->rt, tmp);
- tcg_gen_addi_i32(addr, addr, 4);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- store_reg(s, a->rt + 1, tmp);
- /* LDRD w/ base writeback is undefined if the registers overlap. */
- op_addr_rr_post(s, a, addr, -4);
- return true;
- }
- static bool trans_STRD_rr(DisasContext *s, arg_ldst_rr *a)
- {
- int mem_idx = get_mem_index(s);
- TCGv_i32 addr, tmp;
- if (!ENABLE_ARCH_5TE) {
- return false;
- }
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- addr = op_addr_rr_pre(s, a);
- tmp = load_reg(s, a->rt);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- tcg_temp_free_i32(tmp);
- tcg_gen_addi_i32(addr, addr, 4);
- tmp = load_reg(s, a->rt + 1);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- tcg_temp_free_i32(tmp);
- op_addr_rr_post(s, a, addr, -4);
- return true;
- }
- /*
- * Load/store immediate index
- */
- static TCGv_i32 op_addr_ri_pre(DisasContext *s, arg_ldst_ri *a)
- {
- int ofs = a->imm;
- if (!a->u) {
- ofs = -ofs;
- }
- if (s->v8m_stackcheck && a->rn == 13 && a->w) {
- /*
- * Stackcheck. Here we know 'addr' is the current SP;
- * U is set if we're moving SP up, else down. It is
- * UNKNOWN whether the limit check triggers when SP starts
- * below the limit and ends up above it; we chose to do so.
- */
- if (!a->u) {
- TCGv_i32 newsp = tcg_temp_new_i32();
- tcg_gen_addi_i32(newsp, cpu_R[13], ofs);
- gen_helper_v8m_stackcheck(cpu_env, newsp);
- tcg_temp_free_i32(newsp);
- } else {
- gen_helper_v8m_stackcheck(cpu_env, cpu_R[13]);
- }
- }
- return add_reg_for_lit(s, a->rn, a->p ? ofs : 0);
- }
- static void op_addr_ri_post(DisasContext *s, arg_ldst_ri *a,
- TCGv_i32 addr, int address_offset)
- {
- if (!a->p) {
- if (a->u) {
- address_offset += a->imm;
- } else {
- address_offset -= a->imm;
- }
- } else if (!a->w) {
- tcg_temp_free_i32(addr);
- return;
- }
- tcg_gen_addi_i32(addr, addr, address_offset);
- store_reg(s, a->rn, addr);
- }
- static bool op_load_ri(DisasContext *s, arg_ldst_ri *a,
- MemOp mop, int mem_idx)
- {
- ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w);
- TCGv_i32 addr, tmp;
- addr = op_addr_ri_pre(s, a);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, mop | s->be_data);
- disas_set_da_iss(s, mop, issinfo);
- /*
- * Perform base writeback before the loaded value to
- * ensure correct behavior with overlapping index registers.
- */
- op_addr_ri_post(s, a, addr, 0);
- store_reg_from_load(s, a->rt, tmp);
- return true;
- }
- static bool op_store_ri(DisasContext *s, arg_ldst_ri *a,
- MemOp mop, int mem_idx)
- {
- ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w) | ISSIsWrite;
- TCGv_i32 addr, tmp;
- addr = op_addr_ri_pre(s, a);
- tmp = load_reg(s, a->rt);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, mop | s->be_data);
- disas_set_da_iss(s, mop, issinfo);
- tcg_temp_free_i32(tmp);
- op_addr_ri_post(s, a, addr, 0);
- return true;
- }
- static bool op_ldrd_ri(DisasContext *s, arg_ldst_ri *a, int rt2)
- {
- int mem_idx = get_mem_index(s);
- TCGv_i32 addr, tmp;
- addr = op_addr_ri_pre(s, a);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- store_reg(s, a->rt, tmp);
- tcg_gen_addi_i32(addr, addr, 4);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- store_reg(s, rt2, tmp);
- /* LDRD w/ base writeback is undefined if the registers overlap. */
- op_addr_ri_post(s, a, addr, -4);
- return true;
- }
- static bool trans_LDRD_ri_a32(DisasContext *s, arg_ldst_ri *a)
- {
- if (!ENABLE_ARCH_5TE || (a->rt & 1)) {
- return false;
- }
- return op_ldrd_ri(s, a, a->rt + 1);
- }
- static bool trans_LDRD_ri_t32(DisasContext *s, arg_ldst_ri2 *a)
- {
- arg_ldst_ri b = {
- .u = a->u, .w = a->w, .p = a->p,
- .rn = a->rn, .rt = a->rt, .imm = a->imm
- };
- return op_ldrd_ri(s, &b, a->rt2);
- }
- static bool op_strd_ri(DisasContext *s, arg_ldst_ri *a, int rt2)
- {
- int mem_idx = get_mem_index(s);
- TCGv_i32 addr, tmp;
- addr = op_addr_ri_pre(s, a);
- tmp = load_reg(s, a->rt);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- tcg_temp_free_i32(tmp);
- tcg_gen_addi_i32(addr, addr, 4);
- tmp = load_reg(s, rt2);
- gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | s->be_data);
- tcg_temp_free_i32(tmp);
- op_addr_ri_post(s, a, addr, -4);
- return true;
- }
- static bool trans_STRD_ri_a32(DisasContext *s, arg_ldst_ri *a)
- {
- if (!ENABLE_ARCH_5TE || (a->rt & 1)) {
- return false;
- }
- return op_strd_ri(s, a, a->rt + 1);
- }
- static bool trans_STRD_ri_t32(DisasContext *s, arg_ldst_ri2 *a)
- {
- arg_ldst_ri b = {
- .u = a->u, .w = a->w, .p = a->p,
- .rn = a->rn, .rt = a->rt, .imm = a->imm
- };
- return op_strd_ri(s, &b, a->rt2);
- }
- #define DO_LDST(NAME, WHICH, MEMOP) \
- static bool trans_##NAME##_ri(DisasContext *s, arg_ldst_ri *a) \
- { \
- return op_##WHICH##_ri(s, a, MEMOP, get_mem_index(s)); \
- } \
- static bool trans_##NAME##T_ri(DisasContext *s, arg_ldst_ri *a) \
- { \
- return op_##WHICH##_ri(s, a, MEMOP, get_a32_user_mem_index(s)); \
- } \
- static bool trans_##NAME##_rr(DisasContext *s, arg_ldst_rr *a) \
- { \
- return op_##WHICH##_rr(s, a, MEMOP, get_mem_index(s)); \
- } \
- static bool trans_##NAME##T_rr(DisasContext *s, arg_ldst_rr *a) \
- { \
- return op_##WHICH##_rr(s, a, MEMOP, get_a32_user_mem_index(s)); \
- }
- DO_LDST(LDR, load, MO_UL)
- DO_LDST(LDRB, load, MO_UB)
- DO_LDST(LDRH, load, MO_UW)
- DO_LDST(LDRSB, load, MO_SB)
- DO_LDST(LDRSH, load, MO_SW)
- DO_LDST(STR, store, MO_UL)
- DO_LDST(STRB, store, MO_UB)
- DO_LDST(STRH, store, MO_UW)
- #undef DO_LDST
- /*
- * Synchronization primitives
- */
- static bool op_swp(DisasContext *s, arg_SWP *a, MemOp opc)
- {
- TCGv_i32 addr, tmp;
- TCGv taddr;
- opc |= s->be_data;
- addr = load_reg(s, a->rn);
- taddr = gen_aa32_addr(s, addr, opc);
- tcg_temp_free_i32(addr);
- tmp = load_reg(s, a->rt2);
- tcg_gen_atomic_xchg_i32(tmp, taddr, tmp, get_mem_index(s), opc);
- tcg_temp_free(taddr);
- store_reg(s, a->rt, tmp);
- return true;
- }
- static bool trans_SWP(DisasContext *s, arg_SWP *a)
- {
- return op_swp(s, a, MO_UL | MO_ALIGN);
- }
- static bool trans_SWPB(DisasContext *s, arg_SWP *a)
- {
- return op_swp(s, a, MO_UB);
- }
- /*
- * Load/Store Exclusive and Load-Acquire/Store-Release
- */
- static bool op_strex(DisasContext *s, arg_STREX *a, MemOp mop, bool rel)
- {
- TCGv_i32 addr;
- /* Some cases stopped being UNPREDICTABLE in v8A (but not v8M) */
- bool v8a = ENABLE_ARCH_8 && !arm_dc_feature(s, ARM_FEATURE_M);
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rd == 15 || a->rn == 15 || a->rt == 15
- || a->rd == a->rn || a->rd == a->rt
- || (!v8a && s->thumb && (a->rd == 13 || a->rt == 13))
- || (mop == MO_64
- && (a->rt2 == 15
- || a->rd == a->rt2
- || (!v8a && s->thumb && a->rt2 == 13)))) {
- unallocated_encoding(s);
- return true;
- }
- if (rel) {
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
- }
- addr = tcg_temp_local_new_i32();
- load_reg_var(s, addr, a->rn);
- tcg_gen_addi_i32(addr, addr, a->imm);
- gen_store_exclusive(s, a->rd, a->rt, a->rt2, addr, mop);
- tcg_temp_free_i32(addr);
- return true;
- }
- static bool trans_STREX(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- return op_strex(s, a, MO_32, false);
- }
- static bool trans_STREXD_a32(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_6K) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- a->rt2 = a->rt + 1;
- return op_strex(s, a, MO_64, false);
- }
- static bool trans_STREXD_t32(DisasContext *s, arg_STREX *a)
- {
- return op_strex(s, a, MO_64, false);
- }
- static bool trans_STREXB(DisasContext *s, arg_STREX *a)
- {
- if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
- return false;
- }
- return op_strex(s, a, MO_8, false);
- }
- static bool trans_STREXH(DisasContext *s, arg_STREX *a)
- {
- if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
- return false;
- }
- return op_strex(s, a, MO_16, false);
- }
- static bool trans_STLEX(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_strex(s, a, MO_32, true);
- }
- static bool trans_STLEXD_a32(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- a->rt2 = a->rt + 1;
- return op_strex(s, a, MO_64, true);
- }
- static bool trans_STLEXD_t32(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_strex(s, a, MO_64, true);
- }
- static bool trans_STLEXB(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_strex(s, a, MO_8, true);
- }
- static bool trans_STLEXH(DisasContext *s, arg_STREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_strex(s, a, MO_16, true);
- }
- static bool op_stl(DisasContext *s, arg_STL *a, MemOp mop)
- {
- TCGv_i32 addr, tmp;
- if (!ENABLE_ARCH_8) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rn == 15 || a->rt == 15) {
- unallocated_encoding(s);
- return true;
- }
- addr = load_reg(s, a->rn);
- tmp = load_reg(s, a->rt);
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
- gen_aa32_st_i32(s, tmp, addr, get_mem_index(s), mop | s->be_data);
- disas_set_da_iss(s, mop, a->rt | ISSIsAcqRel | ISSIsWrite);
- tcg_temp_free_i32(tmp);
- tcg_temp_free_i32(addr);
- return true;
- }
- static bool trans_STL(DisasContext *s, arg_STL *a)
- {
- return op_stl(s, a, MO_UL);
- }
- static bool trans_STLB(DisasContext *s, arg_STL *a)
- {
- return op_stl(s, a, MO_UB);
- }
- static bool trans_STLH(DisasContext *s, arg_STL *a)
- {
- return op_stl(s, a, MO_UW);
- }
- static bool op_ldrex(DisasContext *s, arg_LDREX *a, MemOp mop, bool acq)
- {
- TCGv_i32 addr;
- /* Some cases stopped being UNPREDICTABLE in v8A (but not v8M) */
- bool v8a = ENABLE_ARCH_8 && !arm_dc_feature(s, ARM_FEATURE_M);
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rn == 15 || a->rt == 15
- || (!v8a && s->thumb && a->rt == 13)
- || (mop == MO_64
- && (a->rt2 == 15 || a->rt == a->rt2
- || (!v8a && s->thumb && a->rt2 == 13)))) {
- unallocated_encoding(s);
- return true;
- }
- addr = tcg_temp_local_new_i32();
- load_reg_var(s, addr, a->rn);
- tcg_gen_addi_i32(addr, addr, a->imm);
- gen_load_exclusive(s, a->rt, a->rt2, addr, mop);
- tcg_temp_free_i32(addr);
- if (acq) {
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
- }
- return true;
- }
- static bool trans_LDREX(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- return op_ldrex(s, a, MO_32, false);
- }
- static bool trans_LDREXD_a32(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_6K) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- a->rt2 = a->rt + 1;
- return op_ldrex(s, a, MO_64, false);
- }
- static bool trans_LDREXD_t32(DisasContext *s, arg_LDREX *a)
- {
- return op_ldrex(s, a, MO_64, false);
- }
- static bool trans_LDREXB(DisasContext *s, arg_LDREX *a)
- {
- if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
- return false;
- }
- return op_ldrex(s, a, MO_8, false);
- }
- static bool trans_LDREXH(DisasContext *s, arg_LDREX *a)
- {
- if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
- return false;
- }
- return op_ldrex(s, a, MO_16, false);
- }
- static bool trans_LDAEX(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_ldrex(s, a, MO_32, true);
- }
- static bool trans_LDAEXD_a32(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rt & 1) {
- unallocated_encoding(s);
- return true;
- }
- a->rt2 = a->rt + 1;
- return op_ldrex(s, a, MO_64, true);
- }
- static bool trans_LDAEXD_t32(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_ldrex(s, a, MO_64, true);
- }
- static bool trans_LDAEXB(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_ldrex(s, a, MO_8, true);
- }
- static bool trans_LDAEXH(DisasContext *s, arg_LDREX *a)
- {
- if (!ENABLE_ARCH_8) {
- return false;
- }
- return op_ldrex(s, a, MO_16, true);
- }
- static bool op_lda(DisasContext *s, arg_LDA *a, MemOp mop)
- {
- TCGv_i32 addr, tmp;
- if (!ENABLE_ARCH_8) {
- return false;
- }
- /* We UNDEF for these UNPREDICTABLE cases. */
- if (a->rn == 15 || a->rt == 15) {
- unallocated_encoding(s);
- return true;
- }
- addr = load_reg(s, a->rn);
- tmp = tcg_temp_new_i32();
- gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s), mop | s->be_data);
- disas_set_da_iss(s, mop, a->rt | ISSIsAcqRel);
- tcg_temp_free_i32(addr);
- store_reg(s, a->rt, tmp);
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
- return true;
- }
- static bool trans_LDA(DisasContext *s, arg_LDA *a)
- {
- return op_lda(s, a, MO_UL);
- }
- static bool trans_LDAB(DisasContext *s, arg_LDA *a)
- {
- return op_lda(s, a, MO_UB);
- }
- static bool trans_LDAH(DisasContext *s, arg_LDA *a)
- {
- return op_lda(s, a, MO_UW);
- }
- /*
- * Media instructions
- */
- static bool trans_USADA8(DisasContext *s, arg_USADA8 *a)
- {
- TCGv_i32 t1, t2;
- if (!ENABLE_ARCH_6) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- gen_helper_usad8(t1, t1, t2);
- tcg_temp_free_i32(t2);
- if (a->ra != 15) {
- t2 = load_reg(s, a->ra);
- tcg_gen_add_i32(t1, t1, t2);
- tcg_temp_free_i32(t2);
- }
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool op_bfx(DisasContext *s, arg_UBFX *a, bool u)
- {
- TCGv_i32 tmp;
- int width = a->widthm1 + 1;
- int shift = a->lsb;
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- if (shift + width > 32) {
- /* UNPREDICTABLE; we choose to UNDEF */
- unallocated_encoding(s);
- return true;
- }
- tmp = load_reg(s, a->rn);
- if (u) {
- tcg_gen_extract_i32(tmp, tmp, shift, width);
- } else {
- tcg_gen_sextract_i32(tmp, tmp, shift, width);
- }
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_SBFX(DisasContext *s, arg_SBFX *a)
- {
- return op_bfx(s, a, false);
- }
- static bool trans_UBFX(DisasContext *s, arg_UBFX *a)
- {
- return op_bfx(s, a, true);
- }
- static bool trans_BFCI(DisasContext *s, arg_BFCI *a)
- {
- TCGv_i32 tmp;
- int msb = a->msb, lsb = a->lsb;
- int width;
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- if (msb < lsb) {
- /* UNPREDICTABLE; we choose to UNDEF */
- unallocated_encoding(s);
- return true;
- }
- width = msb + 1 - lsb;
- if (a->rn == 15) {
- /* BFC */
- tmp = tcg_const_i32(0);
- } else {
- /* BFI */
- tmp = load_reg(s, a->rn);
- }
- if (width != 32) {
- TCGv_i32 tmp2 = load_reg(s, a->rd);
- tcg_gen_deposit_i32(tmp, tmp2, tmp, lsb, width);
- tcg_temp_free_i32(tmp2);
- }
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_UDF(DisasContext *s, arg_UDF *a)
- {
- unallocated_encoding(s);
- return true;
- }
- /*
- * Parallel addition and subtraction
- */
- static bool op_par_addsub(DisasContext *s, arg_rrr *a,
- void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32))
- {
- TCGv_i32 t0, t1;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- t0 = load_reg(s, a->rn);
- t1 = load_reg(s, a->rm);
- gen(t0, t0, t1);
- tcg_temp_free_i32(t1);
- store_reg(s, a->rd, t0);
- return true;
- }
- static bool op_par_addsub_ge(DisasContext *s, arg_rrr *a,
- void (*gen)(TCGv_i32, TCGv_i32,
- TCGv_i32, TCGv_ptr))
- {
- TCGv_i32 t0, t1;
- TCGv_ptr ge;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- t0 = load_reg(s, a->rn);
- t1 = load_reg(s, a->rm);
- ge = tcg_temp_new_ptr();
- tcg_gen_addi_ptr(ge, cpu_env, offsetof(CPUARMState, GE));
- gen(t0, t0, t1, ge);
- tcg_temp_free_ptr(ge);
- tcg_temp_free_i32(t1);
- store_reg(s, a->rd, t0);
- return true;
- }
- #define DO_PAR_ADDSUB(NAME, helper) \
- static bool trans_##NAME(DisasContext *s, arg_rrr *a) \
- { \
- return op_par_addsub(s, a, helper); \
- }
- #define DO_PAR_ADDSUB_GE(NAME, helper) \
- static bool trans_##NAME(DisasContext *s, arg_rrr *a) \
- { \
- return op_par_addsub_ge(s, a, helper); \
- }
- DO_PAR_ADDSUB_GE(SADD16, gen_helper_sadd16)
- DO_PAR_ADDSUB_GE(SASX, gen_helper_saddsubx)
- DO_PAR_ADDSUB_GE(SSAX, gen_helper_ssubaddx)
- DO_PAR_ADDSUB_GE(SSUB16, gen_helper_ssub16)
- DO_PAR_ADDSUB_GE(SADD8, gen_helper_sadd8)
- DO_PAR_ADDSUB_GE(SSUB8, gen_helper_ssub8)
- DO_PAR_ADDSUB_GE(UADD16, gen_helper_uadd16)
- DO_PAR_ADDSUB_GE(UASX, gen_helper_uaddsubx)
- DO_PAR_ADDSUB_GE(USAX, gen_helper_usubaddx)
- DO_PAR_ADDSUB_GE(USUB16, gen_helper_usub16)
- DO_PAR_ADDSUB_GE(UADD8, gen_helper_uadd8)
- DO_PAR_ADDSUB_GE(USUB8, gen_helper_usub8)
- DO_PAR_ADDSUB(QADD16, gen_helper_qadd16)
- DO_PAR_ADDSUB(QASX, gen_helper_qaddsubx)
- DO_PAR_ADDSUB(QSAX, gen_helper_qsubaddx)
- DO_PAR_ADDSUB(QSUB16, gen_helper_qsub16)
- DO_PAR_ADDSUB(QADD8, gen_helper_qadd8)
- DO_PAR_ADDSUB(QSUB8, gen_helper_qsub8)
- DO_PAR_ADDSUB(UQADD16, gen_helper_uqadd16)
- DO_PAR_ADDSUB(UQASX, gen_helper_uqaddsubx)
- DO_PAR_ADDSUB(UQSAX, gen_helper_uqsubaddx)
- DO_PAR_ADDSUB(UQSUB16, gen_helper_uqsub16)
- DO_PAR_ADDSUB(UQADD8, gen_helper_uqadd8)
- DO_PAR_ADDSUB(UQSUB8, gen_helper_uqsub8)
- DO_PAR_ADDSUB(SHADD16, gen_helper_shadd16)
- DO_PAR_ADDSUB(SHASX, gen_helper_shaddsubx)
- DO_PAR_ADDSUB(SHSAX, gen_helper_shsubaddx)
- DO_PAR_ADDSUB(SHSUB16, gen_helper_shsub16)
- DO_PAR_ADDSUB(SHADD8, gen_helper_shadd8)
- DO_PAR_ADDSUB(SHSUB8, gen_helper_shsub8)
- DO_PAR_ADDSUB(UHADD16, gen_helper_uhadd16)
- DO_PAR_ADDSUB(UHASX, gen_helper_uhaddsubx)
- DO_PAR_ADDSUB(UHSAX, gen_helper_uhsubaddx)
- DO_PAR_ADDSUB(UHSUB16, gen_helper_uhsub16)
- DO_PAR_ADDSUB(UHADD8, gen_helper_uhadd8)
- DO_PAR_ADDSUB(UHSUB8, gen_helper_uhsub8)
- #undef DO_PAR_ADDSUB
- #undef DO_PAR_ADDSUB_GE
- /*
- * Packing, unpacking, saturation, and reversal
- */
- static bool trans_PKH(DisasContext *s, arg_PKH *a)
- {
- TCGv_i32 tn, tm;
- int shift = a->imm;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- tn = load_reg(s, a->rn);
- tm = load_reg(s, a->rm);
- if (a->tb) {
- /* PKHTB */
- if (shift == 0) {
- shift = 31;
- }
- tcg_gen_sari_i32(tm, tm, shift);
- tcg_gen_deposit_i32(tn, tn, tm, 0, 16);
- } else {
- /* PKHBT */
- tcg_gen_shli_i32(tm, tm, shift);
- tcg_gen_deposit_i32(tn, tm, tn, 0, 16);
- }
- tcg_temp_free_i32(tm);
- store_reg(s, a->rd, tn);
- return true;
- }
- static bool op_sat(DisasContext *s, arg_sat *a,
- void (*gen)(TCGv_i32, TCGv_env, TCGv_i32, TCGv_i32))
- {
- TCGv_i32 tmp, satimm;
- int shift = a->imm;
- if (!ENABLE_ARCH_6) {
- return false;
- }
- tmp = load_reg(s, a->rn);
- if (a->sh) {
- tcg_gen_sari_i32(tmp, tmp, shift ? shift : 31);
- } else {
- tcg_gen_shli_i32(tmp, tmp, shift);
- }
- satimm = tcg_const_i32(a->satimm);
- gen(tmp, cpu_env, tmp, satimm);
- tcg_temp_free_i32(satimm);
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_SSAT(DisasContext *s, arg_sat *a)
- {
- return op_sat(s, a, gen_helper_ssat);
- }
- static bool trans_USAT(DisasContext *s, arg_sat *a)
- {
- return op_sat(s, a, gen_helper_usat);
- }
- static bool trans_SSAT16(DisasContext *s, arg_sat *a)
- {
- if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
- return false;
- }
- return op_sat(s, a, gen_helper_ssat16);
- }
- static bool trans_USAT16(DisasContext *s, arg_sat *a)
- {
- if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
- return false;
- }
- return op_sat(s, a, gen_helper_usat16);
- }
- static bool op_xta(DisasContext *s, arg_rrr_rot *a,
- void (*gen_extract)(TCGv_i32, TCGv_i32),
- void (*gen_add)(TCGv_i32, TCGv_i32, TCGv_i32))
- {
- TCGv_i32 tmp;
- if (!ENABLE_ARCH_6) {
- return false;
- }
- tmp = load_reg(s, a->rm);
- /*
- * TODO: In many cases we could do a shift instead of a rotate.
- * Combined with a simple extend, that becomes an extract.
- */
- tcg_gen_rotri_i32(tmp, tmp, a->rot * 8);
- gen_extract(tmp, tmp);
- if (a->rn != 15) {
- TCGv_i32 tmp2 = load_reg(s, a->rn);
- gen_add(tmp, tmp, tmp2);
- tcg_temp_free_i32(tmp2);
- }
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_SXTAB(DisasContext *s, arg_rrr_rot *a)
- {
- return op_xta(s, a, tcg_gen_ext8s_i32, tcg_gen_add_i32);
- }
- static bool trans_SXTAH(DisasContext *s, arg_rrr_rot *a)
- {
- return op_xta(s, a, tcg_gen_ext16s_i32, tcg_gen_add_i32);
- }
- static bool trans_SXTAB16(DisasContext *s, arg_rrr_rot *a)
- {
- if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
- return false;
- }
- return op_xta(s, a, gen_helper_sxtb16, gen_add16);
- }
- static bool trans_UXTAB(DisasContext *s, arg_rrr_rot *a)
- {
- return op_xta(s, a, tcg_gen_ext8u_i32, tcg_gen_add_i32);
- }
- static bool trans_UXTAH(DisasContext *s, arg_rrr_rot *a)
- {
- return op_xta(s, a, tcg_gen_ext16u_i32, tcg_gen_add_i32);
- }
- static bool trans_UXTAB16(DisasContext *s, arg_rrr_rot *a)
- {
- if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
- return false;
- }
- return op_xta(s, a, gen_helper_uxtb16, gen_add16);
- }
- static bool trans_SEL(DisasContext *s, arg_rrr *a)
- {
- TCGv_i32 t1, t2, t3;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- t3 = tcg_temp_new_i32();
- tcg_gen_ld_i32(t3, cpu_env, offsetof(CPUARMState, GE));
- gen_helper_sel_flags(t1, t3, t1, t2);
- tcg_temp_free_i32(t3);
- tcg_temp_free_i32(t2);
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool op_rr(DisasContext *s, arg_rr *a,
- void (*gen)(TCGv_i32, TCGv_i32))
- {
- TCGv_i32 tmp;
- tmp = load_reg(s, a->rm);
- gen(tmp, tmp);
- store_reg(s, a->rd, tmp);
- return true;
- }
- static bool trans_REV(DisasContext *s, arg_rr *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- return op_rr(s, a, tcg_gen_bswap32_i32);
- }
- static bool trans_REV16(DisasContext *s, arg_rr *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- return op_rr(s, a, gen_rev16);
- }
- static bool trans_REVSH(DisasContext *s, arg_rr *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- return op_rr(s, a, gen_revsh);
- }
- static bool trans_RBIT(DisasContext *s, arg_rr *a)
- {
- if (!ENABLE_ARCH_6T2) {
- return false;
- }
- return op_rr(s, a, gen_helper_rbit);
- }
- /*
- * Signed multiply, signed and unsigned divide
- */
- static bool op_smlad(DisasContext *s, arg_rrrr *a, bool m_swap, bool sub)
- {
- TCGv_i32 t1, t2;
- if (!ENABLE_ARCH_6) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- if (m_swap) {
- gen_swap_half(t2, t2);
- }
- gen_smul_dual(t1, t2);
- if (sub) {
- /* This subtraction cannot overflow. */
- tcg_gen_sub_i32(t1, t1, t2);
- } else {
- /*
- * This addition cannot overflow 32 bits; however it may
- * overflow considered as a signed operation, in which case
- * we must set the Q flag.
- */
- gen_helper_add_setq(t1, cpu_env, t1, t2);
- }
- tcg_temp_free_i32(t2);
- if (a->ra != 15) {
- t2 = load_reg(s, a->ra);
- gen_helper_add_setq(t1, cpu_env, t1, t2);
- tcg_temp_free_i32(t2);
- }
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool trans_SMLAD(DisasContext *s, arg_rrrr *a)
- {
- return op_smlad(s, a, false, false);
- }
- static bool trans_SMLADX(DisasContext *s, arg_rrrr *a)
- {
- return op_smlad(s, a, true, false);
- }
- static bool trans_SMLSD(DisasContext *s, arg_rrrr *a)
- {
- return op_smlad(s, a, false, true);
- }
- static bool trans_SMLSDX(DisasContext *s, arg_rrrr *a)
- {
- return op_smlad(s, a, true, true);
- }
- static bool op_smlald(DisasContext *s, arg_rrrr *a, bool m_swap, bool sub)
- {
- TCGv_i32 t1, t2;
- TCGv_i64 l1, l2;
- if (!ENABLE_ARCH_6) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- if (m_swap) {
- gen_swap_half(t2, t2);
- }
- gen_smul_dual(t1, t2);
- l1 = tcg_temp_new_i64();
- l2 = tcg_temp_new_i64();
- tcg_gen_ext_i32_i64(l1, t1);
- tcg_gen_ext_i32_i64(l2, t2);
- tcg_temp_free_i32(t1);
- tcg_temp_free_i32(t2);
- if (sub) {
- tcg_gen_sub_i64(l1, l1, l2);
- } else {
- tcg_gen_add_i64(l1, l1, l2);
- }
- tcg_temp_free_i64(l2);
- gen_addq(s, l1, a->ra, a->rd);
- gen_storeq_reg(s, a->ra, a->rd, l1);
- tcg_temp_free_i64(l1);
- return true;
- }
- static bool trans_SMLALD(DisasContext *s, arg_rrrr *a)
- {
- return op_smlald(s, a, false, false);
- }
- static bool trans_SMLALDX(DisasContext *s, arg_rrrr *a)
- {
- return op_smlald(s, a, true, false);
- }
- static bool trans_SMLSLD(DisasContext *s, arg_rrrr *a)
- {
- return op_smlald(s, a, false, true);
- }
- static bool trans_SMLSLDX(DisasContext *s, arg_rrrr *a)
- {
- return op_smlald(s, a, true, true);
- }
- static bool op_smmla(DisasContext *s, arg_rrrr *a, bool round, bool sub)
- {
- TCGv_i32 t1, t2;
- if (s->thumb
- ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
- : !ENABLE_ARCH_6) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- tcg_gen_muls2_i32(t2, t1, t1, t2);
- if (a->ra != 15) {
- TCGv_i32 t3 = load_reg(s, a->ra);
- if (sub) {
- /*
- * For SMMLS, we need a 64-bit subtract. Borrow caused by
- * a non-zero multiplicand lowpart, and the correct result
- * lowpart for rounding.
- */
- TCGv_i32 zero = tcg_const_i32(0);
- tcg_gen_sub2_i32(t2, t1, zero, t3, t2, t1);
- tcg_temp_free_i32(zero);
- } else {
- tcg_gen_add_i32(t1, t1, t3);
- }
- tcg_temp_free_i32(t3);
- }
- if (round) {
- /*
- * Adding 0x80000000 to the 64-bit quantity means that we have
- * carry in to the high word when the low word has the msb set.
- */
- tcg_gen_shri_i32(t2, t2, 31);
- tcg_gen_add_i32(t1, t1, t2);
- }
- tcg_temp_free_i32(t2);
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool trans_SMMLA(DisasContext *s, arg_rrrr *a)
- {
- return op_smmla(s, a, false, false);
- }
- static bool trans_SMMLAR(DisasContext *s, arg_rrrr *a)
- {
- return op_smmla(s, a, true, false);
- }
- static bool trans_SMMLS(DisasContext *s, arg_rrrr *a)
- {
- return op_smmla(s, a, false, true);
- }
- static bool trans_SMMLSR(DisasContext *s, arg_rrrr *a)
- {
- return op_smmla(s, a, true, true);
- }
- static bool op_div(DisasContext *s, arg_rrr *a, bool u)
- {
- TCGv_i32 t1, t2;
- if (s->thumb
- ? !dc_isar_feature(aa32_thumb_div, s)
- : !dc_isar_feature(aa32_arm_div, s)) {
- return false;
- }
- t1 = load_reg(s, a->rn);
- t2 = load_reg(s, a->rm);
- if (u) {
- gen_helper_udiv(t1, t1, t2);
- } else {
- gen_helper_sdiv(t1, t1, t2);
- }
- tcg_temp_free_i32(t2);
- store_reg(s, a->rd, t1);
- return true;
- }
- static bool trans_SDIV(DisasContext *s, arg_rrr *a)
- {
- return op_div(s, a, false);
- }
- static bool trans_UDIV(DisasContext *s, arg_rrr *a)
- {
- return op_div(s, a, true);
- }
- /*
- * Block data transfer
- */
- static TCGv_i32 op_addr_block_pre(DisasContext *s, arg_ldst_block *a, int n)
- {
- TCGv_i32 addr = load_reg(s, a->rn);
- if (a->b) {
- if (a->i) {
- /* pre increment */
- tcg_gen_addi_i32(addr, addr, 4);
- } else {
- /* pre decrement */
- tcg_gen_addi_i32(addr, addr, -(n * 4));
- }
- } else if (!a->i && n != 1) {
- /* post decrement */
- tcg_gen_addi_i32(addr, addr, -((n - 1) * 4));
- }
- if (s->v8m_stackcheck && a->rn == 13 && a->w) {
- /*
- * If the writeback is incrementing SP rather than
- * decrementing it, and the initial SP is below the
- * stack limit but the final written-back SP would
- * be above, then then we must not perform any memory
- * accesses, but it is IMPDEF whether we generate
- * an exception. We choose to do so in this case.
- * At this point 'addr' is the lowest address, so
- * either the original SP (if incrementing) or our
- * final SP (if decrementing), so that's what we check.
- */
- gen_helper_v8m_stackcheck(cpu_env, addr);
- }
- return addr;
- }
- static void op_addr_block_post(DisasContext *s, arg_ldst_block *a,
- TCGv_i32 addr, int n)
- {
- if (a->w) {
- /* write back */
- if (!a->b) {
- if (a->i) {
- /* post increment */
- tcg_gen_addi_i32(addr, addr, 4);
- } else {
- /* post decrement */
- tcg_gen_addi_i32(addr, addr, -(n * 4));
- }
- } else if (!a->i && n != 1) {
- /* pre decrement */
- tcg_gen_addi_i32(addr, addr, -((n - 1) * 4));
- }
- store_reg(s, a->rn, addr);
- } else {
- tcg_temp_free_i32(addr);
- }
- }
- static bool op_stm(DisasContext *s, arg_ldst_block *a, int min_n)
- {
- int i, j, n, list, mem_idx;
- bool user = a->u;
- TCGv_i32 addr, tmp, tmp2;
- if (user) {
- /* STM (user) */
- if (IS_USER(s)) {
- /* Only usable in supervisor mode. */
- unallocated_encoding(s);
- return true;
- }
- }
- list = a->list;
- n = ctpop16(list);
- if (n < min_n || a->rn == 15) {
- unallocated_encoding(s);
- return true;
- }
- addr = op_addr_block_pre(s, a, n);
- mem_idx = get_mem_index(s);
- for (i = j = 0; i < 16; i++) {
- if (!(list & (1 << i))) {
- continue;
- }
- if (user && i != 15) {
- tmp = tcg_temp_new_i32();
- tmp2 = tcg_const_i32(i);
- gen_helper_get_user_reg(tmp, cpu_env, tmp2);
- tcg_temp_free_i32(tmp2);
- } else {
- tmp = load_reg(s, i);
- }
- gen_aa32_st32(s, tmp, addr, mem_idx);
- tcg_temp_free_i32(tmp);
- /* No need to add after the last transfer. */
- if (++j != n) {
- tcg_gen_addi_i32(addr, addr, 4);
- }
- }
- op_addr_block_post(s, a, addr, n);
- return true;
- }
- static bool trans_STM(DisasContext *s, arg_ldst_block *a)
- {
- /* BitCount(list) < 1 is UNPREDICTABLE */
- return op_stm(s, a, 1);
- }
- static bool trans_STM_t32(DisasContext *s, arg_ldst_block *a)
- {
- /* Writeback register in register list is UNPREDICTABLE for T32. */
- if (a->w && (a->list & (1 << a->rn))) {
- unallocated_encoding(s);
- return true;
- }
- /* BitCount(list) < 2 is UNPREDICTABLE */
- return op_stm(s, a, 2);
- }
- static bool do_ldm(DisasContext *s, arg_ldst_block *a, int min_n)
- {
- int i, j, n, list, mem_idx;
- bool loaded_base;
- bool user = a->u;
- bool exc_return = false;
- TCGv_i32 addr, tmp, tmp2, loaded_var;
- if (user) {
- /* LDM (user), LDM (exception return) */
- if (IS_USER(s)) {
- /* Only usable in supervisor mode. */
- unallocated_encoding(s);
- return true;
- }
- if (extract32(a->list, 15, 1)) {
- exc_return = true;
- user = false;
- } else {
- /* LDM (user) does not allow writeback. */
- if (a->w) {
- unallocated_encoding(s);
- return true;
- }
- }
- }
- list = a->list;
- n = ctpop16(list);
- if (n < min_n || a->rn == 15) {
- unallocated_encoding(s);
- return true;
- }
- addr = op_addr_block_pre(s, a, n);
- mem_idx = get_mem_index(s);
- loaded_base = false;
- loaded_var = NULL;
- for (i = j = 0; i < 16; i++) {
- if (!(list & (1 << i))) {
- continue;
- }
- tmp = tcg_temp_new_i32();
- gen_aa32_ld32u(s, tmp, addr, mem_idx);
- if (user) {
- tmp2 = tcg_const_i32(i);
- gen_helper_set_user_reg(cpu_env, tmp2, tmp);
- tcg_temp_free_i32(tmp2);
- tcg_temp_free_i32(tmp);
- } else if (i == a->rn) {
- loaded_var = tmp;
- loaded_base = true;
- } else if (i == 15 && exc_return) {
- store_pc_exc_ret(s, tmp);
- } else {
- store_reg_from_load(s, i, tmp);
- }
- /* No need to add after the last transfer. */
- if (++j != n) {
- tcg_gen_addi_i32(addr, addr, 4);
- }
- }
- op_addr_block_post(s, a, addr, n);
- if (loaded_base) {
- /* Note that we reject base == pc above. */
- store_reg(s, a->rn, loaded_var);
- }
- if (exc_return) {
- /* Restore CPSR from SPSR. */
- tmp = load_cpu_field(spsr);
- if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
- gen_io_start();
- }
- gen_helper_cpsr_write_eret(cpu_env, tmp);
- tcg_temp_free_i32(tmp);
- /* Must exit loop to check un-masked IRQs */
- s->base.is_jmp = DISAS_EXIT;
- }
- return true;
- }
- static bool trans_LDM_a32(DisasContext *s, arg_ldst_block *a)
- {
- /*
- * Writeback register in register list is UNPREDICTABLE
- * for ArchVersion() >= 7. Prior to v7, A32 would write
- * an UNKNOWN value to the base register.
- */
- if (ENABLE_ARCH_7 && a->w && (a->list & (1 << a->rn))) {
- unallocated_encoding(s);
- return true;
- }
- /* BitCount(list) < 1 is UNPREDICTABLE */
- return do_ldm(s, a, 1);
- }
- static bool trans_LDM_t32(DisasContext *s, arg_ldst_block *a)
- {
- /* Writeback register in register list is UNPREDICTABLE for T32. */
- if (a->w && (a->list & (1 << a->rn))) {
- unallocated_encoding(s);
- return true;
- }
- /* BitCount(list) < 2 is UNPREDICTABLE */
- return do_ldm(s, a, 2);
- }
- static bool trans_LDM_t16(DisasContext *s, arg_ldst_block *a)
- {
- /* Writeback is conditional on the base register not being loaded. */
- a->w = !(a->list & (1 << a->rn));
- /* BitCount(list) < 1 is UNPREDICTABLE */
- return do_ldm(s, a, 1);
- }
- /*
- * Branch, branch with link
- */
- static bool trans_B(DisasContext *s, arg_i *a)
- {
- gen_jmp(s, read_pc(s) + a->imm);
- return true;
- }
- static bool trans_B_cond_thumb(DisasContext *s, arg_ci *a)
- {
- /* This has cond from encoding, required to be outside IT block. */
- if (a->cond >= 0xe) {
- return false;
- }
- if (s->condexec_mask) {
- unallocated_encoding(s);
- return true;
- }
- arm_skip_unless(s, a->cond);
- gen_jmp(s, read_pc(s) + a->imm);
- return true;
- }
- static bool trans_BL(DisasContext *s, arg_i *a)
- {
- tcg_gen_movi_i32(cpu_R[14], s->base.pc_next | s->thumb);
- gen_jmp(s, read_pc(s) + a->imm);
- return true;
- }
- static bool trans_BLX_i(DisasContext *s, arg_BLX_i *a)
- {
- TCGv_i32 tmp;
- /* For A32, ARCH(5) is checked near the start of the uncond block. */
- if (s->thumb && (a->imm & 2)) {
- return false;
- }
- tcg_gen_movi_i32(cpu_R[14], s->base.pc_next | s->thumb);
- tmp = tcg_const_i32(!s->thumb);
- store_cpu_field(tmp, thumb);
- gen_jmp(s, (read_pc(s) & ~3) + a->imm);
- return true;
- }
- static bool trans_BL_BLX_prefix(DisasContext *s, arg_BL_BLX_prefix *a)
- {
- assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
- tcg_gen_movi_i32(cpu_R[14], read_pc(s) + (a->imm << 12));
- return true;
- }
- static bool trans_BL_suffix(DisasContext *s, arg_BL_suffix *a)
- {
- TCGv_i32 tmp = tcg_temp_new_i32();
- assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
- tcg_gen_addi_i32(tmp, cpu_R[14], (a->imm << 1) | 1);
- tcg_gen_movi_i32(cpu_R[14], s->base.pc_next | 1);
- gen_bx(s, tmp);
- return true;
- }
- static bool trans_BLX_suffix(DisasContext *s, arg_BLX_suffix *a)
- {
- TCGv_i32 tmp;
- assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
- if (!ENABLE_ARCH_5) {
- return false;
- }
- tmp = tcg_temp_new_i32();
- tcg_gen_addi_i32(tmp, cpu_R[14], a->imm << 1);
- tcg_gen_andi_i32(tmp, tmp, 0xfffffffc);
- tcg_gen_movi_i32(cpu_R[14], s->base.pc_next | 1);
- gen_bx(s, tmp);
- return true;
- }
- static bool op_tbranch(DisasContext *s, arg_tbranch *a, bool half)
- {
- TCGv_i32 addr, tmp;
- tmp = load_reg(s, a->rm);
- if (half) {
- tcg_gen_add_i32(tmp, tmp, tmp);
- }
- addr = load_reg(s, a->rn);
- tcg_gen_add_i32(addr, addr, tmp);
- gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
- half ? MO_UW | s->be_data : MO_UB);
- tcg_temp_free_i32(addr);
- tcg_gen_add_i32(tmp, tmp, tmp);
- tcg_gen_addi_i32(tmp, tmp, read_pc(s));
- store_reg(s, 15, tmp);
- return true;
- }
- static bool trans_TBB(DisasContext *s, arg_tbranch *a)
- {
- return op_tbranch(s, a, false);
- }
- static bool trans_TBH(DisasContext *s, arg_tbranch *a)
- {
- return op_tbranch(s, a, true);
- }
- static bool trans_CBZ(DisasContext *s, arg_CBZ *a)
- {
- TCGv_i32 tmp = load_reg(s, a->rn);
- arm_gen_condlabel(s);
- tcg_gen_brcondi_i32(a->nz ? TCG_COND_EQ : TCG_COND_NE,
- tmp, 0, s->condlabel);
- tcg_temp_free_i32(tmp);
- gen_jmp(s, read_pc(s) + a->imm);
- return true;
- }
- /*
- * Supervisor call - both T32 & A32 come here so we need to check
- * which mode we are in when checking for semihosting.
- */
- static bool trans_SVC(DisasContext *s, arg_SVC *a)
- {
- const uint32_t semihost_imm = s->thumb ? 0xab : 0x123456;
- if (!arm_dc_feature(s, ARM_FEATURE_M) && semihosting_enabled() &&
- #ifndef CONFIG_USER_ONLY
- !IS_USER(s) &&
- #endif
- (a->imm == semihost_imm)) {
- gen_exception_internal_insn(s, s->pc_curr, EXCP_SEMIHOST);
- } else {
- gen_set_pc_im(s, s->base.pc_next);
- s->svc_imm = a->imm;
- s->base.is_jmp = DISAS_SWI;
- }
- return true;
- }
- /*
- * Unconditional system instructions
- */
- static bool trans_RFE(DisasContext *s, arg_RFE *a)
- {
- static const int8_t pre_offset[4] = {
- /* DA */ -4, /* IA */ 0, /* DB */ -8, /* IB */ 4
- };
- static const int8_t post_offset[4] = {
- /* DA */ -8, /* IA */ 4, /* DB */ -4, /* IB */ 0
- };
- TCGv_i32 addr, t1, t2;
- if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (IS_USER(s)) {
- unallocated_encoding(s);
- return true;
- }
- addr = load_reg(s, a->rn);
- tcg_gen_addi_i32(addr, addr, pre_offset[a->pu]);
- /* Load PC into tmp and CPSR into tmp2. */
- t1 = tcg_temp_new_i32();
- gen_aa32_ld32u(s, t1, addr, get_mem_index(s));
- tcg_gen_addi_i32(addr, addr, 4);
- t2 = tcg_temp_new_i32();
- gen_aa32_ld32u(s, t2, addr, get_mem_index(s));
- if (a->w) {
- /* Base writeback. */
- tcg_gen_addi_i32(addr, addr, post_offset[a->pu]);
- store_reg(s, a->rn, addr);
- } else {
- tcg_temp_free_i32(addr);
- }
- gen_rfe(s, t1, t2);
- return true;
- }
- static bool trans_SRS(DisasContext *s, arg_SRS *a)
- {
- if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- gen_srs(s, a->mode, a->pu, a->w);
- return true;
- }
- static bool trans_CPS(DisasContext *s, arg_CPS *a)
- {
- uint32_t mask, val;
- if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (IS_USER(s)) {
- /* Implemented as NOP in user mode. */
- return true;
- }
- /* TODO: There are quite a lot of UNPREDICTABLE argument combinations. */
- mask = val = 0;
- if (a->imod & 2) {
- if (a->A) {
- mask |= CPSR_A;
- }
- if (a->I) {
- mask |= CPSR_I;
- }
- if (a->F) {
- mask |= CPSR_F;
- }
- if (a->imod & 1) {
- val |= mask;
- }
- }
- if (a->M) {
- mask |= CPSR_M;
- val |= a->mode;
- }
- if (mask) {
- gen_set_psr_im(s, mask, 0, val);
- }
- return true;
- }
- static bool trans_CPS_v7m(DisasContext *s, arg_CPS_v7m *a)
- {
- TCGv_i32 tmp, addr, el;
- if (!arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- if (IS_USER(s)) {
- /* Implemented as NOP in user mode. */
- return true;
- }
- tmp = tcg_const_i32(a->im);
- /* FAULTMASK */
- if (a->F) {
- addr = tcg_const_i32(19);
- gen_helper_v7m_msr(cpu_env, addr, tmp);
- tcg_temp_free_i32(addr);
- }
- /* PRIMASK */
- if (a->I) {
- addr = tcg_const_i32(16);
- gen_helper_v7m_msr(cpu_env, addr, tmp);
- tcg_temp_free_i32(addr);
- }
- el = tcg_const_i32(s->current_el);
- gen_helper_rebuild_hflags_m32(cpu_env, el);
- tcg_temp_free_i32(el);
- tcg_temp_free_i32(tmp);
- gen_lookup_tb(s);
- return true;
- }
- /*
- * Clear-Exclusive, Barriers
- */
- static bool trans_CLREX(DisasContext *s, arg_CLREX *a)
- {
- if (s->thumb
- ? !ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)
- : !ENABLE_ARCH_6K) {
- return false;
- }
- gen_clrex(s);
- return true;
- }
- static bool trans_DSB(DisasContext *s, arg_DSB *a)
- {
- if (!ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
- return true;
- }
- static bool trans_DMB(DisasContext *s, arg_DMB *a)
- {
- return trans_DSB(s, NULL);
- }
- static bool trans_ISB(DisasContext *s, arg_ISB *a)
- {
- if (!ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)) {
- return false;
- }
- /*
- * We need to break the TB after this insn to execute
- * self-modifying code correctly and also to take
- * any pending interrupts immediately.
- */
- gen_goto_tb(s, 0, s->base.pc_next);
- return true;
- }
- static bool trans_SB(DisasContext *s, arg_SB *a)
- {
- if (!dc_isar_feature(aa32_sb, s)) {
- return false;
- }
- /*
- * TODO: There is no speculation barrier opcode
- * for TCG; MB and end the TB instead.
- */
- tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
- gen_goto_tb(s, 0, s->base.pc_next);
- return true;
- }
- static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
- {
- if (!ENABLE_ARCH_6) {
- return false;
- }
- if (a->E != (s->be_data == MO_BE)) {
- gen_helper_setend(cpu_env);
- s->base.is_jmp = DISAS_UPDATE_EXIT;
- }
- return true;
- }
- /*
- * Preload instructions
- * All are nops, contingent on the appropriate arch level.
- */
- static bool trans_PLD(DisasContext *s, arg_PLD *a)
- {
- return ENABLE_ARCH_5TE;
- }
- static bool trans_PLDW(DisasContext *s, arg_PLD *a)
- {
- return arm_dc_feature(s, ARM_FEATURE_V7MP);
- }
- static bool trans_PLI(DisasContext *s, arg_PLD *a)
- {
- return ENABLE_ARCH_7;
- }
- /*
- * If-then
- */
- static bool trans_IT(DisasContext *s, arg_IT *a)
- {
- int cond_mask = a->cond_mask;
- /*
- * No actual code generated for this insn, just setup state.
- *
- * Combinations of firstcond and mask which set up an 0b1111
- * condition are UNPREDICTABLE; we take the CONSTRAINED
- * UNPREDICTABLE choice to treat 0b1111 the same as 0b1110,
- * i.e. both meaning "execute always".
- */
- s->condexec_cond = (cond_mask >> 4) & 0xe;
- s->condexec_mask = cond_mask & 0x1f;
- return true;
- }
- /*
- * Legacy decoder.
- */
- static void disas_arm_insn(DisasContext *s, unsigned int insn)
- {
- unsigned int cond = insn >> 28;
- /* M variants do not implement ARM mode; this must raise the INVSTATE
- * UsageFault exception.
- */
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- gen_exception_insn(s, s->pc_curr, EXCP_INVSTATE, syn_uncategorized(),
- default_exception_el(s));
- return;
- }
- if (cond == 0xf) {
- /* In ARMv3 and v4 the NV condition is UNPREDICTABLE; we
- * choose to UNDEF. In ARMv5 and above the space is used
- * for miscellaneous unconditional instructions.
- */
- ARCH(5);
- /* Unconditional instructions. */
- /* TODO: Perhaps merge these into one decodetree output file. */
- if (disas_a32_uncond(s, insn) ||
- disas_vfp_uncond(s, insn) ||
- disas_neon_dp(s, insn) ||
- disas_neon_ls(s, insn) ||
- disas_neon_shared(s, insn)) {
- return;
- }
- /* fall back to legacy decoder */
- if ((insn & 0x0e000f00) == 0x0c000100) {
- if (arm_dc_feature(s, ARM_FEATURE_IWMMXT)) {
- /* iWMMXt register transfer. */
- if (extract32(s->c15_cpar, 1, 1)) {
- if (!disas_iwmmxt_insn(s, insn)) {
- return;
- }
- }
- }
- }
- goto illegal_op;
- }
- if (cond != 0xe) {
- /* if not always execute, we generate a conditional jump to
- next instruction */
- arm_skip_unless(s, cond);
- }
- /* TODO: Perhaps merge these into one decodetree output file. */
- if (disas_a32(s, insn) ||
- disas_vfp(s, insn)) {
- return;
- }
- /* fall back to legacy decoder */
- switch ((insn >> 24) & 0xf) {
- case 0xc:
- case 0xd:
- case 0xe:
- if (((insn >> 8) & 0xe) == 10) {
- /* VFP, but failed disas_vfp. */
- goto illegal_op;
- }
- if (disas_coproc_insn(s, insn)) {
- /* Coprocessor. */
- goto illegal_op;
- }
- break;
- default:
- illegal_op:
- unallocated_encoding(s);
- break;
- }
- }
- static bool thumb_insn_is_16bit(DisasContext *s, uint32_t pc, uint32_t insn)
- {
- /*
- * Return true if this is a 16 bit instruction. We must be precise
- * about this (matching the decode).
- */
- if ((insn >> 11) < 0x1d) {
- /* Definitely a 16-bit instruction */
- return true;
- }
- /* Top five bits 0b11101 / 0b11110 / 0b11111 : this is the
- * first half of a 32-bit Thumb insn. Thumb-1 cores might
- * end up actually treating this as two 16-bit insns, though,
- * if it's half of a bl/blx pair that might span a page boundary.
- */
- if (arm_dc_feature(s, ARM_FEATURE_THUMB2) ||
- arm_dc_feature(s, ARM_FEATURE_M)) {
- /* Thumb2 cores (including all M profile ones) always treat
- * 32-bit insns as 32-bit.
- */
- return false;
- }
- if ((insn >> 11) == 0x1e && pc - s->page_start < TARGET_PAGE_SIZE - 3) {
- /* 0b1111_0xxx_xxxx_xxxx : BL/BLX prefix, and the suffix
- * is not on the next page; we merge this into a 32-bit
- * insn.
- */
- return false;
- }
- /* 0b1110_1xxx_xxxx_xxxx : BLX suffix (or UNDEF);
- * 0b1111_1xxx_xxxx_xxxx : BL suffix;
- * 0b1111_0xxx_xxxx_xxxx : BL/BLX prefix on the end of a page
- * -- handle as single 16 bit insn
- */
- return true;
- }
- /* Translate a 32-bit thumb instruction. */
- static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
- {
- /*
- * ARMv6-M supports a limited subset of Thumb2 instructions.
- * Other Thumb1 architectures allow only 32-bit
- * combined BL/BLX prefix and suffix.
- */
- if (arm_dc_feature(s, ARM_FEATURE_M) &&
- !arm_dc_feature(s, ARM_FEATURE_V7)) {
- int i;
- bool found = false;
- static const uint32_t armv6m_insn[] = {0xf3808000 /* msr */,
- 0xf3b08040 /* dsb */,
- 0xf3b08050 /* dmb */,
- 0xf3b08060 /* isb */,
- 0xf3e08000 /* mrs */,
- 0xf000d000 /* bl */};
- static const uint32_t armv6m_mask[] = {0xffe0d000,
- 0xfff0d0f0,
- 0xfff0d0f0,
- 0xfff0d0f0,
- 0xffe0d000,
- 0xf800d000};
- for (i = 0; i < ARRAY_SIZE(armv6m_insn); i++) {
- if ((insn & armv6m_mask[i]) == armv6m_insn[i]) {
- found = true;
- break;
- }
- }
- if (!found) {
- goto illegal_op;
- }
- } else if ((insn & 0xf800e800) != 0xf000e800) {
- ARCH(6T2);
- }
- if ((insn & 0xef000000) == 0xef000000) {
- /*
- * T32 encodings 0b111p_1111_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
- * transform into
- * A32 encodings 0b1111_001p_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
- */
- uint32_t a32_insn = (insn & 0xe2ffffff) |
- ((insn & (1 << 28)) >> 4) | (1 << 28);
- if (disas_neon_dp(s, a32_insn)) {
- return;
- }
- }
- if ((insn & 0xff100000) == 0xf9000000) {
- /*
- * T32 encodings 0b1111_1001_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
- * transform into
- * A32 encodings 0b1111_0100_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
- */
- uint32_t a32_insn = (insn & 0x00ffffff) | 0xf4000000;
- if (disas_neon_ls(s, a32_insn)) {
- return;
- }
- }
- /*
- * TODO: Perhaps merge these into one decodetree output file.
- * Note disas_vfp is written for a32 with cond field in the
- * top nibble. The t32 encoding requires 0xe in the top nibble.
- */
- if (disas_t32(s, insn) ||
- disas_vfp_uncond(s, insn) ||
- disas_neon_shared(s, insn) ||
- ((insn >> 28) == 0xe && disas_vfp(s, insn))) {
- return;
- }
- /* fall back to legacy decoder */
- switch ((insn >> 25) & 0xf) {
- case 0: case 1: case 2: case 3:
- /* 16-bit instructions. Should never happen. */
- abort();
- case 6: case 7: case 14: case 15:
- /* Coprocessor. */
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
- /* 0b111x_11xx_xxxx_xxxx_xxxx_xxxx_xxxx_xxxx */
- if (extract32(insn, 24, 2) == 3) {
- goto illegal_op; /* op0 = 0b11 : unallocated */
- }
- if (((insn >> 8) & 0xe) == 10 &&
- dc_isar_feature(aa32_fpsp_v2, s)) {
- /* FP, and the CPU supports it */
- goto illegal_op;
- } else {
- /* All other insns: NOCP */
- gen_exception_insn(s, s->pc_curr, EXCP_NOCP,
- syn_uncategorized(),
- default_exception_el(s));
- }
- break;
- }
- if (((insn >> 24) & 3) == 3) {
- /* Neon DP, but failed disas_neon_dp() */
- goto illegal_op;
- } else if (((insn >> 8) & 0xe) == 10) {
- /* VFP, but failed disas_vfp. */
- goto illegal_op;
- } else {
- if (insn & (1 << 28))
- goto illegal_op;
- if (disas_coproc_insn(s, insn)) {
- goto illegal_op;
- }
- }
- break;
- case 12:
- goto illegal_op;
- default:
- illegal_op:
- unallocated_encoding(s);
- }
- }
- static void disas_thumb_insn(DisasContext *s, uint32_t insn)
- {
- if (!disas_t16(s, insn)) {
- unallocated_encoding(s);
- }
- }
- static bool insn_crosses_page(CPUARMState *env, DisasContext *s)
- {
- /* Return true if the insn at dc->base.pc_next might cross a page boundary.
- * (False positives are OK, false negatives are not.)
- * We know this is a Thumb insn, and our caller ensures we are
- * only called if dc->base.pc_next is less than 4 bytes from the page
- * boundary, so we cross the page if the first 16 bits indicate
- * that this is a 32 bit insn.
- */
- uint16_t insn = arm_lduw_code(env, s->base.pc_next, s->sctlr_b);
- return !thumb_insn_is_16bit(s, s->base.pc_next, insn);
- }
- static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- CPUARMState *env = cs->env_ptr;
- ARMCPU *cpu = env_archcpu(env);
- uint32_t tb_flags = dc->base.tb->flags;
- uint32_t condexec, core_mmu_idx;
- dc->isar = &cpu->isar;
- dc->condjmp = 0;
- dc->aarch64 = 0;
- /* If we are coming from secure EL0 in a system with a 32-bit EL3, then
- * there is no secure EL1, so we route exceptions to EL3.
- */
- dc->secure_routed_to_el3 = arm_feature(env, ARM_FEATURE_EL3) &&
- !arm_el_is_aa64(env, 3);
- dc->thumb = FIELD_EX32(tb_flags, TBFLAG_AM32, THUMB);
- dc->be_data = FIELD_EX32(tb_flags, TBFLAG_ANY, BE_DATA) ? MO_BE : MO_LE;
- condexec = FIELD_EX32(tb_flags, TBFLAG_AM32, CONDEXEC);
- dc->condexec_mask = (condexec & 0xf) << 1;
- dc->condexec_cond = condexec >> 4;
- core_mmu_idx = FIELD_EX32(tb_flags, TBFLAG_ANY, MMUIDX);
- dc->mmu_idx = core_to_arm_mmu_idx(env, core_mmu_idx);
- dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
- #if !defined(CONFIG_USER_ONLY)
- dc->user = (dc->current_el == 0);
- #endif
- dc->fp_excp_el = FIELD_EX32(tb_flags, TBFLAG_ANY, FPEXC_EL);
- if (arm_feature(env, ARM_FEATURE_M)) {
- dc->vfp_enabled = 1;
- dc->be_data = MO_TE;
- dc->v7m_handler_mode = FIELD_EX32(tb_flags, TBFLAG_M32, HANDLER);
- dc->v8m_secure = arm_feature(env, ARM_FEATURE_M_SECURITY) &&
- regime_is_secure(env, dc->mmu_idx);
- dc->v8m_stackcheck = FIELD_EX32(tb_flags, TBFLAG_M32, STACKCHECK);
- dc->v8m_fpccr_s_wrong =
- FIELD_EX32(tb_flags, TBFLAG_M32, FPCCR_S_WRONG);
- dc->v7m_new_fp_ctxt_needed =
- FIELD_EX32(tb_flags, TBFLAG_M32, NEW_FP_CTXT_NEEDED);
- dc->v7m_lspact = FIELD_EX32(tb_flags, TBFLAG_M32, LSPACT);
- } else {
- dc->be_data =
- FIELD_EX32(tb_flags, TBFLAG_ANY, BE_DATA) ? MO_BE : MO_LE;
- dc->debug_target_el =
- FIELD_EX32(tb_flags, TBFLAG_ANY, DEBUG_TARGET_EL);
- dc->sctlr_b = FIELD_EX32(tb_flags, TBFLAG_A32, SCTLR_B);
- dc->hstr_active = FIELD_EX32(tb_flags, TBFLAG_A32, HSTR_ACTIVE);
- dc->ns = FIELD_EX32(tb_flags, TBFLAG_A32, NS);
- dc->vfp_enabled = FIELD_EX32(tb_flags, TBFLAG_A32, VFPEN);
- if (arm_feature(env, ARM_FEATURE_XSCALE)) {
- dc->c15_cpar = FIELD_EX32(tb_flags, TBFLAG_A32, XSCALE_CPAR);
- } else {
- dc->vec_len = FIELD_EX32(tb_flags, TBFLAG_A32, VECLEN);
- dc->vec_stride = FIELD_EX32(tb_flags, TBFLAG_A32, VECSTRIDE);
- }
- }
- dc->cp_regs = cpu->cp_regs;
- dc->features = env->features;
- /* Single step state. The code-generation logic here is:
- * SS_ACTIVE == 0:
- * generate code with no special handling for single-stepping (except
- * that anything that can make us go to SS_ACTIVE == 1 must end the TB;
- * this happens anyway because those changes are all system register or
- * PSTATE writes).
- * SS_ACTIVE == 1, PSTATE.SS == 1: (active-not-pending)
- * emit code for one insn
- * emit code to clear PSTATE.SS
- * emit code to generate software step exception for completed step
- * end TB (as usual for having generated an exception)
- * SS_ACTIVE == 1, PSTATE.SS == 0: (active-pending)
- * emit code to generate a software step exception
- * end the TB
- */
- dc->ss_active = FIELD_EX32(tb_flags, TBFLAG_ANY, SS_ACTIVE);
- dc->pstate_ss = FIELD_EX32(tb_flags, TBFLAG_ANY, PSTATE_SS);
- dc->is_ldex = false;
- dc->page_start = dc->base.pc_first & TARGET_PAGE_MASK;
- /* If architectural single step active, limit to 1. */
- if (is_singlestepping(dc)) {
- dc->base.max_insns = 1;
- }
- /* ARM is a fixed-length ISA. Bound the number of insns to execute
- to those left on the page. */
- if (!dc->thumb) {
- int bound = -(dc->base.pc_first | TARGET_PAGE_MASK) / 4;
- dc->base.max_insns = MIN(dc->base.max_insns, bound);
- }
- cpu_V0 = tcg_temp_new_i64();
- cpu_V1 = tcg_temp_new_i64();
- /* FIXME: cpu_M0 can probably be the same as cpu_V0. */
- cpu_M0 = tcg_temp_new_i64();
- }
- static void arm_tr_tb_start(DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- /* A note on handling of the condexec (IT) bits:
- *
- * We want to avoid the overhead of having to write the updated condexec
- * bits back to the CPUARMState for every instruction in an IT block. So:
- * (1) if the condexec bits are not already zero then we write
- * zero back into the CPUARMState now. This avoids complications trying
- * to do it at the end of the block. (For example if we don't do this
- * it's hard to identify whether we can safely skip writing condexec
- * at the end of the TB, which we definitely want to do for the case
- * where a TB doesn't do anything with the IT state at all.)
- * (2) if we are going to leave the TB then we call gen_set_condexec()
- * which will write the correct value into CPUARMState if zero is wrong.
- * This is done both for leaving the TB at the end, and for leaving
- * it because of an exception we know will happen, which is done in
- * gen_exception_insn(). The latter is necessary because we need to
- * leave the TB with the PC/IT state just prior to execution of the
- * instruction which caused the exception.
- * (3) if we leave the TB unexpectedly (eg a data abort on a load)
- * then the CPUARMState will be wrong and we need to reset it.
- * This is handled in the same way as restoration of the
- * PC in these situations; we save the value of the condexec bits
- * for each PC via tcg_gen_insn_start(), and restore_state_to_opc()
- * then uses this to restore them after an exception.
- *
- * Note that there are no instructions which can read the condexec
- * bits, and none which can write non-static values to them, so
- * we don't need to care about whether CPUARMState is correct in the
- * middle of a TB.
- */
- /* Reset the conditional execution bits immediately. This avoids
- complications trying to do it at the end of the block. */
- if (dc->condexec_mask || dc->condexec_cond) {
- TCGv_i32 tmp = tcg_temp_new_i32();
- tcg_gen_movi_i32(tmp, 0);
- store_cpu_field(tmp, condexec_bits);
- }
- }
- static void arm_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- tcg_gen_insn_start(dc->base.pc_next,
- (dc->condexec_cond << 4) | (dc->condexec_mask >> 1),
- 0);
- dc->insn_start = tcg_last_op();
- }
- static bool arm_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cpu,
- const CPUBreakpoint *bp)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- if (bp->flags & BP_CPU) {
- gen_set_condexec(dc);
- gen_set_pc_im(dc, dc->base.pc_next);
- gen_helper_check_breakpoints(cpu_env);
- /* End the TB early; it's likely not going to be executed */
- dc->base.is_jmp = DISAS_TOO_MANY;
- } else {
- gen_exception_internal_insn(dc, dc->base.pc_next, EXCP_DEBUG);
- /* The address covered by the breakpoint must be
- included in [tb->pc, tb->pc + tb->size) in order
- to for it to be properly cleared -- thus we
- increment the PC here so that the logic setting
- tb->size below does the right thing. */
- /* TODO: Advance PC by correct instruction length to
- * avoid disassembler error messages */
- dc->base.pc_next += 2;
- dc->base.is_jmp = DISAS_NORETURN;
- }
- return true;
- }
- static bool arm_pre_translate_insn(DisasContext *dc)
- {
- #ifdef CONFIG_USER_ONLY
- /* Intercept jump to the magic kernel page. */
- if (dc->base.pc_next >= 0xffff0000) {
- /* We always get here via a jump, so know we are not in a
- conditional execution block. */
- gen_exception_internal(EXCP_KERNEL_TRAP);
- dc->base.is_jmp = DISAS_NORETURN;
- return true;
- }
- #endif
- if (dc->ss_active && !dc->pstate_ss) {
- /* Singlestep state is Active-pending.
- * If we're in this state at the start of a TB then either
- * a) we just took an exception to an EL which is being debugged
- * and this is the first insn in the exception handler
- * b) debug exceptions were masked and we just unmasked them
- * without changing EL (eg by clearing PSTATE.D)
- * In either case we're going to take a swstep exception in the
- * "did not step an insn" case, and so the syndrome ISV and EX
- * bits should be zero.
- */
- assert(dc->base.num_insns == 1);
- gen_swstep_exception(dc, 0, 0);
- dc->base.is_jmp = DISAS_NORETURN;
- return true;
- }
- return false;
- }
- static void arm_post_translate_insn(DisasContext *dc)
- {
- if (dc->condjmp && !dc->base.is_jmp) {
- gen_set_label(dc->condlabel);
- dc->condjmp = 0;
- }
- translator_loop_temp_check(&dc->base);
- }
- static void arm_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- CPUARMState *env = cpu->env_ptr;
- unsigned int insn;
- if (arm_pre_translate_insn(dc)) {
- return;
- }
- dc->pc_curr = dc->base.pc_next;
- insn = arm_ldl_code(env, dc->base.pc_next, dc->sctlr_b);
- dc->insn = insn;
- dc->base.pc_next += 4;
- disas_arm_insn(dc, insn);
- arm_post_translate_insn(dc);
- /* ARM is a fixed-length ISA. We performed the cross-page check
- in init_disas_context by adjusting max_insns. */
- }
- static bool thumb_insn_is_unconditional(DisasContext *s, uint32_t insn)
- {
- /* Return true if this Thumb insn is always unconditional,
- * even inside an IT block. This is true of only a very few
- * instructions: BKPT, HLT, and SG.
- *
- * A larger class of instructions are UNPREDICTABLE if used
- * inside an IT block; we do not need to detect those here, because
- * what we do by default (perform the cc check and update the IT
- * bits state machine) is a permitted CONSTRAINED UNPREDICTABLE
- * choice for those situations.
- *
- * insn is either a 16-bit or a 32-bit instruction; the two are
- * distinguishable because for the 16-bit case the top 16 bits
- * are zeroes, and that isn't a valid 32-bit encoding.
- */
- if ((insn & 0xffffff00) == 0xbe00) {
- /* BKPT */
- return true;
- }
- if ((insn & 0xffffffc0) == 0xba80 && arm_dc_feature(s, ARM_FEATURE_V8) &&
- !arm_dc_feature(s, ARM_FEATURE_M)) {
- /* HLT: v8A only. This is unconditional even when it is going to
- * UNDEF; see the v8A ARM ARM DDI0487B.a H3.3.
- * For v7 cores this was a plain old undefined encoding and so
- * honours its cc check. (We might be using the encoding as
- * a semihosting trap, but we don't change the cc check behaviour
- * on that account, because a debugger connected to a real v7A
- * core and emulating semihosting traps by catching the UNDEF
- * exception would also only see cases where the cc check passed.
- * No guest code should be trying to do a HLT semihosting trap
- * in an IT block anyway.
- */
- return true;
- }
- if (insn == 0xe97fe97f && arm_dc_feature(s, ARM_FEATURE_V8) &&
- arm_dc_feature(s, ARM_FEATURE_M)) {
- /* SG: v8M only */
- return true;
- }
- return false;
- }
- static void thumb_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- CPUARMState *env = cpu->env_ptr;
- uint32_t insn;
- bool is_16bit;
- if (arm_pre_translate_insn(dc)) {
- return;
- }
- dc->pc_curr = dc->base.pc_next;
- insn = arm_lduw_code(env, dc->base.pc_next, dc->sctlr_b);
- is_16bit = thumb_insn_is_16bit(dc, dc->base.pc_next, insn);
- dc->base.pc_next += 2;
- if (!is_16bit) {
- uint32_t insn2 = arm_lduw_code(env, dc->base.pc_next, dc->sctlr_b);
- insn = insn << 16 | insn2;
- dc->base.pc_next += 2;
- }
- dc->insn = insn;
- if (dc->condexec_mask && !thumb_insn_is_unconditional(dc, insn)) {
- uint32_t cond = dc->condexec_cond;
- /*
- * Conditionally skip the insn. Note that both 0xe and 0xf mean
- * "always"; 0xf is not "never".
- */
- if (cond < 0x0e) {
- arm_skip_unless(dc, cond);
- }
- }
- if (is_16bit) {
- disas_thumb_insn(dc, insn);
- } else {
- disas_thumb2_insn(dc, insn);
- }
- /* Advance the Thumb condexec condition. */
- if (dc->condexec_mask) {
- dc->condexec_cond = ((dc->condexec_cond & 0xe) |
- ((dc->condexec_mask >> 4) & 1));
- dc->condexec_mask = (dc->condexec_mask << 1) & 0x1f;
- if (dc->condexec_mask == 0) {
- dc->condexec_cond = 0;
- }
- }
- arm_post_translate_insn(dc);
- /* Thumb is a variable-length ISA. Stop translation when the next insn
- * will touch a new page. This ensures that prefetch aborts occur at
- * the right place.
- *
- * We want to stop the TB if the next insn starts in a new page,
- * or if it spans between this page and the next. This means that
- * if we're looking at the last halfword in the page we need to
- * see if it's a 16-bit Thumb insn (which will fit in this TB)
- * or a 32-bit Thumb insn (which won't).
- * This is to avoid generating a silly TB with a single 16-bit insn
- * in it at the end of this page (which would execute correctly
- * but isn't very efficient).
- */
- if (dc->base.is_jmp == DISAS_NEXT
- && (dc->base.pc_next - dc->page_start >= TARGET_PAGE_SIZE
- || (dc->base.pc_next - dc->page_start >= TARGET_PAGE_SIZE - 3
- && insn_crosses_page(env, dc)))) {
- dc->base.is_jmp = DISAS_TOO_MANY;
- }
- }
- static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- if (tb_cflags(dc->base.tb) & CF_LAST_IO && dc->condjmp) {
- /* FIXME: This can theoretically happen with self-modifying code. */
- cpu_abort(cpu, "IO on conditional branch instruction");
- }
- /* At this stage dc->condjmp will only be set when the skipped
- instruction was a conditional branch or trap, and the PC has
- already been written. */
- gen_set_condexec(dc);
- if (dc->base.is_jmp == DISAS_BX_EXCRET) {
- /* Exception return branches need some special case code at the
- * end of the TB, which is complex enough that it has to
- * handle the single-step vs not and the condition-failed
- * insn codepath itself.
- */
- gen_bx_excret_final_code(dc);
- } else if (unlikely(is_singlestepping(dc))) {
- /* Unconditional and "condition passed" instruction codepath. */
- switch (dc->base.is_jmp) {
- case DISAS_SWI:
- gen_ss_advance(dc);
- gen_exception(EXCP_SWI, syn_aa32_svc(dc->svc_imm, dc->thumb),
- default_exception_el(dc));
- break;
- case DISAS_HVC:
- gen_ss_advance(dc);
- gen_exception(EXCP_HVC, syn_aa32_hvc(dc->svc_imm), 2);
- break;
- case DISAS_SMC:
- gen_ss_advance(dc);
- gen_exception(EXCP_SMC, syn_aa32_smc(), 3);
- break;
- case DISAS_NEXT:
- case DISAS_TOO_MANY:
- case DISAS_UPDATE_EXIT:
- case DISAS_UPDATE_NOCHAIN:
- gen_set_pc_im(dc, dc->base.pc_next);
- /* fall through */
- default:
- /* FIXME: Single stepping a WFI insn will not halt the CPU. */
- gen_singlestep_exception(dc);
- break;
- case DISAS_NORETURN:
- break;
- }
- } else {
- /* While branches must always occur at the end of an IT block,
- there are a few other things that can cause us to terminate
- the TB in the middle of an IT block:
- - Exception generating instructions (bkpt, swi, undefined).
- - Page boundaries.
- - Hardware watchpoints.
- Hardware breakpoints have already been handled and skip this code.
- */
- switch(dc->base.is_jmp) {
- case DISAS_NEXT:
- case DISAS_TOO_MANY:
- gen_goto_tb(dc, 1, dc->base.pc_next);
- break;
- case DISAS_UPDATE_NOCHAIN:
- gen_set_pc_im(dc, dc->base.pc_next);
- /* fall through */
- case DISAS_JUMP:
- gen_goto_ptr();
- break;
- case DISAS_UPDATE_EXIT:
- gen_set_pc_im(dc, dc->base.pc_next);
- /* fall through */
- default:
- /* indicate that the hash table must be used to find the next TB */
- tcg_gen_exit_tb(NULL, 0);
- break;
- case DISAS_NORETURN:
- /* nothing more to generate */
- break;
- case DISAS_WFI:
- {
- TCGv_i32 tmp = tcg_const_i32((dc->thumb &&
- !(dc->insn & (1U << 31))) ? 2 : 4);
- gen_helper_wfi(cpu_env, tmp);
- tcg_temp_free_i32(tmp);
- /* The helper doesn't necessarily throw an exception, but we
- * must go back to the main loop to check for interrupts anyway.
- */
- tcg_gen_exit_tb(NULL, 0);
- break;
- }
- case DISAS_WFE:
- gen_helper_wfe(cpu_env);
- break;
- case DISAS_YIELD:
- gen_helper_yield(cpu_env);
- break;
- case DISAS_SWI:
- gen_exception(EXCP_SWI, syn_aa32_svc(dc->svc_imm, dc->thumb),
- default_exception_el(dc));
- break;
- case DISAS_HVC:
- gen_exception(EXCP_HVC, syn_aa32_hvc(dc->svc_imm), 2);
- break;
- case DISAS_SMC:
- gen_exception(EXCP_SMC, syn_aa32_smc(), 3);
- break;
- }
- }
- if (dc->condjmp) {
- /* "Condition failed" instruction codepath for the branch/trap insn */
- gen_set_label(dc->condlabel);
- gen_set_condexec(dc);
- if (unlikely(is_singlestepping(dc))) {
- gen_set_pc_im(dc, dc->base.pc_next);
- gen_singlestep_exception(dc);
- } else {
- gen_goto_tb(dc, 1, dc->base.pc_next);
- }
- }
- }
- static void arm_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
- {
- DisasContext *dc = container_of(dcbase, DisasContext, base);
- qemu_log("IN: %s\n", lookup_symbol(dc->base.pc_first));
- log_target_disas(cpu, dc->base.pc_first, dc->base.tb->size);
- }
- static const TranslatorOps arm_translator_ops = {
- .init_disas_context = arm_tr_init_disas_context,
- .tb_start = arm_tr_tb_start,
- .insn_start = arm_tr_insn_start,
- .breakpoint_check = arm_tr_breakpoint_check,
- .translate_insn = arm_tr_translate_insn,
- .tb_stop = arm_tr_tb_stop,
- .disas_log = arm_tr_disas_log,
- };
- static const TranslatorOps thumb_translator_ops = {
- .init_disas_context = arm_tr_init_disas_context,
- .tb_start = arm_tr_tb_start,
- .insn_start = arm_tr_insn_start,
- .breakpoint_check = arm_tr_breakpoint_check,
- .translate_insn = thumb_tr_translate_insn,
- .tb_stop = arm_tr_tb_stop,
- .disas_log = arm_tr_disas_log,
- };
- /* generate intermediate code for basic block 'tb'. */
- void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
- {
- DisasContext dc = { };
- const TranslatorOps *ops = &arm_translator_ops;
- if (FIELD_EX32(tb->flags, TBFLAG_AM32, THUMB)) {
- ops = &thumb_translator_ops;
- }
- #ifdef TARGET_AARCH64
- if (FIELD_EX32(tb->flags, TBFLAG_ANY, AARCH64_STATE)) {
- ops = &aarch64_translator_ops;
- }
- #endif
- translator_loop(ops, &dc.base, cpu, tb, max_insns);
- }
- void restore_state_to_opc(CPUARMState *env, TranslationBlock *tb,
- target_ulong *data)
- {
- if (is_a64(env)) {
- env->pc = data[0];
- env->condexec_bits = 0;
- env->exception.syndrome = data[2] << ARM_INSN_START_WORD2_SHIFT;
- } else {
- env->regs[15] = data[0];
- env->condexec_bits = data[1];
- env->exception.syndrome = data[2] << ARM_INSN_START_WORD2_SHIFT;
- }
- }
|