ARMInstrMVE.td 310 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900490149024903490449054906490749084909491049114912491349144915491649174918491949204921492249234924492549264927492849294930493149324933493449354936493749384939494049414942494349444945494649474948494949504951495249534954495549564957495849594960496149624963496449654966496749684969497049714972497349744975497649774978497949804981498249834984498549864987498849894990499149924993499449954996499749984999500050015002500350045005500650075008500950105011501250135014501550165017501850195020502150225023502450255026502750285029503050315032503350345035503650375038503950405041504250435044504550465047504850495050505150525053505450555056505750585059506050615062506350645065506650675068506950705071507250735074507550765077507850795080508150825083508450855086508750885089509050915092509350945095509650975098509951005101510251035104510551065107510851095110511151125113511451155116511751185119512051215122512351245125512651275128512951305131513251335134513551365137513851395140514151425143514451455146514751485149515051515152515351545155515651575158515951605161516251635164516551665167516851695170517151725173517451755176517751785179518051815182518351845185518651875188518951905191519251935194519551965197519851995200520152025203520452055206520752085209521052115212521352145215521652175218521952205221522252235224522552265227522852295230523152325233523452355236523752385239524052415242524352445245524652475248524952505251525252535254525552565257525852595260526152625263526452655266526752685269527052715272527352745275527652775278527952805281528252835284528552865287528852895290529152925293529452955296529752985299530053015302530353045305530653075308530953105311531253135314531553165317531853195320532153225323532453255326532753285329533053315332533353345335533653375338533953405341534253435344534553465347534853495350535153525353535453555356535753585359536053615362536353645365536653675368536953705371537253735374537553765377537853795380538153825383538453855386538753885389539053915392539353945395539653975398539954005401540254035404540554065407540854095410541154125413541454155416541754185419542054215422542354245425542654275428542954305431543254335434543554365437543854395440544154425443544454455446544754485449545054515452545354545455545654575458545954605461546254635464546554665467546854695470547154725473547454755476547754785479548054815482548354845485548654875488548954905491549254935494549554965497549854995500550155025503550455055506550755085509551055115512551355145515551655175518551955205521552255235524552555265527552855295530553155325533553455355536553755385539554055415542554355445545554655475548554955505551555255535554555555565557555855595560556155625563556455655566556755685569557055715572557355745575557655775578557955805581558255835584558555865587558855895590559155925593559455955596559755985599560056015602560356045605560656075608560956105611561256135614561556165617561856195620562156225623562456255626562756285629563056315632563356345635563656375638563956405641564256435644564556465647564856495650565156525653565456555656565756585659566056615662566356645665566656675668566956705671567256735674567556765677567856795680568156825683568456855686568756885689569056915692569356945695569656975698569957005701570257035704570557065707570857095710571157125713571457155716571757185719572057215722572357245725572657275728572957305731573257335734573557365737573857395740574157425743574457455746574757485749575057515752575357545755575657575758575957605761576257635764576557665767576857695770577157725773577457755776577757785779578057815782578357845785578657875788578957905791579257935794579557965797579857995800580158025803580458055806580758085809581058115812581358145815581658175818581958205821582258235824582558265827582858295830583158325833583458355836583758385839584058415842584358445845584658475848584958505851585258535854585558565857585858595860586158625863586458655866586758685869587058715872587358745875587658775878587958805881588258835884588558865887588858895890589158925893589458955896589758985899590059015902590359045905590659075908590959105911591259135914591559165917591859195920592159225923592459255926592759285929593059315932593359345935593659375938593959405941594259435944594559465947594859495950595159525953595459555956595759585959596059615962596359645965596659675968596959705971597259735974597559765977597859795980598159825983598459855986598759885989599059915992599359945995599659975998599960006001600260036004600560066007600860096010601160126013601460156016601760186019602060216022602360246025602660276028602960306031603260336034603560366037603860396040604160426043604460456046604760486049605060516052605360546055605660576058605960606061606260636064606560666067606860696070607160726073607460756076607760786079608060816082608360846085608660876088608960906091609260936094609560966097609860996100610161026103610461056106610761086109611061116112611361146115611661176118611961206121612261236124612561266127612861296130613161326133613461356136613761386139614061416142614361446145614661476148614961506151615261536154615561566157615861596160616161626163616461656166616761686169617061716172617361746175617661776178617961806181618261836184618561866187618861896190619161926193619461956196619761986199620062016202620362046205620662076208620962106211621262136214621562166217621862196220622162226223622462256226622762286229623062316232623362346235623662376238623962406241624262436244624562466247624862496250625162526253625462556256625762586259626062616262626362646265626662676268626962706271627262736274627562766277627862796280628162826283628462856286628762886289629062916292629362946295629662976298629963006301630263036304630563066307630863096310631163126313631463156316631763186319632063216322632363246325632663276328632963306331633263336334633563366337633863396340634163426343634463456346634763486349635063516352635363546355635663576358635963606361636263636364636563666367636863696370637163726373637463756376637763786379638063816382638363846385638663876388638963906391639263936394639563966397639863996400640164026403640464056406640764086409641064116412641364146415641664176418641964206421642264236424642564266427642864296430643164326433643464356436643764386439644064416442644364446445644664476448644964506451645264536454645564566457645864596460646164626463646464656466646764686469647064716472647364746475647664776478647964806481648264836484648564866487648864896490649164926493649464956496649764986499650065016502650365046505650665076508650965106511651265136514651565166517651865196520652165226523652465256526652765286529653065316532653365346535653665376538653965406541654265436544654565466547654865496550655165526553655465556556655765586559656065616562656365646565656665676568656965706571657265736574657565766577657865796580658165826583658465856586658765886589659065916592659365946595659665976598659966006601660266036604660566066607660866096610661166126613661466156616661766186619662066216622662366246625662666276628662966306631663266336634663566366637663866396640664166426643664466456646664766486649665066516652665366546655665666576658665966606661666266636664666566666667666866696670667166726673667466756676667766786679668066816682668366846685668666876688668966906691669266936694669566966697669866996700670167026703670467056706670767086709671067116712671367146715671667176718671967206721672267236724672567266727672867296730673167326733673467356736673767386739674067416742674367446745674667476748674967506751675267536754675567566757675867596760676167626763676467656766676767686769677067716772677367746775677667776778677967806781678267836784678567866787678867896790679167926793679467956796679767986799680068016802680368046805680668076808680968106811681268136814681568166817681868196820682168226823682468256826682768286829683068316832683368346835683668376838683968406841684268436844684568466847684868496850685168526853685468556856685768586859686068616862686368646865686668676868686968706871687268736874687568766877687868796880688168826883688468856886688768886889689068916892689368946895689668976898689969006901690269036904690569066907690869096910691169126913691469156916691769186919692069216922692369246925692669276928692969306931693269336934693569366937693869396940694169426943694469456946694769486949695069516952695369546955695669576958695969606961696269636964696569666967696869696970697169726973697469756976697769786979698069816982698369846985698669876988698969906991699269936994699569966997699869997000700170027003700470057006700770087009701070117012701370147015701670177018701970207021702270237024702570267027702870297030703170327033703470357036703770387039704070417042704370447045704670477048704970507051705270537054705570567057705870597060706170627063706470657066706770687069707070717072707370747075707670777078707970807081708270837084708570867087708870897090709170927093709470957096709770987099710071017102710371047105710671077108710971107111711271137114711571167117711871197120712171227123712471257126712771287129713071317132713371347135713671377138713971407141714271437144714571467147714871497150715171527153715471557156715771587159716071617162716371647165716671677168716971707171717271737174717571767177717871797180718171827183718471857186718771887189719071917192719371947195719671977198719972007201720272037204720572067207720872097210721172127213721472157216721772187219722072217222722372247225722672277228722972307231723272337234723572367237723872397240724172427243724472457246724772487249725072517252725372547255725672577258725972607261726272637264726572667267726872697270727172727273727472757276727772787279728072817282728372847285728672877288728972907291729272937294729572967297729872997300730173027303730473057306730773087309731073117312731373147315731673177318731973207321732273237324732573267327732873297330733173327333733473357336733773387339734073417342734373447345734673477348734973507351735273537354735573567357735873597360736173627363736473657366736773687369737073717372737373747375737673777378737973807381738273837384738573867387738873897390739173927393
  1. //===-- ARMInstrMVE.td - MVE support for ARM ---------------*- tablegen -*-===//
  2. //
  3. // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
  4. // See https://llvm.org/LICENSE.txt for license information.
  5. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
  6. //
  7. //===----------------------------------------------------------------------===//
  8. //
  9. // This file describes the ARM MVE instruction set.
  10. //
  11. //===----------------------------------------------------------------------===//
  12. // VPT condition mask
  13. def vpt_mask : Operand<i32> {
  14. let PrintMethod = "printVPTMask";
  15. let ParserMatchClass = it_mask_asmoperand;
  16. let EncoderMethod = "getVPTMaskOpValue";
  17. let DecoderMethod = "DecodeVPTMaskOperand";
  18. }
  19. // VPT/VCMP restricted predicate for sign invariant types
  20. def pred_restricted_i_asmoperand : AsmOperandClass {
  21. let Name = "CondCodeRestrictedI";
  22. let RenderMethod = "addITCondCodeOperands";
  23. let PredicateMethod = "isITCondCodeRestrictedI";
  24. let ParserMethod = "parseITCondCode";
  25. let DiagnosticString = "condition code for sign-independent integer "#
  26. "comparison must be EQ or NE";
  27. }
  28. // VPT/VCMP restricted predicate for signed types
  29. def pred_restricted_s_asmoperand : AsmOperandClass {
  30. let Name = "CondCodeRestrictedS";
  31. let RenderMethod = "addITCondCodeOperands";
  32. let PredicateMethod = "isITCondCodeRestrictedS";
  33. let ParserMethod = "parseITCondCode";
  34. let DiagnosticString = "condition code for signed integer "#
  35. "comparison must be EQ, NE, LT, GT, LE or GE";
  36. }
  37. // VPT/VCMP restricted predicate for unsigned types
  38. def pred_restricted_u_asmoperand : AsmOperandClass {
  39. let Name = "CondCodeRestrictedU";
  40. let RenderMethod = "addITCondCodeOperands";
  41. let PredicateMethod = "isITCondCodeRestrictedU";
  42. let ParserMethod = "parseITCondCode";
  43. let DiagnosticString = "condition code for unsigned integer "#
  44. "comparison must be EQ, NE, HS or HI";
  45. }
  46. // VPT/VCMP restricted predicate for floating point
  47. def pred_restricted_fp_asmoperand : AsmOperandClass {
  48. let Name = "CondCodeRestrictedFP";
  49. let RenderMethod = "addITCondCodeOperands";
  50. let PredicateMethod = "isITCondCodeRestrictedFP";
  51. let ParserMethod = "parseITCondCode";
  52. let DiagnosticString = "condition code for floating-point "#
  53. "comparison must be EQ, NE, LT, GT, LE or GE";
  54. }
  55. class VCMPPredicateOperand : Operand<i32>;
  56. def pred_basic_i : VCMPPredicateOperand {
  57. let PrintMethod = "printMandatoryRestrictedPredicateOperand";
  58. let ParserMatchClass = pred_restricted_i_asmoperand;
  59. let DecoderMethod = "DecodeRestrictedIPredicateOperand";
  60. let EncoderMethod = "getRestrictedCondCodeOpValue";
  61. }
  62. def pred_basic_u : VCMPPredicateOperand {
  63. let PrintMethod = "printMandatoryRestrictedPredicateOperand";
  64. let ParserMatchClass = pred_restricted_u_asmoperand;
  65. let DecoderMethod = "DecodeRestrictedUPredicateOperand";
  66. let EncoderMethod = "getRestrictedCondCodeOpValue";
  67. }
  68. def pred_basic_s : VCMPPredicateOperand {
  69. let PrintMethod = "printMandatoryRestrictedPredicateOperand";
  70. let ParserMatchClass = pred_restricted_s_asmoperand;
  71. let DecoderMethod = "DecodeRestrictedSPredicateOperand";
  72. let EncoderMethod = "getRestrictedCondCodeOpValue";
  73. }
  74. def pred_basic_fp : VCMPPredicateOperand {
  75. let PrintMethod = "printMandatoryRestrictedPredicateOperand";
  76. let ParserMatchClass = pred_restricted_fp_asmoperand;
  77. let DecoderMethod = "DecodeRestrictedFPPredicateOperand";
  78. let EncoderMethod = "getRestrictedCondCodeOpValue";
  79. }
  80. // Register list operands for interleaving load/stores
  81. def VecList2QAsmOperand : AsmOperandClass {
  82. let Name = "VecListTwoMQ";
  83. let ParserMethod = "parseVectorList";
  84. let RenderMethod = "addMVEVecListOperands";
  85. let DiagnosticString = "operand must be a list of two consecutive "#
  86. "q-registers in range [q0,q7]";
  87. }
  88. def VecList2Q : RegisterOperand<QQPR, "printMVEVectorListTwoQ"> {
  89. let ParserMatchClass = VecList2QAsmOperand;
  90. let PrintMethod = "printMVEVectorList<2>";
  91. }
  92. def VecList4QAsmOperand : AsmOperandClass {
  93. let Name = "VecListFourMQ";
  94. let ParserMethod = "parseVectorList";
  95. let RenderMethod = "addMVEVecListOperands";
  96. let DiagnosticString = "operand must be a list of four consecutive "#
  97. "q-registers in range [q0,q7]";
  98. }
  99. def VecList4Q : RegisterOperand<QQQQPR, "printMVEVectorListFourQ"> {
  100. let ParserMatchClass = VecList4QAsmOperand;
  101. let PrintMethod = "printMVEVectorList<4>";
  102. }
  103. // taddrmode_imm7 := reg[r0-r7] +/- (imm7 << shift)
  104. class TMemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
  105. let Name = "TMemImm7Shift"#shift#"Offset";
  106. let PredicateMethod = "isMemImm7ShiftedOffset<"#shift#",ARM::tGPRRegClassID>";
  107. let RenderMethod = "addMemImmOffsetOperands";
  108. }
  109. class taddrmode_imm7<int shift> : MemOperand,
  110. ComplexPattern<i32, 2, "SelectTAddrModeImm7<"#shift#">", []> {
  111. let ParserMatchClass = TMemImm7ShiftOffsetAsmOperand<shift>;
  112. // They are printed the same way as the T2 imm8 version
  113. let PrintMethod = "printT2AddrModeImm8Operand<false>";
  114. // This can also be the same as the T2 version.
  115. let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
  116. let DecoderMethod = "DecodeTAddrModeImm7<"#shift#">";
  117. let MIOperandInfo = (ops tGPR:$base, i32imm:$offsimm);
  118. }
  119. // t2addrmode_imm7 := reg +/- (imm7)
  120. class MemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
  121. let Name = "MemImm7Shift"#shift#"Offset";
  122. let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
  123. ",ARM::GPRnopcRegClassID>";
  124. let RenderMethod = "addMemImmOffsetOperands";
  125. }
  126. def MemImm7Shift0OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<0>;
  127. def MemImm7Shift1OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<1>;
  128. def MemImm7Shift2OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<2>;
  129. class T2AddrMode_Imm7<int shift> : MemOperand,
  130. ComplexPattern<i32, 2, "SelectT2AddrModeImm7<"#shift#">", []> {
  131. let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
  132. let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 0>";
  133. let ParserMatchClass =
  134. !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetAsmOperand");
  135. let MIOperandInfo = (ops GPRnopc:$base, i32imm:$offsimm);
  136. }
  137. class t2addrmode_imm7<int shift> : T2AddrMode_Imm7<shift> {
  138. // They are printed the same way as the imm8 version
  139. let PrintMethod = "printT2AddrModeImm8Operand<false>";
  140. }
  141. class MemImm7ShiftOffsetWBAsmOperand<int shift> : AsmOperandClass {
  142. let Name = "MemImm7Shift"#shift#"OffsetWB";
  143. let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
  144. ",ARM::rGPRRegClassID>";
  145. let RenderMethod = "addMemImmOffsetOperands";
  146. }
  147. def MemImm7Shift0OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<0>;
  148. def MemImm7Shift1OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<1>;
  149. def MemImm7Shift2OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<2>;
  150. class t2addrmode_imm7_pre<int shift> : T2AddrMode_Imm7<shift> {
  151. // They are printed the same way as the imm8 version
  152. let PrintMethod = "printT2AddrModeImm8Operand<true>";
  153. let ParserMatchClass =
  154. !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetWBAsmOperand");
  155. let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 1>";
  156. let MIOperandInfo = (ops rGPR:$base, i32imm:$offsim);
  157. }
  158. class t2am_imm7shiftOffsetAsmOperand<int shift>
  159. : AsmOperandClass { let Name = "Imm7Shift"#shift; }
  160. def t2am_imm7shift0OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<0>;
  161. def t2am_imm7shift1OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<1>;
  162. def t2am_imm7shift2OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<2>;
  163. class t2am_imm7_offset<int shift> : MemOperand,
  164. ComplexPattern<i32, 1, "SelectT2AddrModeImm7Offset<"#shift#">",
  165. [], [SDNPWantRoot]> {
  166. // They are printed the same way as the imm8 version
  167. let PrintMethod = "printT2AddrModeImm8OffsetOperand";
  168. let ParserMatchClass =
  169. !cast<AsmOperandClass>("t2am_imm7shift"#shift#"OffsetAsmOperand");
  170. let EncoderMethod = "getT2ScaledImmOpValue<7,"#shift#">";
  171. let DecoderMethod = "DecodeT2Imm7<"#shift#">";
  172. }
  173. // Operands for gather/scatter loads of the form [Rbase, Qoffsets]
  174. class MemRegRQOffsetAsmOperand<int shift> : AsmOperandClass {
  175. let Name = "MemRegRQS"#shift#"Offset";
  176. let PredicateMethod = "isMemRegRQOffset<"#shift#">";
  177. let RenderMethod = "addMemRegRQOffsetOperands";
  178. }
  179. def MemRegRQS0OffsetAsmOperand : MemRegRQOffsetAsmOperand<0>;
  180. def MemRegRQS1OffsetAsmOperand : MemRegRQOffsetAsmOperand<1>;
  181. def MemRegRQS2OffsetAsmOperand : MemRegRQOffsetAsmOperand<2>;
  182. def MemRegRQS3OffsetAsmOperand : MemRegRQOffsetAsmOperand<3>;
  183. // mve_addr_rq_shift := reg + vreg{ << UXTW #shift}
  184. class mve_addr_rq_shift<int shift> : MemOperand {
  185. let EncoderMethod = "getMveAddrModeRQOpValue";
  186. let PrintMethod = "printMveAddrModeRQOperand<"#shift#">";
  187. let ParserMatchClass =
  188. !cast<AsmOperandClass>("MemRegRQS"#shift#"OffsetAsmOperand");
  189. let DecoderMethod = "DecodeMveAddrModeRQ";
  190. let MIOperandInfo = (ops GPRnopc:$base, MQPR:$offsreg);
  191. }
  192. class MemRegQOffsetAsmOperand<int shift> : AsmOperandClass {
  193. let Name = "MemRegQS"#shift#"Offset";
  194. let PredicateMethod = "isMemRegQOffset<"#shift#">";
  195. let RenderMethod = "addMemImmOffsetOperands";
  196. }
  197. def MemRegQS2OffsetAsmOperand : MemRegQOffsetAsmOperand<2>;
  198. def MemRegQS3OffsetAsmOperand : MemRegQOffsetAsmOperand<3>;
  199. // mve_addr_q_shift := vreg {+ #imm7s2/4}
  200. class mve_addr_q_shift<int shift> : MemOperand {
  201. let EncoderMethod = "getMveAddrModeQOpValue<"#shift#">";
  202. // Can be printed same way as other reg + imm operands
  203. let PrintMethod = "printT2AddrModeImm8Operand<false>";
  204. let ParserMatchClass =
  205. !cast<AsmOperandClass>("MemRegQS"#shift#"OffsetAsmOperand");
  206. let DecoderMethod = "DecodeMveAddrModeQ<"#shift#">";
  207. let MIOperandInfo = (ops MQPR:$base, i32imm:$imm);
  208. }
  209. // A family of classes wrapping up information about the vector types
  210. // used by MVE.
  211. class MVEVectorVTInfo<ValueType vec, ValueType dblvec,
  212. ValueType pred, ValueType dblpred,
  213. bits<2> size, string suffixletter, bit unsigned> {
  214. // The LLVM ValueType representing the vector, so we can use it in
  215. // ISel patterns.
  216. ValueType Vec = vec;
  217. // The LLVM ValueType representing a vector with elements double the size
  218. // of those in Vec, so we can use it in ISel patterns. It is up to the
  219. // invoker of this class to ensure that this is a correct choice.
  220. ValueType DblVec = dblvec;
  221. // An LLVM ValueType representing a corresponding vector of
  222. // predicate bits, for use in ISel patterns that handle an IR
  223. // intrinsic describing the predicated form of the instruction.
  224. //
  225. // Usually, for a vector of N things, this will be vNi1. But for
  226. // vectors of 2 values, we make an exception, and use v4i1 instead
  227. // of v2i1. Rationale: MVE codegen doesn't support doing all the
  228. // auxiliary operations on v2i1 (vector shuffles etc), and also,
  229. // there's no MVE compare instruction that will _generate_ v2i1
  230. // directly.
  231. ValueType Pred = pred;
  232. // Same as Pred but for DblVec rather than Vec.
  233. ValueType DblPred = dblpred;
  234. // The most common representation of the vector element size in MVE
  235. // instruction encodings: a 2-bit value V representing an (8<<V)-bit
  236. // vector element.
  237. bits<2> Size = size;
  238. // For vectors explicitly mentioning a signedness of integers: 0 for
  239. // signed and 1 for unsigned. For anything else, undefined.
  240. bit Unsigned = unsigned;
  241. // The number of bits in a vector element, in integer form.
  242. int LaneBits = !shl(8, Size);
  243. // The suffix used in assembly language on an instruction operating
  244. // on this lane if it only cares about number of bits.
  245. string BitsSuffix = !if(!eq(suffixletter, "p"),
  246. !if(!eq(unsigned, 0b0), "8", "16"),
  247. !cast<string>(LaneBits));
  248. // The suffix used on an instruction that mentions the whole type.
  249. string Suffix = suffixletter # BitsSuffix;
  250. // The letter part of the suffix only.
  251. string SuffixLetter = suffixletter;
  252. }
  253. // Integer vector types that don't treat signed and unsigned differently.
  254. def MVE_v16i8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "i", ?>;
  255. def MVE_v8i16 : MVEVectorVTInfo<v8i16, v4i32, v8i1, v4i1, 0b01, "i", ?>;
  256. def MVE_v4i32 : MVEVectorVTInfo<v4i32, v2i64, v4i1, v4i1, 0b10, "i", ?>;
  257. def MVE_v2i64 : MVEVectorVTInfo<v2i64, ?, v4i1, ?, 0b11, "i", ?>;
  258. // Explicitly signed and unsigned integer vectors. They map to the
  259. // same set of LLVM ValueTypes as above, but are represented
  260. // differently in assembly and instruction encodings.
  261. def MVE_v16s8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "s", 0b0>;
  262. def MVE_v8s16 : MVEVectorVTInfo<v8i16, v4i32, v8i1, v4i1, 0b01, "s", 0b0>;
  263. def MVE_v4s32 : MVEVectorVTInfo<v4i32, v2i64, v4i1, v4i1, 0b10, "s", 0b0>;
  264. def MVE_v2s64 : MVEVectorVTInfo<v2i64, ?, v4i1, ?, 0b11, "s", 0b0>;
  265. def MVE_v16u8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "u", 0b1>;
  266. def MVE_v8u16 : MVEVectorVTInfo<v8i16, v4i32, v8i1, v4i1, 0b01, "u", 0b1>;
  267. def MVE_v4u32 : MVEVectorVTInfo<v4i32, v2i64, v4i1, v4i1, 0b10, "u", 0b1>;
  268. def MVE_v2u64 : MVEVectorVTInfo<v2i64, ?, v4i1, ?, 0b11, "u", 0b1>;
  269. // FP vector types.
  270. def MVE_v8f16 : MVEVectorVTInfo<v8f16, v4f32, v8i1, v4i1, 0b01, "f", ?>;
  271. def MVE_v4f32 : MVEVectorVTInfo<v4f32, v2f64, v4i1, v4i1, 0b10, "f", ?>;
  272. def MVE_v2f64 : MVEVectorVTInfo<v2f64, ?, v4i1, ?, 0b11, "f", ?>;
  273. // Polynomial vector types.
  274. def MVE_v16p8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b11, "p", 0b0>;
  275. def MVE_v8p16 : MVEVectorVTInfo<v8i16, v4i32, v8i1, v4i1, 0b11, "p", 0b1>;
  276. multiclass MVE_TwoOpPattern<MVEVectorVTInfo VTI, PatFrag Op, Intrinsic PredInt,
  277. dag PredOperands, Instruction Inst,
  278. SDPatternOperator IdentityVec = null_frag> {
  279. // Unpredicated
  280. def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
  281. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  282. // Predicated with select
  283. if !ne(VTI.Size, 0b11) then {
  284. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
  285. (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
  286. (VTI.Vec MQPR:$Qn))),
  287. (VTI.Vec MQPR:$inactive))),
  288. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  289. ARMVCCThen, (VTI.Pred VCCR:$mask),
  290. (VTI.Vec MQPR:$inactive)))>;
  291. // Optionally with the select folded through the op
  292. def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
  293. (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
  294. (VTI.Vec MQPR:$Qn),
  295. (VTI.Vec IdentityVec))))),
  296. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  297. ARMVCCThen, (VTI.Pred VCCR:$mask),
  298. (VTI.Vec MQPR:$Qm)))>;
  299. }
  300. // Predicated with intrinsic
  301. def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)),
  302. PredOperands,
  303. (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
  304. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  305. ARMVCCThen, (VTI.Pred VCCR:$mask),
  306. (VTI.Vec MQPR:$inactive)))>;
  307. }
  308. multiclass MVE_TwoOpPatternDup<MVEVectorVTInfo VTI, PatFrag Op, Intrinsic PredInt,
  309. dag PredOperands, Instruction Inst,
  310. SDPatternOperator IdentityVec = null_frag> {
  311. // Unpredicated
  312. def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn)))),
  313. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn))>;
  314. // Predicated with select
  315. if !ne(VTI.Size, 0b11) then {
  316. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
  317. (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
  318. (VTI.Vec (ARMvdup rGPR:$Rn)))),
  319. (VTI.Vec MQPR:$inactive))),
  320. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
  321. ARMVCCThen, (VTI.Pred VCCR:$mask),
  322. (VTI.Vec MQPR:$inactive)))>;
  323. // Optionally with the select folded through the op
  324. def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
  325. (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
  326. (ARMvdup rGPR:$Rn),
  327. (VTI.Vec IdentityVec))))),
  328. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
  329. ARMVCCThen, (VTI.Pred VCCR:$mask),
  330. (VTI.Vec MQPR:$Qm)))>;
  331. }
  332. // Predicated with intrinsic
  333. def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn))),
  334. PredOperands,
  335. (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
  336. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
  337. ARMVCCThen, (VTI.Pred VCCR:$mask),
  338. (VTI.Vec MQPR:$inactive)))>;
  339. }
  340. // --------- Start of base classes for the instructions themselves
  341. class MVE_MI<dag oops, dag iops, InstrItinClass itin, string asm,
  342. string ops, string cstr, list<dag> pattern>
  343. : Thumb2XI<oops, iops, AddrModeNone, 4, itin, !strconcat(asm, "\t", ops), cstr,
  344. pattern>,
  345. Requires<[HasMVEInt]> {
  346. let D = MVEDomain;
  347. let DecoderNamespace = "MVE";
  348. }
  349. // MVE_p is used for most predicated instructions, to add the cluster
  350. // of input operands that provides the VPT suffix (none, T or E) and
  351. // the input predicate register.
  352. class MVE_p<dag oops, dag iops, InstrItinClass itin, string iname,
  353. string suffix, string ops, vpred_ops vpred, string cstr,
  354. list<dag> pattern=[]>
  355. : MVE_MI<oops, !con(iops, (ins vpred:$vp)), itin,
  356. // If the instruction has a suffix, like vadd.f32, then the
  357. // VPT predication suffix goes before the dot, so the full
  358. // name has to be "vadd${vp}.f32".
  359. !strconcat(iname, "${vp}",
  360. !if(!eq(suffix, ""), "", !strconcat(".", suffix))),
  361. ops, !strconcat(cstr, vpred.vpred_constraint), pattern> {
  362. let Inst{31-29} = 0b111;
  363. let Inst{27-26} = 0b11;
  364. }
  365. class MVE_f<dag oops, dag iops, InstrItinClass itin, string iname,
  366. string suffix, string ops, vpred_ops vpred, string cstr,
  367. list<dag> pattern=[]>
  368. : MVE_p<oops, iops, itin, iname, suffix, ops, vpred, cstr, pattern> {
  369. let Predicates = [HasMVEFloat];
  370. }
  371. class MVE_MI_with_pred<dag oops, dag iops, InstrItinClass itin, string asm,
  372. string ops, string cstr, list<dag> pattern>
  373. : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm, !strconcat("\t", ops), cstr,
  374. pattern>,
  375. Requires<[HasV8_1MMainline, HasMVEInt]> {
  376. let D = MVEDomain;
  377. let DecoderNamespace = "MVE";
  378. }
  379. class MVE_VMOV_lane_base<dag oops, dag iops, InstrItinClass itin, string asm,
  380. string suffix, string ops, string cstr,
  381. list<dag> pattern>
  382. : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm,
  383. !if(!eq(suffix, ""), "", "." # suffix) # "\t" # ops,
  384. cstr, pattern>,
  385. Requires<[HasV8_1MMainline, HasMVEInt]> {
  386. let D = MVEDomain;
  387. let DecoderNamespace = "MVE";
  388. }
  389. class MVE_ScalarShift<string iname, dag oops, dag iops, string asm, string cstr,
  390. list<dag> pattern=[]>
  391. : MVE_MI_with_pred<oops, iops, NoItinerary, iname, asm, cstr, pattern> {
  392. let Inst{31-20} = 0b111010100101;
  393. let Inst{8} = 0b1;
  394. let validForTailPredication=1;
  395. }
  396. class MVE_ScalarShiftSingleReg<string iname, dag iops, string asm, string cstr,
  397. list<dag> pattern=[]>
  398. : MVE_ScalarShift<iname, (outs rGPR:$RdaDest), iops, asm, cstr, pattern> {
  399. bits<4> RdaDest;
  400. let Inst{19-16} = RdaDest{3-0};
  401. }
  402. class MVE_ScalarShiftSRegImm<string iname, bits<2> op5_4>
  403. : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, long_shift:$imm),
  404. "$RdaSrc, $imm", "$RdaDest = $RdaSrc",
  405. [(set rGPR:$RdaDest,
  406. (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
  407. (i32 rGPR:$RdaSrc), (i32 imm:$imm))))]> {
  408. bits<5> imm;
  409. let Inst{15} = 0b0;
  410. let Inst{14-12} = imm{4-2};
  411. let Inst{11-8} = 0b1111;
  412. let Inst{7-6} = imm{1-0};
  413. let Inst{5-4} = op5_4{1-0};
  414. let Inst{3-0} = 0b1111;
  415. }
  416. def MVE_SQSHL : MVE_ScalarShiftSRegImm<"sqshl", 0b11>;
  417. def MVE_SRSHR : MVE_ScalarShiftSRegImm<"srshr", 0b10>;
  418. def MVE_UQSHL : MVE_ScalarShiftSRegImm<"uqshl", 0b00>;
  419. def MVE_URSHR : MVE_ScalarShiftSRegImm<"urshr", 0b01>;
  420. class MVE_ScalarShiftSRegReg<string iname, bits<2> op5_4>
  421. : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, rGPR:$Rm),
  422. "$RdaSrc, $Rm", "$RdaDest = $RdaSrc",
  423. [(set rGPR:$RdaDest,
  424. (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
  425. (i32 rGPR:$RdaSrc), (i32 rGPR:$Rm))))]> {
  426. bits<4> Rm;
  427. let Inst{15-12} = Rm{3-0};
  428. let Inst{11-8} = 0b1111;
  429. let Inst{7-6} = 0b00;
  430. let Inst{5-4} = op5_4{1-0};
  431. let Inst{3-0} = 0b1101;
  432. let Unpredictable{8-6} = 0b111;
  433. }
  434. def MVE_SQRSHR : MVE_ScalarShiftSRegReg<"sqrshr", 0b10>;
  435. def MVE_UQRSHL : MVE_ScalarShiftSRegReg<"uqrshl", 0b00>;
  436. class MVE_ScalarShiftDoubleReg<string iname, dag iops, string asm,
  437. string cstr, list<dag> pattern=[]>
  438. : MVE_ScalarShift<iname, (outs tGPREven:$RdaLo, tGPROdd:$RdaHi),
  439. iops, asm, cstr, pattern> {
  440. bits<4> RdaLo;
  441. bits<4> RdaHi;
  442. let Inst{19-17} = RdaLo{3-1};
  443. let Inst{11-9} = RdaHi{3-1};
  444. let hasSideEffects = 0;
  445. }
  446. class MVE_ScalarShiftDRegImm<string iname, bits<2> op5_4, bit op16,
  447. list<dag> pattern=[]>
  448. : MVE_ScalarShiftDoubleReg<
  449. iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, long_shift:$imm),
  450. "$RdaLo, $RdaHi, $imm", "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
  451. pattern> {
  452. bits<5> imm;
  453. let Inst{16} = op16;
  454. let Inst{15} = 0b0;
  455. let Inst{14-12} = imm{4-2};
  456. let Inst{7-6} = imm{1-0};
  457. let Inst{5-4} = op5_4{1-0};
  458. let Inst{3-0} = 0b1111;
  459. }
  460. class MVE_ScalarShiftDRegRegBase<string iname, dag iops, string asm,
  461. bit op5, bit op16, list<dag> pattern=[]>
  462. : MVE_ScalarShiftDoubleReg<
  463. iname, iops, asm, "@earlyclobber $RdaHi,@earlyclobber $RdaLo,"
  464. "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
  465. pattern> {
  466. bits<4> Rm;
  467. let Inst{16} = op16;
  468. let Inst{15-12} = Rm{3-0};
  469. let Inst{6} = 0b0;
  470. let Inst{5} = op5;
  471. let Inst{4} = 0b0;
  472. let Inst{3-0} = 0b1101;
  473. // Custom decoder method because of the following overlapping encodings:
  474. // ASRL and SQRSHR
  475. // LSLL and UQRSHL
  476. // SQRSHRL and SQRSHR
  477. // UQRSHLL and UQRSHL
  478. let DecoderMethod = "DecodeMVEOverlappingLongShift";
  479. }
  480. class MVE_ScalarShiftDRegReg<string iname, bit op5, list<dag> pattern=[]>
  481. : MVE_ScalarShiftDRegRegBase<
  482. iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm),
  483. "$RdaLo, $RdaHi, $Rm", op5, 0b0, pattern> {
  484. let Inst{7} = 0b0;
  485. }
  486. class MVE_ScalarShiftDRegRegWithSat<string iname, bit op5, list<dag> pattern=[]>
  487. : MVE_ScalarShiftDRegRegBase<
  488. iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm, saturateop:$sat),
  489. "$RdaLo, $RdaHi, $sat, $Rm", op5, 0b1, pattern> {
  490. bit sat;
  491. let Inst{7} = sat;
  492. }
  493. def MVE_ASRLr : MVE_ScalarShiftDRegReg<"asrl", 0b1, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
  494. (ARMasrl tGPREven:$RdaLo_src,
  495. tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
  496. def MVE_ASRLi : MVE_ScalarShiftDRegImm<"asrl", 0b10, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
  497. (ARMasrl tGPREven:$RdaLo_src,
  498. tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
  499. def MVE_LSLLr : MVE_ScalarShiftDRegReg<"lsll", 0b0, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
  500. (ARMlsll tGPREven:$RdaLo_src,
  501. tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
  502. def MVE_LSLLi : MVE_ScalarShiftDRegImm<"lsll", 0b00, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
  503. (ARMlsll tGPREven:$RdaLo_src,
  504. tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
  505. def MVE_LSRL : MVE_ScalarShiftDRegImm<"lsrl", 0b01, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
  506. (ARMlsrl tGPREven:$RdaLo_src,
  507. tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
  508. def MVE_SQRSHRL : MVE_ScalarShiftDRegRegWithSat<"sqrshrl", 0b1>;
  509. def MVE_SQSHLL : MVE_ScalarShiftDRegImm<"sqshll", 0b11, 0b1>;
  510. def MVE_SRSHRL : MVE_ScalarShiftDRegImm<"srshrl", 0b10, 0b1>;
  511. def MVE_UQRSHLL : MVE_ScalarShiftDRegRegWithSat<"uqrshll", 0b0>;
  512. def MVE_UQSHLL : MVE_ScalarShiftDRegImm<"uqshll", 0b00, 0b1>;
  513. def MVE_URSHRL : MVE_ScalarShiftDRegImm<"urshrl", 0b01, 0b1>;
  514. // start of mve_rDest instructions
  515. class MVE_rDest<dag oops, dag iops, InstrItinClass itin,
  516. string iname, string suffix,
  517. string ops, string cstr, list<dag> pattern=[]>
  518. // Always use vpred_n and not vpred_r: with the output register being
  519. // a GPR and not a vector register, there can't be any question of
  520. // what to put in its inactive lanes.
  521. : MVE_p<oops, iops, itin, iname, suffix, ops, vpred_n, cstr, pattern> {
  522. let Inst{25-23} = 0b101;
  523. let Inst{11-9} = 0b111;
  524. let Inst{4} = 0b0;
  525. }
  526. class MVE_VABAV<string suffix, bit U, bits<2> size>
  527. : MVE_rDest<(outs rGPR:$Rda), (ins rGPR:$Rda_src, MQPR:$Qn, MQPR:$Qm),
  528. NoItinerary, "vabav", suffix, "$Rda, $Qn, $Qm", "$Rda = $Rda_src",
  529. []> {
  530. bits<4> Qm;
  531. bits<4> Qn;
  532. bits<4> Rda;
  533. let Inst{28} = U;
  534. let Inst{22} = 0b0;
  535. let Inst{21-20} = size{1-0};
  536. let Inst{19-17} = Qn{2-0};
  537. let Inst{16} = 0b0;
  538. let Inst{15-12} = Rda{3-0};
  539. let Inst{8} = 0b1;
  540. let Inst{7} = Qn{3};
  541. let Inst{6} = 0b0;
  542. let Inst{5} = Qm{3};
  543. let Inst{3-1} = Qm{2-0};
  544. let Inst{0} = 0b1;
  545. let horizontalReduction = 1;
  546. }
  547. multiclass MVE_VABAV_m<MVEVectorVTInfo VTI> {
  548. def "" : MVE_VABAV<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  549. defvar Inst = !cast<Instruction>(NAME);
  550. let Predicates = [HasMVEInt] in {
  551. def : Pat<(i32 (int_arm_mve_vabav
  552. (i32 VTI.Unsigned),
  553. (i32 rGPR:$Rda_src),
  554. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  555. (i32 (Inst (i32 rGPR:$Rda_src),
  556. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
  557. def : Pat<(i32 (int_arm_mve_vabav_predicated
  558. (i32 VTI.Unsigned),
  559. (i32 rGPR:$Rda_src),
  560. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  561. (VTI.Pred VCCR:$mask))),
  562. (i32 (Inst (i32 rGPR:$Rda_src),
  563. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  564. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  565. }
  566. }
  567. defm MVE_VABAVs8 : MVE_VABAV_m<MVE_v16s8>;
  568. defm MVE_VABAVs16 : MVE_VABAV_m<MVE_v8s16>;
  569. defm MVE_VABAVs32 : MVE_VABAV_m<MVE_v4s32>;
  570. defm MVE_VABAVu8 : MVE_VABAV_m<MVE_v16u8>;
  571. defm MVE_VABAVu16 : MVE_VABAV_m<MVE_v8u16>;
  572. defm MVE_VABAVu32 : MVE_VABAV_m<MVE_v4u32>;
  573. class MVE_VADDV<string iname, string suffix, dag iops, string cstr,
  574. bit A, bit U, bits<2> size, list<dag> pattern=[]>
  575. : MVE_rDest<(outs tGPREven:$Rda), iops, NoItinerary,
  576. iname, suffix, "$Rda, $Qm", cstr, pattern> {
  577. bits<3> Qm;
  578. bits<4> Rda;
  579. let Inst{28} = U;
  580. let Inst{22-20} = 0b111;
  581. let Inst{19-18} = size{1-0};
  582. let Inst{17-16} = 0b01;
  583. let Inst{15-13} = Rda{3-1};
  584. let Inst{12} = 0b0;
  585. let Inst{8-6} = 0b100;
  586. let Inst{5} = A;
  587. let Inst{3-1} = Qm{2-0};
  588. let Inst{0} = 0b0;
  589. let horizontalReduction = 1;
  590. let validForTailPredication = 1;
  591. }
  592. def SDTVecReduceP : SDTypeProfile<1, 2, [ // VADDLVp
  593. SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
  594. ]>;
  595. def ARMVADDVs : SDNode<"ARMISD::VADDVs", SDTVecReduce>;
  596. def ARMVADDVu : SDNode<"ARMISD::VADDVu", SDTVecReduce>;
  597. def ARMVADDVps : SDNode<"ARMISD::VADDVps", SDTVecReduceP>;
  598. def ARMVADDVpu : SDNode<"ARMISD::VADDVpu", SDTVecReduceP>;
  599. multiclass MVE_VADDV_A<MVEVectorVTInfo VTI> {
  600. def acc : MVE_VADDV<"vaddva", VTI.Suffix,
  601. (ins tGPREven:$Rda_src, MQPR:$Qm), "$Rda = $Rda_src",
  602. 0b1, VTI.Unsigned, VTI.Size>;
  603. def no_acc : MVE_VADDV<"vaddv", VTI.Suffix,
  604. (ins MQPR:$Qm), "",
  605. 0b0, VTI.Unsigned, VTI.Size>;
  606. defvar InstA = !cast<Instruction>(NAME # "acc");
  607. defvar InstN = !cast<Instruction>(NAME # "no_acc");
  608. let Predicates = [HasMVEInt] in {
  609. if VTI.Unsigned then {
  610. def : Pat<(i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
  611. (i32 (InstN $vec))>;
  612. def : Pat<(i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  613. (VTI.Vec MQPR:$vec),
  614. (VTI.Vec ARMimmAllZerosV))))),
  615. (i32 (InstN $vec, ARMVCCThen, $pred))>;
  616. def : Pat<(i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
  617. (i32 (InstN $vec))>;
  618. def : Pat<(i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
  619. (i32 (InstN $vec, ARMVCCThen, $pred))>;
  620. def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
  621. (i32 tGPREven:$acc))),
  622. (i32 (InstA $acc, $vec))>;
  623. def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  624. (VTI.Vec MQPR:$vec),
  625. (VTI.Vec ARMimmAllZerosV))))),
  626. (i32 tGPREven:$acc))),
  627. (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
  628. def : Pat<(i32 (add (i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
  629. (i32 tGPREven:$acc))),
  630. (i32 (InstA $acc, $vec))>;
  631. def : Pat<(i32 (add (i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
  632. (i32 tGPREven:$acc))),
  633. (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
  634. } else {
  635. def : Pat<(i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
  636. (i32 (InstN $vec))>;
  637. def : Pat<(i32 (add (i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
  638. (i32 tGPREven:$acc))),
  639. (i32 (InstA $acc, $vec))>;
  640. def : Pat<(i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
  641. (i32 (InstN $vec, ARMVCCThen, $pred))>;
  642. def : Pat<(i32 (add (i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
  643. (i32 tGPREven:$acc))),
  644. (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
  645. }
  646. def : Pat<(i32 (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
  647. (i32 VTI.Unsigned),
  648. (VTI.Pred VCCR:$pred))),
  649. (i32 (InstN $vec, ARMVCCThen, $pred))>;
  650. def : Pat<(i32 (add (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
  651. (i32 VTI.Unsigned),
  652. (VTI.Pred VCCR:$pred)),
  653. (i32 tGPREven:$acc))),
  654. (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
  655. }
  656. }
  657. defm MVE_VADDVs8 : MVE_VADDV_A<MVE_v16s8>;
  658. defm MVE_VADDVs16 : MVE_VADDV_A<MVE_v8s16>;
  659. defm MVE_VADDVs32 : MVE_VADDV_A<MVE_v4s32>;
  660. defm MVE_VADDVu8 : MVE_VADDV_A<MVE_v16u8>;
  661. defm MVE_VADDVu16 : MVE_VADDV_A<MVE_v8u16>;
  662. defm MVE_VADDVu32 : MVE_VADDV_A<MVE_v4u32>;
  663. class MVE_VADDLV<string iname, string suffix, dag iops, string cstr,
  664. bit A, bit U, list<dag> pattern=[]>
  665. : MVE_rDest<(outs tGPREven:$RdaLo, tGPROdd:$RdaHi), iops, NoItinerary, iname,
  666. suffix, "$RdaLo, $RdaHi, $Qm", cstr, pattern> {
  667. bits<3> Qm;
  668. bits<4> RdaLo;
  669. bits<4> RdaHi;
  670. let Inst{28} = U;
  671. let Inst{22-20} = RdaHi{3-1};
  672. let Inst{19-18} = 0b10;
  673. let Inst{17-16} = 0b01;
  674. let Inst{15-13} = RdaLo{3-1};
  675. let Inst{12} = 0b0;
  676. let Inst{8-6} = 0b100;
  677. let Inst{5} = A;
  678. let Inst{3-1} = Qm{2-0};
  679. let Inst{0} = 0b0;
  680. let horizontalReduction = 1;
  681. }
  682. def SDTVecReduceL : SDTypeProfile<2, 1, [ // VADDLV
  683. SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
  684. ]>;
  685. def SDTVecReduceLA : SDTypeProfile<2, 3, [ // VADDLVA
  686. SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
  687. SDTCisVec<4>
  688. ]>;
  689. def SDTVecReduceLP : SDTypeProfile<2, 2, [ // VADDLVp
  690. SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<2>
  691. ]>;
  692. def SDTVecReduceLPA : SDTypeProfile<2, 4, [ // VADDLVAp
  693. SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
  694. SDTCisVec<4>, SDTCisVec<5>
  695. ]>;
  696. multiclass MVE_VADDLV_A<MVEVectorVTInfo VTI> {
  697. def acc : MVE_VADDLV<"vaddlva", VTI.Suffix,
  698. (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, MQPR:$Qm),
  699. "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
  700. 0b1, VTI.Unsigned>;
  701. def no_acc : MVE_VADDLV<"vaddlv", VTI.Suffix,
  702. (ins MQPR:$Qm), "",
  703. 0b0, VTI.Unsigned>;
  704. defvar InstA = !cast<Instruction>(NAME # "acc");
  705. defvar InstN = !cast<Instruction>(NAME # "no_acc");
  706. defvar letter = VTI.SuffixLetter;
  707. defvar ARMVADDLV = SDNode<"ARMISD::VADDLV" # letter, SDTVecReduceL>;
  708. defvar ARMVADDLVA = SDNode<"ARMISD::VADDLVA" # letter, SDTVecReduceLA>;
  709. defvar ARMVADDLVp = SDNode<"ARMISD::VADDLVp" # letter, SDTVecReduceLP>;
  710. defvar ARMVADDLVAp = SDNode<"ARMISD::VADDLVAp" # letter, SDTVecReduceLPA>;
  711. let Predicates = [HasMVEInt] in {
  712. def : Pat<(ARMVADDLV (v4i32 MQPR:$vec)),
  713. (InstN (v4i32 MQPR:$vec))>;
  714. def : Pat<(ARMVADDLVA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec)),
  715. (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec))>;
  716. def : Pat<(ARMVADDLVp (v4i32 MQPR:$vec), (VTI.Pred VCCR:$pred)),
  717. (InstN (v4i32 MQPR:$vec), ARMVCCThen, (VTI.Pred VCCR:$pred))>;
  718. def : Pat<(ARMVADDLVAp tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
  719. (VTI.Pred VCCR:$pred)),
  720. (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
  721. ARMVCCThen, (VTI.Pred VCCR:$pred))>;
  722. }
  723. }
  724. defm MVE_VADDLVs32 : MVE_VADDLV_A<MVE_v4s32>;
  725. defm MVE_VADDLVu32 : MVE_VADDLV_A<MVE_v4u32>;
  726. class MVE_VMINMAXNMV<string iname, string suffix, bit sz,
  727. bit bit_17, bit bit_7, list<dag> pattern=[]>
  728. : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm),
  729. NoItinerary, iname, suffix, "$RdaSrc, $Qm",
  730. "$RdaDest = $RdaSrc", pattern> {
  731. bits<3> Qm;
  732. bits<4> RdaDest;
  733. let Inst{28} = sz;
  734. let Inst{22-20} = 0b110;
  735. let Inst{19-18} = 0b11;
  736. let Inst{17} = bit_17;
  737. let Inst{16} = 0b0;
  738. let Inst{15-12} = RdaDest{3-0};
  739. let Inst{8} = 0b1;
  740. let Inst{7} = bit_7;
  741. let Inst{6-5} = 0b00;
  742. let Inst{3-1} = Qm{2-0};
  743. let Inst{0} = 0b0;
  744. let horizontalReduction = 1;
  745. let Predicates = [HasMVEFloat];
  746. let hasSideEffects = 0;
  747. }
  748. multiclass MVE_VMINMAXNMV_p<string iname, bit notAbs, bit isMin,
  749. MVEVectorVTInfo VTI, string intrBaseName,
  750. ValueType Scalar, RegisterClass ScalarReg> {
  751. def "": MVE_VMINMAXNMV<iname, VTI.Suffix, VTI.Size{0}, notAbs, isMin>;
  752. defvar Inst = !cast<Instruction>(NAME);
  753. defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
  754. defvar pred_intr = !cast<Intrinsic>(intrBaseName#"_predicated");
  755. let Predicates = [HasMVEFloat] in {
  756. def : Pat<(Scalar (unpred_intr (Scalar ScalarReg:$prev),
  757. (VTI.Vec MQPR:$vec))),
  758. (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
  759. (VTI.Vec MQPR:$vec)),
  760. ScalarReg)>;
  761. def : Pat<(Scalar (pred_intr (Scalar ScalarReg:$prev),
  762. (VTI.Vec MQPR:$vec),
  763. (VTI.Pred VCCR:$pred))),
  764. (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
  765. (VTI.Vec MQPR:$vec),
  766. ARMVCCThen, (VTI.Pred VCCR:$pred)),
  767. ScalarReg)>;
  768. }
  769. }
  770. multiclass MVE_VMINMAXNMV_fty<string iname, bit notAbs, bit isMin,
  771. string intrBase> {
  772. defm f32 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v4f32, intrBase,
  773. f32, SPR>;
  774. defm f16 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v8f16, intrBase,
  775. f16, HPR>;
  776. }
  777. defm MVE_VMINNMV : MVE_VMINMAXNMV_fty<"vminnmv", 1, 1, "int_arm_mve_minnmv">;
  778. defm MVE_VMAXNMV : MVE_VMINMAXNMV_fty<"vmaxnmv", 1, 0, "int_arm_mve_maxnmv">;
  779. defm MVE_VMINNMAV: MVE_VMINMAXNMV_fty<"vminnmav", 0, 1, "int_arm_mve_minnmav">;
  780. defm MVE_VMAXNMAV: MVE_VMINMAXNMV_fty<"vmaxnmav", 0, 0, "int_arm_mve_maxnmav">;
  781. class MVE_VMINMAXV<string iname, string suffix, bit U, bits<2> size,
  782. bit bit_17, bit bit_7, list<dag> pattern=[]>
  783. : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm), NoItinerary,
  784. iname, suffix, "$RdaSrc, $Qm", "$RdaDest = $RdaSrc", pattern> {
  785. bits<3> Qm;
  786. bits<4> RdaDest;
  787. let Inst{28} = U;
  788. let Inst{22-20} = 0b110;
  789. let Inst{19-18} = size{1-0};
  790. let Inst{17} = bit_17;
  791. let Inst{16} = 0b0;
  792. let Inst{15-12} = RdaDest{3-0};
  793. let Inst{8} = 0b1;
  794. let Inst{7} = bit_7;
  795. let Inst{6-5} = 0b00;
  796. let Inst{3-1} = Qm{2-0};
  797. let Inst{0} = 0b0;
  798. let horizontalReduction = 1;
  799. }
  800. multiclass MVE_VMINMAXV_p<string iname, bit notAbs, bit isMin,
  801. MVEVectorVTInfo VTI, string intrBaseName> {
  802. def "": MVE_VMINMAXV<iname, VTI.Suffix, VTI.Unsigned, VTI.Size,
  803. notAbs, isMin>;
  804. defvar Inst = !cast<Instruction>(NAME);
  805. defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
  806. defvar pred_intr = !cast<Intrinsic>(intrBaseName#"_predicated");
  807. defvar base_args = (? (i32 rGPR:$prev), (VTI.Vec MQPR:$vec));
  808. defvar args = !if(notAbs, !con(base_args, (? (i32 VTI.Unsigned))),
  809. base_args);
  810. let Predicates = [HasMVEInt] in {
  811. def : Pat<(i32 !con(args, (unpred_intr))),
  812. (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec)))>;
  813. def : Pat<(i32 !con(args, (pred_intr (VTI.Pred VCCR:$pred)))),
  814. (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec),
  815. ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
  816. }
  817. }
  818. multiclass MVE_VMINMAXV_ty<string iname, bit isMin, string intrBaseName> {
  819. defm s8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16s8, intrBaseName>;
  820. defm s16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8s16, intrBaseName>;
  821. defm s32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4s32, intrBaseName>;
  822. defm u8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16u8, intrBaseName>;
  823. defm u16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8u16, intrBaseName>;
  824. defm u32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4u32, intrBaseName>;
  825. }
  826. def SDTVecReduceR : SDTypeProfile<1, 2, [ // Reduction of an integer and vector into an integer
  827. SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
  828. ]>;
  829. def ARMVMINVu : SDNode<"ARMISD::VMINVu", SDTVecReduceR>;
  830. def ARMVMINVs : SDNode<"ARMISD::VMINVs", SDTVecReduceR>;
  831. def ARMVMAXVu : SDNode<"ARMISD::VMAXVu", SDTVecReduceR>;
  832. def ARMVMAXVs : SDNode<"ARMISD::VMAXVs", SDTVecReduceR>;
  833. defm MVE_VMINV : MVE_VMINMAXV_ty<"vminv", 1, "int_arm_mve_minv">;
  834. defm MVE_VMAXV : MVE_VMINMAXV_ty<"vmaxv", 0, "int_arm_mve_maxv">;
  835. let Predicates = [HasMVEInt] in {
  836. def : Pat<(i32 (vecreduce_smax (v16i8 MQPR:$src))),
  837. (i32 (MVE_VMAXVs8 (t2MVNi (i32 127)), $src))>;
  838. def : Pat<(i32 (vecreduce_smax (v8i16 MQPR:$src))),
  839. (i32 (MVE_VMAXVs16 (t2MOVi32imm (i32 -32768)), $src))>;
  840. def : Pat<(i32 (vecreduce_smax (v4i32 MQPR:$src))),
  841. (i32 (MVE_VMAXVs32 (t2MOVi (i32 -2147483648)), $src))>;
  842. def : Pat<(i32 (vecreduce_umax (v16i8 MQPR:$src))),
  843. (i32 (MVE_VMAXVu8 (t2MOVi (i32 0)), $src))>;
  844. def : Pat<(i32 (vecreduce_umax (v8i16 MQPR:$src))),
  845. (i32 (MVE_VMAXVu16 (t2MOVi (i32 0)), $src))>;
  846. def : Pat<(i32 (vecreduce_umax (v4i32 MQPR:$src))),
  847. (i32 (MVE_VMAXVu32 (t2MOVi (i32 0)), $src))>;
  848. def : Pat<(i32 (vecreduce_smin (v16i8 MQPR:$src))),
  849. (i32 (MVE_VMINVs8 (t2MOVi (i32 127)), $src))>;
  850. def : Pat<(i32 (vecreduce_smin (v8i16 MQPR:$src))),
  851. (i32 (MVE_VMINVs16 (t2MOVi16 (i32 32767)), $src))>;
  852. def : Pat<(i32 (vecreduce_smin (v4i32 MQPR:$src))),
  853. (i32 (MVE_VMINVs32 (t2MVNi (i32 -2147483648)), $src))>;
  854. def : Pat<(i32 (vecreduce_umin (v16i8 MQPR:$src))),
  855. (i32 (MVE_VMINVu8 (t2MOVi (i32 255)), $src))>;
  856. def : Pat<(i32 (vecreduce_umin (v8i16 MQPR:$src))),
  857. (i32 (MVE_VMINVu16 (t2MOVi16 (i32 65535)), $src))>;
  858. def : Pat<(i32 (vecreduce_umin (v4i32 MQPR:$src))),
  859. (i32 (MVE_VMINVu32 (t2MOVi (i32 4294967295)), $src))>;
  860. def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
  861. (i32 (MVE_VMINVu8 $x, $src))>;
  862. def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
  863. (i32 (MVE_VMINVu16 $x, $src))>;
  864. def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
  865. (i32 (MVE_VMINVu32 $x, $src))>;
  866. def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
  867. (i32 (MVE_VMINVs8 $x, $src))>;
  868. def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
  869. (i32 (MVE_VMINVs16 $x, $src))>;
  870. def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
  871. (i32 (MVE_VMINVs32 $x, $src))>;
  872. def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
  873. (i32 (MVE_VMAXVu8 $x, $src))>;
  874. def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
  875. (i32 (MVE_VMAXVu16 $x, $src))>;
  876. def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
  877. (i32 (MVE_VMAXVu32 $x, $src))>;
  878. def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
  879. (i32 (MVE_VMAXVs8 $x, $src))>;
  880. def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
  881. (i32 (MVE_VMAXVs16 $x, $src))>;
  882. def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
  883. (i32 (MVE_VMAXVs32 $x, $src))>;
  884. }
  885. multiclass MVE_VMINMAXAV_ty<string iname, bit isMin, string intrBaseName> {
  886. defm s8 : MVE_VMINMAXV_p<iname, 0, isMin, MVE_v16s8, intrBaseName>;
  887. defm s16: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v8s16, intrBaseName>;
  888. defm s32: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v4s32, intrBaseName>;
  889. }
  890. defm MVE_VMINAV : MVE_VMINMAXAV_ty<"vminav", 1, "int_arm_mve_minav">;
  891. defm MVE_VMAXAV : MVE_VMINMAXAV_ty<"vmaxav", 0, "int_arm_mve_maxav">;
  892. class MVE_VMLAMLSDAV<string iname, string suffix, dag iops, string cstr,
  893. bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0>
  894. : MVE_rDest<(outs tGPREven:$RdaDest), iops, NoItinerary, iname, suffix,
  895. "$RdaDest, $Qn, $Qm", cstr, []> {
  896. bits<4> RdaDest;
  897. bits<3> Qm;
  898. bits<3> Qn;
  899. let Inst{28} = bit_28;
  900. let Inst{22-20} = 0b111;
  901. let Inst{19-17} = Qn{2-0};
  902. let Inst{16} = sz;
  903. let Inst{15-13} = RdaDest{3-1};
  904. let Inst{12} = X;
  905. let Inst{8} = bit_8;
  906. let Inst{7-6} = 0b00;
  907. let Inst{5} = A;
  908. let Inst{3-1} = Qm{2-0};
  909. let Inst{0} = bit_0;
  910. let horizontalReduction = 1;
  911. // Allow tail predication for non-exchanging versions. As this is also a
  912. // horizontalReduction, ARMLowOverheadLoops will also have to check that
  913. // the vector operands contain zeros in their false lanes for the instruction
  914. // to be properly valid.
  915. let validForTailPredication = !eq(X, 0);
  916. }
  917. multiclass MVE_VMLAMLSDAV_A<string iname, string x, MVEVectorVTInfo VTI,
  918. bit sz, bit bit_28, bit X, bit bit_8, bit bit_0> {
  919. def ""#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # x, VTI.Suffix,
  920. (ins MQPR:$Qn, MQPR:$Qm), "",
  921. sz, bit_28, 0b0, X, bit_8, bit_0>;
  922. def "a"#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # "a" # x, VTI.Suffix,
  923. (ins tGPREven:$RdaSrc, MQPR:$Qn, MQPR:$Qm),
  924. "$RdaDest = $RdaSrc",
  925. sz, bit_28, 0b1, X, bit_8, bit_0>;
  926. let Predicates = [HasMVEInt] in {
  927. def : Pat<(i32 (int_arm_mve_vmldava
  928. (i32 VTI.Unsigned),
  929. (i32 bit_0) /* subtract */,
  930. (i32 X) /* exchange */,
  931. (i32 0) /* accumulator */,
  932. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  933. (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
  934. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
  935. def : Pat<(i32 (int_arm_mve_vmldava_predicated
  936. (i32 VTI.Unsigned),
  937. (i32 bit_0) /* subtract */,
  938. (i32 X) /* exchange */,
  939. (i32 0) /* accumulator */,
  940. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  941. (VTI.Pred VCCR:$mask))),
  942. (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
  943. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  944. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  945. def : Pat<(i32 (int_arm_mve_vmldava
  946. (i32 VTI.Unsigned),
  947. (i32 bit_0) /* subtract */,
  948. (i32 X) /* exchange */,
  949. (i32 tGPREven:$RdaSrc),
  950. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  951. (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
  952. (i32 tGPREven:$RdaSrc),
  953. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
  954. def : Pat<(i32 (int_arm_mve_vmldava_predicated
  955. (i32 VTI.Unsigned),
  956. (i32 bit_0) /* subtract */,
  957. (i32 X) /* exchange */,
  958. (i32 tGPREven:$RdaSrc),
  959. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  960. (VTI.Pred VCCR:$mask))),
  961. (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
  962. (i32 tGPREven:$RdaSrc),
  963. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  964. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  965. }
  966. }
  967. multiclass MVE_VMLAMLSDAV_AX<string iname, MVEVectorVTInfo VTI, bit sz,
  968. bit bit_28, bit bit_8, bit bit_0> {
  969. defm "" : MVE_VMLAMLSDAV_A<iname, "", VTI, sz, bit_28,
  970. 0b0, bit_8, bit_0>;
  971. defm "" : MVE_VMLAMLSDAV_A<iname, "x", VTI, sz, bit_28,
  972. 0b1, bit_8, bit_0>;
  973. }
  974. multiclass MVE_VMLADAV_multi<MVEVectorVTInfo SVTI, MVEVectorVTInfo UVTI,
  975. bit sz, bit bit_8> {
  976. defm "" : MVE_VMLAMLSDAV_AX<"vmladav", SVTI,
  977. sz, 0b0, bit_8, 0b0>;
  978. defm "" : MVE_VMLAMLSDAV_A<"vmladav", "", UVTI,
  979. sz, 0b1, 0b0, bit_8, 0b0>;
  980. }
  981. multiclass MVE_VMLSDAV_multi<MVEVectorVTInfo VTI, bit sz, bit bit_28> {
  982. defm "" : MVE_VMLAMLSDAV_AX<"vmlsdav", VTI,
  983. sz, bit_28, 0b0, 0b1>;
  984. }
  985. defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v16s8, MVE_v16u8, 0b0, 0b1>;
  986. defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v8s16, MVE_v8u16, 0b0, 0b0>;
  987. defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v4s32, MVE_v4u32, 0b1, 0b0>;
  988. defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v16s8, 0b0, 0b1>;
  989. defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v8s16, 0b0, 0b0>;
  990. defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v4s32, 0b1, 0b0>;
  991. def SDTVecReduce2 : SDTypeProfile<1, 2, [ // VMLAV
  992. SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
  993. ]>;
  994. def SDTVecReduce2L : SDTypeProfile<2, 2, [ // VMLALV
  995. SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>
  996. ]>;
  997. def SDTVecReduce2LA : SDTypeProfile<2, 4, [ // VMLALVA
  998. SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
  999. SDTCisVec<4>, SDTCisVec<5>
  1000. ]>;
  1001. def SDTVecReduce2P : SDTypeProfile<1, 3, [ // VMLAV
  1002. SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>, SDTCisVec<3>
  1003. ]>;
  1004. def SDTVecReduce2LP : SDTypeProfile<2, 3, [ // VMLALV
  1005. SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>, SDTCisVec<4>
  1006. ]>;
  1007. def SDTVecReduce2LAP : SDTypeProfile<2, 5, [ // VMLALVA
  1008. SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
  1009. SDTCisVec<4>, SDTCisVec<5>, SDTCisVec<6>
  1010. ]>;
  1011. def ARMVMLAVs : SDNode<"ARMISD::VMLAVs", SDTVecReduce2>;
  1012. def ARMVMLAVu : SDNode<"ARMISD::VMLAVu", SDTVecReduce2>;
  1013. def ARMVMLALVs : SDNode<"ARMISD::VMLALVs", SDTVecReduce2L>;
  1014. def ARMVMLALVu : SDNode<"ARMISD::VMLALVu", SDTVecReduce2L>;
  1015. def ARMVMLALVAs : SDNode<"ARMISD::VMLALVAs", SDTVecReduce2LA>;
  1016. def ARMVMLALVAu : SDNode<"ARMISD::VMLALVAu", SDTVecReduce2LA>;
  1017. def ARMVMLAVps : SDNode<"ARMISD::VMLAVps", SDTVecReduce2P>;
  1018. def ARMVMLAVpu : SDNode<"ARMISD::VMLAVpu", SDTVecReduce2P>;
  1019. def ARMVMLALVps : SDNode<"ARMISD::VMLALVps", SDTVecReduce2LP>;
  1020. def ARMVMLALVpu : SDNode<"ARMISD::VMLALVpu", SDTVecReduce2LP>;
  1021. def ARMVMLALVAps : SDNode<"ARMISD::VMLALVAps", SDTVecReduce2LAP>;
  1022. def ARMVMLALVApu : SDNode<"ARMISD::VMLALVApu", SDTVecReduce2LAP>;
  1023. let Predicates = [HasMVEInt] in {
  1024. def : Pat<(i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
  1025. (i32 (MVE_VMLADAVu32 $src1, $src2))>;
  1026. def : Pat<(i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
  1027. (i32 (MVE_VMLADAVu16 $src1, $src2))>;
  1028. def : Pat<(i32 (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
  1029. (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
  1030. def : Pat<(i32 (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
  1031. (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
  1032. def : Pat<(i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
  1033. (i32 (MVE_VMLADAVu8 $src1, $src2))>;
  1034. def : Pat<(i32 (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
  1035. (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
  1036. def : Pat<(i32 (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
  1037. (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
  1038. def : Pat<(i32 (add (i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
  1039. (i32 tGPREven:$src3))),
  1040. (i32 (MVE_VMLADAVau32 $src3, $src1, $src2))>;
  1041. def : Pat<(i32 (add (i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
  1042. (i32 tGPREven:$src3))),
  1043. (i32 (MVE_VMLADAVau16 $src3, $src1, $src2))>;
  1044. def : Pat<(i32 (add (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
  1045. (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
  1046. def : Pat<(i32 (add (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
  1047. (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
  1048. def : Pat<(i32 (add (i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
  1049. (i32 tGPREven:$src3))),
  1050. (i32 (MVE_VMLADAVau8 $src3, $src1, $src2))>;
  1051. def : Pat<(i32 (add (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
  1052. (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
  1053. def : Pat<(i32 (add (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
  1054. (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
  1055. // Predicated
  1056. def : Pat<(i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
  1057. (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
  1058. (v4i32 ARMimmAllZerosV)))),
  1059. (i32 (MVE_VMLADAVu32 $src1, $src2, ARMVCCThen, $pred))>;
  1060. def : Pat<(i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
  1061. (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
  1062. (v8i16 ARMimmAllZerosV)))),
  1063. (i32 (MVE_VMLADAVu16 $src1, $src2, ARMVCCThen, $pred))>;
  1064. def : Pat<(i32 (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
  1065. (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
  1066. def : Pat<(i32 (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
  1067. (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
  1068. def : Pat<(i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
  1069. (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
  1070. (v16i8 ARMimmAllZerosV)))),
  1071. (i32 (MVE_VMLADAVu8 $src1, $src2, ARMVCCThen, $pred))>;
  1072. def : Pat<(i32 (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
  1073. (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
  1074. def : Pat<(i32 (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
  1075. (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
  1076. def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
  1077. (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
  1078. (v4i32 ARMimmAllZerosV)))),
  1079. (i32 tGPREven:$src3))),
  1080. (i32 (MVE_VMLADAVau32 $src3, $src1, $src2, ARMVCCThen, $pred))>;
  1081. def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
  1082. (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
  1083. (v8i16 ARMimmAllZerosV)))),
  1084. (i32 tGPREven:$src3))),
  1085. (i32 (MVE_VMLADAVau16 $src3, $src1, $src2, ARMVCCThen, $pred))>;
  1086. def : Pat<(i32 (add (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
  1087. (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
  1088. def : Pat<(i32 (add (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
  1089. (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
  1090. def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
  1091. (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
  1092. (v16i8 ARMimmAllZerosV)))),
  1093. (i32 tGPREven:$src3))),
  1094. (i32 (MVE_VMLADAVau8 $src3, $src1, $src2, ARMVCCThen, $pred))>;
  1095. def : Pat<(i32 (add (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
  1096. (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
  1097. def : Pat<(i32 (add (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
  1098. (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
  1099. }
  1100. // vmlav aliases vmladav
  1101. foreach acc = ["", "a"] in {
  1102. foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32"] in {
  1103. def : MVEInstAlias<"vmlav"#acc#"${vp}."#suffix#"\t$RdaDest, $Qn, $Qm",
  1104. (!cast<Instruction>("MVE_VMLADAV"#acc#suffix)
  1105. tGPREven:$RdaDest, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1106. }
  1107. }
  1108. // Base class for VMLALDAV and VMLSLDAV, VRMLALDAVH, VRMLSLDAVH
  1109. class MVE_VMLALDAVBase<string iname, string suffix, dag iops, string cstr,
  1110. bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0,
  1111. list<dag> pattern=[]>
  1112. : MVE_rDest<(outs tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest), iops, NoItinerary,
  1113. iname, suffix, "$RdaLoDest, $RdaHiDest, $Qn, $Qm", cstr, pattern> {
  1114. bits<4> RdaLoDest;
  1115. bits<4> RdaHiDest;
  1116. bits<3> Qm;
  1117. bits<3> Qn;
  1118. let Inst{28} = bit_28;
  1119. let Inst{22-20} = RdaHiDest{3-1};
  1120. let Inst{19-17} = Qn{2-0};
  1121. let Inst{16} = sz;
  1122. let Inst{15-13} = RdaLoDest{3-1};
  1123. let Inst{12} = X;
  1124. let Inst{8} = bit_8;
  1125. let Inst{7-6} = 0b00;
  1126. let Inst{5} = A;
  1127. let Inst{3-1} = Qm{2-0};
  1128. let Inst{0} = bit_0;
  1129. let horizontalReduction = 1;
  1130. // Allow tail predication for non-exchanging versions. As this is also a
  1131. // horizontalReduction, ARMLowOverheadLoops will also have to check that
  1132. // the vector operands contain zeros in their false lanes for the instruction
  1133. // to be properly valid.
  1134. let validForTailPredication = !eq(X, 0);
  1135. let hasSideEffects = 0;
  1136. }
  1137. multiclass MVE_VMLALDAVBase_A<string iname, string x, string suffix,
  1138. bit sz, bit bit_28, bit X, bit bit_8, bit bit_0,
  1139. list<dag> pattern=[]> {
  1140. def ""#x#suffix : MVE_VMLALDAVBase<
  1141. iname # x, suffix, (ins MQPR:$Qn, MQPR:$Qm), "",
  1142. sz, bit_28, 0b0, X, bit_8, bit_0, pattern>;
  1143. def "a"#x#suffix : MVE_VMLALDAVBase<
  1144. iname # "a" # x, suffix,
  1145. (ins tGPREven:$RdaLoSrc, tGPROdd:$RdaHiSrc, MQPR:$Qn, MQPR:$Qm),
  1146. "$RdaLoDest = $RdaLoSrc,$RdaHiDest = $RdaHiSrc",
  1147. sz, bit_28, 0b1, X, bit_8, bit_0, pattern>;
  1148. }
  1149. multiclass MVE_VMLALDAVBase_AX<string iname, string suffix, bit sz, bit bit_28,
  1150. bit bit_8, bit bit_0, list<dag> pattern=[]> {
  1151. defm "" : MVE_VMLALDAVBase_A<iname, "", suffix, sz,
  1152. bit_28, 0b0, bit_8, bit_0, pattern>;
  1153. defm "" : MVE_VMLALDAVBase_A<iname, "x", suffix, sz,
  1154. bit_28, 0b1, bit_8, bit_0, pattern>;
  1155. }
  1156. multiclass MVE_VRMLALDAVH_multi<string suffix, list<dag> pattern=[]> {
  1157. defm "" : MVE_VMLALDAVBase_AX<"vrmlaldavh", "s"#suffix,
  1158. 0b0, 0b0, 0b1, 0b0, pattern>;
  1159. defm "" : MVE_VMLALDAVBase_A<"vrmlaldavh", "", "u"#suffix,
  1160. 0b0, 0b1, 0b0, 0b1, 0b0, pattern>;
  1161. }
  1162. defm MVE_VRMLALDAVH : MVE_VRMLALDAVH_multi<"32">;
  1163. // vrmlalvh aliases for vrmlaldavh
  1164. def : MVEInstAlias<"vrmlalvh${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
  1165. (MVE_VRMLALDAVHs32
  1166. tGPREven:$RdaLo, tGPROdd:$RdaHi,
  1167. MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1168. def : MVEInstAlias<"vrmlalvha${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
  1169. (MVE_VRMLALDAVHas32
  1170. tGPREven:$RdaLo, tGPROdd:$RdaHi,
  1171. MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1172. def : MVEInstAlias<"vrmlalvh${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
  1173. (MVE_VRMLALDAVHu32
  1174. tGPREven:$RdaLo, tGPROdd:$RdaHi,
  1175. MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1176. def : MVEInstAlias<"vrmlalvha${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
  1177. (MVE_VRMLALDAVHau32
  1178. tGPREven:$RdaLo, tGPROdd:$RdaHi,
  1179. MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1180. multiclass MVE_VMLALDAV_multi<string suffix, bit sz, list<dag> pattern=[]> {
  1181. defm "" : MVE_VMLALDAVBase_AX<"vmlaldav", "s"#suffix, sz, 0b0, 0b0, 0b0, pattern>;
  1182. defm "" : MVE_VMLALDAVBase_A<"vmlaldav", "", "u"#suffix,
  1183. sz, 0b1, 0b0, 0b0, 0b0, pattern>;
  1184. }
  1185. defm MVE_VMLALDAV : MVE_VMLALDAV_multi<"16", 0b0>;
  1186. defm MVE_VMLALDAV : MVE_VMLALDAV_multi<"32", 0b1>;
  1187. let Predicates = [HasMVEInt] in {
  1188. def : Pat<(ARMVMLALVs (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
  1189. (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
  1190. def : Pat<(ARMVMLALVu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
  1191. (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
  1192. def : Pat<(ARMVMLALVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
  1193. (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
  1194. def : Pat<(ARMVMLALVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
  1195. (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
  1196. def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
  1197. (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
  1198. def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
  1199. (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
  1200. def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
  1201. (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
  1202. def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
  1203. (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
  1204. // Predicated
  1205. def : Pat<(ARMVMLALVps (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
  1206. (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
  1207. def : Pat<(ARMVMLALVpu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
  1208. (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
  1209. def : Pat<(ARMVMLALVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
  1210. (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
  1211. def : Pat<(ARMVMLALVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
  1212. (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
  1213. def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
  1214. (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
  1215. def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
  1216. (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
  1217. def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
  1218. (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
  1219. def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
  1220. (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
  1221. }
  1222. // vmlalv aliases vmlaldav
  1223. foreach acc = ["", "a"] in {
  1224. foreach suffix = ["s16", "s32", "u16", "u32"] in {
  1225. def : MVEInstAlias<"vmlalv" # acc # "${vp}." # suffix #
  1226. "\t$RdaLoDest, $RdaHiDest, $Qn, $Qm",
  1227. (!cast<Instruction>("MVE_VMLALDAV"#acc#suffix)
  1228. tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest,
  1229. MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  1230. }
  1231. }
  1232. multiclass MVE_VMLSLDAV_multi<string iname, string suffix, bit sz,
  1233. bit bit_28, list<dag> pattern=[]> {
  1234. defm "" : MVE_VMLALDAVBase_AX<iname, suffix, sz, bit_28, 0b0, 0b1, pattern>;
  1235. }
  1236. defm MVE_VMLSLDAV : MVE_VMLSLDAV_multi<"vmlsldav", "s16", 0b0, 0b0>;
  1237. defm MVE_VMLSLDAV : MVE_VMLSLDAV_multi<"vmlsldav", "s32", 0b1, 0b0>;
  1238. defm MVE_VRMLSLDAVH : MVE_VMLSLDAV_multi<"vrmlsldavh", "s32", 0b0, 0b1>;
  1239. // end of mve_rDest instructions
  1240. // start of mve_comp instructions
  1241. class MVE_comp<InstrItinClass itin, string iname, string suffix,
  1242. string cstr, list<dag> pattern=[]>
  1243. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), itin, iname, suffix,
  1244. "$Qd, $Qn, $Qm", vpred_r, cstr, pattern> {
  1245. bits<4> Qd;
  1246. bits<4> Qn;
  1247. bits<4> Qm;
  1248. let Inst{22} = Qd{3};
  1249. let Inst{19-17} = Qn{2-0};
  1250. let Inst{16} = 0b0;
  1251. let Inst{15-13} = Qd{2-0};
  1252. let Inst{12} = 0b0;
  1253. let Inst{10-9} = 0b11;
  1254. let Inst{7} = Qn{3};
  1255. let Inst{5} = Qm{3};
  1256. let Inst{3-1} = Qm{2-0};
  1257. let Inst{0} = 0b0;
  1258. }
  1259. class MVE_VMINMAXNM<string iname, string suffix, bit sz, bit bit_21,
  1260. list<dag> pattern=[]>
  1261. : MVE_comp<NoItinerary, iname, suffix, "", pattern> {
  1262. let Inst{28} = 0b1;
  1263. let Inst{25-24} = 0b11;
  1264. let Inst{23} = 0b0;
  1265. let Inst{21} = bit_21;
  1266. let Inst{20} = sz;
  1267. let Inst{11} = 0b1;
  1268. let Inst{8} = 0b1;
  1269. let Inst{6} = 0b1;
  1270. let Inst{4} = 0b1;
  1271. let Predicates = [HasMVEFloat];
  1272. }
  1273. multiclass MVE_VMINMAXNM_m<string iname, bit bit_4, MVEVectorVTInfo VTI, SDNode Op, Intrinsic PredInt> {
  1274. def "" : MVE_VMINMAXNM<iname, VTI.Suffix, VTI.Size{0}, bit_4>;
  1275. let Predicates = [HasMVEFloat] in {
  1276. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 0)), !cast<Instruction>(NAME)>;
  1277. }
  1278. }
  1279. defm MVE_VMAXNMf32 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v4f32, fmaxnum, int_arm_mve_max_predicated>;
  1280. defm MVE_VMAXNMf16 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v8f16, fmaxnum, int_arm_mve_max_predicated>;
  1281. defm MVE_VMINNMf32 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v4f32, fminnum, int_arm_mve_min_predicated>;
  1282. defm MVE_VMINNMf16 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v8f16, fminnum, int_arm_mve_min_predicated>;
  1283. class MVE_VMINMAX<string iname, string suffix, bit U, bits<2> size,
  1284. bit bit_4, list<dag> pattern=[]>
  1285. : MVE_comp<NoItinerary, iname, suffix, "", pattern> {
  1286. let Inst{28} = U;
  1287. let Inst{25-24} = 0b11;
  1288. let Inst{23} = 0b0;
  1289. let Inst{21-20} = size{1-0};
  1290. let Inst{11} = 0b0;
  1291. let Inst{8} = 0b0;
  1292. let Inst{6} = 0b1;
  1293. let Inst{4} = bit_4;
  1294. let validForTailPredication = 1;
  1295. }
  1296. multiclass MVE_VMINMAX_m<string iname, bit bit_4, MVEVectorVTInfo VTI,
  1297. SDNode Op, Intrinsic PredInt> {
  1298. def "" : MVE_VMINMAX<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, bit_4>;
  1299. let Predicates = [HasMVEInt] in {
  1300. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
  1301. }
  1302. }
  1303. multiclass MVE_VMAX<MVEVectorVTInfo VTI>
  1304. : MVE_VMINMAX_m<"vmax", 0b0, VTI, !if(VTI.Unsigned, umax, smax), int_arm_mve_max_predicated>;
  1305. multiclass MVE_VMIN<MVEVectorVTInfo VTI>
  1306. : MVE_VMINMAX_m<"vmin", 0b1, VTI, !if(VTI.Unsigned, umin, smin), int_arm_mve_min_predicated>;
  1307. defm MVE_VMINs8 : MVE_VMIN<MVE_v16s8>;
  1308. defm MVE_VMINs16 : MVE_VMIN<MVE_v8s16>;
  1309. defm MVE_VMINs32 : MVE_VMIN<MVE_v4s32>;
  1310. defm MVE_VMINu8 : MVE_VMIN<MVE_v16u8>;
  1311. defm MVE_VMINu16 : MVE_VMIN<MVE_v8u16>;
  1312. defm MVE_VMINu32 : MVE_VMIN<MVE_v4u32>;
  1313. defm MVE_VMAXs8 : MVE_VMAX<MVE_v16s8>;
  1314. defm MVE_VMAXs16 : MVE_VMAX<MVE_v8s16>;
  1315. defm MVE_VMAXs32 : MVE_VMAX<MVE_v4s32>;
  1316. defm MVE_VMAXu8 : MVE_VMAX<MVE_v16u8>;
  1317. defm MVE_VMAXu16 : MVE_VMAX<MVE_v8u16>;
  1318. defm MVE_VMAXu32 : MVE_VMAX<MVE_v4u32>;
  1319. // end of mve_comp instructions
  1320. // start of mve_bit instructions
  1321. class MVE_bit_arith<dag oops, dag iops, string iname, string suffix,
  1322. string ops, string cstr, list<dag> pattern=[]>
  1323. : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred_r, cstr, pattern> {
  1324. bits<4> Qd;
  1325. bits<4> Qm;
  1326. let Inst{22} = Qd{3};
  1327. let Inst{15-13} = Qd{2-0};
  1328. let Inst{5} = Qm{3};
  1329. let Inst{3-1} = Qm{2-0};
  1330. }
  1331. def MVE_VBIC : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
  1332. "vbic", "", "$Qd, $Qn, $Qm", ""> {
  1333. bits<4> Qn;
  1334. let Inst{28} = 0b0;
  1335. let Inst{25-23} = 0b110;
  1336. let Inst{21-20} = 0b01;
  1337. let Inst{19-17} = Qn{2-0};
  1338. let Inst{16} = 0b0;
  1339. let Inst{12-8} = 0b00001;
  1340. let Inst{7} = Qn{3};
  1341. let Inst{6} = 0b1;
  1342. let Inst{4} = 0b1;
  1343. let Inst{0} = 0b0;
  1344. let validForTailPredication = 1;
  1345. }
  1346. class MVE_VREV<string iname, string suffix, bits<2> size, bits<2> bit_8_7, string cstr="">
  1347. : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm), iname,
  1348. suffix, "$Qd, $Qm", cstr> {
  1349. let Inst{28} = 0b1;
  1350. let Inst{25-23} = 0b111;
  1351. let Inst{21-20} = 0b11;
  1352. let Inst{19-18} = size;
  1353. let Inst{17-16} = 0b00;
  1354. let Inst{12-9} = 0b0000;
  1355. let Inst{8-7} = bit_8_7;
  1356. let Inst{6} = 0b1;
  1357. let Inst{4} = 0b0;
  1358. let Inst{0} = 0b0;
  1359. }
  1360. def MVE_VREV64_8 : MVE_VREV<"vrev64", "8", 0b00, 0b00, "@earlyclobber $Qd">;
  1361. def MVE_VREV64_16 : MVE_VREV<"vrev64", "16", 0b01, 0b00, "@earlyclobber $Qd">;
  1362. def MVE_VREV64_32 : MVE_VREV<"vrev64", "32", 0b10, 0b00, "@earlyclobber $Qd">;
  1363. def MVE_VREV32_8 : MVE_VREV<"vrev32", "8", 0b00, 0b01>;
  1364. def MVE_VREV32_16 : MVE_VREV<"vrev32", "16", 0b01, 0b01>;
  1365. def MVE_VREV16_8 : MVE_VREV<"vrev16", "8", 0b00, 0b10>;
  1366. let Predicates = [HasMVEInt] in {
  1367. def : Pat<(v8i16 (bswap (v8i16 MQPR:$src))),
  1368. (v8i16 (MVE_VREV16_8 (v8i16 MQPR:$src)))>;
  1369. def : Pat<(v4i32 (bswap (v4i32 MQPR:$src))),
  1370. (v4i32 (MVE_VREV32_8 (v4i32 MQPR:$src)))>;
  1371. }
  1372. multiclass MVE_VREV_basic_patterns<int revbits, list<MVEVectorVTInfo> VTIs,
  1373. Instruction Inst> {
  1374. defvar unpred_op = !cast<SDNode>("ARMvrev" # revbits);
  1375. foreach VTI = VTIs in {
  1376. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src))),
  1377. (VTI.Vec (Inst (VTI.Vec MQPR:$src)))>;
  1378. def : Pat<(VTI.Vec (int_arm_mve_vrev_predicated (VTI.Vec MQPR:$src),
  1379. revbits, (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
  1380. (VTI.Vec (Inst (VTI.Vec MQPR:$src), ARMVCCThen,
  1381. (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
  1382. }
  1383. }
  1384. let Predicates = [HasMVEInt] in {
  1385. defm: MVE_VREV_basic_patterns<64, [MVE_v4i32, MVE_v4f32], MVE_VREV64_32>;
  1386. defm: MVE_VREV_basic_patterns<64, [MVE_v8i16, MVE_v8f16], MVE_VREV64_16>;
  1387. defm: MVE_VREV_basic_patterns<64, [MVE_v16i8 ], MVE_VREV64_8>;
  1388. defm: MVE_VREV_basic_patterns<32, [MVE_v8i16, MVE_v8f16], MVE_VREV32_16>;
  1389. defm: MVE_VREV_basic_patterns<32, [MVE_v16i8 ], MVE_VREV32_8>;
  1390. defm: MVE_VREV_basic_patterns<16, [MVE_v16i8 ], MVE_VREV16_8>;
  1391. }
  1392. def MVE_VMVN : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm),
  1393. "vmvn", "", "$Qd, $Qm", ""> {
  1394. let Inst{28} = 0b1;
  1395. let Inst{25-23} = 0b111;
  1396. let Inst{21-16} = 0b110000;
  1397. let Inst{12-6} = 0b0010111;
  1398. let Inst{4} = 0b0;
  1399. let Inst{0} = 0b0;
  1400. let validForTailPredication = 1;
  1401. }
  1402. let Predicates = [HasMVEInt] in {
  1403. foreach VTI = [ MVE_v16i8, MVE_v8i16, MVE_v4i32, MVE_v2i64 ] in {
  1404. def : Pat<(VTI.Vec (vnotq (VTI.Vec MQPR:$val1))),
  1405. (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1)))>;
  1406. def : Pat<(VTI.Vec (int_arm_mve_mvn_predicated (VTI.Vec MQPR:$val1),
  1407. (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
  1408. (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1), ARMVCCThen,
  1409. (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
  1410. }
  1411. }
  1412. class MVE_bit_ops<string iname, bits<2> bit_21_20, bit bit_28>
  1413. : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
  1414. iname, "", "$Qd, $Qn, $Qm", ""> {
  1415. bits<4> Qn;
  1416. let Inst{28} = bit_28;
  1417. let Inst{25-23} = 0b110;
  1418. let Inst{21-20} = bit_21_20;
  1419. let Inst{19-17} = Qn{2-0};
  1420. let Inst{16} = 0b0;
  1421. let Inst{12-8} = 0b00001;
  1422. let Inst{7} = Qn{3};
  1423. let Inst{6} = 0b1;
  1424. let Inst{4} = 0b1;
  1425. let Inst{0} = 0b0;
  1426. let validForTailPredication = 1;
  1427. }
  1428. def MVE_VEOR : MVE_bit_ops<"veor", 0b00, 0b1>;
  1429. def MVE_VORN : MVE_bit_ops<"vorn", 0b11, 0b0>;
  1430. def MVE_VORR : MVE_bit_ops<"vorr", 0b10, 0b0>;
  1431. def MVE_VAND : MVE_bit_ops<"vand", 0b00, 0b0>;
  1432. // add ignored suffixes as aliases
  1433. foreach s=["s8", "s16", "s32", "u8", "u16", "u32", "i8", "i16", "i32", "f16", "f32"] in {
  1434. def : MVEInstAlias<"vbic${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
  1435. (MVE_VBIC MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
  1436. def : MVEInstAlias<"veor${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
  1437. (MVE_VEOR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
  1438. def : MVEInstAlias<"vorn${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
  1439. (MVE_VORN MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
  1440. def : MVEInstAlias<"vorr${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
  1441. (MVE_VORR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
  1442. def : MVEInstAlias<"vand${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
  1443. (MVE_VAND MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
  1444. }
  1445. let Predicates = [HasMVEInt] in {
  1446. defm : MVE_TwoOpPattern<MVE_v16i8, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
  1447. defm : MVE_TwoOpPattern<MVE_v8i16, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
  1448. defm : MVE_TwoOpPattern<MVE_v4i32, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
  1449. defm : MVE_TwoOpPattern<MVE_v2i64, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
  1450. defm : MVE_TwoOpPattern<MVE_v16i8, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
  1451. defm : MVE_TwoOpPattern<MVE_v8i16, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
  1452. defm : MVE_TwoOpPattern<MVE_v4i32, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
  1453. defm : MVE_TwoOpPattern<MVE_v2i64, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
  1454. defm : MVE_TwoOpPattern<MVE_v16i8, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
  1455. defm : MVE_TwoOpPattern<MVE_v8i16, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
  1456. defm : MVE_TwoOpPattern<MVE_v4i32, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
  1457. defm : MVE_TwoOpPattern<MVE_v2i64, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
  1458. defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
  1459. int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
  1460. defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
  1461. int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
  1462. defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
  1463. int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
  1464. defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
  1465. int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
  1466. defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
  1467. int_arm_mve_orn_predicated, (? ), MVE_VORN>;
  1468. defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
  1469. int_arm_mve_orn_predicated, (? ), MVE_VORN>;
  1470. defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
  1471. int_arm_mve_orn_predicated, (? ), MVE_VORN>;
  1472. defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
  1473. int_arm_mve_orn_predicated, (? ), MVE_VORN>;
  1474. }
  1475. class MVE_bit_cmode<string iname, string suffix, bit halfword, dag inOps>
  1476. : MVE_p<(outs MQPR:$Qd), inOps, NoItinerary,
  1477. iname, suffix, "$Qd, $imm", vpred_n, "$Qd = $Qd_src"> {
  1478. bits<12> imm;
  1479. bits<4> Qd;
  1480. let Inst{28} = imm{7};
  1481. let Inst{27-23} = 0b11111;
  1482. let Inst{22} = Qd{3};
  1483. let Inst{21-19} = 0b000;
  1484. let Inst{18-16} = imm{6-4};
  1485. let Inst{15-13} = Qd{2-0};
  1486. let Inst{12} = 0b0;
  1487. let Inst{11} = halfword;
  1488. let Inst{10} = !if(halfword, 0, imm{10});
  1489. let Inst{9} = imm{9};
  1490. let Inst{8} = 0b1;
  1491. let Inst{7-6} = 0b01;
  1492. let Inst{4} = 0b1;
  1493. let Inst{3-0} = imm{3-0};
  1494. }
  1495. multiclass MVE_bit_cmode_p<string iname, bit opcode,
  1496. MVEVectorVTInfo VTI, Operand imm_type, SDNode op> {
  1497. def "" : MVE_bit_cmode<iname, VTI.Suffix, VTI.Size{0},
  1498. (ins MQPR:$Qd_src, imm_type:$imm)> {
  1499. let Inst{5} = opcode;
  1500. let validForTailPredication = 1;
  1501. }
  1502. defvar Inst = !cast<Instruction>(NAME);
  1503. defvar UnpredPat = (VTI.Vec (op (VTI.Vec MQPR:$src), timm:$simm));
  1504. let Predicates = [HasMVEInt] in {
  1505. def : Pat<UnpredPat,
  1506. (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm))>;
  1507. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  1508. UnpredPat, (VTI.Vec MQPR:$src))),
  1509. (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm,
  1510. ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
  1511. }
  1512. }
  1513. multiclass MVE_VORRimm<MVEVectorVTInfo VTI, Operand imm_type> {
  1514. defm "": MVE_bit_cmode_p<"vorr", 0, VTI, imm_type, ARMvorrImm>;
  1515. }
  1516. multiclass MVE_VBICimm<MVEVectorVTInfo VTI, Operand imm_type> {
  1517. defm "": MVE_bit_cmode_p<"vbic", 1, VTI, imm_type, ARMvbicImm>;
  1518. }
  1519. defm MVE_VORRimmi16 : MVE_VORRimm<MVE_v8i16, nImmSplatI16>;
  1520. defm MVE_VORRimmi32 : MVE_VORRimm<MVE_v4i32, nImmSplatI32>;
  1521. defm MVE_VBICimmi16 : MVE_VBICimm<MVE_v8i16, nImmSplatI16>;
  1522. defm MVE_VBICimmi32 : MVE_VBICimm<MVE_v4i32, nImmSplatI32>;
  1523. def MVE_VORNimmi16 : MVEInstAlias<"vorn${vp}.i16\t$Qd, $imm",
  1524. (MVE_VORRimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
  1525. def MVE_VORNimmi32 : MVEInstAlias<"vorn${vp}.i32\t$Qd, $imm",
  1526. (MVE_VORRimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
  1527. def MVE_VANDimmi16 : MVEInstAlias<"vand${vp}.i16\t$Qd, $imm",
  1528. (MVE_VBICimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
  1529. def MVE_VANDimmi32 : MVEInstAlias<"vand${vp}.i32\t$Qd, $imm",
  1530. (MVE_VBICimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
  1531. def MVE_VMOV : MVEInstAlias<"vmov${vp}\t$Qd, $Qm",
  1532. (MVE_VORR MQPR:$Qd, MQPR:$Qm, MQPR:$Qm, vpred_r:$vp)>;
  1533. class MVE_VMOV_lane_direction {
  1534. bit bit_20;
  1535. dag oops;
  1536. dag iops;
  1537. string ops;
  1538. string cstr;
  1539. }
  1540. def MVE_VMOV_from_lane : MVE_VMOV_lane_direction {
  1541. let bit_20 = 0b1;
  1542. let oops = (outs rGPR:$Rt);
  1543. let iops = (ins MQPR:$Qd);
  1544. let ops = "$Rt, $Qd$Idx";
  1545. let cstr = "";
  1546. }
  1547. def MVE_VMOV_to_lane : MVE_VMOV_lane_direction {
  1548. let bit_20 = 0b0;
  1549. let oops = (outs MQPR:$Qd);
  1550. let iops = (ins MQPR:$Qd_src, rGPR:$Rt);
  1551. let ops = "$Qd$Idx, $Rt";
  1552. let cstr = "$Qd = $Qd_src";
  1553. }
  1554. class MVE_VMOV_lane<string suffix, bit U, dag indexop,
  1555. MVE_VMOV_lane_direction dir>
  1556. : MVE_VMOV_lane_base<dir.oops, !con(dir.iops, indexop), NoItinerary,
  1557. "vmov", suffix, dir.ops, dir.cstr, []> {
  1558. bits<4> Qd;
  1559. bits<4> Rt;
  1560. let Inst{31-24} = 0b11101110;
  1561. let Inst{23} = U;
  1562. let Inst{20} = dir.bit_20;
  1563. let Inst{19-17} = Qd{2-0};
  1564. let Inst{15-12} = Rt{3-0};
  1565. let Inst{11-8} = 0b1011;
  1566. let Inst{7} = Qd{3};
  1567. let Inst{4-0} = 0b10000;
  1568. let hasSideEffects = 0;
  1569. }
  1570. class MVE_VMOV_lane_32<MVE_VMOV_lane_direction dir>
  1571. : MVE_VMOV_lane<"32", 0b0, (ins MVEVectorIndex<4>:$Idx), dir> {
  1572. bits<2> Idx;
  1573. let Inst{22} = 0b0;
  1574. let Inst{6-5} = 0b00;
  1575. let Inst{16} = Idx{1};
  1576. let Inst{21} = Idx{0};
  1577. let Predicates = [HasFPRegsV8_1M];
  1578. }
  1579. class MVE_VMOV_lane_16<string suffix, bit U, MVE_VMOV_lane_direction dir>
  1580. : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<8>:$Idx), dir> {
  1581. bits<3> Idx;
  1582. let Inst{22} = 0b0;
  1583. let Inst{5} = 0b1;
  1584. let Inst{16} = Idx{2};
  1585. let Inst{21} = Idx{1};
  1586. let Inst{6} = Idx{0};
  1587. }
  1588. class MVE_VMOV_lane_8<string suffix, bit U, MVE_VMOV_lane_direction dir>
  1589. : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<16>:$Idx), dir> {
  1590. bits<4> Idx;
  1591. let Inst{22} = 0b1;
  1592. let Inst{16} = Idx{3};
  1593. let Inst{21} = Idx{2};
  1594. let Inst{6} = Idx{1};
  1595. let Inst{5} = Idx{0};
  1596. }
  1597. def MVE_VMOV_from_lane_32 : MVE_VMOV_lane_32< MVE_VMOV_from_lane>;
  1598. def MVE_VMOV_to_lane_32 : MVE_VMOV_lane_32< MVE_VMOV_to_lane>;
  1599. def MVE_VMOV_from_lane_s16 : MVE_VMOV_lane_16<"s16", 0b0, MVE_VMOV_from_lane>;
  1600. def MVE_VMOV_from_lane_u16 : MVE_VMOV_lane_16<"u16", 0b1, MVE_VMOV_from_lane>;
  1601. def MVE_VMOV_to_lane_16 : MVE_VMOV_lane_16< "16", 0b0, MVE_VMOV_to_lane>;
  1602. def MVE_VMOV_from_lane_s8 : MVE_VMOV_lane_8 < "s8", 0b0, MVE_VMOV_from_lane>;
  1603. def MVE_VMOV_from_lane_u8 : MVE_VMOV_lane_8 < "u8", 0b1, MVE_VMOV_from_lane>;
  1604. def MVE_VMOV_to_lane_8 : MVE_VMOV_lane_8 < "8", 0b0, MVE_VMOV_to_lane>;
  1605. let Predicates = [HasMVEInt] in {
  1606. def : Pat<(extractelt (v2f64 MQPR:$src), imm:$lane),
  1607. (f64 (EXTRACT_SUBREG MQPR:$src, (DSubReg_f64_reg imm:$lane)))>;
  1608. def : Pat<(insertelt (v2f64 MQPR:$src1), DPR:$src2, imm:$lane),
  1609. (INSERT_SUBREG (v2f64 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), DPR:$src2, (DSubReg_f64_reg imm:$lane))>;
  1610. def : Pat<(extractelt (v4i32 MQPR:$src), imm:$lane),
  1611. (COPY_TO_REGCLASS
  1612. (i32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), rGPR)>;
  1613. def : Pat<(insertelt (v4i32 MQPR:$src1), rGPR:$src2, imm:$lane),
  1614. (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$src2, imm:$lane)>;
  1615. def : Pat<(vector_insert (v16i8 MQPR:$src1), rGPR:$src2, imm:$lane),
  1616. (MVE_VMOV_to_lane_8 MQPR:$src1, rGPR:$src2, imm:$lane)>;
  1617. def : Pat<(vector_insert (v8i16 MQPR:$src1), rGPR:$src2, imm:$lane),
  1618. (MVE_VMOV_to_lane_16 MQPR:$src1, rGPR:$src2, imm:$lane)>;
  1619. def : Pat<(ARMvgetlanes (v16i8 MQPR:$src), imm:$lane),
  1620. (MVE_VMOV_from_lane_s8 MQPR:$src, imm:$lane)>;
  1621. def : Pat<(ARMvgetlanes (v8i16 MQPR:$src), imm:$lane),
  1622. (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
  1623. def : Pat<(ARMvgetlanes (v8f16 MQPR:$src), imm:$lane),
  1624. (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
  1625. def : Pat<(ARMvgetlaneu (v16i8 MQPR:$src), imm:$lane),
  1626. (MVE_VMOV_from_lane_u8 MQPR:$src, imm:$lane)>;
  1627. def : Pat<(ARMvgetlaneu (v8i16 MQPR:$src), imm:$lane),
  1628. (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
  1629. def : Pat<(ARMvgetlaneu (v8f16 MQPR:$src), imm:$lane),
  1630. (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
  1631. def : Pat<(v16i8 (scalar_to_vector GPR:$src)),
  1632. (MVE_VMOV_to_lane_8 (v16i8 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
  1633. def : Pat<(v8i16 (scalar_to_vector GPR:$src)),
  1634. (MVE_VMOV_to_lane_16 (v8i16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
  1635. def : Pat<(v4i32 (scalar_to_vector GPR:$src)),
  1636. (MVE_VMOV_to_lane_32 (v4i32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
  1637. // Floating point patterns, still enabled under HasMVEInt
  1638. def : Pat<(extractelt (v4f32 MQPR:$src), imm:$lane),
  1639. (COPY_TO_REGCLASS (f32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), SPR)>;
  1640. def : Pat<(insertelt (v4f32 MQPR:$src1), (f32 SPR:$src2), imm:$lane),
  1641. (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), SPR:$src2, (SSubReg_f32_reg imm:$lane))>;
  1642. def : Pat<(insertelt (v8f16 MQPR:$src1), (f16 HPR:$src2), imm:$lane),
  1643. (MVE_VMOV_to_lane_16 MQPR:$src1, (COPY_TO_REGCLASS (f16 HPR:$src2), rGPR), imm:$lane)>;
  1644. def : Pat<(extractelt (v8f16 MQPR:$src), imm_even:$lane),
  1645. (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_even:$lane))>;
  1646. def : Pat<(extractelt (v8f16 MQPR:$src), imm_odd:$lane),
  1647. (COPY_TO_REGCLASS
  1648. (VMOVH (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_odd:$lane))),
  1649. HPR)>;
  1650. def : Pat<(v4f32 (scalar_to_vector SPR:$src)),
  1651. (INSERT_SUBREG (v4f32 (IMPLICIT_DEF)), SPR:$src, ssub_0)>;
  1652. def : Pat<(v4f32 (scalar_to_vector GPR:$src)),
  1653. (MVE_VMOV_to_lane_32 (v4f32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
  1654. def : Pat<(v8f16 (scalar_to_vector (f16 HPR:$src))),
  1655. (INSERT_SUBREG (v8f16 (IMPLICIT_DEF)), (f16 HPR:$src), ssub_0)>;
  1656. def : Pat<(v8f16 (scalar_to_vector GPR:$src)),
  1657. (MVE_VMOV_to_lane_16 (v8f16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
  1658. }
  1659. // end of mve_bit instructions
  1660. // start of MVE Integer instructions
  1661. class MVE_int<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
  1662. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
  1663. iname, suffix, "$Qd, $Qn, $Qm", vpred_r, "", pattern> {
  1664. bits<4> Qd;
  1665. bits<4> Qn;
  1666. bits<4> Qm;
  1667. let Inst{22} = Qd{3};
  1668. let Inst{21-20} = size;
  1669. let Inst{19-17} = Qn{2-0};
  1670. let Inst{15-13} = Qd{2-0};
  1671. let Inst{7} = Qn{3};
  1672. let Inst{6} = 0b1;
  1673. let Inst{5} = Qm{3};
  1674. let Inst{3-1} = Qm{2-0};
  1675. }
  1676. class MVE_VMULt1<string iname, string suffix, bits<2> size,
  1677. list<dag> pattern=[]>
  1678. : MVE_int<iname, suffix, size, pattern> {
  1679. let Inst{28} = 0b0;
  1680. let Inst{25-23} = 0b110;
  1681. let Inst{16} = 0b0;
  1682. let Inst{12-8} = 0b01001;
  1683. let Inst{4} = 0b1;
  1684. let Inst{0} = 0b0;
  1685. let validForTailPredication = 1;
  1686. }
  1687. multiclass MVE_VMUL_m<MVEVectorVTInfo VTI> {
  1688. def "" : MVE_VMULt1<"vmul", VTI.Suffix, VTI.Size>;
  1689. let Predicates = [HasMVEInt] in {
  1690. defm : MVE_TwoOpPattern<VTI, mul, int_arm_mve_mul_predicated, (? ),
  1691. !cast<Instruction>(NAME), ARMimmOneV>;
  1692. }
  1693. }
  1694. defm MVE_VMULi8 : MVE_VMUL_m<MVE_v16i8>;
  1695. defm MVE_VMULi16 : MVE_VMUL_m<MVE_v8i16>;
  1696. defm MVE_VMULi32 : MVE_VMUL_m<MVE_v4i32>;
  1697. class MVE_VQxDMULH_Base<string iname, string suffix, bits<2> size, bit rounding,
  1698. list<dag> pattern=[]>
  1699. : MVE_int<iname, suffix, size, pattern> {
  1700. let Inst{28} = rounding;
  1701. let Inst{25-23} = 0b110;
  1702. let Inst{16} = 0b0;
  1703. let Inst{12-8} = 0b01011;
  1704. let Inst{4} = 0b0;
  1705. let Inst{0} = 0b0;
  1706. let validForTailPredication = 1;
  1707. }
  1708. def MVEvqdmulh : SDNode<"ARMISD::VQDMULH", SDTIntBinOp>;
  1709. multiclass MVE_VQxDMULH_m<string iname, MVEVectorVTInfo VTI,
  1710. SDNode Op, Intrinsic unpred_int, Intrinsic pred_int,
  1711. bit rounding> {
  1712. def "" : MVE_VQxDMULH_Base<iname, VTI.Suffix, VTI.Size, rounding>;
  1713. defvar Inst = !cast<Instruction>(NAME);
  1714. let Predicates = [HasMVEInt] in {
  1715. defm : MVE_TwoOpPattern<VTI, Op, pred_int, (? ), Inst>;
  1716. // Extra unpredicated multiply intrinsic patterns
  1717. def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
  1718. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  1719. }
  1720. }
  1721. multiclass MVE_VQxDMULH<string iname, MVEVectorVTInfo VTI, bit rounding>
  1722. : MVE_VQxDMULH_m<iname, VTI, !if(rounding, null_frag,
  1723. MVEvqdmulh),
  1724. !if(rounding, int_arm_mve_vqrdmulh,
  1725. int_arm_mve_vqdmulh),
  1726. !if(rounding, int_arm_mve_qrdmulh_predicated,
  1727. int_arm_mve_qdmulh_predicated),
  1728. rounding>;
  1729. defm MVE_VQDMULHi8 : MVE_VQxDMULH<"vqdmulh", MVE_v16s8, 0b0>;
  1730. defm MVE_VQDMULHi16 : MVE_VQxDMULH<"vqdmulh", MVE_v8s16, 0b0>;
  1731. defm MVE_VQDMULHi32 : MVE_VQxDMULH<"vqdmulh", MVE_v4s32, 0b0>;
  1732. defm MVE_VQRDMULHi8 : MVE_VQxDMULH<"vqrdmulh", MVE_v16s8, 0b1>;
  1733. defm MVE_VQRDMULHi16 : MVE_VQxDMULH<"vqrdmulh", MVE_v8s16, 0b1>;
  1734. defm MVE_VQRDMULHi32 : MVE_VQxDMULH<"vqrdmulh", MVE_v4s32, 0b1>;
  1735. class MVE_VADDSUB<string iname, string suffix, bits<2> size, bit subtract,
  1736. list<dag> pattern=[]>
  1737. : MVE_int<iname, suffix, size, pattern> {
  1738. let Inst{28} = subtract;
  1739. let Inst{25-23} = 0b110;
  1740. let Inst{16} = 0b0;
  1741. let Inst{12-8} = 0b01000;
  1742. let Inst{4} = 0b0;
  1743. let Inst{0} = 0b0;
  1744. let validForTailPredication = 1;
  1745. }
  1746. multiclass MVE_VADDSUB_m<string iname, MVEVectorVTInfo VTI, bit subtract,
  1747. SDNode Op, Intrinsic PredInt> {
  1748. def "" : MVE_VADDSUB<iname, VTI.Suffix, VTI.Size, subtract>;
  1749. defvar Inst = !cast<Instruction>(NAME);
  1750. let Predicates = [HasMVEInt] in {
  1751. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
  1752. }
  1753. }
  1754. multiclass MVE_VADD<MVEVectorVTInfo VTI>
  1755. : MVE_VADDSUB_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
  1756. multiclass MVE_VSUB<MVEVectorVTInfo VTI>
  1757. : MVE_VADDSUB_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
  1758. defm MVE_VADDi8 : MVE_VADD<MVE_v16i8>;
  1759. defm MVE_VADDi16 : MVE_VADD<MVE_v8i16>;
  1760. defm MVE_VADDi32 : MVE_VADD<MVE_v4i32>;
  1761. defm MVE_VSUBi8 : MVE_VSUB<MVE_v16i8>;
  1762. defm MVE_VSUBi16 : MVE_VSUB<MVE_v8i16>;
  1763. defm MVE_VSUBi32 : MVE_VSUB<MVE_v4i32>;
  1764. class MVE_VQADDSUB<string iname, string suffix, bit U, bit subtract,
  1765. bits<2> size>
  1766. : MVE_int<iname, suffix, size, []> {
  1767. let Inst{28} = U;
  1768. let Inst{25-23} = 0b110;
  1769. let Inst{16} = 0b0;
  1770. let Inst{12-10} = 0b000;
  1771. let Inst{9} = subtract;
  1772. let Inst{8} = 0b0;
  1773. let Inst{4} = 0b1;
  1774. let Inst{0} = 0b0;
  1775. let validForTailPredication = 1;
  1776. }
  1777. class MVE_VQADD_<string suffix, bit U, bits<2> size>
  1778. : MVE_VQADDSUB<"vqadd", suffix, U, 0b0, size>;
  1779. class MVE_VQSUB_<string suffix, bit U, bits<2> size>
  1780. : MVE_VQADDSUB<"vqsub", suffix, U, 0b1, size>;
  1781. multiclass MVE_VQADD_m<MVEVectorVTInfo VTI,
  1782. SDNode Op, Intrinsic PredInt> {
  1783. def "" : MVE_VQADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1784. defvar Inst = !cast<Instruction>(NAME);
  1785. let Predicates = [HasMVEInt] in {
  1786. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
  1787. !cast<Instruction>(NAME)>;
  1788. }
  1789. }
  1790. multiclass MVE_VQADD<MVEVectorVTInfo VTI, SDNode unpred_op>
  1791. : MVE_VQADD_m<VTI, unpred_op, int_arm_mve_qadd_predicated>;
  1792. defm MVE_VQADDs8 : MVE_VQADD<MVE_v16s8, saddsat>;
  1793. defm MVE_VQADDs16 : MVE_VQADD<MVE_v8s16, saddsat>;
  1794. defm MVE_VQADDs32 : MVE_VQADD<MVE_v4s32, saddsat>;
  1795. defm MVE_VQADDu8 : MVE_VQADD<MVE_v16u8, uaddsat>;
  1796. defm MVE_VQADDu16 : MVE_VQADD<MVE_v8u16, uaddsat>;
  1797. defm MVE_VQADDu32 : MVE_VQADD<MVE_v4u32, uaddsat>;
  1798. multiclass MVE_VQSUB_m<MVEVectorVTInfo VTI,
  1799. SDNode Op, Intrinsic PredInt> {
  1800. def "" : MVE_VQSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1801. defvar Inst = !cast<Instruction>(NAME);
  1802. let Predicates = [HasMVEInt] in {
  1803. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
  1804. !cast<Instruction>(NAME)>;
  1805. }
  1806. }
  1807. multiclass MVE_VQSUB<MVEVectorVTInfo VTI, SDNode unpred_op>
  1808. : MVE_VQSUB_m<VTI, unpred_op, int_arm_mve_qsub_predicated>;
  1809. defm MVE_VQSUBs8 : MVE_VQSUB<MVE_v16s8, ssubsat>;
  1810. defm MVE_VQSUBs16 : MVE_VQSUB<MVE_v8s16, ssubsat>;
  1811. defm MVE_VQSUBs32 : MVE_VQSUB<MVE_v4s32, ssubsat>;
  1812. defm MVE_VQSUBu8 : MVE_VQSUB<MVE_v16u8, usubsat>;
  1813. defm MVE_VQSUBu16 : MVE_VQSUB<MVE_v8u16, usubsat>;
  1814. defm MVE_VQSUBu32 : MVE_VQSUB<MVE_v4u32, usubsat>;
  1815. class MVE_VABD_int<string suffix, bit U, bits<2> size,
  1816. list<dag> pattern=[]>
  1817. : MVE_int<"vabd", suffix, size, pattern> {
  1818. let Inst{28} = U;
  1819. let Inst{25-23} = 0b110;
  1820. let Inst{16} = 0b0;
  1821. let Inst{12-8} = 0b00111;
  1822. let Inst{4} = 0b0;
  1823. let Inst{0} = 0b0;
  1824. let validForTailPredication = 1;
  1825. }
  1826. multiclass MVE_VABD_m<MVEVectorVTInfo VTI,
  1827. Intrinsic unpred_int, Intrinsic pred_int> {
  1828. def "" : MVE_VABD_int<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1829. defvar Inst = !cast<Instruction>(NAME);
  1830. let Predicates = [HasMVEInt] in {
  1831. // Unpredicated absolute difference
  1832. def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1833. (i32 VTI.Unsigned))),
  1834. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  1835. // Predicated absolute difference
  1836. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1837. (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
  1838. (VTI.Vec MQPR:$inactive))),
  1839. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1840. ARMVCCThen, (VTI.Pred VCCR:$mask),
  1841. (VTI.Vec MQPR:$inactive)))>;
  1842. }
  1843. }
  1844. multiclass MVE_VABD<MVEVectorVTInfo VTI>
  1845. : MVE_VABD_m<VTI, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
  1846. defm MVE_VABDs8 : MVE_VABD<MVE_v16s8>;
  1847. defm MVE_VABDs16 : MVE_VABD<MVE_v8s16>;
  1848. defm MVE_VABDs32 : MVE_VABD<MVE_v4s32>;
  1849. defm MVE_VABDu8 : MVE_VABD<MVE_v16u8>;
  1850. defm MVE_VABDu16 : MVE_VABD<MVE_v8u16>;
  1851. defm MVE_VABDu32 : MVE_VABD<MVE_v4u32>;
  1852. class MVE_VRHADD_Base<string suffix, bit U, bits<2> size, list<dag> pattern=[]>
  1853. : MVE_int<"vrhadd", suffix, size, pattern> {
  1854. let Inst{28} = U;
  1855. let Inst{25-23} = 0b110;
  1856. let Inst{16} = 0b0;
  1857. let Inst{12-8} = 0b00001;
  1858. let Inst{4} = 0b0;
  1859. let Inst{0} = 0b0;
  1860. let validForTailPredication = 1;
  1861. }
  1862. def addnuw : PatFrag<(ops node:$lhs, node:$rhs),
  1863. (add node:$lhs, node:$rhs), [{
  1864. return N->getFlags().hasNoUnsignedWrap();
  1865. }]>;
  1866. def addnsw : PatFrag<(ops node:$lhs, node:$rhs),
  1867. (add node:$lhs, node:$rhs), [{
  1868. return N->getFlags().hasNoSignedWrap();
  1869. }]>;
  1870. def subnuw : PatFrag<(ops node:$lhs, node:$rhs),
  1871. (sub node:$lhs, node:$rhs), [{
  1872. return N->getFlags().hasNoUnsignedWrap();
  1873. }]>;
  1874. def subnsw : PatFrag<(ops node:$lhs, node:$rhs),
  1875. (sub node:$lhs, node:$rhs), [{
  1876. return N->getFlags().hasNoSignedWrap();
  1877. }]>;
  1878. multiclass MVE_VRHADD_m<MVEVectorVTInfo VTI,
  1879. SDNode unpred_op, Intrinsic pred_int> {
  1880. def "" : MVE_VRHADD_Base<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1881. defvar Inst = !cast<Instruction>(NAME);
  1882. let Predicates = [HasMVEInt] in {
  1883. // Unpredicated rounding add-with-divide-by-two
  1884. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1885. (i32 VTI.Unsigned))),
  1886. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  1887. // Predicated add-with-divide-by-two
  1888. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1889. (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
  1890. (VTI.Vec MQPR:$inactive))),
  1891. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1892. ARMVCCThen, (VTI.Pred VCCR:$mask),
  1893. (VTI.Vec MQPR:$inactive)))>;
  1894. }
  1895. }
  1896. multiclass MVE_VRHADD<MVEVectorVTInfo VTI>
  1897. : MVE_VRHADD_m<VTI, int_arm_mve_vrhadd, int_arm_mve_rhadd_predicated>;
  1898. defm MVE_VRHADDs8 : MVE_VRHADD<MVE_v16s8>;
  1899. defm MVE_VRHADDs16 : MVE_VRHADD<MVE_v8s16>;
  1900. defm MVE_VRHADDs32 : MVE_VRHADD<MVE_v4s32>;
  1901. defm MVE_VRHADDu8 : MVE_VRHADD<MVE_v16u8>;
  1902. defm MVE_VRHADDu16 : MVE_VRHADD<MVE_v8u16>;
  1903. defm MVE_VRHADDu32 : MVE_VRHADD<MVE_v4u32>;
  1904. // Rounding Halving Add perform the arithemtic operation with an extra bit of
  1905. // precision, before performing the shift, to void clipping errors. We're not
  1906. // modelling that here with these patterns, but we're using no wrap forms of
  1907. // add to ensure that the extra bit of information is not needed for the
  1908. // arithmetic or the rounding.
  1909. let Predicates = [HasMVEInt] in {
  1910. def : Pat<(v16i8 (ARMvshrsImm (addnsw (addnsw (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)),
  1911. (v16i8 (ARMvmovImm (i32 3585)))),
  1912. (i32 1))),
  1913. (MVE_VRHADDs8 MQPR:$Qm, MQPR:$Qn)>;
  1914. def : Pat<(v8i16 (ARMvshrsImm (addnsw (addnsw (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)),
  1915. (v8i16 (ARMvmovImm (i32 2049)))),
  1916. (i32 1))),
  1917. (MVE_VRHADDs16 MQPR:$Qm, MQPR:$Qn)>;
  1918. def : Pat<(v4i32 (ARMvshrsImm (addnsw (addnsw (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)),
  1919. (v4i32 (ARMvmovImm (i32 1)))),
  1920. (i32 1))),
  1921. (MVE_VRHADDs32 MQPR:$Qm, MQPR:$Qn)>;
  1922. def : Pat<(v16i8 (ARMvshruImm (addnuw (addnuw (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)),
  1923. (v16i8 (ARMvmovImm (i32 3585)))),
  1924. (i32 1))),
  1925. (MVE_VRHADDu8 MQPR:$Qm, MQPR:$Qn)>;
  1926. def : Pat<(v8i16 (ARMvshruImm (addnuw (addnuw (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)),
  1927. (v8i16 (ARMvmovImm (i32 2049)))),
  1928. (i32 1))),
  1929. (MVE_VRHADDu16 MQPR:$Qm, MQPR:$Qn)>;
  1930. def : Pat<(v4i32 (ARMvshruImm (addnuw (addnuw (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)),
  1931. (v4i32 (ARMvmovImm (i32 1)))),
  1932. (i32 1))),
  1933. (MVE_VRHADDu32 MQPR:$Qm, MQPR:$Qn)>;
  1934. }
  1935. class MVE_VHADDSUB<string iname, string suffix, bit U, bit subtract,
  1936. bits<2> size, list<dag> pattern=[]>
  1937. : MVE_int<iname, suffix, size, pattern> {
  1938. let Inst{28} = U;
  1939. let Inst{25-23} = 0b110;
  1940. let Inst{16} = 0b0;
  1941. let Inst{12-10} = 0b000;
  1942. let Inst{9} = subtract;
  1943. let Inst{8} = 0b0;
  1944. let Inst{4} = 0b0;
  1945. let Inst{0} = 0b0;
  1946. let validForTailPredication = 1;
  1947. }
  1948. class MVE_VHADD_<string suffix, bit U, bits<2> size,
  1949. list<dag> pattern=[]>
  1950. : MVE_VHADDSUB<"vhadd", suffix, U, 0b0, size, pattern>;
  1951. class MVE_VHSUB_<string suffix, bit U, bits<2> size,
  1952. list<dag> pattern=[]>
  1953. : MVE_VHADDSUB<"vhsub", suffix, U, 0b1, size, pattern>;
  1954. multiclass MVE_VHADD_m<MVEVectorVTInfo VTI,
  1955. SDNode unpred_op, Intrinsic pred_int, PatFrag add_op,
  1956. SDNode shift_op> {
  1957. def "" : MVE_VHADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1958. defvar Inst = !cast<Instruction>(NAME);
  1959. let Predicates = [HasMVEInt] in {
  1960. // Unpredicated add-and-divide-by-two
  1961. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn), (i32 VTI.Unsigned))),
  1962. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  1963. def : Pat<(VTI.Vec (shift_op (add_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)), (i32 1))),
  1964. (Inst MQPR:$Qm, MQPR:$Qn)>;
  1965. // Predicated add-and-divide-by-two
  1966. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn), (i32 VTI.Unsigned),
  1967. (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
  1968. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1969. ARMVCCThen, (VTI.Pred VCCR:$mask),
  1970. (VTI.Vec MQPR:$inactive)))>;
  1971. }
  1972. }
  1973. multiclass MVE_VHADD<MVEVectorVTInfo VTI, PatFrag add_op, SDNode shift_op>
  1974. : MVE_VHADD_m<VTI, int_arm_mve_vhadd, int_arm_mve_hadd_predicated, add_op,
  1975. shift_op>;
  1976. // Halving add/sub perform the arithemtic operation with an extra bit of
  1977. // precision, before performing the shift, to void clipping errors. We're not
  1978. // modelling that here with these patterns, but we're using no wrap forms of
  1979. // add/sub to ensure that the extra bit of information is not needed.
  1980. defm MVE_VHADDs8 : MVE_VHADD<MVE_v16s8, addnsw, ARMvshrsImm>;
  1981. defm MVE_VHADDs16 : MVE_VHADD<MVE_v8s16, addnsw, ARMvshrsImm>;
  1982. defm MVE_VHADDs32 : MVE_VHADD<MVE_v4s32, addnsw, ARMvshrsImm>;
  1983. defm MVE_VHADDu8 : MVE_VHADD<MVE_v16u8, addnuw, ARMvshruImm>;
  1984. defm MVE_VHADDu16 : MVE_VHADD<MVE_v8u16, addnuw, ARMvshruImm>;
  1985. defm MVE_VHADDu32 : MVE_VHADD<MVE_v4u32, addnuw, ARMvshruImm>;
  1986. multiclass MVE_VHSUB_m<MVEVectorVTInfo VTI,
  1987. SDNode unpred_op, Intrinsic pred_int, PatFrag sub_op,
  1988. SDNode shift_op> {
  1989. def "" : MVE_VHSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
  1990. defvar Inst = !cast<Instruction>(NAME);
  1991. let Predicates = [HasMVEInt] in {
  1992. // Unpredicated subtract-and-divide-by-two
  1993. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  1994. (i32 VTI.Unsigned))),
  1995. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  1996. def : Pat<(VTI.Vec (shift_op (sub_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)), (i32 1))),
  1997. (Inst MQPR:$Qm, MQPR:$Qn)>;
  1998. // Predicated subtract-and-divide-by-two
  1999. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  2000. (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
  2001. (VTI.Vec MQPR:$inactive))),
  2002. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  2003. ARMVCCThen, (VTI.Pred VCCR:$mask),
  2004. (VTI.Vec MQPR:$inactive)))>;
  2005. }
  2006. }
  2007. multiclass MVE_VHSUB<MVEVectorVTInfo VTI, PatFrag sub_op, SDNode shift_op>
  2008. : MVE_VHSUB_m<VTI, int_arm_mve_vhsub, int_arm_mve_hsub_predicated, sub_op,
  2009. shift_op>;
  2010. defm MVE_VHSUBs8 : MVE_VHSUB<MVE_v16s8, subnsw, ARMvshrsImm>;
  2011. defm MVE_VHSUBs16 : MVE_VHSUB<MVE_v8s16, subnsw, ARMvshrsImm>;
  2012. defm MVE_VHSUBs32 : MVE_VHSUB<MVE_v4s32, subnsw, ARMvshrsImm>;
  2013. defm MVE_VHSUBu8 : MVE_VHSUB<MVE_v16u8, subnuw, ARMvshruImm>;
  2014. defm MVE_VHSUBu16 : MVE_VHSUB<MVE_v8u16, subnuw, ARMvshruImm>;
  2015. defm MVE_VHSUBu32 : MVE_VHSUB<MVE_v4u32, subnuw, ARMvshruImm>;
  2016. class MVE_VDUP<string suffix, bit B, bit E, list<dag> pattern=[]>
  2017. : MVE_p<(outs MQPR:$Qd), (ins rGPR:$Rt), NoItinerary,
  2018. "vdup", suffix, "$Qd, $Rt", vpred_r, "", pattern> {
  2019. bits<4> Qd;
  2020. bits<4> Rt;
  2021. let Inst{28} = 0b0;
  2022. let Inst{25-23} = 0b101;
  2023. let Inst{22} = B;
  2024. let Inst{21-20} = 0b10;
  2025. let Inst{19-17} = Qd{2-0};
  2026. let Inst{16} = 0b0;
  2027. let Inst{15-12} = Rt;
  2028. let Inst{11-8} = 0b1011;
  2029. let Inst{7} = Qd{3};
  2030. let Inst{6} = 0b0;
  2031. let Inst{5} = E;
  2032. let Inst{4-0} = 0b10000;
  2033. let validForTailPredication = 1;
  2034. }
  2035. def MVE_VDUP32 : MVE_VDUP<"32", 0b0, 0b0>;
  2036. def MVE_VDUP16 : MVE_VDUP<"16", 0b0, 0b1>;
  2037. def MVE_VDUP8 : MVE_VDUP<"8", 0b1, 0b0>;
  2038. let Predicates = [HasMVEInt] in {
  2039. def : Pat<(v16i8 (ARMvdup (i32 rGPR:$elem))),
  2040. (MVE_VDUP8 rGPR:$elem)>;
  2041. def : Pat<(v8i16 (ARMvdup (i32 rGPR:$elem))),
  2042. (MVE_VDUP16 rGPR:$elem)>;
  2043. def : Pat<(v4i32 (ARMvdup (i32 rGPR:$elem))),
  2044. (MVE_VDUP32 rGPR:$elem)>;
  2045. def : Pat<(v8f16 (ARMvdup (i32 rGPR:$elem))),
  2046. (MVE_VDUP16 rGPR:$elem)>;
  2047. def : Pat<(v4f32 (ARMvdup (i32 rGPR:$elem))),
  2048. (MVE_VDUP32 rGPR:$elem)>;
  2049. // Match a vselect with an ARMvdup as a predicated MVE_VDUP
  2050. def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred),
  2051. (v16i8 (ARMvdup (i32 rGPR:$elem))),
  2052. (v16i8 MQPR:$inactive))),
  2053. (MVE_VDUP8 rGPR:$elem, ARMVCCThen, (v16i1 VCCR:$pred),
  2054. (v16i8 MQPR:$inactive))>;
  2055. def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred),
  2056. (v8i16 (ARMvdup (i32 rGPR:$elem))),
  2057. (v8i16 MQPR:$inactive))),
  2058. (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred),
  2059. (v8i16 MQPR:$inactive))>;
  2060. def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred),
  2061. (v4i32 (ARMvdup (i32 rGPR:$elem))),
  2062. (v4i32 MQPR:$inactive))),
  2063. (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred),
  2064. (v4i32 MQPR:$inactive))>;
  2065. def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred),
  2066. (v4f32 (ARMvdup (i32 rGPR:$elem))),
  2067. (v4f32 MQPR:$inactive))),
  2068. (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred),
  2069. (v4f32 MQPR:$inactive))>;
  2070. def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred),
  2071. (v8f16 (ARMvdup (i32 rGPR:$elem))),
  2072. (v8f16 MQPR:$inactive))),
  2073. (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred),
  2074. (v8f16 MQPR:$inactive))>;
  2075. }
  2076. class MVEIntSingleSrc<string iname, string suffix, bits<2> size,
  2077. list<dag> pattern=[]>
  2078. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm), NoItinerary,
  2079. iname, suffix, "$Qd, $Qm", vpred_r, "", pattern> {
  2080. bits<4> Qd;
  2081. bits<4> Qm;
  2082. let Inst{22} = Qd{3};
  2083. let Inst{19-18} = size{1-0};
  2084. let Inst{15-13} = Qd{2-0};
  2085. let Inst{5} = Qm{3};
  2086. let Inst{3-1} = Qm{2-0};
  2087. }
  2088. class MVE_VCLSCLZ<string iname, string suffix, bits<2> size,
  2089. bit count_zeroes, list<dag> pattern=[]>
  2090. : MVEIntSingleSrc<iname, suffix, size, pattern> {
  2091. let Inst{28} = 0b1;
  2092. let Inst{25-23} = 0b111;
  2093. let Inst{21-20} = 0b11;
  2094. let Inst{17-16} = 0b00;
  2095. let Inst{12-8} = 0b00100;
  2096. let Inst{7} = count_zeroes;
  2097. let Inst{6} = 0b1;
  2098. let Inst{4} = 0b0;
  2099. let Inst{0} = 0b0;
  2100. let validForTailPredication = 1;
  2101. }
  2102. multiclass MVE_VCLSCLZ_p<string opname, bit opcode, MVEVectorVTInfo VTI,
  2103. SDNode unpred_op> {
  2104. def "": MVE_VCLSCLZ<"v"#opname, VTI.Suffix, VTI.Size, opcode>;
  2105. defvar Inst = !cast<Instruction>(NAME);
  2106. defvar pred_int = !cast<Intrinsic>("int_arm_mve_"#opname#"_predicated");
  2107. let Predicates = [HasMVEInt] in {
  2108. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
  2109. (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
  2110. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
  2111. (VTI.Vec MQPR:$inactive))),
  2112. (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
  2113. (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
  2114. }
  2115. }
  2116. defm MVE_VCLSs8 : MVE_VCLSCLZ_p<"cls", 0, MVE_v16s8, int_arm_mve_vcls>;
  2117. defm MVE_VCLSs16 : MVE_VCLSCLZ_p<"cls", 0, MVE_v8s16, int_arm_mve_vcls>;
  2118. defm MVE_VCLSs32 : MVE_VCLSCLZ_p<"cls", 0, MVE_v4s32, int_arm_mve_vcls>;
  2119. defm MVE_VCLZs8 : MVE_VCLSCLZ_p<"clz", 1, MVE_v16i8, ctlz>;
  2120. defm MVE_VCLZs16 : MVE_VCLSCLZ_p<"clz", 1, MVE_v8i16, ctlz>;
  2121. defm MVE_VCLZs32 : MVE_VCLSCLZ_p<"clz", 1, MVE_v4i32, ctlz>;
  2122. class MVE_VABSNEG_int<string iname, string suffix, bits<2> size, bit negate,
  2123. bit saturate, list<dag> pattern=[]>
  2124. : MVEIntSingleSrc<iname, suffix, size, pattern> {
  2125. let Inst{28} = 0b1;
  2126. let Inst{25-23} = 0b111;
  2127. let Inst{21-20} = 0b11;
  2128. let Inst{17} = 0b0;
  2129. let Inst{16} = !eq(saturate, 0);
  2130. let Inst{12-11} = 0b00;
  2131. let Inst{10} = saturate;
  2132. let Inst{9-8} = 0b11;
  2133. let Inst{7} = negate;
  2134. let Inst{6} = 0b1;
  2135. let Inst{4} = 0b0;
  2136. let Inst{0} = 0b0;
  2137. let validForTailPredication = 1;
  2138. }
  2139. multiclass MVE_VABSNEG_int_m<string iname, bit negate, bit saturate,
  2140. SDNode unpred_op, Intrinsic pred_int,
  2141. MVEVectorVTInfo VTI> {
  2142. def "" : MVE_VABSNEG_int<iname, VTI.Suffix, VTI.Size, negate, saturate>;
  2143. defvar Inst = !cast<Instruction>(NAME);
  2144. let Predicates = [HasMVEInt] in {
  2145. // VQABS and VQNEG have more difficult isel patterns defined elsewhere
  2146. if !not(saturate) then {
  2147. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
  2148. (VTI.Vec (Inst $v))>;
  2149. }
  2150. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
  2151. (VTI.Vec MQPR:$inactive))),
  2152. (VTI.Vec (Inst $v, ARMVCCThen, $mask, $inactive))>;
  2153. }
  2154. }
  2155. foreach VTI = [ MVE_v16s8, MVE_v8s16, MVE_v4s32 ] in {
  2156. defm "MVE_VABS" # VTI.Suffix : MVE_VABSNEG_int_m<
  2157. "vabs", 0, 0, abs, int_arm_mve_abs_predicated, VTI>;
  2158. defm "MVE_VQABS" # VTI.Suffix : MVE_VABSNEG_int_m<
  2159. "vqabs", 0, 1, ?, int_arm_mve_qabs_predicated, VTI>;
  2160. defm "MVE_VNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
  2161. "vneg", 1, 0, vnegq, int_arm_mve_neg_predicated, VTI>;
  2162. defm "MVE_VQNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
  2163. "vqneg", 1, 1, ?, int_arm_mve_qneg_predicated, VTI>;
  2164. }
  2165. // int_min/int_max: vector containing INT_MIN/INT_MAX VTI.Size times
  2166. // zero_vec: v4i32-initialized zero vector, potentially wrapped in a bitconvert
  2167. multiclass vqabsneg_pattern<MVEVectorVTInfo VTI, dag int_min, dag int_max,
  2168. dag zero_vec, MVE_VABSNEG_int vqabs_instruction,
  2169. MVE_VABSNEG_int vqneg_instruction> {
  2170. let Predicates = [HasMVEInt] in {
  2171. // The below tree can be replaced by a vqabs instruction, as it represents
  2172. // the following vectorized expression (r being the value in $reg):
  2173. // r > 0 ? r : (r == INT_MIN ? INT_MAX : -r)
  2174. def : Pat<(VTI.Vec (vselect
  2175. (VTI.Pred (ARMvcmpz (VTI.Vec MQPR:$reg), ARMCCgt)),
  2176. (VTI.Vec MQPR:$reg),
  2177. (VTI.Vec (vselect
  2178. (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
  2179. int_max,
  2180. (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))))),
  2181. (VTI.Vec (vqabs_instruction (VTI.Vec MQPR:$reg)))>;
  2182. // Similarly, this tree represents vqneg, i.e. the following vectorized expression:
  2183. // r == INT_MIN ? INT_MAX : -r
  2184. def : Pat<(VTI.Vec (vselect
  2185. (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
  2186. int_max,
  2187. (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))),
  2188. (VTI.Vec (vqneg_instruction (VTI.Vec MQPR:$reg)))>;
  2189. }
  2190. }
  2191. defm MVE_VQABSNEG_Ps8 : vqabsneg_pattern<MVE_v16i8,
  2192. (v16i8 (ARMvmovImm (i32 3712))),
  2193. (v16i8 (ARMvmovImm (i32 3711))),
  2194. (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
  2195. MVE_VQABSs8, MVE_VQNEGs8>;
  2196. defm MVE_VQABSNEG_Ps16 : vqabsneg_pattern<MVE_v8i16,
  2197. (v8i16 (ARMvmovImm (i32 2688))),
  2198. (v8i16 (ARMvmvnImm (i32 2688))),
  2199. (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
  2200. MVE_VQABSs16, MVE_VQNEGs16>;
  2201. defm MVE_VQABSNEG_Ps32 : vqabsneg_pattern<MVE_v4i32,
  2202. (v4i32 (ARMvmovImm (i32 1664))),
  2203. (v4i32 (ARMvmvnImm (i32 1664))),
  2204. (ARMvmovImm (i32 0)),
  2205. MVE_VQABSs32, MVE_VQNEGs32>;
  2206. class MVE_mod_imm<string iname, string suffix, bits<4> cmode, bit op,
  2207. dag iops, list<dag> pattern=[]>
  2208. : MVE_p<(outs MQPR:$Qd), iops, NoItinerary, iname, suffix, "$Qd, $imm",
  2209. vpred_r, "", pattern> {
  2210. bits<13> imm;
  2211. bits<4> Qd;
  2212. let Inst{28} = imm{7};
  2213. let Inst{25-23} = 0b111;
  2214. let Inst{22} = Qd{3};
  2215. let Inst{21-19} = 0b000;
  2216. let Inst{18-16} = imm{6-4};
  2217. let Inst{15-13} = Qd{2-0};
  2218. let Inst{12} = 0b0;
  2219. let Inst{11-8} = cmode{3-0};
  2220. let Inst{7-6} = 0b01;
  2221. let Inst{5} = op;
  2222. let Inst{4} = 0b1;
  2223. let Inst{3-0} = imm{3-0};
  2224. let DecoderMethod = "DecodeMVEModImmInstruction";
  2225. let validForTailPredication = 1;
  2226. }
  2227. let isReMaterializable = 1 in {
  2228. let isAsCheapAsAMove = 1 in {
  2229. def MVE_VMOVimmi8 : MVE_mod_imm<"vmov", "i8", {1,1,1,0}, 0b0, (ins nImmSplatI8:$imm)>;
  2230. def MVE_VMOVimmi16 : MVE_mod_imm<"vmov", "i16", {1,0,?,0}, 0b0, (ins nImmSplatI16:$imm)> {
  2231. let Inst{9} = imm{9};
  2232. }
  2233. def MVE_VMOVimmi32 : MVE_mod_imm<"vmov", "i32", {?,?,?,?}, 0b0, (ins nImmVMOVI32:$imm)> {
  2234. let Inst{11-8} = imm{11-8};
  2235. }
  2236. def MVE_VMOVimmi64 : MVE_mod_imm<"vmov", "i64", {1,1,1,0}, 0b1, (ins nImmSplatI64:$imm)>;
  2237. def MVE_VMOVimmf32 : MVE_mod_imm<"vmov", "f32", {1,1,1,1}, 0b0, (ins nImmVMOVF32:$imm)>;
  2238. } // let isAsCheapAsAMove = 1
  2239. def MVE_VMVNimmi16 : MVE_mod_imm<"vmvn", "i16", {1,0,?,0}, 0b1, (ins nImmSplatI16:$imm)> {
  2240. let Inst{9} = imm{9};
  2241. }
  2242. def MVE_VMVNimmi32 : MVE_mod_imm<"vmvn", "i32", {?,?,?,?}, 0b1, (ins nImmVMOVI32:$imm)> {
  2243. let Inst{11-8} = imm{11-8};
  2244. }
  2245. } // let isReMaterializable = 1
  2246. let Predicates = [HasMVEInt] in {
  2247. def : Pat<(v16i8 (ARMvmovImm timm:$simm)),
  2248. (v16i8 (MVE_VMOVimmi8 nImmSplatI8:$simm))>;
  2249. def : Pat<(v8i16 (ARMvmovImm timm:$simm)),
  2250. (v8i16 (MVE_VMOVimmi16 nImmSplatI16:$simm))>;
  2251. def : Pat<(v4i32 (ARMvmovImm timm:$simm)),
  2252. (v4i32 (MVE_VMOVimmi32 nImmVMOVI32:$simm))>;
  2253. def : Pat<(v2i64 (ARMvmovImm timm:$simm)),
  2254. (v2i64 (MVE_VMOVimmi64 nImmSplatI64:$simm))>;
  2255. def : Pat<(v8i16 (ARMvmvnImm timm:$simm)),
  2256. (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm))>;
  2257. def : Pat<(v4i32 (ARMvmvnImm timm:$simm)),
  2258. (v4i32 (MVE_VMVNimmi32 nImmVMOVI32:$simm))>;
  2259. def : Pat<(v4f32 (ARMvmovFPImm timm:$simm)),
  2260. (v4f32 (MVE_VMOVimmf32 nImmVMOVF32:$simm))>;
  2261. def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
  2262. MQPR:$inactive)),
  2263. (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm,
  2264. ARMVCCThen, VCCR:$pred, MQPR:$inactive))>;
  2265. def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
  2266. MQPR:$inactive)),
  2267. (v4i32 (MVE_VMVNimmi32 nImmSplatI32:$simm,
  2268. ARMVCCThen, VCCR:$pred, MQPR:$inactive))>;
  2269. }
  2270. class MVE_VMINMAXA<string iname, string suffix, bits<2> size,
  2271. bit bit_12, list<dag> pattern=[]>
  2272. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
  2273. NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
  2274. pattern> {
  2275. bits<4> Qd;
  2276. bits<4> Qm;
  2277. let Inst{28} = 0b0;
  2278. let Inst{25-23} = 0b100;
  2279. let Inst{22} = Qd{3};
  2280. let Inst{21-20} = 0b11;
  2281. let Inst{19-18} = size;
  2282. let Inst{17-16} = 0b11;
  2283. let Inst{15-13} = Qd{2-0};
  2284. let Inst{12} = bit_12;
  2285. let Inst{11-6} = 0b111010;
  2286. let Inst{5} = Qm{3};
  2287. let Inst{4} = 0b0;
  2288. let Inst{3-1} = Qm{2-0};
  2289. let Inst{0} = 0b1;
  2290. let validForTailPredication = 1;
  2291. }
  2292. multiclass MVE_VMINMAXA_m<string iname, MVEVectorVTInfo VTI,
  2293. SDNode unpred_op, Intrinsic pred_int, bit bit_12> {
  2294. def "" : MVE_VMINMAXA<iname, VTI.Suffix, VTI.Size, bit_12>;
  2295. defvar Inst = !cast<Instruction>(NAME);
  2296. let Predicates = [HasMVEInt] in {
  2297. // Unpredicated v(min|max)a
  2298. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qd), (abs (VTI.Vec MQPR:$Qm)))),
  2299. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
  2300. // Predicated v(min|max)a
  2301. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
  2302. (VTI.Pred VCCR:$mask))),
  2303. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
  2304. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  2305. }
  2306. }
  2307. multiclass MVE_VMINA<MVEVectorVTInfo VTI>
  2308. : MVE_VMINMAXA_m<"vmina", VTI, umin, int_arm_mve_vmina_predicated, 0b1>;
  2309. defm MVE_VMINAs8 : MVE_VMINA<MVE_v16s8>;
  2310. defm MVE_VMINAs16 : MVE_VMINA<MVE_v8s16>;
  2311. defm MVE_VMINAs32 : MVE_VMINA<MVE_v4s32>;
  2312. multiclass MVE_VMAXA<MVEVectorVTInfo VTI>
  2313. : MVE_VMINMAXA_m<"vmaxa", VTI, umax, int_arm_mve_vmaxa_predicated, 0b0>;
  2314. defm MVE_VMAXAs8 : MVE_VMAXA<MVE_v16s8>;
  2315. defm MVE_VMAXAs16 : MVE_VMAXA<MVE_v8s16>;
  2316. defm MVE_VMAXAs32 : MVE_VMAXA<MVE_v4s32>;
  2317. // end of MVE Integer instructions
  2318. // start of mve_imm_shift instructions
  2319. def MVE_VSHLC : MVE_p<(outs rGPR:$RdmDest, MQPR:$Qd),
  2320. (ins MQPR:$QdSrc, rGPR:$RdmSrc, long_shift:$imm),
  2321. NoItinerary, "vshlc", "", "$QdSrc, $RdmSrc, $imm",
  2322. vpred_n, "$RdmDest = $RdmSrc,$Qd = $QdSrc"> {
  2323. bits<5> imm;
  2324. bits<4> Qd;
  2325. bits<4> RdmDest;
  2326. let Inst{28} = 0b0;
  2327. let Inst{25-23} = 0b101;
  2328. let Inst{22} = Qd{3};
  2329. let Inst{21} = 0b1;
  2330. let Inst{20-16} = imm{4-0};
  2331. let Inst{15-13} = Qd{2-0};
  2332. let Inst{12-4} = 0b011111100;
  2333. let Inst{3-0} = RdmDest{3-0};
  2334. }
  2335. class MVE_shift_imm<dag oops, dag iops, string iname, string suffix,
  2336. string ops, vpred_ops vpred, string cstr,
  2337. list<dag> pattern=[]>
  2338. : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
  2339. bits<4> Qd;
  2340. bits<4> Qm;
  2341. let Inst{22} = Qd{3};
  2342. let Inst{15-13} = Qd{2-0};
  2343. let Inst{5} = Qm{3};
  2344. let Inst{3-1} = Qm{2-0};
  2345. }
  2346. class MVE_VMOVL<string iname, string suffix, bits<2> sz, bit U, bit top,
  2347. list<dag> pattern=[]>
  2348. : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
  2349. iname, suffix, "$Qd, $Qm", vpred_r, "",
  2350. pattern> {
  2351. let Inst{28} = U;
  2352. let Inst{25-23} = 0b101;
  2353. let Inst{21} = 0b1;
  2354. let Inst{20-19} = sz{1-0};
  2355. let Inst{18-16} = 0b000;
  2356. let Inst{12} = top;
  2357. let Inst{11-6} = 0b111101;
  2358. let Inst{4} = 0b0;
  2359. let Inst{0} = 0b0;
  2360. let doubleWidthResult = 1;
  2361. }
  2362. multiclass MVE_VMOVL_m<bit top, string chr, MVEVectorVTInfo OutVTI,
  2363. MVEVectorVTInfo InVTI> {
  2364. def "": MVE_VMOVL<"vmovl" # chr, InVTI.Suffix, OutVTI.Size,
  2365. InVTI.Unsigned, top>;
  2366. defvar Inst = !cast<Instruction>(NAME);
  2367. def : Pat<(OutVTI.Vec (int_arm_mve_vmovl_predicated (InVTI.Vec MQPR:$src),
  2368. (i32 InVTI.Unsigned), (i32 top),
  2369. (OutVTI.Pred VCCR:$pred),
  2370. (OutVTI.Vec MQPR:$inactive))),
  2371. (OutVTI.Vec (Inst (InVTI.Vec MQPR:$src), ARMVCCThen,
  2372. (OutVTI.Pred VCCR:$pred),
  2373. (OutVTI.Vec MQPR:$inactive)))>;
  2374. }
  2375. defm MVE_VMOVLs8bh : MVE_VMOVL_m<0, "b", MVE_v8s16, MVE_v16s8>;
  2376. defm MVE_VMOVLs8th : MVE_VMOVL_m<1, "t", MVE_v8s16, MVE_v16s8>;
  2377. defm MVE_VMOVLu8bh : MVE_VMOVL_m<0, "b", MVE_v8u16, MVE_v16u8>;
  2378. defm MVE_VMOVLu8th : MVE_VMOVL_m<1, "t", MVE_v8u16, MVE_v16u8>;
  2379. defm MVE_VMOVLs16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8s16>;
  2380. defm MVE_VMOVLs16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8s16>;
  2381. defm MVE_VMOVLu16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8u16>;
  2382. defm MVE_VMOVLu16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8u16>;
  2383. let Predicates = [HasMVEInt] in {
  2384. def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i16),
  2385. (MVE_VMOVLs16bh MQPR:$src)>;
  2386. def : Pat<(sext_inreg (v8i16 MQPR:$src), v8i8),
  2387. (MVE_VMOVLs8bh MQPR:$src)>;
  2388. def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i8),
  2389. (MVE_VMOVLs16bh (MVE_VMOVLs8bh MQPR:$src))>;
  2390. def : Pat<(sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), v8i8),
  2391. (MVE_VMOVLs8th MQPR:$src)>;
  2392. def : Pat<(sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))), v4i16),
  2393. (MVE_VMOVLs16th MQPR:$src)>;
  2394. // zext_inreg 8 -> 16
  2395. def : Pat<(ARMvbicImm (v8i16 MQPR:$src), (i32 0xAFF)),
  2396. (MVE_VMOVLu8bh MQPR:$src)>;
  2397. // zext_inreg 16 -> 32
  2398. def : Pat<(and (v4i32 MQPR:$src), (v4i32 (ARMvmovImm (i32 0xCFF)))),
  2399. (MVE_VMOVLu16bh MQPR:$src)>;
  2400. // Same zext_inreg with vrevs, picking the top half
  2401. def : Pat<(ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), (i32 0xAFF)),
  2402. (MVE_VMOVLu8th MQPR:$src)>;
  2403. def : Pat<(and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))),
  2404. (v4i32 (ARMvmovImm (i32 0xCFF)))),
  2405. (MVE_VMOVLu16th MQPR:$src)>;
  2406. }
  2407. class MVE_VSHLL_imm<string iname, string suffix, bit U, bit th,
  2408. Operand immtype, list<dag> pattern=[]>
  2409. : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm, immtype:$imm),
  2410. iname, suffix, "$Qd, $Qm, $imm", vpred_r, "", pattern> {
  2411. let Inst{28} = U;
  2412. let Inst{25-23} = 0b101;
  2413. let Inst{21} = 0b1;
  2414. let Inst{12} = th;
  2415. let Inst{11-6} = 0b111101;
  2416. let Inst{4} = 0b0;
  2417. let Inst{0} = 0b0;
  2418. // For the MVE_VSHLL_patterns multiclass to refer to
  2419. Operand immediateType = immtype;
  2420. let doubleWidthResult = 1;
  2421. }
  2422. // The immediate VSHLL instructions accept shift counts from 1 up to
  2423. // the lane width (8 or 16), but the full-width shifts have an
  2424. // entirely separate encoding, given below with 'lw' in the name.
  2425. class MVE_VSHLL_imm8<string iname, string suffix,
  2426. bit U, bit th, list<dag> pattern=[]>
  2427. : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_7, pattern> {
  2428. bits<3> imm;
  2429. let Inst{20-19} = 0b01;
  2430. let Inst{18-16} = imm;
  2431. }
  2432. class MVE_VSHLL_imm16<string iname, string suffix,
  2433. bit U, bit th, list<dag> pattern=[]>
  2434. : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_15, pattern> {
  2435. bits<4> imm;
  2436. let Inst{20} = 0b1;
  2437. let Inst{19-16} = imm;
  2438. }
  2439. def MVE_VSHLL_imms8bh : MVE_VSHLL_imm8 <"vshllb", "s8", 0b0, 0b0>;
  2440. def MVE_VSHLL_imms8th : MVE_VSHLL_imm8 <"vshllt", "s8", 0b0, 0b1>;
  2441. def MVE_VSHLL_immu8bh : MVE_VSHLL_imm8 <"vshllb", "u8", 0b1, 0b0>;
  2442. def MVE_VSHLL_immu8th : MVE_VSHLL_imm8 <"vshllt", "u8", 0b1, 0b1>;
  2443. def MVE_VSHLL_imms16bh : MVE_VSHLL_imm16<"vshllb", "s16", 0b0, 0b0>;
  2444. def MVE_VSHLL_imms16th : MVE_VSHLL_imm16<"vshllt", "s16", 0b0, 0b1>;
  2445. def MVE_VSHLL_immu16bh : MVE_VSHLL_imm16<"vshllb", "u16", 0b1, 0b0>;
  2446. def MVE_VSHLL_immu16th : MVE_VSHLL_imm16<"vshllt", "u16", 0b1, 0b1>;
  2447. class MVE_VSHLL_by_lane_width<string iname, string suffix, bits<2> size,
  2448. bit U, string ops, list<dag> pattern=[]>
  2449. : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
  2450. iname, suffix, ops, vpred_r, "", pattern> {
  2451. let Inst{28} = U;
  2452. let Inst{25-23} = 0b100;
  2453. let Inst{21-20} = 0b11;
  2454. let Inst{19-18} = size{1-0};
  2455. let Inst{17-16} = 0b01;
  2456. let Inst{11-6} = 0b111000;
  2457. let Inst{4} = 0b0;
  2458. let Inst{0} = 0b1;
  2459. let doubleWidthResult = 1;
  2460. }
  2461. multiclass MVE_VSHLL_lw<string iname, string suffix, bits<2> sz, bit U,
  2462. string ops, list<dag> pattern=[]> {
  2463. def bh : MVE_VSHLL_by_lane_width<iname#"b", suffix, sz, U, ops, pattern> {
  2464. let Inst{12} = 0b0;
  2465. }
  2466. def th : MVE_VSHLL_by_lane_width<iname#"t", suffix, sz, U, ops, pattern> {
  2467. let Inst{12} = 0b1;
  2468. }
  2469. }
  2470. defm MVE_VSHLL_lws8 : MVE_VSHLL_lw<"vshll", "s8", 0b00, 0b0, "$Qd, $Qm, #8">;
  2471. defm MVE_VSHLL_lws16 : MVE_VSHLL_lw<"vshll", "s16", 0b01, 0b0, "$Qd, $Qm, #16">;
  2472. defm MVE_VSHLL_lwu8 : MVE_VSHLL_lw<"vshll", "u8", 0b00, 0b1, "$Qd, $Qm, #8">;
  2473. defm MVE_VSHLL_lwu16 : MVE_VSHLL_lw<"vshll", "u16", 0b01, 0b1, "$Qd, $Qm, #16">;
  2474. multiclass MVE_VSHLL_patterns<MVEVectorVTInfo VTI, int top> {
  2475. defvar suffix = !strconcat(VTI.Suffix, !if(top, "th", "bh"));
  2476. defvar inst_imm = !cast<MVE_VSHLL_imm>("MVE_VSHLL_imm" # suffix);
  2477. defvar inst_lw = !cast<MVE_VSHLL_by_lane_width>("MVE_VSHLL_lw" # suffix);
  2478. defvar unpred_int = int_arm_mve_vshll_imm;
  2479. defvar pred_int = int_arm_mve_vshll_imm_predicated;
  2480. defvar imm = inst_imm.immediateType;
  2481. def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), imm:$imm,
  2482. (i32 VTI.Unsigned), (i32 top))),
  2483. (VTI.DblVec (inst_imm (VTI.Vec MQPR:$src), imm:$imm))>;
  2484. def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
  2485. (i32 VTI.Unsigned), (i32 top))),
  2486. (VTI.DblVec (inst_lw (VTI.Vec MQPR:$src)))>;
  2487. def : Pat<(VTI.DblVec (pred_int (VTI.Vec MQPR:$src), imm:$imm,
  2488. (i32 VTI.Unsigned), (i32 top),
  2489. (VTI.DblPred VCCR:$mask),
  2490. (VTI.DblVec MQPR:$inactive))),
  2491. (VTI.DblVec (inst_imm (VTI.Vec MQPR:$src), imm:$imm,
  2492. ARMVCCThen, (VTI.DblPred VCCR:$mask),
  2493. (VTI.DblVec MQPR:$inactive)))>;
  2494. def : Pat<(VTI.DblVec (pred_int (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
  2495. (i32 VTI.Unsigned), (i32 top),
  2496. (VTI.DblPred VCCR:$mask),
  2497. (VTI.DblVec MQPR:$inactive))),
  2498. (VTI.DblVec (inst_lw (VTI.Vec MQPR:$src), ARMVCCThen,
  2499. (VTI.DblPred VCCR:$mask),
  2500. (VTI.DblVec MQPR:$inactive)))>;
  2501. }
  2502. foreach VTI = [MVE_v16s8, MVE_v8s16, MVE_v16u8, MVE_v8u16] in
  2503. foreach top = [0, 1] in
  2504. defm : MVE_VSHLL_patterns<VTI, top>;
  2505. class MVE_shift_imm_partial<Operand imm, string iname, string suffix>
  2506. : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$QdSrc, MQPR:$Qm, imm:$imm),
  2507. iname, suffix, "$Qd, $Qm, $imm", vpred_n, "$Qd = $QdSrc"> {
  2508. Operand immediateType = imm;
  2509. }
  2510. class MVE_VxSHRN<string iname, string suffix, bit bit_12, bit bit_28,
  2511. Operand imm, list<dag> pattern=[]>
  2512. : MVE_shift_imm_partial<imm, iname, suffix> {
  2513. bits<5> imm;
  2514. let Inst{28} = bit_28;
  2515. let Inst{25-23} = 0b101;
  2516. let Inst{21} = 0b0;
  2517. let Inst{20-16} = imm{4-0};
  2518. let Inst{12} = bit_12;
  2519. let Inst{11-6} = 0b111111;
  2520. let Inst{4} = 0b0;
  2521. let Inst{0} = 0b1;
  2522. let validForTailPredication = 1;
  2523. let retainsPreviousHalfElement = 1;
  2524. }
  2525. def MVE_VRSHRNi16bh : MVE_VxSHRN<"vrshrnb", "i16", 0b0, 0b1, shr_imm8> {
  2526. let Inst{20-19} = 0b01;
  2527. }
  2528. def MVE_VRSHRNi16th : MVE_VxSHRN<"vrshrnt", "i16", 0b1, 0b1, shr_imm8> {
  2529. let Inst{20-19} = 0b01;
  2530. }
  2531. def MVE_VRSHRNi32bh : MVE_VxSHRN<"vrshrnb", "i32", 0b0, 0b1, shr_imm16> {
  2532. let Inst{20} = 0b1;
  2533. }
  2534. def MVE_VRSHRNi32th : MVE_VxSHRN<"vrshrnt", "i32", 0b1, 0b1, shr_imm16> {
  2535. let Inst{20} = 0b1;
  2536. }
  2537. def MVE_VSHRNi16bh : MVE_VxSHRN<"vshrnb", "i16", 0b0, 0b0, shr_imm8> {
  2538. let Inst{20-19} = 0b01;
  2539. }
  2540. def MVE_VSHRNi16th : MVE_VxSHRN<"vshrnt", "i16", 0b1, 0b0, shr_imm8> {
  2541. let Inst{20-19} = 0b01;
  2542. }
  2543. def MVE_VSHRNi32bh : MVE_VxSHRN<"vshrnb", "i32", 0b0, 0b0, shr_imm16> {
  2544. let Inst{20} = 0b1;
  2545. }
  2546. def MVE_VSHRNi32th : MVE_VxSHRN<"vshrnt", "i32", 0b1, 0b0, shr_imm16> {
  2547. let Inst{20} = 0b1;
  2548. }
  2549. class MVE_VxQRSHRUN<string iname, string suffix, bit bit_28, bit bit_12,
  2550. Operand imm, list<dag> pattern=[]>
  2551. : MVE_shift_imm_partial<imm, iname, suffix> {
  2552. bits<5> imm;
  2553. let Inst{28} = bit_28;
  2554. let Inst{25-23} = 0b101;
  2555. let Inst{21} = 0b0;
  2556. let Inst{20-16} = imm{4-0};
  2557. let Inst{12} = bit_12;
  2558. let Inst{11-6} = 0b111111;
  2559. let Inst{4} = 0b0;
  2560. let Inst{0} = 0b0;
  2561. let validForTailPredication = 1;
  2562. let retainsPreviousHalfElement = 1;
  2563. }
  2564. def MVE_VQRSHRUNs16bh : MVE_VxQRSHRUN<
  2565. "vqrshrunb", "s16", 0b1, 0b0, shr_imm8> {
  2566. let Inst{20-19} = 0b01;
  2567. }
  2568. def MVE_VQRSHRUNs16th : MVE_VxQRSHRUN<
  2569. "vqrshrunt", "s16", 0b1, 0b1, shr_imm8> {
  2570. let Inst{20-19} = 0b01;
  2571. }
  2572. def MVE_VQRSHRUNs32bh : MVE_VxQRSHRUN<
  2573. "vqrshrunb", "s32", 0b1, 0b0, shr_imm16> {
  2574. let Inst{20} = 0b1;
  2575. }
  2576. def MVE_VQRSHRUNs32th : MVE_VxQRSHRUN<
  2577. "vqrshrunt", "s32", 0b1, 0b1, shr_imm16> {
  2578. let Inst{20} = 0b1;
  2579. }
  2580. def MVE_VQSHRUNs16bh : MVE_VxQRSHRUN<
  2581. "vqshrunb", "s16", 0b0, 0b0, shr_imm8> {
  2582. let Inst{20-19} = 0b01;
  2583. }
  2584. def MVE_VQSHRUNs16th : MVE_VxQRSHRUN<
  2585. "vqshrunt", "s16", 0b0, 0b1, shr_imm8> {
  2586. let Inst{20-19} = 0b01;
  2587. }
  2588. def MVE_VQSHRUNs32bh : MVE_VxQRSHRUN<
  2589. "vqshrunb", "s32", 0b0, 0b0, shr_imm16> {
  2590. let Inst{20} = 0b1;
  2591. }
  2592. def MVE_VQSHRUNs32th : MVE_VxQRSHRUN<
  2593. "vqshrunt", "s32", 0b0, 0b1, shr_imm16> {
  2594. let Inst{20} = 0b1;
  2595. }
  2596. class MVE_VxQRSHRN<string iname, string suffix, bit bit_0, bit bit_12,
  2597. Operand imm, list<dag> pattern=[]>
  2598. : MVE_shift_imm_partial<imm, iname, suffix> {
  2599. bits<5> imm;
  2600. let Inst{25-23} = 0b101;
  2601. let Inst{21} = 0b0;
  2602. let Inst{20-16} = imm{4-0};
  2603. let Inst{12} = bit_12;
  2604. let Inst{11-6} = 0b111101;
  2605. let Inst{4} = 0b0;
  2606. let Inst{0} = bit_0;
  2607. let validForTailPredication = 1;
  2608. let retainsPreviousHalfElement = 1;
  2609. }
  2610. multiclass MVE_VxQRSHRN_types<string iname, bit bit_0, bit bit_12> {
  2611. def s16 : MVE_VxQRSHRN<iname, "s16", bit_0, bit_12, shr_imm8> {
  2612. let Inst{28} = 0b0;
  2613. let Inst{20-19} = 0b01;
  2614. }
  2615. def u16 : MVE_VxQRSHRN<iname, "u16", bit_0, bit_12, shr_imm8> {
  2616. let Inst{28} = 0b1;
  2617. let Inst{20-19} = 0b01;
  2618. }
  2619. def s32 : MVE_VxQRSHRN<iname, "s32", bit_0, bit_12, shr_imm16> {
  2620. let Inst{28} = 0b0;
  2621. let Inst{20} = 0b1;
  2622. }
  2623. def u32 : MVE_VxQRSHRN<iname, "u32", bit_0, bit_12, shr_imm16> {
  2624. let Inst{28} = 0b1;
  2625. let Inst{20} = 0b1;
  2626. }
  2627. }
  2628. defm MVE_VQRSHRNbh : MVE_VxQRSHRN_types<"vqrshrnb", 0b1, 0b0>;
  2629. defm MVE_VQRSHRNth : MVE_VxQRSHRN_types<"vqrshrnt", 0b1, 0b1>;
  2630. defm MVE_VQSHRNbh : MVE_VxQRSHRN_types<"vqshrnb", 0b0, 0b0>;
  2631. defm MVE_VQSHRNth : MVE_VxQRSHRN_types<"vqshrnt", 0b0, 0b1>;
  2632. multiclass MVE_VSHRN_patterns<MVE_shift_imm_partial inst,
  2633. MVEVectorVTInfo OutVTI, MVEVectorVTInfo InVTI,
  2634. bit q, bit r, bit top> {
  2635. defvar inparams = (? (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
  2636. (inst.immediateType:$imm), (i32 q), (i32 r),
  2637. (i32 OutVTI.Unsigned), (i32 InVTI.Unsigned), (i32 top));
  2638. defvar outparams = (inst (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
  2639. (imm:$imm));
  2640. def : Pat<(OutVTI.Vec !setdagop(inparams, int_arm_mve_vshrn)),
  2641. (OutVTI.Vec outparams)>;
  2642. def : Pat<(OutVTI.Vec !con(inparams, (int_arm_mve_vshrn_predicated
  2643. (InVTI.Pred VCCR:$pred)))),
  2644. (OutVTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred)))>;
  2645. }
  2646. defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh, MVE_v16s8, MVE_v8s16, 0,0,0>;
  2647. defm : MVE_VSHRN_patterns<MVE_VSHRNi16th, MVE_v16s8, MVE_v8s16, 0,0,1>;
  2648. defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh, MVE_v8s16, MVE_v4s32, 0,0,0>;
  2649. defm : MVE_VSHRN_patterns<MVE_VSHRNi32th, MVE_v8s16, MVE_v4s32, 0,0,1>;
  2650. defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh, MVE_v16u8, MVE_v8u16, 0,0,0>;
  2651. defm : MVE_VSHRN_patterns<MVE_VSHRNi16th, MVE_v16u8, MVE_v8u16, 0,0,1>;
  2652. defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh, MVE_v8u16, MVE_v4u32, 0,0,0>;
  2653. defm : MVE_VSHRN_patterns<MVE_VSHRNi32th, MVE_v8u16, MVE_v4u32, 0,0,1>;
  2654. defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh, MVE_v16s8, MVE_v8s16, 0,1,0>;
  2655. defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th, MVE_v16s8, MVE_v8s16, 0,1,1>;
  2656. defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh, MVE_v8s16, MVE_v4s32, 0,1,0>;
  2657. defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th, MVE_v8s16, MVE_v4s32, 0,1,1>;
  2658. defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh, MVE_v16u8, MVE_v8u16, 0,1,0>;
  2659. defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th, MVE_v16u8, MVE_v8u16, 0,1,1>;
  2660. defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh, MVE_v8u16, MVE_v4u32, 0,1,0>;
  2661. defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th, MVE_v8u16, MVE_v4u32, 0,1,1>;
  2662. defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs16, MVE_v16s8, MVE_v8s16, 1,0,0>;
  2663. defm : MVE_VSHRN_patterns<MVE_VQSHRNths16, MVE_v16s8, MVE_v8s16, 1,0,1>;
  2664. defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs32, MVE_v8s16, MVE_v4s32, 1,0,0>;
  2665. defm : MVE_VSHRN_patterns<MVE_VQSHRNths32, MVE_v8s16, MVE_v4s32, 1,0,1>;
  2666. defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu16, MVE_v16u8, MVE_v8u16, 1,0,0>;
  2667. defm : MVE_VSHRN_patterns<MVE_VQSHRNthu16, MVE_v16u8, MVE_v8u16, 1,0,1>;
  2668. defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu32, MVE_v8u16, MVE_v4u32, 1,0,0>;
  2669. defm : MVE_VSHRN_patterns<MVE_VQSHRNthu32, MVE_v8u16, MVE_v4u32, 1,0,1>;
  2670. defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs16, MVE_v16s8, MVE_v8s16, 1,1,0>;
  2671. defm : MVE_VSHRN_patterns<MVE_VQRSHRNths16, MVE_v16s8, MVE_v8s16, 1,1,1>;
  2672. defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs32, MVE_v8s16, MVE_v4s32, 1,1,0>;
  2673. defm : MVE_VSHRN_patterns<MVE_VQRSHRNths32, MVE_v8s16, MVE_v4s32, 1,1,1>;
  2674. defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu16, MVE_v16u8, MVE_v8u16, 1,1,0>;
  2675. defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu16, MVE_v16u8, MVE_v8u16, 1,1,1>;
  2676. defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu32, MVE_v8u16, MVE_v4u32, 1,1,0>;
  2677. defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu32, MVE_v8u16, MVE_v4u32, 1,1,1>;
  2678. defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16bh, MVE_v16u8, MVE_v8s16, 1,0,0>;
  2679. defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16th, MVE_v16u8, MVE_v8s16, 1,0,1>;
  2680. defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32bh, MVE_v8u16, MVE_v4s32, 1,0,0>;
  2681. defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32th, MVE_v8u16, MVE_v4s32, 1,0,1>;
  2682. defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16bh, MVE_v16u8, MVE_v8s16, 1,1,0>;
  2683. defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16th, MVE_v16u8, MVE_v8s16, 1,1,1>;
  2684. defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32bh, MVE_v8u16, MVE_v4s32, 1,1,0>;
  2685. defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32th, MVE_v8u16, MVE_v4s32, 1,1,1>;
  2686. // end of mve_imm_shift instructions
  2687. // start of mve_shift instructions
  2688. class MVE_shift_by_vec<string iname, string suffix, bit U,
  2689. bits<2> size, bit bit_4, bit bit_8>
  2690. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm, MQPR:$Qn), NoItinerary,
  2691. iname, suffix, "$Qd, $Qm, $Qn", vpred_r, "", []> {
  2692. // Shift instructions which take a vector of shift counts
  2693. bits<4> Qd;
  2694. bits<4> Qm;
  2695. bits<4> Qn;
  2696. let Inst{28} = U;
  2697. let Inst{25-24} = 0b11;
  2698. let Inst{23} = 0b0;
  2699. let Inst{22} = Qd{3};
  2700. let Inst{21-20} = size;
  2701. let Inst{19-17} = Qn{2-0};
  2702. let Inst{16} = 0b0;
  2703. let Inst{15-13} = Qd{2-0};
  2704. let Inst{12-9} = 0b0010;
  2705. let Inst{8} = bit_8;
  2706. let Inst{7} = Qn{3};
  2707. let Inst{6} = 0b1;
  2708. let Inst{5} = Qm{3};
  2709. let Inst{4} = bit_4;
  2710. let Inst{3-1} = Qm{2-0};
  2711. let Inst{0} = 0b0;
  2712. let validForTailPredication = 1;
  2713. }
  2714. multiclass MVE_shift_by_vec_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
  2715. def "" : MVE_shift_by_vec<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
  2716. defvar Inst = !cast<Instruction>(NAME);
  2717. def : Pat<(VTI.Vec (int_arm_mve_vshl_vector
  2718. (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
  2719. (i32 q), (i32 r), (i32 VTI.Unsigned))),
  2720. (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh)))>;
  2721. def : Pat<(VTI.Vec (int_arm_mve_vshl_vector_predicated
  2722. (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
  2723. (i32 q), (i32 r), (i32 VTI.Unsigned),
  2724. (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
  2725. (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
  2726. ARMVCCThen, (VTI.Pred VCCR:$mask),
  2727. (VTI.Vec MQPR:$inactive)))>;
  2728. }
  2729. multiclass mve_shift_by_vec_multi<string iname, bit bit_4, bit bit_8> {
  2730. defm s8 : MVE_shift_by_vec_p<iname, MVE_v16s8, bit_4, bit_8>;
  2731. defm s16 : MVE_shift_by_vec_p<iname, MVE_v8s16, bit_4, bit_8>;
  2732. defm s32 : MVE_shift_by_vec_p<iname, MVE_v4s32, bit_4, bit_8>;
  2733. defm u8 : MVE_shift_by_vec_p<iname, MVE_v16u8, bit_4, bit_8>;
  2734. defm u16 : MVE_shift_by_vec_p<iname, MVE_v8u16, bit_4, bit_8>;
  2735. defm u32 : MVE_shift_by_vec_p<iname, MVE_v4u32, bit_4, bit_8>;
  2736. }
  2737. defm MVE_VSHL_by_vec : mve_shift_by_vec_multi<"vshl", 0b0, 0b0>;
  2738. defm MVE_VQSHL_by_vec : mve_shift_by_vec_multi<"vqshl", 0b1, 0b0>;
  2739. defm MVE_VQRSHL_by_vec : mve_shift_by_vec_multi<"vqrshl", 0b1, 0b1>;
  2740. defm MVE_VRSHL_by_vec : mve_shift_by_vec_multi<"vrshl", 0b0, 0b1>;
  2741. let Predicates = [HasMVEInt] in {
  2742. def : Pat<(v4i32 (ARMvshlu (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn))),
  2743. (v4i32 (MVE_VSHL_by_vecu32 (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)))>;
  2744. def : Pat<(v8i16 (ARMvshlu (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn))),
  2745. (v8i16 (MVE_VSHL_by_vecu16 (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)))>;
  2746. def : Pat<(v16i8 (ARMvshlu (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn))),
  2747. (v16i8 (MVE_VSHL_by_vecu8 (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)))>;
  2748. def : Pat<(v4i32 (ARMvshls (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn))),
  2749. (v4i32 (MVE_VSHL_by_vecs32 (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)))>;
  2750. def : Pat<(v8i16 (ARMvshls (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn))),
  2751. (v8i16 (MVE_VSHL_by_vecs16 (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)))>;
  2752. def : Pat<(v16i8 (ARMvshls (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn))),
  2753. (v16i8 (MVE_VSHL_by_vecs8 (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)))>;
  2754. }
  2755. class MVE_shift_with_imm<string iname, string suffix, dag oops, dag iops,
  2756. string ops, vpred_ops vpred, string cstr,
  2757. list<dag> pattern=[]>
  2758. : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
  2759. bits<4> Qd;
  2760. bits<4> Qm;
  2761. let Inst{23} = 0b1;
  2762. let Inst{22} = Qd{3};
  2763. let Inst{15-13} = Qd{2-0};
  2764. let Inst{12-11} = 0b00;
  2765. let Inst{7-6} = 0b01;
  2766. let Inst{5} = Qm{3};
  2767. let Inst{4} = 0b1;
  2768. let Inst{3-1} = Qm{2-0};
  2769. let Inst{0} = 0b0;
  2770. let validForTailPredication = 1;
  2771. // For the MVE_shift_imm_patterns multiclass to refer to
  2772. MVEVectorVTInfo VTI;
  2773. Operand immediateType;
  2774. Intrinsic unpred_int;
  2775. Intrinsic pred_int;
  2776. dag unsignedFlag = (?);
  2777. }
  2778. class MVE_VSxI_imm<string iname, string suffix, bit bit_8, Operand immType>
  2779. : MVE_shift_with_imm<iname, suffix, (outs MQPR:$Qd),
  2780. (ins MQPR:$Qd_src, MQPR:$Qm, immType:$imm),
  2781. "$Qd, $Qm, $imm", vpred_n, "$Qd = $Qd_src"> {
  2782. bits<6> imm;
  2783. let Inst{28} = 0b1;
  2784. let Inst{25-24} = 0b11;
  2785. let Inst{21-16} = imm;
  2786. let Inst{10-9} = 0b10;
  2787. let Inst{8} = bit_8;
  2788. let validForTailPredication = 1;
  2789. Operand immediateType = immType;
  2790. }
  2791. def MVE_VSRIimm8 : MVE_VSxI_imm<"vsri", "8", 0b0, shr_imm8> {
  2792. let Inst{21-19} = 0b001;
  2793. }
  2794. def MVE_VSRIimm16 : MVE_VSxI_imm<"vsri", "16", 0b0, shr_imm16> {
  2795. let Inst{21-20} = 0b01;
  2796. }
  2797. def MVE_VSRIimm32 : MVE_VSxI_imm<"vsri", "32", 0b0, shr_imm32> {
  2798. let Inst{21} = 0b1;
  2799. }
  2800. def MVE_VSLIimm8 : MVE_VSxI_imm<"vsli", "8", 0b1, imm0_7> {
  2801. let Inst{21-19} = 0b001;
  2802. }
  2803. def MVE_VSLIimm16 : MVE_VSxI_imm<"vsli", "16", 0b1, imm0_15> {
  2804. let Inst{21-20} = 0b01;
  2805. }
  2806. def MVE_VSLIimm32 : MVE_VSxI_imm<"vsli", "32", 0b1,imm0_31> {
  2807. let Inst{21} = 0b1;
  2808. }
  2809. multiclass MVE_VSxI_patterns<MVE_VSxI_imm inst, string name,
  2810. MVEVectorVTInfo VTI> {
  2811. defvar inparams = (? (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
  2812. (inst.immediateType:$imm));
  2813. defvar outparams = (inst (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
  2814. (inst.immediateType:$imm));
  2815. defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # name);
  2816. defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # name # "_predicated");
  2817. def : Pat<(VTI.Vec !setdagop(inparams, unpred_int)),
  2818. (VTI.Vec outparams)>;
  2819. def : Pat<(VTI.Vec !con(inparams, (pred_int (VTI.Pred VCCR:$pred)))),
  2820. (VTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred)))>;
  2821. }
  2822. defm : MVE_VSxI_patterns<MVE_VSLIimm8, "vsli", MVE_v16i8>;
  2823. defm : MVE_VSxI_patterns<MVE_VSLIimm16, "vsli", MVE_v8i16>;
  2824. defm : MVE_VSxI_patterns<MVE_VSLIimm32, "vsli", MVE_v4i32>;
  2825. defm : MVE_VSxI_patterns<MVE_VSRIimm8, "vsri", MVE_v16i8>;
  2826. defm : MVE_VSxI_patterns<MVE_VSRIimm16, "vsri", MVE_v8i16>;
  2827. defm : MVE_VSxI_patterns<MVE_VSRIimm32, "vsri", MVE_v4i32>;
  2828. class MVE_VQSHL_imm<MVEVectorVTInfo VTI_, Operand immType>
  2829. : MVE_shift_with_imm<"vqshl", VTI_.Suffix, (outs MQPR:$Qd),
  2830. (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
  2831. vpred_r, ""> {
  2832. bits<6> imm;
  2833. let Inst{28} = VTI_.Unsigned;
  2834. let Inst{25-24} = 0b11;
  2835. let Inst{21-16} = imm;
  2836. let Inst{10-8} = 0b111;
  2837. let VTI = VTI_;
  2838. let immediateType = immType;
  2839. let unsignedFlag = (? (i32 VTI.Unsigned));
  2840. }
  2841. let unpred_int = int_arm_mve_vqshl_imm,
  2842. pred_int = int_arm_mve_vqshl_imm_predicated in {
  2843. def MVE_VQSHLimms8 : MVE_VQSHL_imm<MVE_v16s8, imm0_7> {
  2844. let Inst{21-19} = 0b001;
  2845. }
  2846. def MVE_VQSHLimmu8 : MVE_VQSHL_imm<MVE_v16u8, imm0_7> {
  2847. let Inst{21-19} = 0b001;
  2848. }
  2849. def MVE_VQSHLimms16 : MVE_VQSHL_imm<MVE_v8s16, imm0_15> {
  2850. let Inst{21-20} = 0b01;
  2851. }
  2852. def MVE_VQSHLimmu16 : MVE_VQSHL_imm<MVE_v8u16, imm0_15> {
  2853. let Inst{21-20} = 0b01;
  2854. }
  2855. def MVE_VQSHLimms32 : MVE_VQSHL_imm<MVE_v4s32, imm0_31> {
  2856. let Inst{21} = 0b1;
  2857. }
  2858. def MVE_VQSHLimmu32 : MVE_VQSHL_imm<MVE_v4u32, imm0_31> {
  2859. let Inst{21} = 0b1;
  2860. }
  2861. }
  2862. class MVE_VQSHLU_imm<MVEVectorVTInfo VTI_, Operand immType>
  2863. : MVE_shift_with_imm<"vqshlu", VTI_.Suffix, (outs MQPR:$Qd),
  2864. (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
  2865. vpred_r, ""> {
  2866. bits<6> imm;
  2867. let Inst{28} = 0b1;
  2868. let Inst{25-24} = 0b11;
  2869. let Inst{21-16} = imm;
  2870. let Inst{10-8} = 0b110;
  2871. let VTI = VTI_;
  2872. let immediateType = immType;
  2873. }
  2874. let unpred_int = int_arm_mve_vqshlu_imm,
  2875. pred_int = int_arm_mve_vqshlu_imm_predicated in {
  2876. def MVE_VQSHLU_imms8 : MVE_VQSHLU_imm<MVE_v16s8, imm0_7> {
  2877. let Inst{21-19} = 0b001;
  2878. }
  2879. def MVE_VQSHLU_imms16 : MVE_VQSHLU_imm<MVE_v8s16, imm0_15> {
  2880. let Inst{21-20} = 0b01;
  2881. }
  2882. def MVE_VQSHLU_imms32 : MVE_VQSHLU_imm<MVE_v4s32, imm0_31> {
  2883. let Inst{21} = 0b1;
  2884. }
  2885. }
  2886. class MVE_VRSHR_imm<MVEVectorVTInfo VTI_, Operand immType>
  2887. : MVE_shift_with_imm<"vrshr", VTI_.Suffix, (outs MQPR:$Qd),
  2888. (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
  2889. vpred_r, ""> {
  2890. bits<6> imm;
  2891. let Inst{28} = VTI_.Unsigned;
  2892. let Inst{25-24} = 0b11;
  2893. let Inst{21-16} = imm;
  2894. let Inst{10-8} = 0b010;
  2895. let VTI = VTI_;
  2896. let immediateType = immType;
  2897. let unsignedFlag = (? (i32 VTI.Unsigned));
  2898. }
  2899. let unpred_int = int_arm_mve_vrshr_imm,
  2900. pred_int = int_arm_mve_vrshr_imm_predicated in {
  2901. def MVE_VRSHR_imms8 : MVE_VRSHR_imm<MVE_v16s8, shr_imm8> {
  2902. let Inst{21-19} = 0b001;
  2903. }
  2904. def MVE_VRSHR_immu8 : MVE_VRSHR_imm<MVE_v16u8, shr_imm8> {
  2905. let Inst{21-19} = 0b001;
  2906. }
  2907. def MVE_VRSHR_imms16 : MVE_VRSHR_imm<MVE_v8s16, shr_imm16> {
  2908. let Inst{21-20} = 0b01;
  2909. }
  2910. def MVE_VRSHR_immu16 : MVE_VRSHR_imm<MVE_v8u16, shr_imm16> {
  2911. let Inst{21-20} = 0b01;
  2912. }
  2913. def MVE_VRSHR_imms32 : MVE_VRSHR_imm<MVE_v4s32, shr_imm32> {
  2914. let Inst{21} = 0b1;
  2915. }
  2916. def MVE_VRSHR_immu32 : MVE_VRSHR_imm<MVE_v4u32, shr_imm32> {
  2917. let Inst{21} = 0b1;
  2918. }
  2919. }
  2920. multiclass MVE_shift_imm_patterns<MVE_shift_with_imm inst> {
  2921. def : Pat<(inst.VTI.Vec !con((inst.unpred_int (inst.VTI.Vec MQPR:$src),
  2922. inst.immediateType:$imm),
  2923. inst.unsignedFlag)),
  2924. (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
  2925. inst.immediateType:$imm))>;
  2926. def : Pat<(inst.VTI.Vec !con((inst.pred_int (inst.VTI.Vec MQPR:$src),
  2927. inst.immediateType:$imm),
  2928. inst.unsignedFlag,
  2929. (? (inst.VTI.Pred VCCR:$mask),
  2930. (inst.VTI.Vec MQPR:$inactive)))),
  2931. (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
  2932. inst.immediateType:$imm,
  2933. ARMVCCThen, (inst.VTI.Pred VCCR:$mask),
  2934. (inst.VTI.Vec MQPR:$inactive)))>;
  2935. }
  2936. defm : MVE_shift_imm_patterns<MVE_VQSHLimms8>;
  2937. defm : MVE_shift_imm_patterns<MVE_VQSHLimmu8>;
  2938. defm : MVE_shift_imm_patterns<MVE_VQSHLimms16>;
  2939. defm : MVE_shift_imm_patterns<MVE_VQSHLimmu16>;
  2940. defm : MVE_shift_imm_patterns<MVE_VQSHLimms32>;
  2941. defm : MVE_shift_imm_patterns<MVE_VQSHLimmu32>;
  2942. defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms8>;
  2943. defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms16>;
  2944. defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms32>;
  2945. defm : MVE_shift_imm_patterns<MVE_VRSHR_imms8>;
  2946. defm : MVE_shift_imm_patterns<MVE_VRSHR_immu8>;
  2947. defm : MVE_shift_imm_patterns<MVE_VRSHR_imms16>;
  2948. defm : MVE_shift_imm_patterns<MVE_VRSHR_immu16>;
  2949. defm : MVE_shift_imm_patterns<MVE_VRSHR_imms32>;
  2950. defm : MVE_shift_imm_patterns<MVE_VRSHR_immu32>;
  2951. class MVE_VSHR_imm<string suffix, dag imm>
  2952. : MVE_shift_with_imm<"vshr", suffix, (outs MQPR:$Qd),
  2953. !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
  2954. vpred_r, ""> {
  2955. bits<6> imm;
  2956. let Inst{25-24} = 0b11;
  2957. let Inst{21-16} = imm;
  2958. let Inst{10-8} = 0b000;
  2959. }
  2960. def MVE_VSHR_imms8 : MVE_VSHR_imm<"s8", (ins shr_imm8:$imm)> {
  2961. let Inst{28} = 0b0;
  2962. let Inst{21-19} = 0b001;
  2963. }
  2964. def MVE_VSHR_immu8 : MVE_VSHR_imm<"u8", (ins shr_imm8:$imm)> {
  2965. let Inst{28} = 0b1;
  2966. let Inst{21-19} = 0b001;
  2967. }
  2968. def MVE_VSHR_imms16 : MVE_VSHR_imm<"s16", (ins shr_imm16:$imm)> {
  2969. let Inst{28} = 0b0;
  2970. let Inst{21-20} = 0b01;
  2971. }
  2972. def MVE_VSHR_immu16 : MVE_VSHR_imm<"u16", (ins shr_imm16:$imm)> {
  2973. let Inst{28} = 0b1;
  2974. let Inst{21-20} = 0b01;
  2975. }
  2976. def MVE_VSHR_imms32 : MVE_VSHR_imm<"s32", (ins shr_imm32:$imm)> {
  2977. let Inst{28} = 0b0;
  2978. let Inst{21} = 0b1;
  2979. }
  2980. def MVE_VSHR_immu32 : MVE_VSHR_imm<"u32", (ins shr_imm32:$imm)> {
  2981. let Inst{28} = 0b1;
  2982. let Inst{21} = 0b1;
  2983. }
  2984. class MVE_VSHL_imm<string suffix, dag imm>
  2985. : MVE_shift_with_imm<"vshl", suffix, (outs MQPR:$Qd),
  2986. !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
  2987. vpred_r, ""> {
  2988. bits<6> imm;
  2989. let Inst{28} = 0b0;
  2990. let Inst{25-24} = 0b11;
  2991. let Inst{21-16} = imm;
  2992. let Inst{10-8} = 0b101;
  2993. }
  2994. def MVE_VSHL_immi8 : MVE_VSHL_imm<"i8", (ins imm0_7:$imm)> {
  2995. let Inst{21-19} = 0b001;
  2996. }
  2997. def MVE_VSHL_immi16 : MVE_VSHL_imm<"i16", (ins imm0_15:$imm)> {
  2998. let Inst{21-20} = 0b01;
  2999. }
  3000. def MVE_VSHL_immi32 : MVE_VSHL_imm<"i32", (ins imm0_31:$imm)> {
  3001. let Inst{21} = 0b1;
  3002. }
  3003. multiclass MVE_immediate_shift_patterns_inner<
  3004. MVEVectorVTInfo VTI, Operand imm_operand_type, SDNode unpred_op,
  3005. Intrinsic pred_int, Instruction inst, list<int> unsignedFlag = []> {
  3006. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src), imm_operand_type:$imm)),
  3007. (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm))>;
  3008. def : Pat<(VTI.Vec !con((pred_int (VTI.Vec MQPR:$src), imm_operand_type:$imm),
  3009. !dag(pred_int, unsignedFlag, ?),
  3010. (pred_int (VTI.Pred VCCR:$mask),
  3011. (VTI.Vec MQPR:$inactive)))),
  3012. (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm,
  3013. ARMVCCThen, (VTI.Pred VCCR:$mask),
  3014. (VTI.Vec MQPR:$inactive)))>;
  3015. }
  3016. multiclass MVE_immediate_shift_patterns<MVEVectorVTInfo VTI,
  3017. Operand imm_operand_type> {
  3018. defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
  3019. ARMvshlImm, int_arm_mve_shl_imm_predicated,
  3020. !cast<Instruction>("MVE_VSHL_immi" # VTI.BitsSuffix)>;
  3021. defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
  3022. ARMvshruImm, int_arm_mve_shr_imm_predicated,
  3023. !cast<Instruction>("MVE_VSHR_immu" # VTI.BitsSuffix), [1]>;
  3024. defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
  3025. ARMvshrsImm, int_arm_mve_shr_imm_predicated,
  3026. !cast<Instruction>("MVE_VSHR_imms" # VTI.BitsSuffix), [0]>;
  3027. }
  3028. let Predicates = [HasMVEInt] in {
  3029. defm : MVE_immediate_shift_patterns<MVE_v16i8, imm0_7>;
  3030. defm : MVE_immediate_shift_patterns<MVE_v8i16, imm0_15>;
  3031. defm : MVE_immediate_shift_patterns<MVE_v4i32, imm0_31>;
  3032. }
  3033. // end of mve_shift instructions
  3034. // start of MVE Floating Point instructions
  3035. class MVE_float<string iname, string suffix, dag oops, dag iops, string ops,
  3036. vpred_ops vpred, string cstr, list<dag> pattern=[]>
  3037. : MVE_f<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
  3038. bits<4> Qm;
  3039. let Inst{12} = 0b0;
  3040. let Inst{6} = 0b1;
  3041. let Inst{5} = Qm{3};
  3042. let Inst{3-1} = Qm{2-0};
  3043. let Inst{0} = 0b0;
  3044. }
  3045. class MVE_VRINT<string rmode, bits<3> op, string suffix, bits<2> size,
  3046. list<dag> pattern=[]>
  3047. : MVE_float<!strconcat("vrint", rmode), suffix, (outs MQPR:$Qd),
  3048. (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
  3049. bits<4> Qd;
  3050. let Inst{28} = 0b1;
  3051. let Inst{25-23} = 0b111;
  3052. let Inst{22} = Qd{3};
  3053. let Inst{21-20} = 0b11;
  3054. let Inst{19-18} = size;
  3055. let Inst{17-16} = 0b10;
  3056. let Inst{15-13} = Qd{2-0};
  3057. let Inst{11-10} = 0b01;
  3058. let Inst{9-7} = op{2-0};
  3059. let Inst{4} = 0b0;
  3060. let validForTailPredication = 1;
  3061. }
  3062. multiclass MVE_VRINT_m<MVEVectorVTInfo VTI, string suffix, bits<3> opcode,
  3063. SDNode unpred_op> {
  3064. def "": MVE_VRINT<suffix, opcode, VTI.Suffix, VTI.Size>;
  3065. defvar Inst = !cast<Instruction>(NAME);
  3066. defvar pred_int = !cast<Intrinsic>("int_arm_mve_vrint"#suffix#"_predicated");
  3067. let Predicates = [HasMVEFloat] in {
  3068. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
  3069. (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
  3070. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
  3071. (VTI.Vec MQPR:$inactive))),
  3072. (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
  3073. (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
  3074. }
  3075. }
  3076. multiclass MVE_VRINT_ops<MVEVectorVTInfo VTI> {
  3077. defm N : MVE_VRINT_m<VTI, "n", 0b000, int_arm_mve_vrintn>;
  3078. defm X : MVE_VRINT_m<VTI, "x", 0b001, frint>;
  3079. defm A : MVE_VRINT_m<VTI, "a", 0b010, fround>;
  3080. defm Z : MVE_VRINT_m<VTI, "z", 0b011, ftrunc>;
  3081. defm M : MVE_VRINT_m<VTI, "m", 0b101, ffloor>;
  3082. defm P : MVE_VRINT_m<VTI, "p", 0b111, fceil>;
  3083. }
  3084. defm MVE_VRINTf16 : MVE_VRINT_ops<MVE_v8f16>;
  3085. defm MVE_VRINTf32 : MVE_VRINT_ops<MVE_v4f32>;
  3086. class MVEFloatArithNeon<string iname, string suffix, bit size,
  3087. dag oops, dag iops, string ops,
  3088. vpred_ops vpred, string cstr, list<dag> pattern=[]>
  3089. : MVE_float<iname, suffix, oops, iops, ops, vpred, cstr, pattern> {
  3090. let Inst{20} = size;
  3091. let Inst{16} = 0b0;
  3092. }
  3093. class MVE_VMUL_fp<string iname, string suffix, bit size, list<dag> pattern=[]>
  3094. : MVEFloatArithNeon<iname, suffix, size, (outs MQPR:$Qd),
  3095. (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm", vpred_r, "",
  3096. pattern> {
  3097. bits<4> Qd;
  3098. bits<4> Qn;
  3099. let Inst{28} = 0b1;
  3100. let Inst{25-23} = 0b110;
  3101. let Inst{22} = Qd{3};
  3102. let Inst{21} = 0b0;
  3103. let Inst{19-17} = Qn{2-0};
  3104. let Inst{15-13} = Qd{2-0};
  3105. let Inst{12-8} = 0b01101;
  3106. let Inst{7} = Qn{3};
  3107. let Inst{4} = 0b1;
  3108. let validForTailPredication = 1;
  3109. }
  3110. multiclass MVE_VMULT_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
  3111. SDNode Op, Intrinsic PredInt> {
  3112. def "" : MVE_VMUL_fp<iname, VTI.Suffix, VTI.Size{0}>;
  3113. defvar Inst = !cast<Instruction>(NAME);
  3114. let Predicates = [HasMVEFloat] in {
  3115. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME)>;
  3116. }
  3117. }
  3118. multiclass MVE_VMUL_fp_m<MVEVectorVTInfo VTI>
  3119. : MVE_VMULT_fp_m<"vmul", 0, VTI, fmul, int_arm_mve_mul_predicated>;
  3120. defm MVE_VMULf32 : MVE_VMUL_fp_m<MVE_v4f32>;
  3121. defm MVE_VMULf16 : MVE_VMUL_fp_m<MVE_v8f16>;
  3122. class MVE_VCMLA<string suffix, bit size>
  3123. : MVEFloatArithNeon<"vcmla", suffix, size, (outs MQPR:$Qd),
  3124. (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
  3125. "$Qd, $Qn, $Qm, $rot", vpred_n, "$Qd = $Qd_src", []> {
  3126. bits<4> Qd;
  3127. bits<4> Qn;
  3128. bits<2> rot;
  3129. let Inst{28} = 0b1;
  3130. let Inst{25} = 0b0;
  3131. let Inst{24-23} = rot;
  3132. let Inst{22} = Qd{3};
  3133. let Inst{21} = 0b1;
  3134. let Inst{19-17} = Qn{2-0};
  3135. let Inst{15-13} = Qd{2-0};
  3136. let Inst{12-8} = 0b01000;
  3137. let Inst{7} = Qn{3};
  3138. let Inst{4} = 0b0;
  3139. }
  3140. multiclass MVE_VCMLA_m<MVEVectorVTInfo VTI, bit size> {
  3141. def "" : MVE_VCMLA<VTI.Suffix, size>;
  3142. defvar Inst = !cast<Instruction>(NAME);
  3143. let Predicates = [HasMVEFloat] in {
  3144. def : Pat<(VTI.Vec (int_arm_mve_vcmlaq
  3145. imm:$rot, (VTI.Vec MQPR:$Qd_src),
  3146. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  3147. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
  3148. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3149. imm:$rot))>;
  3150. def : Pat<(VTI.Vec (int_arm_mve_vcmlaq_predicated
  3151. imm:$rot, (VTI.Vec MQPR:$Qd_src),
  3152. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3153. (VTI.Pred VCCR:$mask))),
  3154. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qn),
  3155. (VTI.Vec MQPR:$Qm), imm:$rot,
  3156. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  3157. }
  3158. }
  3159. defm MVE_VCMLAf16 : MVE_VCMLA_m<MVE_v8f16, 0b0>;
  3160. defm MVE_VCMLAf32 : MVE_VCMLA_m<MVE_v4f32, 0b1>;
  3161. class MVE_VADDSUBFMA_fp<string iname, string suffix, bit size, bit bit_4,
  3162. bit bit_8, bit bit_21, dag iops=(ins),
  3163. vpred_ops vpred=vpred_r, string cstr="",
  3164. list<dag> pattern=[]>
  3165. : MVEFloatArithNeon<iname, suffix, size, (outs MQPR:$Qd),
  3166. !con(iops, (ins MQPR:$Qn, MQPR:$Qm)), "$Qd, $Qn, $Qm",
  3167. vpred, cstr, pattern> {
  3168. bits<4> Qd;
  3169. bits<4> Qn;
  3170. let Inst{28} = 0b0;
  3171. let Inst{25-23} = 0b110;
  3172. let Inst{22} = Qd{3};
  3173. let Inst{21} = bit_21;
  3174. let Inst{19-17} = Qn{2-0};
  3175. let Inst{15-13} = Qd{2-0};
  3176. let Inst{11-9} = 0b110;
  3177. let Inst{8} = bit_8;
  3178. let Inst{7} = Qn{3};
  3179. let Inst{4} = bit_4;
  3180. let validForTailPredication = 1;
  3181. }
  3182. multiclass MVE_VFMA_fp_multi<string iname, bit fms, MVEVectorVTInfo VTI> {
  3183. def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size{0}, 0b1, 0b0, fms,
  3184. (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
  3185. defvar Inst = !cast<Instruction>(NAME);
  3186. defvar pred_int = int_arm_mve_fma_predicated;
  3187. defvar m1 = (VTI.Vec MQPR:$m1);
  3188. defvar m2 = (VTI.Vec MQPR:$m2);
  3189. defvar add = (VTI.Vec MQPR:$add);
  3190. defvar pred = (VTI.Pred VCCR:$pred);
  3191. let Predicates = [HasMVEFloat] in {
  3192. if fms then {
  3193. def : Pat<(VTI.Vec (fma (fneg m1), m2, add)),
  3194. (Inst $add, $m1, $m2)>;
  3195. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  3196. (VTI.Vec (fma (fneg m1), m2, add)),
  3197. add)),
  3198. (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
  3199. def : Pat<(VTI.Vec (pred_int (fneg m1), m2, add, pred)),
  3200. (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
  3201. def : Pat<(VTI.Vec (pred_int m1, (fneg m2), add, pred)),
  3202. (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
  3203. } else {
  3204. def : Pat<(VTI.Vec (fma m1, m2, add)),
  3205. (Inst $add, $m1, $m2)>;
  3206. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  3207. (VTI.Vec (fma m1, m2, add)),
  3208. add)),
  3209. (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
  3210. def : Pat<(VTI.Vec (pred_int m1, m2, add, pred)),
  3211. (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
  3212. }
  3213. }
  3214. }
  3215. defm MVE_VFMAf32 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v4f32>;
  3216. defm MVE_VFMAf16 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v8f16>;
  3217. defm MVE_VFMSf32 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v4f32>;
  3218. defm MVE_VFMSf16 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v8f16>;
  3219. multiclass MVE_VADDSUB_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
  3220. SDNode Op, Intrinsic PredInt> {
  3221. def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size{0}, 0, 1, bit_21> {
  3222. let validForTailPredication = 1;
  3223. }
  3224. defvar Inst = !cast<Instruction>(NAME);
  3225. let Predicates = [HasMVEFloat] in {
  3226. defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME)>;
  3227. }
  3228. }
  3229. multiclass MVE_VADD_fp_m<MVEVectorVTInfo VTI>
  3230. : MVE_VADDSUB_fp_m<"vadd", 0, VTI, fadd, int_arm_mve_add_predicated>;
  3231. multiclass MVE_VSUB_fp_m<MVEVectorVTInfo VTI>
  3232. : MVE_VADDSUB_fp_m<"vsub", 1, VTI, fsub, int_arm_mve_sub_predicated>;
  3233. defm MVE_VADDf32 : MVE_VADD_fp_m<MVE_v4f32>;
  3234. defm MVE_VADDf16 : MVE_VADD_fp_m<MVE_v8f16>;
  3235. defm MVE_VSUBf32 : MVE_VSUB_fp_m<MVE_v4f32>;
  3236. defm MVE_VSUBf16 : MVE_VSUB_fp_m<MVE_v8f16>;
  3237. class MVE_VCADD<string suffix, bit size, string cstr="">
  3238. : MVEFloatArithNeon<"vcadd", suffix, size, (outs MQPR:$Qd),
  3239. (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
  3240. "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
  3241. bits<4> Qd;
  3242. bits<4> Qn;
  3243. bit rot;
  3244. let Inst{28} = 0b1;
  3245. let Inst{25} = 0b0;
  3246. let Inst{24} = rot;
  3247. let Inst{23} = 0b1;
  3248. let Inst{22} = Qd{3};
  3249. let Inst{21} = 0b0;
  3250. let Inst{19-17} = Qn{2-0};
  3251. let Inst{15-13} = Qd{2-0};
  3252. let Inst{12-8} = 0b01000;
  3253. let Inst{7} = Qn{3};
  3254. let Inst{4} = 0b0;
  3255. }
  3256. multiclass MVE_VCADD_m<MVEVectorVTInfo VTI, bit size, string cstr=""> {
  3257. def "" : MVE_VCADD<VTI.Suffix, size, cstr>;
  3258. defvar Inst = !cast<Instruction>(NAME);
  3259. let Predicates = [HasMVEFloat] in {
  3260. def : Pat<(VTI.Vec (int_arm_mve_vcaddq (i32 1),
  3261. imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  3262. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3263. imm:$rot))>;
  3264. def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated (i32 1),
  3265. imm:$rot, (VTI.Vec MQPR:$inactive),
  3266. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3267. (VTI.Pred VCCR:$mask))),
  3268. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3269. imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
  3270. (VTI.Vec MQPR:$inactive)))>;
  3271. }
  3272. }
  3273. defm MVE_VCADDf16 : MVE_VCADD_m<MVE_v8f16, 0b0>;
  3274. defm MVE_VCADDf32 : MVE_VCADD_m<MVE_v4f32, 0b1, "@earlyclobber $Qd">;
  3275. class MVE_VABD_fp<string suffix, bit size>
  3276. : MVE_float<"vabd", suffix, (outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
  3277. "$Qd, $Qn, $Qm", vpred_r, ""> {
  3278. bits<4> Qd;
  3279. bits<4> Qn;
  3280. let Inst{28} = 0b1;
  3281. let Inst{25-23} = 0b110;
  3282. let Inst{22} = Qd{3};
  3283. let Inst{21} = 0b1;
  3284. let Inst{20} = size;
  3285. let Inst{19-17} = Qn{2-0};
  3286. let Inst{16} = 0b0;
  3287. let Inst{15-13} = Qd{2-0};
  3288. let Inst{11-8} = 0b1101;
  3289. let Inst{7} = Qn{3};
  3290. let Inst{4} = 0b0;
  3291. let validForTailPredication = 1;
  3292. }
  3293. multiclass MVE_VABDT_fp_m<MVEVectorVTInfo VTI,
  3294. Intrinsic unpred_int, Intrinsic pred_int> {
  3295. def "" : MVE_VABD_fp<VTI.Suffix, VTI.Size{0}>;
  3296. defvar Inst = !cast<Instruction>(NAME);
  3297. let Predicates = [HasMVEFloat] in {
  3298. def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  3299. (i32 0))),
  3300. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  3301. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  3302. (i32 0), (VTI.Pred VCCR:$mask),
  3303. (VTI.Vec MQPR:$inactive))),
  3304. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  3305. ARMVCCThen, (VTI.Pred VCCR:$mask),
  3306. (VTI.Vec MQPR:$inactive)))>;
  3307. }
  3308. }
  3309. multiclass MVE_VABD_fp_m<MVEVectorVTInfo VTI>
  3310. : MVE_VABDT_fp_m<VTI, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
  3311. defm MVE_VABDf32 : MVE_VABD_fp_m<MVE_v4f32>;
  3312. defm MVE_VABDf16 : MVE_VABD_fp_m<MVE_v8f16>;
  3313. let Predicates = [HasMVEFloat] in {
  3314. def : Pat<(v8f16 (fabs (fsub (v8f16 MQPR:$Qm), (v8f16 MQPR:$Qn)))),
  3315. (MVE_VABDf16 MQPR:$Qm, MQPR:$Qn)>;
  3316. def : Pat<(v4f32 (fabs (fsub (v4f32 MQPR:$Qm), (v4f32 MQPR:$Qn)))),
  3317. (MVE_VABDf32 MQPR:$Qm, MQPR:$Qn)>;
  3318. }
  3319. class MVE_VCVT_fix<string suffix, bit fsi, bit U, bit op,
  3320. Operand imm_operand_type>
  3321. : MVE_float<"vcvt", suffix,
  3322. (outs MQPR:$Qd), (ins MQPR:$Qm, imm_operand_type:$imm6),
  3323. "$Qd, $Qm, $imm6", vpred_r, "", []> {
  3324. bits<4> Qd;
  3325. bits<6> imm6;
  3326. let Inst{28} = U;
  3327. let Inst{25-23} = 0b111;
  3328. let Inst{22} = Qd{3};
  3329. let Inst{21} = 0b1;
  3330. let Inst{19-16} = imm6{3-0};
  3331. let Inst{15-13} = Qd{2-0};
  3332. let Inst{11-10} = 0b11;
  3333. let Inst{9} = fsi;
  3334. let Inst{8} = op;
  3335. let Inst{7} = 0b0;
  3336. let Inst{4} = 0b1;
  3337. let DecoderMethod = "DecodeMVEVCVTt1fp";
  3338. let validForTailPredication = 1;
  3339. }
  3340. class MVE_VCVT_imm_asmop<int Bits> : AsmOperandClass {
  3341. let PredicateMethod = "isImmediate<1," # Bits # ">";
  3342. let DiagnosticString =
  3343. "MVE fixed-point immediate operand must be between 1 and " # Bits;
  3344. let Name = "MVEVcvtImm" # Bits;
  3345. let RenderMethod = "addImmOperands";
  3346. }
  3347. class MVE_VCVT_imm<int Bits>: Operand<i32> {
  3348. let ParserMatchClass = MVE_VCVT_imm_asmop<Bits>;
  3349. let EncoderMethod = "getNEONVcvtImm32OpValue";
  3350. let DecoderMethod = "DecodeVCVTImmOperand";
  3351. }
  3352. class MVE_VCVT_fix_f32<string suffix, bit U, bit op>
  3353. : MVE_VCVT_fix<suffix, 0b1, U, op, MVE_VCVT_imm<32>> {
  3354. let Inst{20} = imm6{4};
  3355. }
  3356. class MVE_VCVT_fix_f16<string suffix, bit U, bit op>
  3357. : MVE_VCVT_fix<suffix, 0b0, U, op, MVE_VCVT_imm<16>> {
  3358. let Inst{20} = 0b1;
  3359. }
  3360. multiclass MVE_VCVT_fix_patterns<Instruction Inst, bit U, MVEVectorVTInfo DestVTI,
  3361. MVEVectorVTInfo SrcVTI> {
  3362. let Predicates = [HasMVEFloat] in {
  3363. def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix
  3364. (i32 U), (SrcVTI.Vec MQPR:$Qm), imm:$scale)),
  3365. (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale))>;
  3366. def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix_predicated (i32 U),
  3367. (DestVTI.Vec MQPR:$inactive),
  3368. (SrcVTI.Vec MQPR:$Qm),
  3369. imm:$scale,
  3370. (DestVTI.Pred VCCR:$mask))),
  3371. (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale,
  3372. ARMVCCThen, (DestVTI.Pred VCCR:$mask),
  3373. (DestVTI.Vec MQPR:$inactive)))>;
  3374. }
  3375. }
  3376. multiclass MVE_VCVT_fix_f32_m<bit U, bit op,
  3377. MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
  3378. def "" : MVE_VCVT_fix_f32<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
  3379. defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
  3380. }
  3381. multiclass MVE_VCVT_fix_f16_m<bit U, bit op,
  3382. MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
  3383. def "" : MVE_VCVT_fix_f16<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
  3384. defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
  3385. }
  3386. defm MVE_VCVTf16s16_fix : MVE_VCVT_fix_f16_m<0b0, 0b0, MVE_v8f16, MVE_v8s16>;
  3387. defm MVE_VCVTs16f16_fix : MVE_VCVT_fix_f16_m<0b0, 0b1, MVE_v8s16, MVE_v8f16>;
  3388. defm MVE_VCVTf16u16_fix : MVE_VCVT_fix_f16_m<0b1, 0b0, MVE_v8f16, MVE_v8u16>;
  3389. defm MVE_VCVTu16f16_fix : MVE_VCVT_fix_f16_m<0b1, 0b1, MVE_v8u16, MVE_v8f16>;
  3390. defm MVE_VCVTf32s32_fix : MVE_VCVT_fix_f32_m<0b0, 0b0, MVE_v4f32, MVE_v4s32>;
  3391. defm MVE_VCVTs32f32_fix : MVE_VCVT_fix_f32_m<0b0, 0b1, MVE_v4s32, MVE_v4f32>;
  3392. defm MVE_VCVTf32u32_fix : MVE_VCVT_fix_f32_m<0b1, 0b0, MVE_v4f32, MVE_v4u32>;
  3393. defm MVE_VCVTu32f32_fix : MVE_VCVT_fix_f32_m<0b1, 0b1, MVE_v4u32, MVE_v4f32>;
  3394. class MVE_VCVT_fp_int_anpm<string suffix, bits<2> size, bit op, string anpm,
  3395. bits<2> rm, list<dag> pattern=[]>
  3396. : MVE_float<!strconcat("vcvt", anpm), suffix, (outs MQPR:$Qd),
  3397. (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
  3398. bits<4> Qd;
  3399. let Inst{28} = 0b1;
  3400. let Inst{25-23} = 0b111;
  3401. let Inst{22} = Qd{3};
  3402. let Inst{21-20} = 0b11;
  3403. let Inst{19-18} = size;
  3404. let Inst{17-16} = 0b11;
  3405. let Inst{15-13} = Qd{2-0};
  3406. let Inst{12-10} = 0b000;
  3407. let Inst{9-8} = rm;
  3408. let Inst{7} = op;
  3409. let Inst{4} = 0b0;
  3410. let validForTailPredication = 1;
  3411. }
  3412. multiclass MVE_VCVT_fp_int_anpm_inner<MVEVectorVTInfo Int, MVEVectorVTInfo Flt,
  3413. string anpm, bits<2> rm> {
  3414. def "": MVE_VCVT_fp_int_anpm<Int.Suffix # "." # Flt.Suffix, Int.Size,
  3415. Int.Unsigned, anpm, rm>;
  3416. defvar Inst = !cast<Instruction>(NAME);
  3417. defvar IntrBaseName = "int_arm_mve_vcvt" # anpm;
  3418. defvar UnpredIntr = !cast<Intrinsic>(IntrBaseName);
  3419. defvar PredIntr = !cast<Intrinsic>(IntrBaseName # "_predicated");
  3420. let Predicates = [HasMVEFloat] in {
  3421. def : Pat<(Int.Vec (UnpredIntr (i32 Int.Unsigned), (Flt.Vec MQPR:$in))),
  3422. (Int.Vec (Inst (Flt.Vec MQPR:$in)))>;
  3423. def : Pat<(Int.Vec (PredIntr (i32 Int.Unsigned), (Int.Vec MQPR:$inactive),
  3424. (Flt.Vec MQPR:$in), (Flt.Pred VCCR:$pred))),
  3425. (Int.Vec (Inst (Flt.Vec MQPR:$in), ARMVCCThen,
  3426. (Flt.Pred VCCR:$pred), (Int.Vec MQPR:$inactive)))>;
  3427. }
  3428. }
  3429. multiclass MVE_VCVT_fp_int_anpm_outer<MVEVectorVTInfo Int,
  3430. MVEVectorVTInfo Flt> {
  3431. defm a : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "a", 0b00>;
  3432. defm n : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "n", 0b01>;
  3433. defm p : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "p", 0b10>;
  3434. defm m : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "m", 0b11>;
  3435. }
  3436. // This defines instructions such as MVE_VCVTu16f16a, with an explicit
  3437. // rounding-mode suffix on the mnemonic. The class below will define
  3438. // the bare MVE_VCVTu16f16 (with implied rounding toward zero).
  3439. defm MVE_VCVTs16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8s16, MVE_v8f16>;
  3440. defm MVE_VCVTu16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8u16, MVE_v8f16>;
  3441. defm MVE_VCVTs32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4s32, MVE_v4f32>;
  3442. defm MVE_VCVTu32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4u32, MVE_v4f32>;
  3443. class MVE_VCVT_fp_int<string suffix, bits<2> size, bit toint, bit unsigned,
  3444. list<dag> pattern=[]>
  3445. : MVE_float<"vcvt", suffix, (outs MQPR:$Qd),
  3446. (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
  3447. bits<4> Qd;
  3448. let Inst{28} = 0b1;
  3449. let Inst{25-23} = 0b111;
  3450. let Inst{22} = Qd{3};
  3451. let Inst{21-20} = 0b11;
  3452. let Inst{19-18} = size;
  3453. let Inst{17-16} = 0b11;
  3454. let Inst{15-13} = Qd{2-0};
  3455. let Inst{12-9} = 0b0011;
  3456. let Inst{8} = toint;
  3457. let Inst{7} = unsigned;
  3458. let Inst{4} = 0b0;
  3459. let validForTailPredication = 1;
  3460. }
  3461. multiclass MVE_VCVT_fp_int_m<MVEVectorVTInfo Dest, MVEVectorVTInfo Src,
  3462. SDNode unpred_op> {
  3463. defvar Unsigned = !or(!eq(Dest.SuffixLetter,"u"), !eq(Src.SuffixLetter,"u"));
  3464. defvar ToInt = !eq(Src.SuffixLetter,"f");
  3465. def "" : MVE_VCVT_fp_int<Dest.Suffix # "." # Src.Suffix, Dest.Size,
  3466. ToInt, Unsigned>;
  3467. defvar Inst = !cast<Instruction>(NAME);
  3468. let Predicates = [HasMVEFloat] in {
  3469. def : Pat<(Dest.Vec (unpred_op (Src.Vec MQPR:$src))),
  3470. (Dest.Vec (Inst (Src.Vec MQPR:$src)))>;
  3471. def : Pat<(Dest.Vec (int_arm_mve_vcvt_fp_int_predicated
  3472. (Src.Vec MQPR:$src), (i32 Unsigned),
  3473. (Src.Pred VCCR:$mask), (Dest.Vec MQPR:$inactive))),
  3474. (Dest.Vec (Inst (Src.Vec MQPR:$src), ARMVCCThen,
  3475. (Src.Pred VCCR:$mask),
  3476. (Dest.Vec MQPR:$inactive)))>;
  3477. }
  3478. }
  3479. // The unsuffixed VCVT for float->int implicitly rounds toward zero,
  3480. // which I reflect here in the llvm instruction names
  3481. defm MVE_VCVTs16f16z : MVE_VCVT_fp_int_m<MVE_v8s16, MVE_v8f16, fp_to_sint>;
  3482. defm MVE_VCVTu16f16z : MVE_VCVT_fp_int_m<MVE_v8u16, MVE_v8f16, fp_to_uint>;
  3483. defm MVE_VCVTs32f32z : MVE_VCVT_fp_int_m<MVE_v4s32, MVE_v4f32, fp_to_sint>;
  3484. defm MVE_VCVTu32f32z : MVE_VCVT_fp_int_m<MVE_v4u32, MVE_v4f32, fp_to_uint>;
  3485. // Whereas VCVT for int->float rounds to nearest
  3486. defm MVE_VCVTf16s16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8s16, sint_to_fp>;
  3487. defm MVE_VCVTf16u16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8u16, uint_to_fp>;
  3488. defm MVE_VCVTf32s32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4s32, sint_to_fp>;
  3489. defm MVE_VCVTf32u32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4u32, uint_to_fp>;
  3490. class MVE_VABSNEG_fp<string iname, string suffix, bits<2> size, bit negate,
  3491. list<dag> pattern=[]>
  3492. : MVE_float<iname, suffix, (outs MQPR:$Qd),
  3493. (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
  3494. bits<4> Qd;
  3495. let Inst{28} = 0b1;
  3496. let Inst{25-23} = 0b111;
  3497. let Inst{22} = Qd{3};
  3498. let Inst{21-20} = 0b11;
  3499. let Inst{19-18} = size;
  3500. let Inst{17-16} = 0b01;
  3501. let Inst{15-13} = Qd{2-0};
  3502. let Inst{11-8} = 0b0111;
  3503. let Inst{7} = negate;
  3504. let Inst{4} = 0b0;
  3505. let validForTailPredication = 1;
  3506. }
  3507. multiclass MVE_VABSNEG_fp_m<string iname, SDNode unpred_op, Intrinsic pred_int,
  3508. MVEVectorVTInfo VTI, bit opcode> {
  3509. def "" : MVE_VABSNEG_fp<iname, VTI.Suffix, VTI.Size, opcode>;
  3510. defvar Inst = !cast<Instruction>(NAME);
  3511. let Predicates = [HasMVEInt] in {
  3512. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
  3513. (VTI.Vec (Inst $v))>;
  3514. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
  3515. (VTI.Vec MQPR:$inactive))),
  3516. (VTI.Vec (Inst $v, ARMVCCThen, $mask, $inactive))>;
  3517. }
  3518. }
  3519. defm MVE_VABSf16 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
  3520. MVE_v8f16, 0>;
  3521. defm MVE_VABSf32 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
  3522. MVE_v4f32, 0>;
  3523. defm MVE_VNEGf16 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
  3524. MVE_v8f16, 1>;
  3525. defm MVE_VNEGf32 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
  3526. MVE_v4f32, 1>;
  3527. class MVE_VMAXMINNMA<string iname, string suffix, bit size, bit bit_12,
  3528. list<dag> pattern=[]>
  3529. : MVE_f<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
  3530. NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
  3531. pattern> {
  3532. bits<4> Qd;
  3533. bits<4> Qm;
  3534. let Inst{28} = size;
  3535. let Inst{25-23} = 0b100;
  3536. let Inst{22} = Qd{3};
  3537. let Inst{21-16} = 0b111111;
  3538. let Inst{15-13} = Qd{2-0};
  3539. let Inst{12} = bit_12;
  3540. let Inst{11-6} = 0b111010;
  3541. let Inst{5} = Qm{3};
  3542. let Inst{4} = 0b0;
  3543. let Inst{3-1} = Qm{2-0};
  3544. let Inst{0} = 0b1;
  3545. let isCommutable = 1;
  3546. }
  3547. multiclass MVE_VMAXMINNMA_m<string iname, MVEVectorVTInfo VTI,
  3548. SDNode unpred_op, Intrinsic pred_int,
  3549. bit bit_12> {
  3550. def "" : MVE_VMAXMINNMA<iname, VTI.Suffix, VTI.Size{0}, bit_12>;
  3551. defvar Inst = !cast<Instruction>(NAME);
  3552. let Predicates = [HasMVEInt] in {
  3553. // Unpredicated v(max|min)nma
  3554. def : Pat<(VTI.Vec (unpred_op (fabs (VTI.Vec MQPR:$Qd)),
  3555. (fabs (VTI.Vec MQPR:$Qm)))),
  3556. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
  3557. // Predicated v(max|min)nma
  3558. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
  3559. (VTI.Pred VCCR:$mask))),
  3560. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
  3561. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  3562. }
  3563. }
  3564. multiclass MVE_VMAXNMA<MVEVectorVTInfo VTI, bit bit_12>
  3565. : MVE_VMAXMINNMA_m<"vmaxnma", VTI, fmaxnum, int_arm_mve_vmaxnma_predicated, bit_12>;
  3566. defm MVE_VMAXNMAf32 : MVE_VMAXNMA<MVE_v4f32, 0b0>;
  3567. defm MVE_VMAXNMAf16 : MVE_VMAXNMA<MVE_v8f16, 0b0>;
  3568. multiclass MVE_VMINNMA<MVEVectorVTInfo VTI, bit bit_12>
  3569. : MVE_VMAXMINNMA_m<"vminnma", VTI, fminnum, int_arm_mve_vminnma_predicated, bit_12>;
  3570. defm MVE_VMINNMAf32 : MVE_VMINNMA<MVE_v4f32, 0b1>;
  3571. defm MVE_VMINNMAf16 : MVE_VMINNMA<MVE_v8f16, 0b1>;
  3572. // end of MVE Floating Point instructions
  3573. // start of MVE compares
  3574. class MVE_VCMPqq<string suffix, bit bit_28, bits<2> bits_21_20,
  3575. VCMPPredicateOperand predtype, list<dag> pattern=[]>
  3576. : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, MQPR:$Qm, predtype:$fc),
  3577. NoItinerary, "vcmp", suffix, "$fc, $Qn, $Qm", vpred_n, "", pattern> {
  3578. // Base class for comparing two vector registers
  3579. bits<3> fc;
  3580. bits<4> Qn;
  3581. bits<4> Qm;
  3582. let Inst{28} = bit_28;
  3583. let Inst{25-22} = 0b1000;
  3584. let Inst{21-20} = bits_21_20;
  3585. let Inst{19-17} = Qn{2-0};
  3586. let Inst{16-13} = 0b1000;
  3587. let Inst{12} = fc{2};
  3588. let Inst{11-8} = 0b1111;
  3589. let Inst{7} = fc{0};
  3590. let Inst{6} = 0b0;
  3591. let Inst{5} = Qm{3};
  3592. let Inst{4} = 0b0;
  3593. let Inst{3-1} = Qm{2-0};
  3594. let Inst{0} = fc{1};
  3595. let Constraints = "";
  3596. // We need a custom decoder method for these instructions because of
  3597. // the output VCCR operand, which isn't encoded in the instruction
  3598. // bits anywhere (there is only one choice for it) but has to be
  3599. // included in the MC operands so that codegen will be able to track
  3600. // its data flow between instructions, spill/reload it when
  3601. // necessary, etc. There seems to be no way to get the Tablegen
  3602. // decoder to emit an operand that isn't affected by any instruction
  3603. // bit.
  3604. let DecoderMethod = "DecodeMVEVCMP<false," # predtype.DecoderMethod # ">";
  3605. let validForTailPredication = 1;
  3606. }
  3607. class MVE_VCMPqqf<string suffix, bit size>
  3608. : MVE_VCMPqq<suffix, size, 0b11, pred_basic_fp> {
  3609. let Predicates = [HasMVEFloat];
  3610. }
  3611. class MVE_VCMPqqi<string suffix, bits<2> size>
  3612. : MVE_VCMPqq<suffix, 0b1, size, pred_basic_i> {
  3613. let Inst{12} = 0b0;
  3614. let Inst{0} = 0b0;
  3615. }
  3616. class MVE_VCMPqqu<string suffix, bits<2> size>
  3617. : MVE_VCMPqq<suffix, 0b1, size, pred_basic_u> {
  3618. let Inst{12} = 0b0;
  3619. let Inst{0} = 0b1;
  3620. }
  3621. class MVE_VCMPqqs<string suffix, bits<2> size>
  3622. : MVE_VCMPqq<suffix, 0b1, size, pred_basic_s> {
  3623. let Inst{12} = 0b1;
  3624. }
  3625. def MVE_VCMPf32 : MVE_VCMPqqf<"f32", 0b0>;
  3626. def MVE_VCMPf16 : MVE_VCMPqqf<"f16", 0b1>;
  3627. def MVE_VCMPi8 : MVE_VCMPqqi<"i8", 0b00>;
  3628. def MVE_VCMPi16 : MVE_VCMPqqi<"i16", 0b01>;
  3629. def MVE_VCMPi32 : MVE_VCMPqqi<"i32", 0b10>;
  3630. def MVE_VCMPu8 : MVE_VCMPqqu<"u8", 0b00>;
  3631. def MVE_VCMPu16 : MVE_VCMPqqu<"u16", 0b01>;
  3632. def MVE_VCMPu32 : MVE_VCMPqqu<"u32", 0b10>;
  3633. def MVE_VCMPs8 : MVE_VCMPqqs<"s8", 0b00>;
  3634. def MVE_VCMPs16 : MVE_VCMPqqs<"s16", 0b01>;
  3635. def MVE_VCMPs32 : MVE_VCMPqqs<"s32", 0b10>;
  3636. class MVE_VCMPqr<string suffix, bit bit_28, bits<2> bits_21_20,
  3637. VCMPPredicateOperand predtype, list<dag> pattern=[]>
  3638. : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, GPRwithZR:$Rm, predtype:$fc),
  3639. NoItinerary, "vcmp", suffix, "$fc, $Qn, $Rm", vpred_n, "", pattern> {
  3640. // Base class for comparing a vector register with a scalar
  3641. bits<3> fc;
  3642. bits<4> Qn;
  3643. bits<4> Rm;
  3644. let Inst{28} = bit_28;
  3645. let Inst{25-22} = 0b1000;
  3646. let Inst{21-20} = bits_21_20;
  3647. let Inst{19-17} = Qn{2-0};
  3648. let Inst{16-13} = 0b1000;
  3649. let Inst{12} = fc{2};
  3650. let Inst{11-8} = 0b1111;
  3651. let Inst{7} = fc{0};
  3652. let Inst{6} = 0b1;
  3653. let Inst{5} = fc{1};
  3654. let Inst{4} = 0b0;
  3655. let Inst{3-0} = Rm{3-0};
  3656. let Constraints = "";
  3657. // Custom decoder method, for the same reason as MVE_VCMPqq
  3658. let DecoderMethod = "DecodeMVEVCMP<true," # predtype.DecoderMethod # ">";
  3659. let validForTailPredication = 1;
  3660. }
  3661. class MVE_VCMPqrf<string suffix, bit size>
  3662. : MVE_VCMPqr<suffix, size, 0b11, pred_basic_fp> {
  3663. let Predicates = [HasMVEFloat];
  3664. }
  3665. class MVE_VCMPqri<string suffix, bits<2> size>
  3666. : MVE_VCMPqr<suffix, 0b1, size, pred_basic_i> {
  3667. let Inst{12} = 0b0;
  3668. let Inst{5} = 0b0;
  3669. }
  3670. class MVE_VCMPqru<string suffix, bits<2> size>
  3671. : MVE_VCMPqr<suffix, 0b1, size, pred_basic_u> {
  3672. let Inst{12} = 0b0;
  3673. let Inst{5} = 0b1;
  3674. }
  3675. class MVE_VCMPqrs<string suffix, bits<2> size>
  3676. : MVE_VCMPqr<suffix, 0b1, size, pred_basic_s> {
  3677. let Inst{12} = 0b1;
  3678. }
  3679. def MVE_VCMPf32r : MVE_VCMPqrf<"f32", 0b0>;
  3680. def MVE_VCMPf16r : MVE_VCMPqrf<"f16", 0b1>;
  3681. def MVE_VCMPi8r : MVE_VCMPqri<"i8", 0b00>;
  3682. def MVE_VCMPi16r : MVE_VCMPqri<"i16", 0b01>;
  3683. def MVE_VCMPi32r : MVE_VCMPqri<"i32", 0b10>;
  3684. def MVE_VCMPu8r : MVE_VCMPqru<"u8", 0b00>;
  3685. def MVE_VCMPu16r : MVE_VCMPqru<"u16", 0b01>;
  3686. def MVE_VCMPu32r : MVE_VCMPqru<"u32", 0b10>;
  3687. def MVE_VCMPs8r : MVE_VCMPqrs<"s8", 0b00>;
  3688. def MVE_VCMPs16r : MVE_VCMPqrs<"s16", 0b01>;
  3689. def MVE_VCMPs32r : MVE_VCMPqrs<"s32", 0b10>;
  3690. multiclass unpred_vcmp_z<string suffix, PatLeaf fc> {
  3691. def i8 : Pat<(v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)),
  3692. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc))>;
  3693. def i16 : Pat<(v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)),
  3694. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc))>;
  3695. def i32 : Pat<(v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)),
  3696. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc))>;
  3697. def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)))),
  3698. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
  3699. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)))),
  3700. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
  3701. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)))),
  3702. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
  3703. }
  3704. multiclass unpred_vcmp_r<string suffix, PatLeaf fc> {
  3705. def i8 : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)),
  3706. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc))>;
  3707. def i16 : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)),
  3708. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc))>;
  3709. def i32 : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)),
  3710. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc))>;
  3711. def i8r : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)),
  3712. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc))>;
  3713. def i16r : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)),
  3714. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
  3715. def i32r : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)),
  3716. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
  3717. def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)))),
  3718. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3719. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)))),
  3720. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3721. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)))),
  3722. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3723. def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)))),
  3724. (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3725. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)))),
  3726. (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3727. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)))),
  3728. (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3729. }
  3730. multiclass unpred_vcmpf_z<PatLeaf fc> {
  3731. def f16 : Pat<(v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)),
  3732. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc))>;
  3733. def f32 : Pat<(v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)),
  3734. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc))>;
  3735. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)))),
  3736. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
  3737. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)))),
  3738. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
  3739. }
  3740. multiclass unpred_vcmpf_r<int fc> {
  3741. def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)),
  3742. (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc))>;
  3743. def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)),
  3744. (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc))>;
  3745. def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)),
  3746. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
  3747. def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)),
  3748. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
  3749. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)))),
  3750. (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3751. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)))),
  3752. (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3753. def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)))),
  3754. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3755. def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)))),
  3756. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
  3757. }
  3758. let Predicates = [HasMVEInt] in {
  3759. defm MVE_VCEQZ : unpred_vcmp_z<"i", ARMCCeq>;
  3760. defm MVE_VCNEZ : unpred_vcmp_z<"i", ARMCCne>;
  3761. defm MVE_VCGEZ : unpred_vcmp_z<"s", ARMCCge>;
  3762. defm MVE_VCLTZ : unpred_vcmp_z<"s", ARMCClt>;
  3763. defm MVE_VCGTZ : unpred_vcmp_z<"s", ARMCCgt>;
  3764. defm MVE_VCLEZ : unpred_vcmp_z<"s", ARMCCle>;
  3765. defm MVE_VCGTUZ : unpred_vcmp_z<"u", ARMCChi>;
  3766. defm MVE_VCGEUZ : unpred_vcmp_z<"u", ARMCChs>;
  3767. defm MVE_VCEQ : unpred_vcmp_r<"i", ARMCCeq>;
  3768. defm MVE_VCNE : unpred_vcmp_r<"i", ARMCCne>;
  3769. defm MVE_VCGE : unpred_vcmp_r<"s", ARMCCge>;
  3770. defm MVE_VCLT : unpred_vcmp_r<"s", ARMCClt>;
  3771. defm MVE_VCGT : unpred_vcmp_r<"s", ARMCCgt>;
  3772. defm MVE_VCLE : unpred_vcmp_r<"s", ARMCCle>;
  3773. defm MVE_VCGTU : unpred_vcmp_r<"u", ARMCChi>;
  3774. defm MVE_VCGEU : unpred_vcmp_r<"u", ARMCChs>;
  3775. }
  3776. let Predicates = [HasMVEFloat] in {
  3777. defm MVE_VFCEQZ : unpred_vcmpf_z<ARMCCeq>;
  3778. defm MVE_VFCNEZ : unpred_vcmpf_z<ARMCCne>;
  3779. defm MVE_VFCGEZ : unpred_vcmpf_z<ARMCCge>;
  3780. defm MVE_VFCLTZ : unpred_vcmpf_z<ARMCClt>;
  3781. defm MVE_VFCGTZ : unpred_vcmpf_z<ARMCCgt>;
  3782. defm MVE_VFCLEZ : unpred_vcmpf_z<ARMCCle>;
  3783. defm MVE_VFCEQ : unpred_vcmpf_r<ARMCCeq>;
  3784. defm MVE_VFCNE : unpred_vcmpf_r<ARMCCne>;
  3785. defm MVE_VFCGE : unpred_vcmpf_r<ARMCCge>;
  3786. defm MVE_VFCLT : unpred_vcmpf_r<ARMCClt>;
  3787. defm MVE_VFCGT : unpred_vcmpf_r<ARMCCgt>;
  3788. defm MVE_VFCLE : unpred_vcmpf_r<ARMCCle>;
  3789. }
  3790. // Extra "worst case" and/or/xor patterns, going into and out of GRP
  3791. multiclass two_predops<SDPatternOperator opnode, Instruction insn> {
  3792. def v16i1 : Pat<(v16i1 (opnode (v16i1 VCCR:$p1), (v16i1 VCCR:$p2))),
  3793. (v16i1 (COPY_TO_REGCLASS
  3794. (insn (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p1), rGPR)),
  3795. (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p2), rGPR))),
  3796. VCCR))>;
  3797. def v8i1 : Pat<(v8i1 (opnode (v8i1 VCCR:$p1), (v8i1 VCCR:$p2))),
  3798. (v8i1 (COPY_TO_REGCLASS
  3799. (insn (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p1), rGPR)),
  3800. (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p2), rGPR))),
  3801. VCCR))>;
  3802. def v4i1 : Pat<(v4i1 (opnode (v4i1 VCCR:$p1), (v4i1 VCCR:$p2))),
  3803. (v4i1 (COPY_TO_REGCLASS
  3804. (insn (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p1), rGPR)),
  3805. (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p2), rGPR))),
  3806. VCCR))>;
  3807. }
  3808. let Predicates = [HasMVEInt] in {
  3809. defm POR : two_predops<or, t2ORRrr>;
  3810. defm PAND : two_predops<and, t2ANDrr>;
  3811. defm PEOR : two_predops<xor, t2EORrr>;
  3812. }
  3813. // Occasionally we need to cast between a i32 and a boolean vector, for
  3814. // example when moving between rGPR and VPR.P0 as part of predicate vector
  3815. // shuffles. We also sometimes need to cast between different predicate
  3816. // vector types (v4i1<>v8i1, etc.) also as part of lowering vector shuffles.
  3817. def predicate_cast : SDNode<"ARMISD::PREDICATE_CAST", SDTUnaryOp>;
  3818. def load_align4 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
  3819. return cast<LoadSDNode>(N)->getAlignment() >= 4;
  3820. }]>;
  3821. let Predicates = [HasMVEInt] in {
  3822. foreach VT = [ v4i1, v8i1, v16i1 ] in {
  3823. def : Pat<(i32 (predicate_cast (VT VCCR:$src))),
  3824. (i32 (COPY_TO_REGCLASS (VT VCCR:$src), VCCR))>;
  3825. def : Pat<(VT (predicate_cast (i32 VCCR:$src))),
  3826. (VT (COPY_TO_REGCLASS (i32 VCCR:$src), VCCR))>;
  3827. foreach VT2 = [ v4i1, v8i1, v16i1 ] in
  3828. def : Pat<(VT (predicate_cast (VT2 VCCR:$src))),
  3829. (VT (COPY_TO_REGCLASS (VT2 VCCR:$src), VCCR))>;
  3830. }
  3831. // If we happen to be casting from a load we can convert that straight
  3832. // into a predicate load, so long as the load is of the correct type.
  3833. foreach VT = [ v4i1, v8i1, v16i1 ] in {
  3834. def : Pat<(VT (predicate_cast (i32 (load_align4 taddrmode_imm7<2>:$addr)))),
  3835. (VT (VLDR_P0_off taddrmode_imm7<2>:$addr))>;
  3836. }
  3837. // Here we match the specific SDNode type 'ARMVectorRegCastImpl'
  3838. // rather than the more general 'ARMVectorRegCast' which would also
  3839. // match some bitconverts. If we use the latter in cases where the
  3840. // input and output types are the same, the bitconvert gets elided
  3841. // and we end up generating a nonsense match of nothing.
  3842. foreach VT = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
  3843. foreach VT2 = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
  3844. def : Pat<(VT (ARMVectorRegCastImpl (VT2 MQPR:$src))),
  3845. (VT MQPR:$src)>;
  3846. }
  3847. // end of MVE compares
  3848. // start of MVE_qDest_qSrc
  3849. class MVE_qDest_qSrc<string iname, string suffix, dag oops, dag iops,
  3850. string ops, vpred_ops vpred, string cstr,
  3851. list<dag> pattern=[]>
  3852. : MVE_p<oops, iops, NoItinerary, iname, suffix,
  3853. ops, vpred, cstr, pattern> {
  3854. bits<4> Qd;
  3855. bits<4> Qm;
  3856. let Inst{25-23} = 0b100;
  3857. let Inst{22} = Qd{3};
  3858. let Inst{15-13} = Qd{2-0};
  3859. let Inst{11-9} = 0b111;
  3860. let Inst{6} = 0b0;
  3861. let Inst{5} = Qm{3};
  3862. let Inst{4} = 0b0;
  3863. let Inst{3-1} = Qm{2-0};
  3864. }
  3865. class MVE_VQxDMLxDH<string iname, bit exch, bit round, bit subtract,
  3866. string suffix, bits<2> size, string cstr="", list<dag> pattern=[]>
  3867. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  3868. (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
  3869. vpred_n, "$Qd = $Qd_src"#cstr, pattern> {
  3870. bits<4> Qn;
  3871. let Inst{28} = subtract;
  3872. let Inst{21-20} = size;
  3873. let Inst{19-17} = Qn{2-0};
  3874. let Inst{16} = 0b0;
  3875. let Inst{12} = exch;
  3876. let Inst{8} = 0b0;
  3877. let Inst{7} = Qn{3};
  3878. let Inst{0} = round;
  3879. }
  3880. multiclass MVE_VQxDMLxDH_p<string iname, bit exch, bit round, bit subtract,
  3881. MVEVectorVTInfo VTI> {
  3882. def "": MVE_VQxDMLxDH<iname, exch, round, subtract, VTI.Suffix, VTI.Size,
  3883. !if(!eq(VTI.LaneBits, 32), ",@earlyclobber $Qd", "")>;
  3884. defvar Inst = !cast<Instruction>(NAME);
  3885. defvar ConstParams = (? (i32 exch), (i32 round), (i32 subtract));
  3886. defvar unpred_intr = int_arm_mve_vqdmlad;
  3887. defvar pred_intr = int_arm_mve_vqdmlad_predicated;
  3888. def : Pat<(VTI.Vec !con((unpred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
  3889. (VTI.Vec MQPR:$c)), ConstParams)),
  3890. (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
  3891. (VTI.Vec MQPR:$c)))>;
  3892. def : Pat<(VTI.Vec !con((pred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
  3893. (VTI.Vec MQPR:$c)), ConstParams,
  3894. (? (VTI.Pred VCCR:$pred)))),
  3895. (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
  3896. (VTI.Vec MQPR:$c),
  3897. ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
  3898. }
  3899. multiclass MVE_VQxDMLxDH_multi<string iname, bit exch,
  3900. bit round, bit subtract> {
  3901. defm s8 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v16s8>;
  3902. defm s16 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v8s16>;
  3903. defm s32 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v4s32>;
  3904. }
  3905. defm MVE_VQDMLADH : MVE_VQxDMLxDH_multi<"vqdmladh", 0b0, 0b0, 0b0>;
  3906. defm MVE_VQDMLADHX : MVE_VQxDMLxDH_multi<"vqdmladhx", 0b1, 0b0, 0b0>;
  3907. defm MVE_VQRDMLADH : MVE_VQxDMLxDH_multi<"vqrdmladh", 0b0, 0b1, 0b0>;
  3908. defm MVE_VQRDMLADHX : MVE_VQxDMLxDH_multi<"vqrdmladhx", 0b1, 0b1, 0b0>;
  3909. defm MVE_VQDMLSDH : MVE_VQxDMLxDH_multi<"vqdmlsdh", 0b0, 0b0, 0b1>;
  3910. defm MVE_VQDMLSDHX : MVE_VQxDMLxDH_multi<"vqdmlsdhx", 0b1, 0b0, 0b1>;
  3911. defm MVE_VQRDMLSDH : MVE_VQxDMLxDH_multi<"vqrdmlsdh", 0b0, 0b1, 0b1>;
  3912. defm MVE_VQRDMLSDHX : MVE_VQxDMLxDH_multi<"vqrdmlsdhx", 0b1, 0b1, 0b1>;
  3913. class MVE_VCMUL<string iname, string suffix, bit size, string cstr="">
  3914. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  3915. (ins MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
  3916. "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
  3917. bits<4> Qn;
  3918. bits<2> rot;
  3919. let Inst{28} = size;
  3920. let Inst{21-20} = 0b11;
  3921. let Inst{19-17} = Qn{2-0};
  3922. let Inst{16} = 0b0;
  3923. let Inst{12} = rot{1};
  3924. let Inst{8} = 0b0;
  3925. let Inst{7} = Qn{3};
  3926. let Inst{0} = rot{0};
  3927. let Predicates = [HasMVEFloat];
  3928. }
  3929. multiclass MVE_VCMUL_m<string iname, MVEVectorVTInfo VTI,
  3930. bit size, string cstr=""> {
  3931. def "" : MVE_VCMUL<iname, VTI.Suffix, size, cstr>;
  3932. defvar Inst = !cast<Instruction>(NAME);
  3933. let Predicates = [HasMVEFloat] in {
  3934. def : Pat<(VTI.Vec (int_arm_mve_vcmulq
  3935. imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  3936. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3937. imm:$rot))>;
  3938. def : Pat<(VTI.Vec (int_arm_mve_vcmulq_predicated
  3939. imm:$rot, (VTI.Vec MQPR:$inactive),
  3940. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3941. (VTI.Pred VCCR:$mask))),
  3942. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  3943. imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
  3944. (VTI.Vec MQPR:$inactive)))>;
  3945. }
  3946. }
  3947. defm MVE_VCMULf16 : MVE_VCMUL_m<"vcmul", MVE_v8f16, 0b0>;
  3948. defm MVE_VCMULf32 : MVE_VCMUL_m<"vcmul", MVE_v4f32, 0b1, "@earlyclobber $Qd">;
  3949. class MVE_VMULL<string iname, string suffix, bit bit_28, bits<2> bits_21_20,
  3950. bit T, string cstr, list<dag> pattern=[]>
  3951. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  3952. (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
  3953. vpred_r, cstr, pattern> {
  3954. bits<4> Qd;
  3955. bits<4> Qn;
  3956. bits<4> Qm;
  3957. let Inst{28} = bit_28;
  3958. let Inst{21-20} = bits_21_20;
  3959. let Inst{19-17} = Qn{2-0};
  3960. let Inst{16} = 0b1;
  3961. let Inst{12} = T;
  3962. let Inst{8} = 0b0;
  3963. let Inst{7} = Qn{3};
  3964. let Inst{0} = 0b0;
  3965. let validForTailPredication = 1;
  3966. let doubleWidthResult = 1;
  3967. }
  3968. multiclass MVE_VMULL_m<MVEVectorVTInfo VTI,
  3969. SDNode unpred_op, Intrinsic pred_int,
  3970. bit Top, string cstr=""> {
  3971. def "" : MVE_VMULL<"vmull" # !if(Top, "t", "b"), VTI.Suffix, VTI.Unsigned,
  3972. VTI.Size, Top, cstr>;
  3973. defvar Inst = !cast<Instruction>(NAME);
  3974. let Predicates = [HasMVEInt] in {
  3975. defvar uflag = !if(!eq(VTI.SuffixLetter, "p"), (?), (? (i32 VTI.Unsigned)));
  3976. // Unpredicated multiply
  3977. def : Pat<(VTI.DblVec !con((unpred_op (VTI.Vec MQPR:$Qm),
  3978. (VTI.Vec MQPR:$Qn)),
  3979. uflag, (? (i32 Top)))),
  3980. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  3981. // Predicated multiply
  3982. def : Pat<(VTI.DblVec !con((pred_int (VTI.Vec MQPR:$Qm),
  3983. (VTI.Vec MQPR:$Qn)),
  3984. uflag, (? (i32 Top), (VTI.DblPred VCCR:$mask),
  3985. (VTI.DblVec MQPR:$inactive)))),
  3986. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  3987. ARMVCCThen, (VTI.DblPred VCCR:$mask),
  3988. (VTI.DblVec MQPR:$inactive)))>;
  3989. }
  3990. }
  3991. // For polynomial multiplies, the size bits take the unused value 0b11, and
  3992. // the unsigned bit switches to encoding the size.
  3993. defm MVE_VMULLBs8 : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
  3994. int_arm_mve_mull_int_predicated, 0b0>;
  3995. defm MVE_VMULLTs8 : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
  3996. int_arm_mve_mull_int_predicated, 0b1>;
  3997. defm MVE_VMULLBs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
  3998. int_arm_mve_mull_int_predicated, 0b0>;
  3999. defm MVE_VMULLTs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
  4000. int_arm_mve_mull_int_predicated, 0b1>;
  4001. defm MVE_VMULLBs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
  4002. int_arm_mve_mull_int_predicated, 0b0,
  4003. "@earlyclobber $Qd">;
  4004. defm MVE_VMULLTs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
  4005. int_arm_mve_mull_int_predicated, 0b1,
  4006. "@earlyclobber $Qd">;
  4007. defm MVE_VMULLBu8 : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
  4008. int_arm_mve_mull_int_predicated, 0b0>;
  4009. defm MVE_VMULLTu8 : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
  4010. int_arm_mve_mull_int_predicated, 0b1>;
  4011. defm MVE_VMULLBu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
  4012. int_arm_mve_mull_int_predicated, 0b0>;
  4013. defm MVE_VMULLTu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
  4014. int_arm_mve_mull_int_predicated, 0b1>;
  4015. defm MVE_VMULLBu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
  4016. int_arm_mve_mull_int_predicated, 0b0,
  4017. "@earlyclobber $Qd">;
  4018. defm MVE_VMULLTu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
  4019. int_arm_mve_mull_int_predicated, 0b1,
  4020. "@earlyclobber $Qd">;
  4021. defm MVE_VMULLBp8 : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
  4022. int_arm_mve_mull_poly_predicated, 0b0>;
  4023. defm MVE_VMULLTp8 : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
  4024. int_arm_mve_mull_poly_predicated, 0b1>;
  4025. defm MVE_VMULLBp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
  4026. int_arm_mve_mull_poly_predicated, 0b0>;
  4027. defm MVE_VMULLTp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
  4028. int_arm_mve_mull_poly_predicated, 0b1>;
  4029. let Predicates = [HasMVEInt] in {
  4030. def : Pat<(v2i64 (ARMvmulls (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
  4031. (MVE_VMULLBs32 MQPR:$src1, MQPR:$src2)>;
  4032. def : Pat<(v2i64 (ARMvmulls (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
  4033. (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
  4034. (MVE_VMULLTs32 MQPR:$src1, MQPR:$src2)>;
  4035. def : Pat<(mul (sext_inreg (v4i32 MQPR:$src1), v4i16),
  4036. (sext_inreg (v4i32 MQPR:$src2), v4i16)),
  4037. (MVE_VMULLBs16 MQPR:$src1, MQPR:$src2)>;
  4038. def : Pat<(mul (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))), v4i16),
  4039. (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))), v4i16)),
  4040. (MVE_VMULLTs16 MQPR:$src1, MQPR:$src2)>;
  4041. def : Pat<(mul (sext_inreg (v8i16 MQPR:$src1), v8i8),
  4042. (sext_inreg (v8i16 MQPR:$src2), v8i8)),
  4043. (MVE_VMULLBs8 MQPR:$src1, MQPR:$src2)>;
  4044. def : Pat<(mul (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), v8i8),
  4045. (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), v8i8)),
  4046. (MVE_VMULLTs8 MQPR:$src1, MQPR:$src2)>;
  4047. def : Pat<(v2i64 (ARMvmullu (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
  4048. (MVE_VMULLBu32 MQPR:$src1, MQPR:$src2)>;
  4049. def : Pat<(v2i64 (ARMvmullu (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
  4050. (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
  4051. (MVE_VMULLTu32 MQPR:$src1, MQPR:$src2)>;
  4052. def : Pat<(mul (and (v4i32 MQPR:$src1), (v4i32 (ARMvmovImm (i32 0xCFF)))),
  4053. (and (v4i32 MQPR:$src2), (v4i32 (ARMvmovImm (i32 0xCFF))))),
  4054. (MVE_VMULLBu16 MQPR:$src1, MQPR:$src2)>;
  4055. def : Pat<(mul (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))),
  4056. (v4i32 (ARMvmovImm (i32 0xCFF)))),
  4057. (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))),
  4058. (v4i32 (ARMvmovImm (i32 0xCFF))))),
  4059. (MVE_VMULLTu16 MQPR:$src1, MQPR:$src2)>;
  4060. def : Pat<(mul (ARMvbicImm (v8i16 MQPR:$src1), (i32 0xAFF)),
  4061. (ARMvbicImm (v8i16 MQPR:$src2), (i32 0xAFF))),
  4062. (MVE_VMULLBu8 MQPR:$src1, MQPR:$src2)>;
  4063. def : Pat<(mul (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), (i32 0xAFF)),
  4064. (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), (i32 0xAFF))),
  4065. (MVE_VMULLTu8 MQPR:$src1, MQPR:$src2)>;
  4066. }
  4067. class MVE_VxMULH<string iname, string suffix, bit U, bits<2> size, bit round,
  4068. list<dag> pattern=[]>
  4069. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  4070. (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
  4071. vpred_r, "", pattern> {
  4072. bits<4> Qn;
  4073. let Inst{28} = U;
  4074. let Inst{21-20} = size;
  4075. let Inst{19-17} = Qn{2-0};
  4076. let Inst{16} = 0b1;
  4077. let Inst{12} = round;
  4078. let Inst{8} = 0b0;
  4079. let Inst{7} = Qn{3};
  4080. let Inst{0} = 0b1;
  4081. }
  4082. multiclass MVE_VxMULH_m<string iname, MVEVectorVTInfo VTI, SDNode unpred_op,
  4083. Intrinsic pred_int, bit round> {
  4084. def "" : MVE_VxMULH<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, round>;
  4085. defvar Inst = !cast<Instruction>(NAME);
  4086. let Predicates = [HasMVEInt] in {
  4087. // Unpredicated multiply returning high bits
  4088. def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  4089. (i32 VTI.Unsigned))),
  4090. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  4091. // Predicated multiply returning high bits
  4092. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  4093. (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
  4094. (VTI.Vec MQPR:$inactive))),
  4095. (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  4096. ARMVCCThen, (VTI.Pred VCCR:$mask),
  4097. (VTI.Vec MQPR:$inactive)))>;
  4098. }
  4099. }
  4100. multiclass MVE_VMULT<string iname, MVEVectorVTInfo VTI, bit round>
  4101. : MVE_VxMULH_m<iname, VTI, !if(round, int_arm_mve_vrmulh, int_arm_mve_vmulh),
  4102. !if(round, int_arm_mve_rmulh_predicated,
  4103. int_arm_mve_mulh_predicated),
  4104. round>;
  4105. defm MVE_VMULHs8 : MVE_VMULT<"vmulh", MVE_v16s8, 0b0>;
  4106. defm MVE_VMULHs16 : MVE_VMULT<"vmulh", MVE_v8s16, 0b0>;
  4107. defm MVE_VMULHs32 : MVE_VMULT<"vmulh", MVE_v4s32, 0b0>;
  4108. defm MVE_VMULHu8 : MVE_VMULT<"vmulh", MVE_v16u8, 0b0>;
  4109. defm MVE_VMULHu16 : MVE_VMULT<"vmulh", MVE_v8u16, 0b0>;
  4110. defm MVE_VMULHu32 : MVE_VMULT<"vmulh", MVE_v4u32, 0b0>;
  4111. defm MVE_VRMULHs8 : MVE_VMULT<"vrmulh", MVE_v16s8, 0b1>;
  4112. defm MVE_VRMULHs16 : MVE_VMULT<"vrmulh", MVE_v8s16, 0b1>;
  4113. defm MVE_VRMULHs32 : MVE_VMULT<"vrmulh", MVE_v4s32, 0b1>;
  4114. defm MVE_VRMULHu8 : MVE_VMULT<"vrmulh", MVE_v16u8, 0b1>;
  4115. defm MVE_VRMULHu16 : MVE_VMULT<"vrmulh", MVE_v8u16, 0b1>;
  4116. defm MVE_VRMULHu32 : MVE_VMULT<"vrmulh", MVE_v4u32, 0b1>;
  4117. class MVE_VxMOVxN<string iname, string suffix, bit bit_28, bit bit_17,
  4118. bits<2> size, bit T, list<dag> pattern=[]>
  4119. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  4120. (ins MQPR:$Qd_src, MQPR:$Qm), "$Qd, $Qm",
  4121. vpred_n, "$Qd = $Qd_src", pattern> {
  4122. let Inst{28} = bit_28;
  4123. let Inst{21-20} = 0b11;
  4124. let Inst{19-18} = size;
  4125. let Inst{17} = bit_17;
  4126. let Inst{16} = 0b1;
  4127. let Inst{12} = T;
  4128. let Inst{8} = 0b0;
  4129. let Inst{7} = !not(bit_17);
  4130. let Inst{0} = 0b1;
  4131. let validForTailPredication = 1;
  4132. let retainsPreviousHalfElement = 1;
  4133. }
  4134. multiclass MVE_VxMOVxN_halves<string iname, string suffix,
  4135. bit bit_28, bit bit_17, bits<2> size> {
  4136. def bh : MVE_VxMOVxN<iname # "b", suffix, bit_28, bit_17, size, 0b0>;
  4137. def th : MVE_VxMOVxN<iname # "t", suffix, bit_28, bit_17, size, 0b1>;
  4138. }
  4139. defm MVE_VMOVNi16 : MVE_VxMOVxN_halves<"vmovn", "i16", 0b1, 0b0, 0b00>;
  4140. defm MVE_VMOVNi32 : MVE_VxMOVxN_halves<"vmovn", "i32", 0b1, 0b0, 0b01>;
  4141. defm MVE_VQMOVNs16 : MVE_VxMOVxN_halves<"vqmovn", "s16", 0b0, 0b1, 0b00>;
  4142. defm MVE_VQMOVNs32 : MVE_VxMOVxN_halves<"vqmovn", "s32", 0b0, 0b1, 0b01>;
  4143. defm MVE_VQMOVNu16 : MVE_VxMOVxN_halves<"vqmovn", "u16", 0b1, 0b1, 0b00>;
  4144. defm MVE_VQMOVNu32 : MVE_VxMOVxN_halves<"vqmovn", "u32", 0b1, 0b1, 0b01>;
  4145. defm MVE_VQMOVUNs16 : MVE_VxMOVxN_halves<"vqmovun", "s16", 0b0, 0b0, 0b00>;
  4146. defm MVE_VQMOVUNs32 : MVE_VxMOVxN_halves<"vqmovun", "s32", 0b0, 0b0, 0b01>;
  4147. def MVEvmovn : SDNode<"ARMISD::VMOVN", SDTARMVEXT>;
  4148. multiclass MVE_VMOVN_p<Instruction Inst, bit top,
  4149. MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
  4150. // Match the most obvious MVEvmovn(a,b,t), which overwrites the odd or even
  4151. // lanes of a (depending on t) with the even lanes of b.
  4152. def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qd_src),
  4153. (VTI.Vec MQPR:$Qm), (i32 top))),
  4154. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
  4155. if !not(top) then {
  4156. // If we see MVEvmovn(a,ARMvrev(b),1), that wants to overwrite the odd
  4157. // lanes of a with the odd lanes of b. In other words, the lanes we're
  4158. // _keeping_ from a are the even ones. So we can flip it round and say that
  4159. // this is the same as overwriting the even lanes of b with the even lanes
  4160. // of a, i.e. it's a VMOVNB with the operands reversed.
  4161. defvar vrev = !cast<SDNode>("ARMvrev" # InVTI.LaneBits);
  4162. def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qm),
  4163. (VTI.Vec (vrev MQPR:$Qd_src)), (i32 1))),
  4164. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
  4165. }
  4166. // Match the IR intrinsic for a predicated VMOVN. This regards the Qm input
  4167. // as having wider lanes that we're narrowing, instead of already-narrow
  4168. // lanes that we're taking every other one of.
  4169. def : Pat<(VTI.Vec (int_arm_mve_vmovn_predicated (VTI.Vec MQPR:$Qd_src),
  4170. (InVTI.Vec MQPR:$Qm), (i32 top),
  4171. (InVTI.Pred VCCR:$pred))),
  4172. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
  4173. (InVTI.Vec MQPR:$Qm),
  4174. ARMVCCThen, (InVTI.Pred VCCR:$pred)))>;
  4175. }
  4176. defm : MVE_VMOVN_p<MVE_VMOVNi32bh, 0, MVE_v8i16, MVE_v4i32>;
  4177. defm : MVE_VMOVN_p<MVE_VMOVNi32th, 1, MVE_v8i16, MVE_v4i32>;
  4178. defm : MVE_VMOVN_p<MVE_VMOVNi16bh, 0, MVE_v16i8, MVE_v8i16>;
  4179. defm : MVE_VMOVN_p<MVE_VMOVNi16th, 1, MVE_v16i8, MVE_v8i16>;
  4180. multiclass MVE_VQMOVN_p<Instruction Inst, bit outU, bit inU, bit top,
  4181. MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
  4182. def : Pat<(VTI.Vec (int_arm_mve_vqmovn (VTI.Vec MQPR:$Qd_src),
  4183. (InVTI.Vec MQPR:$Qm),
  4184. (i32 outU), (i32 inU), (i32 top))),
  4185. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
  4186. (InVTI.Vec MQPR:$Qm)))>;
  4187. def : Pat<(VTI.Vec (int_arm_mve_vqmovn_predicated (VTI.Vec MQPR:$Qd_src),
  4188. (InVTI.Vec MQPR:$Qm),
  4189. (i32 outU), (i32 inU), (i32 top),
  4190. (InVTI.Pred VCCR:$pred))),
  4191. (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
  4192. (InVTI.Vec MQPR:$Qm),
  4193. ARMVCCThen, (InVTI.Pred VCCR:$pred)))>;
  4194. }
  4195. defm : MVE_VQMOVN_p<MVE_VQMOVNs32bh, 0, 0, 0, MVE_v8i16, MVE_v4i32>;
  4196. defm : MVE_VQMOVN_p<MVE_VQMOVNs32th, 0, 0, 1, MVE_v8i16, MVE_v4i32>;
  4197. defm : MVE_VQMOVN_p<MVE_VQMOVNs16bh, 0, 0, 0, MVE_v16i8, MVE_v8i16>;
  4198. defm : MVE_VQMOVN_p<MVE_VQMOVNs16th, 0, 0, 1, MVE_v16i8, MVE_v8i16>;
  4199. defm : MVE_VQMOVN_p<MVE_VQMOVNu32bh, 1, 1, 0, MVE_v8i16, MVE_v4i32>;
  4200. defm : MVE_VQMOVN_p<MVE_VQMOVNu32th, 1, 1, 1, MVE_v8i16, MVE_v4i32>;
  4201. defm : MVE_VQMOVN_p<MVE_VQMOVNu16bh, 1, 1, 0, MVE_v16i8, MVE_v8i16>;
  4202. defm : MVE_VQMOVN_p<MVE_VQMOVNu16th, 1, 1, 1, MVE_v16i8, MVE_v8i16>;
  4203. defm : MVE_VQMOVN_p<MVE_VQMOVUNs32bh, 1, 0, 0, MVE_v8i16, MVE_v4i32>;
  4204. defm : MVE_VQMOVN_p<MVE_VQMOVUNs32th, 1, 0, 1, MVE_v8i16, MVE_v4i32>;
  4205. defm : MVE_VQMOVN_p<MVE_VQMOVUNs16bh, 1, 0, 0, MVE_v16i8, MVE_v8i16>;
  4206. defm : MVE_VQMOVN_p<MVE_VQMOVUNs16th, 1, 0, 1, MVE_v16i8, MVE_v8i16>;
  4207. def SDTARMVMOVNQ : SDTypeProfile<1, 3, [SDTCisVec<0>, SDTCisSameAs<0, 1>,
  4208. SDTCisVec<2>, SDTCisVT<3, i32>]>;
  4209. def MVEvqmovns : SDNode<"ARMISD::VQMOVNs", SDTARMVMOVNQ>;
  4210. def MVEvqmovnu : SDNode<"ARMISD::VQMOVNu", SDTARMVMOVNQ>;
  4211. let Predicates = [HasMVEInt] in {
  4212. def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
  4213. (v8i16 (MVE_VQMOVNs32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
  4214. def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
  4215. (v8i16 (MVE_VQMOVNs32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
  4216. def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
  4217. (v16i8 (MVE_VQMOVNs16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
  4218. def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
  4219. (v16i8 (MVE_VQMOVNs16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
  4220. def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
  4221. (v8i16 (MVE_VQMOVNu32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
  4222. def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
  4223. (v8i16 (MVE_VQMOVNu32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
  4224. def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
  4225. (v16i8 (MVE_VQMOVNu16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
  4226. def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
  4227. (v16i8 (MVE_VQMOVNu16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
  4228. def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
  4229. (v8i16 (MVE_VQSHRNbhs32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
  4230. def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
  4231. (v16i8 (MVE_VQSHRNbhs16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
  4232. def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
  4233. (v8i16 (MVE_VQSHRNths32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
  4234. def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
  4235. (v16i8 (MVE_VQSHRNths16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
  4236. def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
  4237. (v8i16 (MVE_VQSHRNbhu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
  4238. def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
  4239. (v16i8 (MVE_VQSHRNbhu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
  4240. def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
  4241. (v8i16 (MVE_VQSHRNthu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
  4242. def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
  4243. (v16i8 (MVE_VQSHRNthu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
  4244. }
  4245. class MVE_VCVT_ff<string iname, string suffix, bit op, bit T,
  4246. dag iops_extra, vpred_ops vpred, string cstr>
  4247. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  4248. !con(iops_extra, (ins MQPR:$Qm)), "$Qd, $Qm",
  4249. vpred, cstr, []> {
  4250. let Inst{28} = op;
  4251. let Inst{21-16} = 0b111111;
  4252. let Inst{12} = T;
  4253. let Inst{8-7} = 0b00;
  4254. let Inst{0} = 0b1;
  4255. let Predicates = [HasMVEFloat];
  4256. let retainsPreviousHalfElement = 1;
  4257. }
  4258. def SDTARMVCVTL : SDTypeProfile<1, 2, [SDTCisVec<0>, SDTCisVec<1>,
  4259. SDTCisVT<2, i32>]>;
  4260. def MVEvcvtn : SDNode<"ARMISD::VCVTN", SDTARMVMOVNQ>;
  4261. def MVEvcvtl : SDNode<"ARMISD::VCVTL", SDTARMVCVTL>;
  4262. multiclass MVE_VCVT_f2h_m<string iname, int half> {
  4263. def "": MVE_VCVT_ff<iname, "f16.f32", 0b0, half,
  4264. (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
  4265. defvar Inst = !cast<Instruction>(NAME);
  4266. let Predicates = [HasMVEFloat] in {
  4267. def : Pat<(v8f16 (int_arm_mve_vcvt_narrow
  4268. (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
  4269. (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
  4270. def : Pat<(v8f16 (int_arm_mve_vcvt_narrow_predicated
  4271. (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half),
  4272. (v4i1 VCCR:$mask))),
  4273. (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm),
  4274. ARMVCCThen, (v4i1 VCCR:$mask)))>;
  4275. def : Pat<(v8f16 (MVEvcvtn (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
  4276. (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
  4277. }
  4278. }
  4279. multiclass MVE_VCVT_h2f_m<string iname, int half> {
  4280. def "": MVE_VCVT_ff<iname, "f32.f16", 0b1, half, (ins), vpred_r, "">;
  4281. defvar Inst = !cast<Instruction>(NAME);
  4282. let Predicates = [HasMVEFloat] in {
  4283. def : Pat<(v4f32 (int_arm_mve_vcvt_widen (v8f16 MQPR:$Qm), (i32 half))),
  4284. (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
  4285. def : Pat<(v4f32 (int_arm_mve_vcvt_widen_predicated
  4286. (v4f32 MQPR:$inactive), (v8f16 MQPR:$Qm), (i32 half),
  4287. (v4i1 VCCR:$mask))),
  4288. (v4f32 (Inst (v8f16 MQPR:$Qm), ARMVCCThen,
  4289. (v4i1 VCCR:$mask), (v4f32 MQPR:$inactive)))>;
  4290. def : Pat<(v4f32 (MVEvcvtl (v8f16 MQPR:$Qm), (i32 half))),
  4291. (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
  4292. }
  4293. }
  4294. defm MVE_VCVTf16f32bh : MVE_VCVT_f2h_m<"vcvtb", 0b0>;
  4295. defm MVE_VCVTf16f32th : MVE_VCVT_f2h_m<"vcvtt", 0b1>;
  4296. defm MVE_VCVTf32f16bh : MVE_VCVT_h2f_m<"vcvtb", 0b0>;
  4297. defm MVE_VCVTf32f16th : MVE_VCVT_h2f_m<"vcvtt", 0b1>;
  4298. class MVE_VxCADD<string iname, string suffix, bits<2> size, bit halve,
  4299. string cstr="">
  4300. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  4301. (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
  4302. "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
  4303. bits<4> Qn;
  4304. bit rot;
  4305. let Inst{28} = halve;
  4306. let Inst{21-20} = size;
  4307. let Inst{19-17} = Qn{2-0};
  4308. let Inst{16} = 0b0;
  4309. let Inst{12} = rot;
  4310. let Inst{8} = 0b1;
  4311. let Inst{7} = Qn{3};
  4312. let Inst{0} = 0b0;
  4313. }
  4314. multiclass MVE_VxCADD_m<string iname, MVEVectorVTInfo VTI,
  4315. bit halve, string cstr=""> {
  4316. def "" : MVE_VxCADD<iname, VTI.Suffix, VTI.Size, halve, cstr>;
  4317. defvar Inst = !cast<Instruction>(NAME);
  4318. let Predicates = [HasMVEInt] in {
  4319. def : Pat<(VTI.Vec (int_arm_mve_vcaddq halve,
  4320. imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
  4321. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  4322. imm:$rot))>;
  4323. def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated halve,
  4324. imm:$rot, (VTI.Vec MQPR:$inactive),
  4325. (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  4326. (VTI.Pred VCCR:$mask))),
  4327. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
  4328. imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
  4329. (VTI.Vec MQPR:$inactive)))>;
  4330. }
  4331. }
  4332. defm MVE_VCADDi8 : MVE_VxCADD_m<"vcadd", MVE_v16i8, 0b1>;
  4333. defm MVE_VCADDi16 : MVE_VxCADD_m<"vcadd", MVE_v8i16, 0b1>;
  4334. defm MVE_VCADDi32 : MVE_VxCADD_m<"vcadd", MVE_v4i32, 0b1, "@earlyclobber $Qd">;
  4335. defm MVE_VHCADDs8 : MVE_VxCADD_m<"vhcadd", MVE_v16s8, 0b0>;
  4336. defm MVE_VHCADDs16 : MVE_VxCADD_m<"vhcadd", MVE_v8s16, 0b0>;
  4337. defm MVE_VHCADDs32 : MVE_VxCADD_m<"vhcadd", MVE_v4s32, 0b0, "@earlyclobber $Qd">;
  4338. class MVE_VADCSBC<string iname, bit I, bit subtract,
  4339. dag carryin, list<dag> pattern=[]>
  4340. : MVE_qDest_qSrc<iname, "i32", (outs MQPR:$Qd, cl_FPSCR_NZCV:$carryout),
  4341. !con((ins MQPR:$Qn, MQPR:$Qm), carryin),
  4342. "$Qd, $Qn, $Qm", vpred_r, "", pattern> {
  4343. bits<4> Qn;
  4344. let Inst{28} = subtract;
  4345. let Inst{21-20} = 0b11;
  4346. let Inst{19-17} = Qn{2-0};
  4347. let Inst{16} = 0b0;
  4348. let Inst{12} = I;
  4349. let Inst{8} = 0b1;
  4350. let Inst{7} = Qn{3};
  4351. let Inst{0} = 0b0;
  4352. // Custom decoder method in order to add the FPSCR operand(s), which
  4353. // Tablegen won't do right
  4354. let DecoderMethod = "DecodeMVEVADCInstruction";
  4355. }
  4356. def MVE_VADC : MVE_VADCSBC<"vadc", 0b0, 0b0, (ins cl_FPSCR_NZCV:$carryin)>;
  4357. def MVE_VADCI : MVE_VADCSBC<"vadci", 0b1, 0b0, (ins)>;
  4358. def MVE_VSBC : MVE_VADCSBC<"vsbc", 0b0, 0b1, (ins cl_FPSCR_NZCV:$carryin)>;
  4359. def MVE_VSBCI : MVE_VADCSBC<"vsbci", 0b1, 0b1, (ins)>;
  4360. class MVE_VQDMULL<string iname, string suffix, bit size, bit T,
  4361. string cstr="", list<dag> pattern=[]>
  4362. : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
  4363. (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
  4364. vpred_r, cstr, pattern> {
  4365. bits<4> Qn;
  4366. let Inst{28} = size;
  4367. let Inst{21-20} = 0b11;
  4368. let Inst{19-17} = Qn{2-0};
  4369. let Inst{16} = 0b0;
  4370. let Inst{12} = T;
  4371. let Inst{8} = 0b1;
  4372. let Inst{7} = Qn{3};
  4373. let Inst{0} = 0b1;
  4374. let validForTailPredication = 1;
  4375. let doubleWidthResult = 1;
  4376. }
  4377. multiclass MVE_VQDMULL_m<string iname, MVEVectorVTInfo VTI, bit size, bit T,
  4378. string cstr> {
  4379. def "" : MVE_VQDMULL<iname, VTI.Suffix, size, T, cstr>;
  4380. defvar Inst = !cast<Instruction>(NAME);
  4381. let Predicates = [HasMVEInt] in {
  4382. // Unpredicated saturating multiply
  4383. def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
  4384. (VTI.Vec MQPR:$Qn), (i32 T))),
  4385. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
  4386. // Predicated saturating multiply
  4387. def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
  4388. (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  4389. (i32 T), (VTI.DblPred VCCR:$mask),
  4390. (VTI.DblVec MQPR:$inactive))),
  4391. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
  4392. ARMVCCThen, (VTI.DblPred VCCR:$mask),
  4393. (VTI.DblVec MQPR:$inactive)))>;
  4394. }
  4395. }
  4396. multiclass MVE_VQDMULL_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
  4397. defm bh : MVE_VQDMULL_m<"vqdmullb", VTI, size, 0b0, cstr>;
  4398. defm th : MVE_VQDMULL_m<"vqdmullt", VTI, size, 0b1, cstr>;
  4399. }
  4400. defm MVE_VQDMULLs16 : MVE_VQDMULL_halves<MVE_v8s16, 0b0>;
  4401. defm MVE_VQDMULLs32 : MVE_VQDMULL_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
  4402. // end of mve_qDest_qSrc
  4403. // start of mve_qDest_rSrc
  4404. class MVE_qr_base<dag oops, dag iops, InstrItinClass itin, string iname,
  4405. string suffix, string ops, vpred_ops vpred, string cstr,
  4406. list<dag> pattern=[]>
  4407. : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
  4408. bits<4> Qd;
  4409. bits<4> Qn;
  4410. bits<4> Rm;
  4411. let Inst{25-23} = 0b100;
  4412. let Inst{22} = Qd{3};
  4413. let Inst{19-17} = Qn{2-0};
  4414. let Inst{15-13} = Qd{2-0};
  4415. let Inst{11-9} = 0b111;
  4416. let Inst{7} = Qn{3};
  4417. let Inst{6} = 0b1;
  4418. let Inst{4} = 0b0;
  4419. let Inst{3-0} = Rm{3-0};
  4420. }
  4421. class MVE_qDest_rSrc<string iname, string suffix, string cstr="", list<dag> pattern=[]>
  4422. : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qn, rGPR:$Rm),
  4423. NoItinerary, iname, suffix, "$Qd, $Qn, $Rm", vpred_r, cstr,
  4424. pattern>;
  4425. class MVE_qDestSrc_rSrc<string iname, string suffix, list<dag> pattern=[]>
  4426. : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qn, rGPR:$Rm),
  4427. NoItinerary, iname, suffix, "$Qd, $Qn, $Rm", vpred_n, "$Qd = $Qd_src",
  4428. pattern>;
  4429. class MVE_qDest_single_rSrc<string iname, string suffix, list<dag> pattern=[]>
  4430. : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, rGPR:$Rm), NoItinerary, iname,
  4431. suffix, "$Qd, $Rm", vpred_n, "$Qd = $Qd_src", pattern> {
  4432. bits<4> Qd;
  4433. bits<4> Rm;
  4434. let Inst{22} = Qd{3};
  4435. let Inst{15-13} = Qd{2-0};
  4436. let Inst{3-0} = Rm{3-0};
  4437. }
  4438. // Patterns for vector-scalar instructions with integer operands
  4439. multiclass MVE_vec_scalar_int_pat_m<Instruction inst, MVEVectorVTInfo VTI,
  4440. SDNode unpred_op, SDNode pred_op,
  4441. bit unpred_has_sign = 0,
  4442. bit pred_has_sign = 0> {
  4443. defvar UnpredSign = !if(unpred_has_sign, (? (i32 VTI.Unsigned)), (?));
  4444. defvar PredSign = !if(pred_has_sign, (? (i32 VTI.Unsigned)), (?));
  4445. let Predicates = [HasMVEInt] in {
  4446. // Unpredicated version
  4447. def : Pat<(VTI.Vec !con((unpred_op (VTI.Vec MQPR:$Qm),
  4448. (VTI.Vec (ARMvdup rGPR:$val))),
  4449. UnpredSign)),
  4450. (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
  4451. // Predicated version
  4452. def : Pat<(VTI.Vec !con((pred_op (VTI.Vec MQPR:$Qm),
  4453. (VTI.Vec (ARMvdup rGPR:$val))),
  4454. PredSign,
  4455. (pred_op (VTI.Pred VCCR:$mask),
  4456. (VTI.Vec MQPR:$inactive)))),
  4457. (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
  4458. ARMVCCThen, (VTI.Pred VCCR:$mask),
  4459. (VTI.Vec MQPR:$inactive)))>;
  4460. }
  4461. }
  4462. class MVE_VADDSUB_qr<string iname, string suffix, bits<2> size,
  4463. bit bit_5, bit bit_12, bit bit_16, bit bit_28>
  4464. : MVE_qDest_rSrc<iname, suffix, ""> {
  4465. let Inst{28} = bit_28;
  4466. let Inst{21-20} = size;
  4467. let Inst{16} = bit_16;
  4468. let Inst{12} = bit_12;
  4469. let Inst{8} = 0b1;
  4470. let Inst{5} = bit_5;
  4471. let validForTailPredication = 1;
  4472. }
  4473. // Vector-scalar add/sub
  4474. multiclass MVE_VADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
  4475. SDNode Op, Intrinsic PredInt> {
  4476. def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b0, subtract, 0b1, 0b0>;
  4477. let Predicates = [HasMVEInt] in {
  4478. defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
  4479. }
  4480. }
  4481. multiclass MVE_VADD_qr_m<MVEVectorVTInfo VTI>
  4482. : MVE_VADDSUB_qr_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
  4483. multiclass MVE_VSUB_qr_m<MVEVectorVTInfo VTI>
  4484. : MVE_VADDSUB_qr_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
  4485. defm MVE_VADD_qr_i8 : MVE_VADD_qr_m<MVE_v16i8>;
  4486. defm MVE_VADD_qr_i16 : MVE_VADD_qr_m<MVE_v8i16>;
  4487. defm MVE_VADD_qr_i32 : MVE_VADD_qr_m<MVE_v4i32>;
  4488. defm MVE_VSUB_qr_i8 : MVE_VSUB_qr_m<MVE_v16i8>;
  4489. defm MVE_VSUB_qr_i16 : MVE_VSUB_qr_m<MVE_v8i16>;
  4490. defm MVE_VSUB_qr_i32 : MVE_VSUB_qr_m<MVE_v4i32>;
  4491. // Vector-scalar saturating add/sub
  4492. multiclass MVE_VQADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
  4493. SDNode Op, Intrinsic PredInt> {
  4494. def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b1, subtract,
  4495. 0b0, VTI.Unsigned>;
  4496. let Predicates = [HasMVEInt] in {
  4497. defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
  4498. !cast<Instruction>(NAME)>;
  4499. }
  4500. }
  4501. multiclass MVE_VQADD_qr_m<MVEVectorVTInfo VTI, SDNode Op>
  4502. : MVE_VQADDSUB_qr_m<"vqadd", VTI, 0b0, Op, int_arm_mve_qadd_predicated>;
  4503. multiclass MVE_VQSUB_qr_m<MVEVectorVTInfo VTI, SDNode Op>
  4504. : MVE_VQADDSUB_qr_m<"vqsub", VTI, 0b1, Op, int_arm_mve_qsub_predicated>;
  4505. defm MVE_VQADD_qr_s8 : MVE_VQADD_qr_m<MVE_v16s8, saddsat>;
  4506. defm MVE_VQADD_qr_s16 : MVE_VQADD_qr_m<MVE_v8s16, saddsat>;
  4507. defm MVE_VQADD_qr_s32 : MVE_VQADD_qr_m<MVE_v4s32, saddsat>;
  4508. defm MVE_VQADD_qr_u8 : MVE_VQADD_qr_m<MVE_v16u8, uaddsat>;
  4509. defm MVE_VQADD_qr_u16 : MVE_VQADD_qr_m<MVE_v8u16, uaddsat>;
  4510. defm MVE_VQADD_qr_u32 : MVE_VQADD_qr_m<MVE_v4u32, uaddsat>;
  4511. defm MVE_VQSUB_qr_s8 : MVE_VQSUB_qr_m<MVE_v16s8, ssubsat>;
  4512. defm MVE_VQSUB_qr_s16 : MVE_VQSUB_qr_m<MVE_v8s16, ssubsat>;
  4513. defm MVE_VQSUB_qr_s32 : MVE_VQSUB_qr_m<MVE_v4s32, ssubsat>;
  4514. defm MVE_VQSUB_qr_u8 : MVE_VQSUB_qr_m<MVE_v16u8, usubsat>;
  4515. defm MVE_VQSUB_qr_u16 : MVE_VQSUB_qr_m<MVE_v8u16, usubsat>;
  4516. defm MVE_VQSUB_qr_u32 : MVE_VQSUB_qr_m<MVE_v4u32, usubsat>;
  4517. class MVE_VQDMULL_qr<string iname, string suffix, bit size,
  4518. bit T, string cstr="", list<dag> pattern=[]>
  4519. : MVE_qDest_rSrc<iname, suffix, cstr, pattern> {
  4520. let Inst{28} = size;
  4521. let Inst{21-20} = 0b11;
  4522. let Inst{16} = 0b0;
  4523. let Inst{12} = T;
  4524. let Inst{8} = 0b1;
  4525. let Inst{5} = 0b1;
  4526. let validForTailPredication = 1;
  4527. let doubleWidthResult = 1;
  4528. }
  4529. multiclass MVE_VQDMULL_qr_m<string iname, MVEVectorVTInfo VTI, bit size,
  4530. bit T, string cstr> {
  4531. def "" : MVE_VQDMULL_qr<iname, VTI.Suffix, size, T, cstr>;
  4532. defvar Inst = !cast<Instruction>(NAME);
  4533. let Predicates = [HasMVEInt] in {
  4534. // Unpredicated saturating multiply
  4535. def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
  4536. (VTI.Vec (ARMvdup rGPR:$val)),
  4537. (i32 T))),
  4538. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
  4539. // Predicated saturating multiply
  4540. def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
  4541. (VTI.Vec MQPR:$Qm),
  4542. (VTI.Vec (ARMvdup rGPR:$val)),
  4543. (i32 T),
  4544. (VTI.DblPred VCCR:$mask),
  4545. (VTI.DblVec MQPR:$inactive))),
  4546. (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
  4547. ARMVCCThen, (VTI.DblPred VCCR:$mask),
  4548. (VTI.DblVec MQPR:$inactive)))>;
  4549. }
  4550. }
  4551. multiclass MVE_VQDMULL_qr_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
  4552. defm bh : MVE_VQDMULL_qr_m<"vqdmullb", VTI, size, 0b0, cstr>;
  4553. defm th : MVE_VQDMULL_qr_m<"vqdmullt", VTI, size, 0b1, cstr>;
  4554. }
  4555. defm MVE_VQDMULL_qr_s16 : MVE_VQDMULL_qr_halves<MVE_v8s16, 0b0>;
  4556. defm MVE_VQDMULL_qr_s32 : MVE_VQDMULL_qr_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
  4557. class MVE_VxADDSUB_qr<string iname, string suffix,
  4558. bit bit_28, bits<2> bits_21_20, bit subtract,
  4559. list<dag> pattern=[]>
  4560. : MVE_qDest_rSrc<iname, suffix, "", pattern> {
  4561. let Inst{28} = bit_28;
  4562. let Inst{21-20} = bits_21_20;
  4563. let Inst{16} = 0b0;
  4564. let Inst{12} = subtract;
  4565. let Inst{8} = 0b1;
  4566. let Inst{5} = 0b0;
  4567. let validForTailPredication = 1;
  4568. }
  4569. multiclass MVE_VHADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
  4570. Intrinsic unpred_int, Intrinsic pred_int> {
  4571. def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, subtract>;
  4572. defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME),
  4573. VTI, unpred_int, pred_int, 1, 1>;
  4574. }
  4575. multiclass MVE_VHADD_qr_m<MVEVectorVTInfo VTI> :
  4576. MVE_VHADDSUB_qr_m<"vhadd", VTI, 0b0, int_arm_mve_vhadd,
  4577. int_arm_mve_hadd_predicated>;
  4578. multiclass MVE_VHSUB_qr_m<MVEVectorVTInfo VTI> :
  4579. MVE_VHADDSUB_qr_m<"vhsub", VTI, 0b1, int_arm_mve_vhsub,
  4580. int_arm_mve_hsub_predicated>;
  4581. defm MVE_VHADD_qr_s8 : MVE_VHADD_qr_m<MVE_v16s8>;
  4582. defm MVE_VHADD_qr_s16 : MVE_VHADD_qr_m<MVE_v8s16>;
  4583. defm MVE_VHADD_qr_s32 : MVE_VHADD_qr_m<MVE_v4s32>;
  4584. defm MVE_VHADD_qr_u8 : MVE_VHADD_qr_m<MVE_v16u8>;
  4585. defm MVE_VHADD_qr_u16 : MVE_VHADD_qr_m<MVE_v8u16>;
  4586. defm MVE_VHADD_qr_u32 : MVE_VHADD_qr_m<MVE_v4u32>;
  4587. defm MVE_VHSUB_qr_s8 : MVE_VHSUB_qr_m<MVE_v16s8>;
  4588. defm MVE_VHSUB_qr_s16 : MVE_VHSUB_qr_m<MVE_v8s16>;
  4589. defm MVE_VHSUB_qr_s32 : MVE_VHSUB_qr_m<MVE_v4s32>;
  4590. defm MVE_VHSUB_qr_u8 : MVE_VHSUB_qr_m<MVE_v16u8>;
  4591. defm MVE_VHSUB_qr_u16 : MVE_VHSUB_qr_m<MVE_v8u16>;
  4592. defm MVE_VHSUB_qr_u32 : MVE_VHSUB_qr_m<MVE_v4u32>;
  4593. multiclass MVE_VADDSUB_qr_f<string iname, MVEVectorVTInfo VTI, bit subtract,
  4594. SDNode Op, Intrinsic PredInt> {
  4595. def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, subtract>;
  4596. defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ),
  4597. !cast<Instruction>(NAME)>;
  4598. }
  4599. let Predicates = [HasMVEFloat] in {
  4600. defm MVE_VADD_qr_f32 : MVE_VADDSUB_qr_f<"vadd", MVE_v4f32, 0b0, fadd,
  4601. int_arm_mve_add_predicated>;
  4602. defm MVE_VADD_qr_f16 : MVE_VADDSUB_qr_f<"vadd", MVE_v8f16, 0b0, fadd,
  4603. int_arm_mve_add_predicated>;
  4604. defm MVE_VSUB_qr_f32 : MVE_VADDSUB_qr_f<"vsub", MVE_v4f32, 0b1, fsub,
  4605. int_arm_mve_sub_predicated>;
  4606. defm MVE_VSUB_qr_f16 : MVE_VADDSUB_qr_f<"vsub", MVE_v8f16, 0b1, fsub,
  4607. int_arm_mve_sub_predicated>;
  4608. }
  4609. class MVE_VxSHL_qr<string iname, string suffix, bit U, bits<2> size,
  4610. bit bit_7, bit bit_17, list<dag> pattern=[]>
  4611. : MVE_qDest_single_rSrc<iname, suffix, pattern> {
  4612. let Inst{28} = U;
  4613. let Inst{25-23} = 0b100;
  4614. let Inst{21-20} = 0b11;
  4615. let Inst{19-18} = size;
  4616. let Inst{17} = bit_17;
  4617. let Inst{16} = 0b1;
  4618. let Inst{12-8} = 0b11110;
  4619. let Inst{7} = bit_7;
  4620. let Inst{6-4} = 0b110;
  4621. let validForTailPredication = 1;
  4622. }
  4623. multiclass MVE_VxSHL_qr_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
  4624. def "" : MVE_VxSHL_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
  4625. defvar Inst = !cast<Instruction>(NAME);
  4626. def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar
  4627. (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
  4628. (i32 q), (i32 r), (i32 VTI.Unsigned))),
  4629. (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh)))>;
  4630. def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar_predicated
  4631. (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
  4632. (i32 q), (i32 r), (i32 VTI.Unsigned),
  4633. (VTI.Pred VCCR:$mask))),
  4634. (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
  4635. ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
  4636. }
  4637. multiclass MVE_VxSHL_qr_types<string iname, bit bit_7, bit bit_17> {
  4638. defm s8 : MVE_VxSHL_qr_p<iname, MVE_v16s8, bit_7, bit_17>;
  4639. defm s16 : MVE_VxSHL_qr_p<iname, MVE_v8s16, bit_7, bit_17>;
  4640. defm s32 : MVE_VxSHL_qr_p<iname, MVE_v4s32, bit_7, bit_17>;
  4641. defm u8 : MVE_VxSHL_qr_p<iname, MVE_v16u8, bit_7, bit_17>;
  4642. defm u16 : MVE_VxSHL_qr_p<iname, MVE_v8u16, bit_7, bit_17>;
  4643. defm u32 : MVE_VxSHL_qr_p<iname, MVE_v4u32, bit_7, bit_17>;
  4644. }
  4645. defm MVE_VSHL_qr : MVE_VxSHL_qr_types<"vshl", 0b0, 0b0>;
  4646. defm MVE_VRSHL_qr : MVE_VxSHL_qr_types<"vrshl", 0b0, 0b1>;
  4647. defm MVE_VQSHL_qr : MVE_VxSHL_qr_types<"vqshl", 0b1, 0b0>;
  4648. defm MVE_VQRSHL_qr : MVE_VxSHL_qr_types<"vqrshl", 0b1, 0b1>;
  4649. let Predicates = [HasMVEInt] in {
  4650. def : Pat<(v4i32 (ARMvshlu (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
  4651. (v4i32 (MVE_VSHL_qru32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
  4652. def : Pat<(v8i16 (ARMvshlu (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
  4653. (v8i16 (MVE_VSHL_qru16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
  4654. def : Pat<(v16i8 (ARMvshlu (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
  4655. (v16i8 (MVE_VSHL_qru8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
  4656. def : Pat<(v4i32 (ARMvshls (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
  4657. (v4i32 (MVE_VSHL_qrs32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
  4658. def : Pat<(v8i16 (ARMvshls (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
  4659. (v8i16 (MVE_VSHL_qrs16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
  4660. def : Pat<(v16i8 (ARMvshls (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
  4661. (v16i8 (MVE_VSHL_qrs8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
  4662. }
  4663. class MVE_VBRSR<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
  4664. : MVE_qDest_rSrc<iname, suffix, "", pattern> {
  4665. let Inst{28} = 0b1;
  4666. let Inst{21-20} = size;
  4667. let Inst{16} = 0b1;
  4668. let Inst{12} = 0b1;
  4669. let Inst{8} = 0b0;
  4670. let Inst{5} = 0b1;
  4671. let validForTailPredication = 1;
  4672. }
  4673. def MVE_VBRSR8 : MVE_VBRSR<"vbrsr", "8", 0b00>;
  4674. def MVE_VBRSR16 : MVE_VBRSR<"vbrsr", "16", 0b01>;
  4675. def MVE_VBRSR32 : MVE_VBRSR<"vbrsr", "32", 0b10>;
  4676. multiclass MVE_VBRSR_pat_m<MVEVectorVTInfo VTI, Instruction Inst> {
  4677. // Unpredicated
  4678. def : Pat<(VTI.Vec (int_arm_mve_vbrsr (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm))),
  4679. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm)))>;
  4680. // Predicated
  4681. def : Pat<(VTI.Vec (int_arm_mve_vbrsr_predicated
  4682. (VTI.Vec MQPR:$inactive),
  4683. (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
  4684. (VTI.Pred VCCR:$mask))),
  4685. (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
  4686. ARMVCCThen, (VTI.Pred VCCR:$mask),
  4687. (VTI.Vec MQPR:$inactive)))>;
  4688. }
  4689. let Predicates = [HasMVEInt] in {
  4690. def : Pat<(v16i8 ( bitreverse (v16i8 MQPR:$val1))),
  4691. (v16i8 ( MVE_VBRSR8 (v16i8 MQPR:$val1), (t2MOVi (i32 8)) ))>;
  4692. def : Pat<(v4i32 ( bitreverse (v4i32 MQPR:$val1))),
  4693. (v4i32 ( MVE_VBRSR32 (v4i32 MQPR:$val1), (t2MOVi (i32 32)) ))>;
  4694. def : Pat<(v8i16 ( bitreverse (v8i16 MQPR:$val1))),
  4695. (v8i16 ( MVE_VBRSR16 (v8i16 MQPR:$val1), (t2MOVi (i32 16)) ))>;
  4696. defm : MVE_VBRSR_pat_m<MVE_v16i8, MVE_VBRSR8>;
  4697. defm : MVE_VBRSR_pat_m<MVE_v8i16, MVE_VBRSR16>;
  4698. defm : MVE_VBRSR_pat_m<MVE_v4i32, MVE_VBRSR32>;
  4699. }
  4700. let Predicates = [HasMVEFloat] in {
  4701. defm : MVE_VBRSR_pat_m<MVE_v8f16, MVE_VBRSR16>;
  4702. defm : MVE_VBRSR_pat_m<MVE_v4f32, MVE_VBRSR32>;
  4703. }
  4704. class MVE_VMUL_qr_int<string iname, string suffix, bits<2> size>
  4705. : MVE_qDest_rSrc<iname, suffix, ""> {
  4706. let Inst{28} = 0b0;
  4707. let Inst{21-20} = size;
  4708. let Inst{16} = 0b1;
  4709. let Inst{12} = 0b1;
  4710. let Inst{8} = 0b0;
  4711. let Inst{5} = 0b1;
  4712. let validForTailPredication = 1;
  4713. }
  4714. multiclass MVE_VMUL_qr_int_m<MVEVectorVTInfo VTI> {
  4715. def "" : MVE_VMUL_qr_int<"vmul", VTI.Suffix, VTI.Size>;
  4716. let Predicates = [HasMVEInt] in {
  4717. defm : MVE_TwoOpPatternDup<VTI, mul, int_arm_mve_mul_predicated, (? ),
  4718. !cast<Instruction>(NAME), ARMimmOneV>;
  4719. }
  4720. }
  4721. defm MVE_VMUL_qr_i8 : MVE_VMUL_qr_int_m<MVE_v16i8>;
  4722. defm MVE_VMUL_qr_i16 : MVE_VMUL_qr_int_m<MVE_v8i16>;
  4723. defm MVE_VMUL_qr_i32 : MVE_VMUL_qr_int_m<MVE_v4i32>;
  4724. class MVE_VxxMUL_qr<string iname, string suffix,
  4725. bit bit_28, bits<2> bits_21_20, list<dag> pattern=[]>
  4726. : MVE_qDest_rSrc<iname, suffix, "", pattern> {
  4727. let Inst{28} = bit_28;
  4728. let Inst{21-20} = bits_21_20;
  4729. let Inst{16} = 0b1;
  4730. let Inst{12} = 0b0;
  4731. let Inst{8} = 0b0;
  4732. let Inst{5} = 0b1;
  4733. let validForTailPredication = 1;
  4734. }
  4735. multiclass MVE_VxxMUL_qr_m<string iname, MVEVectorVTInfo VTI, bit bit_28,
  4736. PatFrag Op, Intrinsic int_unpred, Intrinsic int_pred> {
  4737. def "" : MVE_VxxMUL_qr<iname, VTI.Suffix, bit_28, VTI.Size>;
  4738. let Predicates = [HasMVEInt] in {
  4739. defm : MVE_TwoOpPatternDup<VTI, Op, int_pred, (? ), !cast<Instruction>(NAME)>;
  4740. }
  4741. defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME), VTI, int_unpred, int_pred>;
  4742. }
  4743. multiclass MVE_VQDMULH_qr_m<MVEVectorVTInfo VTI> :
  4744. MVE_VxxMUL_qr_m<"vqdmulh", VTI, 0b0, MVEvqdmulh,
  4745. int_arm_mve_vqdmulh, int_arm_mve_qdmulh_predicated>;
  4746. multiclass MVE_VQRDMULH_qr_m<MVEVectorVTInfo VTI> :
  4747. MVE_VxxMUL_qr_m<"vqrdmulh", VTI, 0b1, null_frag,
  4748. int_arm_mve_vqrdmulh, int_arm_mve_qrdmulh_predicated>;
  4749. defm MVE_VQDMULH_qr_s8 : MVE_VQDMULH_qr_m<MVE_v16s8>;
  4750. defm MVE_VQDMULH_qr_s16 : MVE_VQDMULH_qr_m<MVE_v8s16>;
  4751. defm MVE_VQDMULH_qr_s32 : MVE_VQDMULH_qr_m<MVE_v4s32>;
  4752. defm MVE_VQRDMULH_qr_s8 : MVE_VQRDMULH_qr_m<MVE_v16s8>;
  4753. defm MVE_VQRDMULH_qr_s16 : MVE_VQRDMULH_qr_m<MVE_v8s16>;
  4754. defm MVE_VQRDMULH_qr_s32 : MVE_VQRDMULH_qr_m<MVE_v4s32>;
  4755. multiclass MVE_VxxMUL_qr_f_m<MVEVectorVTInfo VTI> {
  4756. let validForTailPredication = 1 in
  4757. def "" : MVE_VxxMUL_qr<"vmul", VTI.Suffix, VTI.Size{0}, 0b11>;
  4758. defm : MVE_TwoOpPatternDup<VTI, fmul, int_arm_mve_mul_predicated, (? ),
  4759. !cast<Instruction>(NAME)>;
  4760. }
  4761. let Predicates = [HasMVEFloat] in {
  4762. defm MVE_VMUL_qr_f16 : MVE_VxxMUL_qr_f_m<MVE_v8f16>;
  4763. defm MVE_VMUL_qr_f32 : MVE_VxxMUL_qr_f_m<MVE_v4f32>;
  4764. }
  4765. class MVE_VFMAMLA_qr<string iname, string suffix,
  4766. bit bit_28, bits<2> bits_21_20, bit S,
  4767. list<dag> pattern=[]>
  4768. : MVE_qDestSrc_rSrc<iname, suffix, pattern> {
  4769. let Inst{28} = bit_28;
  4770. let Inst{21-20} = bits_21_20;
  4771. let Inst{16} = 0b1;
  4772. let Inst{12} = S;
  4773. let Inst{8} = 0b0;
  4774. let Inst{5} = 0b0;
  4775. let validForTailPredication = 1;
  4776. let hasSideEffects = 0;
  4777. }
  4778. multiclass MVE_VMLA_qr_multi<string iname, MVEVectorVTInfo VTI,
  4779. bit scalar_addend> {
  4780. def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size,
  4781. scalar_addend>;
  4782. defvar Inst = !cast<Instruction>(NAME);
  4783. defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_n_predicated");
  4784. defvar v1 = (VTI.Vec MQPR:$v1);
  4785. defvar v2 = (VTI.Vec MQPR:$v2);
  4786. defvar vs = (VTI.Vec (ARMvdup rGPR:$s));
  4787. defvar s = (i32 rGPR:$s);
  4788. defvar pred = (VTI.Pred VCCR:$pred);
  4789. // The signed and unsigned variants of this instruction have different
  4790. // encodings, but they're functionally identical. For the sake of
  4791. // determinism, we generate only the unsigned variant.
  4792. if VTI.Unsigned then let Predicates = [HasMVEInt] in {
  4793. if scalar_addend then {
  4794. def : Pat<(VTI.Vec (add (mul v1, v2), vs)),
  4795. (VTI.Vec (Inst v1, v2, s))>;
  4796. } else {
  4797. def : Pat<(VTI.Vec (add (mul v2, vs), v1)),
  4798. (VTI.Vec (Inst v1, v2, s))>;
  4799. }
  4800. def : Pat<(VTI.Vec (pred_int v1, v2, s, pred)),
  4801. (VTI.Vec (Inst v1, v2, s, ARMVCCThen, pred))>;
  4802. }
  4803. }
  4804. defm MVE_VMLA_qr_s8 : MVE_VMLA_qr_multi<"vmla", MVE_v16s8, 0b0>;
  4805. defm MVE_VMLA_qr_s16 : MVE_VMLA_qr_multi<"vmla", MVE_v8s16, 0b0>;
  4806. defm MVE_VMLA_qr_s32 : MVE_VMLA_qr_multi<"vmla", MVE_v4s32, 0b0>;
  4807. defm MVE_VMLA_qr_u8 : MVE_VMLA_qr_multi<"vmla", MVE_v16u8, 0b0>;
  4808. defm MVE_VMLA_qr_u16 : MVE_VMLA_qr_multi<"vmla", MVE_v8u16, 0b0>;
  4809. defm MVE_VMLA_qr_u32 : MVE_VMLA_qr_multi<"vmla", MVE_v4u32, 0b0>;
  4810. defm MVE_VMLAS_qr_s8 : MVE_VMLA_qr_multi<"vmlas", MVE_v16s8, 0b1>;
  4811. defm MVE_VMLAS_qr_s16 : MVE_VMLA_qr_multi<"vmlas", MVE_v8s16, 0b1>;
  4812. defm MVE_VMLAS_qr_s32 : MVE_VMLA_qr_multi<"vmlas", MVE_v4s32, 0b1>;
  4813. defm MVE_VMLAS_qr_u8 : MVE_VMLA_qr_multi<"vmlas", MVE_v16u8, 0b1>;
  4814. defm MVE_VMLAS_qr_u16 : MVE_VMLA_qr_multi<"vmlas", MVE_v8u16, 0b1>;
  4815. defm MVE_VMLAS_qr_u32 : MVE_VMLA_qr_multi<"vmlas", MVE_v4u32, 0b1>;
  4816. multiclass MVE_VFMA_qr_multi<string iname, MVEVectorVTInfo VTI,
  4817. bit scalar_addend> {
  4818. def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, scalar_addend>;
  4819. defvar Inst = !cast<Instruction>(NAME);
  4820. defvar pred_int = int_arm_mve_fma_predicated;
  4821. defvar v1 = (VTI.Vec MQPR:$v1);
  4822. defvar v2 = (VTI.Vec MQPR:$v2);
  4823. defvar vs = (VTI.Vec (ARMvdup (i32 rGPR:$s)));
  4824. defvar is = (i32 rGPR:$s);
  4825. defvar pred = (VTI.Pred VCCR:$pred);
  4826. let Predicates = [HasMVEFloat] in {
  4827. if scalar_addend then {
  4828. def : Pat<(VTI.Vec (fma v1, v2, vs)),
  4829. (VTI.Vec (Inst v1, v2, is))>;
  4830. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  4831. (VTI.Vec (fma v1, v2, vs)),
  4832. v1)),
  4833. (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
  4834. def : Pat<(VTI.Vec (pred_int v1, v2, vs, pred)),
  4835. (VTI.Vec (Inst v1, v2, is, ARMVCCThen, pred))>;
  4836. } else {
  4837. def : Pat<(VTI.Vec (fma v1, vs, v2)),
  4838. (VTI.Vec (Inst v2, v1, is))>;
  4839. def : Pat<(VTI.Vec (fma vs, v1, v2)),
  4840. (VTI.Vec (Inst v2, v1, is))>;
  4841. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  4842. (VTI.Vec (fma vs, v2, v1)),
  4843. v1)),
  4844. (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
  4845. def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
  4846. (VTI.Vec (fma v2, vs, v1)),
  4847. v1)),
  4848. (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
  4849. def : Pat<(VTI.Vec (pred_int v1, vs, v2, pred)),
  4850. (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred))>;
  4851. def : Pat<(VTI.Vec (pred_int vs, v1, v2, pred)),
  4852. (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred))>;
  4853. }
  4854. }
  4855. }
  4856. let Predicates = [HasMVEFloat] in {
  4857. defm MVE_VFMA_qr_f16 : MVE_VFMA_qr_multi<"vfma", MVE_v8f16, 0>;
  4858. defm MVE_VFMA_qr_f32 : MVE_VFMA_qr_multi<"vfma", MVE_v4f32, 0>;
  4859. defm MVE_VFMA_qr_Sf16 : MVE_VFMA_qr_multi<"vfmas", MVE_v8f16, 1>;
  4860. defm MVE_VFMA_qr_Sf32 : MVE_VFMA_qr_multi<"vfmas", MVE_v4f32, 1>;
  4861. }
  4862. class MVE_VQDMLAH_qr<string iname, string suffix, bit U, bits<2> size,
  4863. bit bit_5, bit bit_12, list<dag> pattern=[]>
  4864. : MVE_qDestSrc_rSrc<iname, suffix, pattern> {
  4865. let Inst{28} = U;
  4866. let Inst{21-20} = size;
  4867. let Inst{16} = 0b0;
  4868. let Inst{12} = bit_12;
  4869. let Inst{8} = 0b0;
  4870. let Inst{5} = bit_5;
  4871. }
  4872. multiclass MVE_VQDMLAH_qr_multi<string iname, MVEVectorVTInfo VTI,
  4873. bit bit_5, bit bit_12> {
  4874. def "": MVE_VQDMLAH_qr<iname, VTI.Suffix, 0b0, VTI.Size, bit_5, bit_12>;
  4875. defvar Inst = !cast<Instruction>(NAME);
  4876. defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # iname);
  4877. defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_predicated");
  4878. let Predicates = [HasMVEInt] in {
  4879. def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
  4880. (i32 rGPR:$s))),
  4881. (VTI.Vec (Inst (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
  4882. (i32 rGPR:$s)))>;
  4883. def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
  4884. (i32 rGPR:$s), (VTI.Pred VCCR:$pred))),
  4885. (VTI.Vec (Inst (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
  4886. (i32 rGPR:$s), ARMVCCThen,
  4887. (VTI.Pred VCCR:$pred)))>;
  4888. }
  4889. }
  4890. multiclass MVE_VQDMLAH_qr_types<string iname, bit bit_5, bit bit_12> {
  4891. defm s8 : MVE_VQDMLAH_qr_multi<iname, MVE_v16s8, bit_5, bit_12>;
  4892. defm s16 : MVE_VQDMLAH_qr_multi<iname, MVE_v8s16, bit_5, bit_12>;
  4893. defm s32 : MVE_VQDMLAH_qr_multi<iname, MVE_v4s32, bit_5, bit_12>;
  4894. }
  4895. defm MVE_VQDMLAH_qr : MVE_VQDMLAH_qr_types<"vqdmlah", 0b1, 0b0>;
  4896. defm MVE_VQRDMLAH_qr : MVE_VQDMLAH_qr_types<"vqrdmlah", 0b0, 0b0>;
  4897. defm MVE_VQDMLASH_qr : MVE_VQDMLAH_qr_types<"vqdmlash", 0b1, 0b1>;
  4898. defm MVE_VQRDMLASH_qr : MVE_VQDMLAH_qr_types<"vqrdmlash", 0b0, 0b1>;
  4899. class MVE_VxDUP<string iname, string suffix, bits<2> size, bit bit_12,
  4900. list<dag> pattern=[]>
  4901. : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
  4902. (ins tGPREven:$Rn_src, MVE_VIDUP_imm:$imm), NoItinerary,
  4903. iname, suffix, "$Qd, $Rn, $imm", vpred_r, "$Rn = $Rn_src",
  4904. pattern> {
  4905. bits<4> Qd;
  4906. bits<4> Rn;
  4907. bits<2> imm;
  4908. let Inst{28} = 0b0;
  4909. let Inst{25-23} = 0b100;
  4910. let Inst{22} = Qd{3};
  4911. let Inst{21-20} = size;
  4912. let Inst{19-17} = Rn{3-1};
  4913. let Inst{16} = 0b1;
  4914. let Inst{15-13} = Qd{2-0};
  4915. let Inst{12} = bit_12;
  4916. let Inst{11-8} = 0b1111;
  4917. let Inst{7} = imm{1};
  4918. let Inst{6-1} = 0b110111;
  4919. let Inst{0} = imm{0};
  4920. let validForTailPredication = 1;
  4921. let hasSideEffects = 0;
  4922. }
  4923. def MVE_VIDUPu8 : MVE_VxDUP<"vidup", "u8", 0b00, 0b0>;
  4924. def MVE_VIDUPu16 : MVE_VxDUP<"vidup", "u16", 0b01, 0b0>;
  4925. def MVE_VIDUPu32 : MVE_VxDUP<"vidup", "u32", 0b10, 0b0>;
  4926. def MVE_VDDUPu8 : MVE_VxDUP<"vddup", "u8", 0b00, 0b1>;
  4927. def MVE_VDDUPu16 : MVE_VxDUP<"vddup", "u16", 0b01, 0b1>;
  4928. def MVE_VDDUPu32 : MVE_VxDUP<"vddup", "u32", 0b10, 0b1>;
  4929. class MVE_VxWDUP<string iname, string suffix, bits<2> size, bit bit_12,
  4930. list<dag> pattern=[]>
  4931. : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
  4932. (ins tGPREven:$Rn_src, tGPROdd:$Rm, MVE_VIDUP_imm:$imm), NoItinerary,
  4933. iname, suffix, "$Qd, $Rn, $Rm, $imm", vpred_r, "$Rn = $Rn_src",
  4934. pattern> {
  4935. bits<4> Qd;
  4936. bits<4> Rm;
  4937. bits<4> Rn;
  4938. bits<2> imm;
  4939. let Inst{28} = 0b0;
  4940. let Inst{25-23} = 0b100;
  4941. let Inst{22} = Qd{3};
  4942. let Inst{21-20} = size;
  4943. let Inst{19-17} = Rn{3-1};
  4944. let Inst{16} = 0b1;
  4945. let Inst{15-13} = Qd{2-0};
  4946. let Inst{12} = bit_12;
  4947. let Inst{11-8} = 0b1111;
  4948. let Inst{7} = imm{1};
  4949. let Inst{6-4} = 0b110;
  4950. let Inst{3-1} = Rm{3-1};
  4951. let Inst{0} = imm{0};
  4952. let validForTailPredication = 1;
  4953. let hasSideEffects = 0;
  4954. }
  4955. def MVE_VIWDUPu8 : MVE_VxWDUP<"viwdup", "u8", 0b00, 0b0>;
  4956. def MVE_VIWDUPu16 : MVE_VxWDUP<"viwdup", "u16", 0b01, 0b0>;
  4957. def MVE_VIWDUPu32 : MVE_VxWDUP<"viwdup", "u32", 0b10, 0b0>;
  4958. def MVE_VDWDUPu8 : MVE_VxWDUP<"vdwdup", "u8", 0b00, 0b1>;
  4959. def MVE_VDWDUPu16 : MVE_VxWDUP<"vdwdup", "u16", 0b01, 0b1>;
  4960. def MVE_VDWDUPu32 : MVE_VxWDUP<"vdwdup", "u32", 0b10, 0b1>;
  4961. let isReMaterializable = 1 in
  4962. class MVE_VCTPInst<string suffix, bits<2> size, list<dag> pattern=[]>
  4963. : MVE_p<(outs VCCR:$P0), (ins rGPR:$Rn), NoItinerary, "vctp", suffix,
  4964. "$Rn", vpred_n, "", pattern> {
  4965. bits<4> Rn;
  4966. let Inst{28-27} = 0b10;
  4967. let Inst{26-22} = 0b00000;
  4968. let Inst{21-20} = size;
  4969. let Inst{19-16} = Rn{3-0};
  4970. let Inst{15-11} = 0b11101;
  4971. let Inst{10-0} = 0b00000000001;
  4972. let Unpredictable{10-0} = 0b11111111111;
  4973. let Constraints = "";
  4974. let DecoderMethod = "DecodeMveVCTP";
  4975. let validForTailPredication = 1;
  4976. }
  4977. multiclass MVE_VCTP<MVEVectorVTInfo VTI, Intrinsic intr> {
  4978. def "": MVE_VCTPInst<VTI.BitsSuffix, VTI.Size>;
  4979. defvar Inst = !cast<Instruction>(NAME);
  4980. let Predicates = [HasMVEInt] in {
  4981. def : Pat<(intr rGPR:$Rn),
  4982. (VTI.Pred (Inst rGPR:$Rn))>;
  4983. def : Pat<(and (intr rGPR:$Rn), (VTI.Pred VCCR:$mask)),
  4984. (VTI.Pred (Inst rGPR:$Rn, ARMVCCThen, VCCR:$mask))>;
  4985. }
  4986. }
  4987. defm MVE_VCTP8 : MVE_VCTP<MVE_v16i8, int_arm_mve_vctp8>;
  4988. defm MVE_VCTP16 : MVE_VCTP<MVE_v8i16, int_arm_mve_vctp16>;
  4989. defm MVE_VCTP32 : MVE_VCTP<MVE_v4i32, int_arm_mve_vctp32>;
  4990. defm MVE_VCTP64 : MVE_VCTP<MVE_v2i64, int_arm_mve_vctp64>;
  4991. // end of mve_qDest_rSrc
  4992. // start of coproc mov
  4993. class MVE_VMOV_64bit<dag oops, dag iops, bit to_qreg, string ops, string cstr>
  4994. : MVE_VMOV_lane_base<oops, !con(iops, (ins MVEPairVectorIndex2:$idx,
  4995. MVEPairVectorIndex0:$idx2)),
  4996. NoItinerary, "vmov", "", ops, cstr, []> {
  4997. bits<5> Rt;
  4998. bits<5> Rt2;
  4999. bits<4> Qd;
  5000. bit idx;
  5001. bit idx2;
  5002. let Inst{31-23} = 0b111011000;
  5003. let Inst{22} = Qd{3};
  5004. let Inst{21} = 0b0;
  5005. let Inst{20} = to_qreg;
  5006. let Inst{19-16} = Rt2{3-0};
  5007. let Inst{15-13} = Qd{2-0};
  5008. let Inst{12-5} = 0b01111000;
  5009. let Inst{4} = idx2;
  5010. let Inst{3-0} = Rt{3-0};
  5011. let hasSideEffects = 0;
  5012. }
  5013. // The assembly syntax for these instructions mentions the vector
  5014. // register name twice, e.g.
  5015. //
  5016. // vmov q2[2], q2[0], r0, r1
  5017. // vmov r0, r1, q2[2], q2[0]
  5018. //
  5019. // which needs a bit of juggling with MC operand handling.
  5020. //
  5021. // For the move _into_ a vector register, the MC operand list also has
  5022. // to mention the register name twice: once as the output, and once as
  5023. // an extra input to represent where the unchanged half of the output
  5024. // register comes from (when this instruction is used in code
  5025. // generation). So we arrange that the first mention of the vector reg
  5026. // in the instruction is considered by the AsmMatcher to be the output
  5027. // ($Qd), and the second one is the input ($QdSrc). Binding them
  5028. // together with the existing 'tie' constraint is enough to enforce at
  5029. // register allocation time that they have to be the same register.
  5030. //
  5031. // For the move _from_ a vector register, there's no way to get round
  5032. // the fact that both instances of that register name have to be
  5033. // inputs. They have to be the same register again, but this time, we
  5034. // can't use a tie constraint, because that has to be between an
  5035. // output and an input operand. So this time, we have to arrange that
  5036. // the q-reg appears just once in the MC operand list, in spite of
  5037. // being mentioned twice in the asm syntax - which needs a custom
  5038. // AsmMatchConverter.
  5039. def MVE_VMOV_q_rr : MVE_VMOV_64bit<(outs MQPR:$Qd),
  5040. (ins MQPR:$QdSrc, rGPR:$Rt, rGPR:$Rt2),
  5041. 0b1, "$Qd$idx, $QdSrc$idx2, $Rt, $Rt2",
  5042. "$Qd = $QdSrc"> {
  5043. let DecoderMethod = "DecodeMVEVMOVDRegtoQ";
  5044. }
  5045. def MVE_VMOV_rr_q : MVE_VMOV_64bit<(outs rGPR:$Rt, rGPR:$Rt2), (ins MQPR:$Qd),
  5046. 0b0, "$Rt, $Rt2, $Qd$idx, $Qd$idx2", ""> {
  5047. let DecoderMethod = "DecodeMVEVMOVQtoDReg";
  5048. let AsmMatchConverter = "cvtMVEVMOVQtoDReg";
  5049. }
  5050. let Predicates = [HasMVEInt] in {
  5051. // Double lane moves. There are a number of patterns here. We know that the
  5052. // insertelt's will be in descending order by index, and need to match the 5
  5053. // patterns that might contain 2-0 or 3-1 pairs. These are:
  5054. // 3 2 1 0 -> vmovqrr 31; vmovqrr 20
  5055. // 3 2 1 -> vmovqrr 31; vmov 2
  5056. // 3 1 -> vmovqrr 31
  5057. // 2 1 0 -> vmovqrr 20; vmov 1
  5058. // 2 0 -> vmovqrr 20
  5059. // The other potential patterns will be handled by single lane inserts.
  5060. def : Pat<(insertelt (insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
  5061. rGPR:$srcA, (i32 0)),
  5062. rGPR:$srcB, (i32 1)),
  5063. rGPR:$srcC, (i32 2)),
  5064. rGPR:$srcD, (i32 3)),
  5065. (MVE_VMOV_q_rr (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcC, (i32 2), (i32 0)),
  5066. rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
  5067. def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
  5068. rGPR:$srcB, (i32 1)),
  5069. rGPR:$srcC, (i32 2)),
  5070. rGPR:$srcD, (i32 3)),
  5071. (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 2)),
  5072. rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
  5073. def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 1)), rGPR:$srcB, (i32 3)),
  5074. (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 3), (i32 1))>;
  5075. def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
  5076. rGPR:$srcB, (i32 0)),
  5077. rGPR:$srcC, (i32 1)),
  5078. rGPR:$srcD, (i32 2)),
  5079. (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 1)),
  5080. rGPR:$srcB, rGPR:$srcD, (i32 2), (i32 0))>;
  5081. def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 0)), rGPR:$srcB, (i32 2)),
  5082. (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 2), (i32 0))>;
  5083. }
  5084. // end of coproc mov
  5085. // start of MVE interleaving load/store
  5086. // Base class for the family of interleaving/deinterleaving
  5087. // load/stores with names like VLD20.8 and VST43.32.
  5088. class MVE_vldst24_base<bit writeback, bit fourregs, bits<2> stage, bits<2> size,
  5089. bit load, dag Oops, dag loadIops, dag wbIops,
  5090. string iname, string ops,
  5091. string cstr, list<dag> pattern=[]>
  5092. : MVE_MI<Oops, !con(loadIops, wbIops), NoItinerary, iname, ops, cstr, pattern> {
  5093. bits<4> VQd;
  5094. bits<4> Rn;
  5095. let Inst{31-22} = 0b1111110010;
  5096. let Inst{21} = writeback;
  5097. let Inst{20} = load;
  5098. let Inst{19-16} = Rn;
  5099. let Inst{15-13} = VQd{2-0};
  5100. let Inst{12-9} = 0b1111;
  5101. let Inst{8-7} = size;
  5102. let Inst{6-5} = stage;
  5103. let Inst{4-1} = 0b0000;
  5104. let Inst{0} = fourregs;
  5105. let mayLoad = load;
  5106. let mayStore = !eq(load,0);
  5107. let hasSideEffects = 0;
  5108. let validForTailPredication = load;
  5109. }
  5110. // A parameter class used to encapsulate all the ways the writeback
  5111. // variants of VLD20 and friends differ from the non-writeback ones.
  5112. class MVE_vldst24_writeback<bit b, dag Oo, dag Io,
  5113. string sy="", string c="", string n=""> {
  5114. bit writeback = b;
  5115. dag Oops = Oo;
  5116. dag Iops = Io;
  5117. string syntax = sy;
  5118. string cstr = c;
  5119. string id_suffix = n;
  5120. }
  5121. // Another parameter class that encapsulates the differences between VLD2x
  5122. // and VLD4x.
  5123. class MVE_vldst24_nvecs<int n, list<int> s, bit b, RegisterOperand vl> {
  5124. int nvecs = n;
  5125. list<int> stages = s;
  5126. bit bit0 = b;
  5127. RegisterOperand VecList = vl;
  5128. }
  5129. // A third parameter class that distinguishes VLDnn.8 from .16 from .32.
  5130. class MVE_vldst24_lanesize<int i, bits<2> b> {
  5131. int lanesize = i;
  5132. bits<2> sizebits = b;
  5133. }
  5134. // A base class for each direction of transfer: one for load, one for
  5135. // store. I can't make these a fourth independent parametric tuple
  5136. // class, because they have to take the nvecs tuple class as a
  5137. // parameter, in order to find the right VecList operand type.
  5138. class MVE_vld24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
  5139. MVE_vldst24_writeback wb, string iname,
  5140. list<dag> pattern=[]>
  5141. : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 1,
  5142. !con((outs n.VecList:$VQd), wb.Oops),
  5143. (ins n.VecList:$VQdSrc), wb.Iops,
  5144. iname, "$VQd, $Rn" # wb.syntax,
  5145. wb.cstr # ",$VQdSrc = $VQd", pattern>;
  5146. class MVE_vst24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
  5147. MVE_vldst24_writeback wb, string iname,
  5148. list<dag> pattern=[]>
  5149. : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 0,
  5150. wb.Oops, (ins n.VecList:$VQd), wb.Iops,
  5151. iname, "$VQd, $Rn" # wb.syntax,
  5152. wb.cstr, pattern>;
  5153. // Actually define all the interleaving loads and stores, by a series
  5154. // of nested foreaches over number of vectors (VLD2/VLD4); stage
  5155. // within one of those series (VLDx0/VLDx1/VLDx2/VLDx3); size of
  5156. // vector lane; writeback or no writeback.
  5157. foreach n = [MVE_vldst24_nvecs<2, [0,1], 0, VecList2Q>,
  5158. MVE_vldst24_nvecs<4, [0,1,2,3], 1, VecList4Q>] in
  5159. foreach stage = n.stages in
  5160. foreach s = [MVE_vldst24_lanesize< 8, 0b00>,
  5161. MVE_vldst24_lanesize<16, 0b01>,
  5162. MVE_vldst24_lanesize<32, 0b10>] in
  5163. foreach wb = [MVE_vldst24_writeback<
  5164. 1, (outs rGPR:$wb), (ins t2_nosp_addr_offset_none:$Rn),
  5165. "!", "$Rn.base = $wb", "_wb">,
  5166. MVE_vldst24_writeback<0, (outs), (ins t2_addr_offset_none:$Rn)>] in {
  5167. // For each case within all of those foreaches, define the actual
  5168. // instructions. The def names are made by gluing together pieces
  5169. // from all the parameter classes, and will end up being things like
  5170. // MVE_VLD20_8 and MVE_VST43_16_wb.
  5171. def "MVE_VLD" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
  5172. : MVE_vld24_base<n, stage, s.sizebits, wb,
  5173. "vld" # n.nvecs # stage # "." # s.lanesize>;
  5174. def "MVE_VST" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
  5175. : MVE_vst24_base<n, stage, s.sizebits, wb,
  5176. "vst" # n.nvecs # stage # "." # s.lanesize>;
  5177. }
  5178. def SDTARMVST2 : SDTypeProfile<1, 5, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
  5179. SDTCisSameAs<3, 4>, SDTCisVT<5, i32>]>;
  5180. def SDTARMVST4 : SDTypeProfile<1, 7, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
  5181. SDTCisSameAs<3, 4>, SDTCisSameAs<3, 5>,
  5182. SDTCisSameAs<3, 6>, SDTCisVT<7, i32>]>;
  5183. def MVEVST2UPD : SDNode<"ARMISD::VST2_UPD", SDTARMVST2, [SDNPHasChain]>;
  5184. def MVEVST4UPD : SDNode<"ARMISD::VST4_UPD", SDTARMVST4, [SDNPHasChain]>;
  5185. multiclass MVE_vst24_patterns<int lanesize, ValueType VT> {
  5186. foreach stage = [0,1] in
  5187. def : Pat<(int_arm_mve_vst2q i32:$addr,
  5188. (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage)),
  5189. (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize)
  5190. (REG_SEQUENCE QQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
  5191. t2_addr_offset_none:$addr)>;
  5192. foreach stage = [0,1] in
  5193. def : Pat<(i32 (MVEVST2UPD i32:$addr, (i32 32),
  5194. (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage))),
  5195. (i32 (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize#_wb)
  5196. (REG_SEQUENCE QQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
  5197. t2_addr_offset_none:$addr))>;
  5198. foreach stage = [0,1,2,3] in
  5199. def : Pat<(int_arm_mve_vst4q i32:$addr,
  5200. (VT MQPR:$v0), (VT MQPR:$v1),
  5201. (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage)),
  5202. (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize)
  5203. (REG_SEQUENCE QQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
  5204. VT:$v2, qsub_2, VT:$v3, qsub_3),
  5205. t2_addr_offset_none:$addr)>;
  5206. foreach stage = [0,1,2,3] in
  5207. def : Pat<(i32 (MVEVST4UPD i32:$addr, (i32 64),
  5208. (VT MQPR:$v0), (VT MQPR:$v1),
  5209. (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage))),
  5210. (i32 (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize#_wb)
  5211. (REG_SEQUENCE QQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
  5212. VT:$v2, qsub_2, VT:$v3, qsub_3),
  5213. t2_addr_offset_none:$addr))>;
  5214. }
  5215. defm : MVE_vst24_patterns<8, v16i8>;
  5216. defm : MVE_vst24_patterns<16, v8i16>;
  5217. defm : MVE_vst24_patterns<32, v4i32>;
  5218. defm : MVE_vst24_patterns<16, v8f16>;
  5219. defm : MVE_vst24_patterns<32, v4f32>;
  5220. // end of MVE interleaving load/store
  5221. // start of MVE predicable load/store
  5222. // A parameter class for the direction of transfer.
  5223. class MVE_ldst_direction<bit b, dag Oo, dag Io, string c=""> {
  5224. bit load = b;
  5225. dag Oops = Oo;
  5226. dag Iops = Io;
  5227. string cstr = c;
  5228. }
  5229. def MVE_ld: MVE_ldst_direction<1, (outs MQPR:$Qd), (ins), ",@earlyclobber $Qd">;
  5230. def MVE_st: MVE_ldst_direction<0, (outs), (ins MQPR:$Qd)>;
  5231. // A parameter class for the size of memory access in a load.
  5232. class MVE_memsz<bits<2> e, int s, AddrMode m, string mn, list<string> types> {
  5233. bits<2> encoding = e; // opcode bit(s) for encoding
  5234. int shift = s; // shift applied to immediate load offset
  5235. AddrMode AM = m;
  5236. // For instruction aliases: define the complete list of type
  5237. // suffixes at this size, and the canonical ones for loads and
  5238. // stores.
  5239. string MnemonicLetter = mn;
  5240. int TypeBits = !shl(8, s);
  5241. string CanonLoadSuffix = ".u" # TypeBits;
  5242. string CanonStoreSuffix = "." # TypeBits;
  5243. list<string> suffixes = !foreach(letter, types, "." # letter # TypeBits);
  5244. }
  5245. // Instances of MVE_memsz.
  5246. //
  5247. // (memD doesn't need an AddrMode, because those are only for
  5248. // contiguous loads, and memD is only used by gather/scatters.)
  5249. def MVE_memB: MVE_memsz<0b00, 0, AddrModeT2_i7, "b", ["", "u", "s"]>;
  5250. def MVE_memH: MVE_memsz<0b01, 1, AddrModeT2_i7s2, "h", ["", "u", "s", "f"]>;
  5251. def MVE_memW: MVE_memsz<0b10, 2, AddrModeT2_i7s4, "w", ["", "u", "s", "f"]>;
  5252. def MVE_memD: MVE_memsz<0b11, 3, ?, "d", ["", "u", "s", "f"]>;
  5253. // This is the base class for all the MVE loads and stores other than
  5254. // the interleaving ones. All the non-interleaving loads/stores share
  5255. // the characteristic that they operate on just one vector register,
  5256. // so they are VPT-predicable.
  5257. //
  5258. // The predication operand is vpred_n, for both loads and stores. For
  5259. // store instructions, the reason is obvious: if there is no output
  5260. // register, there can't be a need for an input parameter giving the
  5261. // output register's previous value. Load instructions also don't need
  5262. // that input parameter, because unlike MVE data processing
  5263. // instructions, predicated loads are defined to set the inactive
  5264. // lanes of the output register to zero, instead of preserving their
  5265. // input values.
  5266. class MVE_VLDRSTR_base<MVE_ldst_direction dir, bit U, bit P, bit W, bit opc,
  5267. dag oops, dag iops, string asm, string suffix,
  5268. string ops, string cstr, list<dag> pattern=[]>
  5269. : MVE_p<oops, iops, NoItinerary, asm, suffix, ops, vpred_n, cstr, pattern> {
  5270. bits<3> Qd;
  5271. let Inst{28} = U;
  5272. let Inst{25} = 0b0;
  5273. let Inst{24} = P;
  5274. let Inst{22} = 0b0;
  5275. let Inst{21} = W;
  5276. let Inst{20} = dir.load;
  5277. let Inst{15-13} = Qd{2-0};
  5278. let Inst{12} = opc;
  5279. let Inst{11-9} = 0b111;
  5280. let mayLoad = dir.load;
  5281. let mayStore = !eq(dir.load,0);
  5282. let hasSideEffects = 0;
  5283. let validForTailPredication = 1;
  5284. }
  5285. // Contiguous load and store instructions. These come in two main
  5286. // categories: same-size loads/stores in which 128 bits of vector
  5287. // register is transferred to or from 128 bits of memory in the most
  5288. // obvious way, and widening loads / narrowing stores, in which the
  5289. // size of memory accessed is less than the size of a vector register,
  5290. // so the load instructions sign- or zero-extend each memory value
  5291. // into a wider vector lane, and the store instructions truncate
  5292. // correspondingly.
  5293. //
  5294. // The instruction mnemonics for these two classes look reasonably
  5295. // similar, but the actual encodings are different enough to need two
  5296. // separate base classes.
  5297. // Contiguous, same size
  5298. class MVE_VLDRSTR_cs<MVE_ldst_direction dir, MVE_memsz memsz, bit P, bit W,
  5299. dag oops, dag iops, string asm, string suffix,
  5300. IndexMode im, string ops, string cstr>
  5301. : MVE_VLDRSTR_base<dir, 0, P, W, 1, oops, iops, asm, suffix, ops, cstr> {
  5302. bits<12> addr;
  5303. let Inst{23} = addr{7};
  5304. let Inst{19-16} = addr{11-8};
  5305. let Inst{8-7} = memsz.encoding;
  5306. let Inst{6-0} = addr{6-0};
  5307. }
  5308. // Contiguous, widening/narrowing
  5309. class MVE_VLDRSTR_cw<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
  5310. bit P, bit W, bits<2> size, dag oops, dag iops,
  5311. string asm, string suffix, IndexMode im,
  5312. string ops, string cstr>
  5313. : MVE_VLDRSTR_base<dir, U, P, W, 0, oops, iops, asm, suffix, ops, cstr> {
  5314. bits<11> addr;
  5315. let Inst{23} = addr{7};
  5316. let Inst{19} = memsz.encoding{0}; // enough to tell 16- from 32-bit
  5317. let Inst{18-16} = addr{10-8};
  5318. let Inst{8-7} = size;
  5319. let Inst{6-0} = addr{6-0};
  5320. let IM = im;
  5321. }
  5322. // Multiclass wrapper on each of the _cw and _cs base classes, to
  5323. // generate three writeback modes (none, preindex, postindex).
  5324. multiclass MVE_VLDRSTR_cw_m<MVE_ldst_direction dir, MVE_memsz memsz,
  5325. string asm, string suffix, bit U, bits<2> size> {
  5326. let AM = memsz.AM in {
  5327. def "" : MVE_VLDRSTR_cw<
  5328. dir, memsz, U, 1, 0, size,
  5329. dir.Oops, !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
  5330. asm, suffix, IndexModeNone, "$Qd, $addr", "">;
  5331. def _pre : MVE_VLDRSTR_cw<
  5332. dir, memsz, U, 1, 1, size,
  5333. !con((outs tGPR:$wb), dir.Oops),
  5334. !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
  5335. asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
  5336. let DecoderMethod = "DecodeMVE_MEM_1_pre<"#memsz.shift#">";
  5337. }
  5338. def _post : MVE_VLDRSTR_cw<
  5339. dir, memsz, U, 0, 1, size,
  5340. !con((outs tGPR:$wb), dir.Oops),
  5341. !con(dir.Iops, (ins t_addr_offset_none:$Rn,
  5342. t2am_imm7_offset<memsz.shift>:$addr)),
  5343. asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
  5344. bits<4> Rn;
  5345. let Inst{18-16} = Rn{2-0};
  5346. }
  5347. }
  5348. }
  5349. multiclass MVE_VLDRSTR_cs_m<MVE_ldst_direction dir, MVE_memsz memsz,
  5350. string asm, string suffix> {
  5351. let AM = memsz.AM in {
  5352. def "" : MVE_VLDRSTR_cs<
  5353. dir, memsz, 1, 0,
  5354. dir.Oops, !con(dir.Iops, (ins t2addrmode_imm7<memsz.shift>:$addr)),
  5355. asm, suffix, IndexModeNone, "$Qd, $addr", "">;
  5356. def _pre : MVE_VLDRSTR_cs<
  5357. dir, memsz, 1, 1,
  5358. !con((outs rGPR:$wb), dir.Oops),
  5359. !con(dir.Iops, (ins t2addrmode_imm7_pre<memsz.shift>:$addr)),
  5360. asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
  5361. let DecoderMethod = "DecodeMVE_MEM_2_pre<"#memsz.shift#">";
  5362. }
  5363. def _post : MVE_VLDRSTR_cs<
  5364. dir, memsz, 0, 1,
  5365. !con((outs rGPR:$wb), dir.Oops),
  5366. // We need an !if here to select the base register class,
  5367. // because it's legal to write back to SP in a load of this
  5368. // type, but not in a store.
  5369. !con(dir.Iops, (ins !if(dir.load, t2_addr_offset_none,
  5370. t2_nosp_addr_offset_none):$Rn,
  5371. t2am_imm7_offset<memsz.shift>:$addr)),
  5372. asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
  5373. bits<4> Rn;
  5374. let Inst{19-16} = Rn{3-0};
  5375. }
  5376. }
  5377. }
  5378. // Now actually declare all the contiguous load/stores, via those
  5379. // multiclasses. The instruction ids coming out of this are the bare
  5380. // names shown in the defm, with _pre or _post appended for writeback,
  5381. // e.g. MVE_VLDRBS16, MVE_VSTRB16_pre, MVE_VSTRHU16_post.
  5382. defm MVE_VLDRBS16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s16", 0, 0b01>;
  5383. defm MVE_VLDRBS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s32", 0, 0b10>;
  5384. defm MVE_VLDRBU16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u16", 1, 0b01>;
  5385. defm MVE_VLDRBU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u32", 1, 0b10>;
  5386. defm MVE_VLDRHS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "s32", 0, 0b10>;
  5387. defm MVE_VLDRHU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "u32", 1, 0b10>;
  5388. defm MVE_VLDRBU8: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memB, "vldrb", "u8">;
  5389. defm MVE_VLDRHU16: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memH, "vldrh", "u16">;
  5390. defm MVE_VLDRWU32: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memW, "vldrw", "u32">;
  5391. defm MVE_VSTRB16: MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "16", 0, 0b01>;
  5392. defm MVE_VSTRB32: MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "32", 0, 0b10>;
  5393. defm MVE_VSTRH32: MVE_VLDRSTR_cw_m<MVE_st, MVE_memH, "vstrh", "32", 0, 0b10>;
  5394. defm MVE_VSTRBU8 : MVE_VLDRSTR_cs_m<MVE_st, MVE_memB, "vstrb", "8">;
  5395. defm MVE_VSTRHU16: MVE_VLDRSTR_cs_m<MVE_st, MVE_memH, "vstrh", "16">;
  5396. defm MVE_VSTRWU32: MVE_VLDRSTR_cs_m<MVE_st, MVE_memW, "vstrw", "32">;
  5397. // Gather loads / scatter stores whose address operand is of the form
  5398. // [Rn,Qm], i.e. a single GPR as the common base address, plus a
  5399. // vector of offset from it. ('Load/store this sequence of elements of
  5400. // the same array.')
  5401. //
  5402. // Like the contiguous family, these loads and stores can widen the
  5403. // loaded values / truncate the stored ones, or they can just
  5404. // load/store the same size of memory and vector lane. But unlike the
  5405. // contiguous family, there's no particular difference in encoding
  5406. // between those two cases.
  5407. //
  5408. // This family also comes with the option to scale the offset values
  5409. // in Qm by the size of the loaded memory (i.e. to treat them as array
  5410. // indices), or not to scale them (to treat them as plain byte offsets
  5411. // in memory, so that perhaps the loaded values are unaligned). The
  5412. // scaled instructions' address operand in assembly looks like
  5413. // [Rn,Qm,UXTW #2] or similar.
  5414. // Base class.
  5415. class MVE_VLDRSTR_rq<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
  5416. bits<2> size, bit os, string asm, string suffix, int shift>
  5417. : MVE_VLDRSTR_base<dir, U, 0b0, 0b0, 0, dir.Oops,
  5418. !con(dir.Iops, (ins mve_addr_rq_shift<shift>:$addr)),
  5419. asm, suffix, "$Qd, $addr", dir.cstr> {
  5420. bits<7> addr;
  5421. let Inst{23} = 0b1;
  5422. let Inst{19-16} = addr{6-3};
  5423. let Inst{8-7} = size;
  5424. let Inst{6} = memsz.encoding{1};
  5425. let Inst{5} = 0;
  5426. let Inst{4} = memsz.encoding{0};
  5427. let Inst{3-1} = addr{2-0};
  5428. let Inst{0} = os;
  5429. }
  5430. // Multiclass that defines the scaled and unscaled versions of an
  5431. // instruction, when the memory size is wider than a byte. The scaled
  5432. // version gets the default name like MVE_VLDRBU16_rq; the unscaled /
  5433. // potentially unaligned version gets a "_u" suffix, e.g.
  5434. // MVE_VLDRBU16_rq_u.
  5435. multiclass MVE_VLDRSTR_rq_w<MVE_ldst_direction dir, MVE_memsz memsz,
  5436. string asm, string suffix, bit U, bits<2> size> {
  5437. def _u : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
  5438. def "" : MVE_VLDRSTR_rq<dir, memsz, U, size, 1, asm, suffix, memsz.shift>;
  5439. }
  5440. // Subclass of MVE_VLDRSTR_rq with the same API as that multiclass,
  5441. // for use when the memory size is one byte, so there's no 'scaled'
  5442. // version of the instruction at all. (This is encoded as if it were
  5443. // unscaled, but named in the default way with no _u suffix.)
  5444. class MVE_VLDRSTR_rq_b<MVE_ldst_direction dir, MVE_memsz memsz,
  5445. string asm, string suffix, bit U, bits<2> size>
  5446. : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
  5447. // Multiclasses wrapping that to add ISel patterns for intrinsics.
  5448. multiclass MVE_VLDR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
  5449. defm "": MVE_VLDRSTR_rq_w<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
  5450. VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
  5451. defvar Inst = !cast<Instruction>(NAME);
  5452. defvar InstU = !cast<Instruction>(NAME # "_u");
  5453. foreach VTI = VTIs in
  5454. foreach UnsignedFlag = !if(!eq(VTI.Size, memsz.encoding),
  5455. [0,1], [VTI.Unsigned]) in {
  5456. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag)),
  5457. (VTI.Vec (InstU GPR:$base, MQPR:$offsets))>;
  5458. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag)),
  5459. (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
  5460. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag, (VTI.Pred VCCR:$pred))),
  5461. (VTI.Vec (InstU GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
  5462. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag, (VTI.Pred VCCR:$pred))),
  5463. (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
  5464. }
  5465. }
  5466. multiclass MVE_VLDR_rq_b<list<MVEVectorVTInfo> VTIs> {
  5467. def "": MVE_VLDRSTR_rq_b<MVE_ld, MVE_memB, "vldrb",
  5468. VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
  5469. defvar Inst = !cast<Instruction>(NAME);
  5470. foreach VTI = VTIs in {
  5471. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned)),
  5472. (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
  5473. def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned, (VTI.Pred VCCR:$pred))),
  5474. (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
  5475. }
  5476. }
  5477. multiclass MVE_VSTR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
  5478. defm "": MVE_VLDRSTR_rq_w<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
  5479. VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
  5480. defvar Inst = !cast<Instruction>(NAME);
  5481. defvar InstU = !cast<Instruction>(NAME # "_u");
  5482. foreach VTI = VTIs in {
  5483. def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0),
  5484. (InstU MQPR:$data, GPR:$base, MQPR:$offsets)>;
  5485. def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift),
  5486. (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
  5487. def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0, (VTI.Pred VCCR:$pred)),
  5488. (InstU MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
  5489. def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift, (VTI.Pred VCCR:$pred)),
  5490. (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
  5491. }
  5492. }
  5493. multiclass MVE_VSTR_rq_b<list<MVEVectorVTInfo> VTIs> {
  5494. def "": MVE_VLDRSTR_rq_b<MVE_st, MVE_memB, "vstrb",
  5495. VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
  5496. defvar Inst = !cast<Instruction>(NAME);
  5497. foreach VTI = VTIs in {
  5498. def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0),
  5499. (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
  5500. def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0, (VTI.Pred VCCR:$pred)),
  5501. (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
  5502. }
  5503. }
  5504. // Actually define all the loads and stores in this family.
  5505. defm MVE_VLDRBU8_rq : MVE_VLDR_rq_b<[MVE_v16u8,MVE_v16s8]>;
  5506. defm MVE_VLDRBU16_rq: MVE_VLDR_rq_b<[MVE_v8u16]>;
  5507. defm MVE_VLDRBS16_rq: MVE_VLDR_rq_b<[MVE_v8s16]>;
  5508. defm MVE_VLDRBU32_rq: MVE_VLDR_rq_b<[MVE_v4u32]>;
  5509. defm MVE_VLDRBS32_rq: MVE_VLDR_rq_b<[MVE_v4s32]>;
  5510. defm MVE_VLDRHU16_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v8u16,MVE_v8s16,MVE_v8f16]>;
  5511. defm MVE_VLDRHU32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4u32]>;
  5512. defm MVE_VLDRHS32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4s32]>;
  5513. defm MVE_VLDRWU32_rq: MVE_VLDR_rq_w<MVE_memW, [MVE_v4u32,MVE_v4s32,MVE_v4f32]>;
  5514. defm MVE_VLDRDU64_rq: MVE_VLDR_rq_w<MVE_memD, [MVE_v2u64,MVE_v2s64]>;
  5515. defm MVE_VSTRB8_rq : MVE_VSTR_rq_b<[MVE_v16i8]>;
  5516. defm MVE_VSTRB16_rq : MVE_VSTR_rq_b<[MVE_v8i16]>;
  5517. defm MVE_VSTRB32_rq : MVE_VSTR_rq_b<[MVE_v4i32]>;
  5518. defm MVE_VSTRH16_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v8i16,MVE_v8f16]>;
  5519. defm MVE_VSTRH32_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v4i32]>;
  5520. defm MVE_VSTRW32_rq : MVE_VSTR_rq_w<MVE_memW, [MVE_v4i32,MVE_v4f32]>;
  5521. defm MVE_VSTRD64_rq : MVE_VSTR_rq_w<MVE_memD, [MVE_v2i64]>;
  5522. // Gather loads / scatter stores whose address operand is of the form
  5523. // [Qm,#imm], i.e. a vector containing a full base address for each
  5524. // loaded item, plus an immediate offset applied consistently to all
  5525. // of them. ('Load/store the same field from this vector of pointers
  5526. // to a structure type.')
  5527. //
  5528. // This family requires the vector lane size to be at least 32 bits
  5529. // (so there's room for an address in each lane at all). It has no
  5530. // widening/narrowing variants. But it does support preindex
  5531. // writeback, in which the address vector is updated to hold the
  5532. // addresses actually loaded from.
  5533. // Base class.
  5534. class MVE_VLDRSTR_qi<MVE_ldst_direction dir, MVE_memsz memsz, bit W, dag wbops,
  5535. string asm, string wbAsm, string suffix, string cstr = "">
  5536. : MVE_VLDRSTR_base<dir, 1, 1, W, 1, !con(wbops, dir.Oops),
  5537. !con(dir.Iops, (ins mve_addr_q_shift<memsz.shift>:$addr)),
  5538. asm, suffix, "$Qd, $addr" # wbAsm, cstr # dir.cstr> {
  5539. bits<11> addr;
  5540. let Inst{23} = addr{7};
  5541. let Inst{19-17} = addr{10-8};
  5542. let Inst{16} = 0;
  5543. let Inst{8} = memsz.encoding{0}; // enough to distinguish 32- from 64-bit
  5544. let Inst{7} = 0;
  5545. let Inst{6-0} = addr{6-0};
  5546. }
  5547. // Multiclass that generates the non-writeback and writeback variants.
  5548. multiclass MVE_VLDRSTR_qi_m<MVE_ldst_direction dir, MVE_memsz memsz,
  5549. string asm, string suffix> {
  5550. def "" : MVE_VLDRSTR_qi<dir, memsz, 0, (outs), asm, "", suffix>;
  5551. def _pre : MVE_VLDRSTR_qi<dir, memsz, 1, (outs MQPR:$wb), asm, "!", suffix,
  5552. "$addr.base = $wb"> {
  5553. let DecoderMethod="DecodeMVE_MEM_3_pre<"#memsz.shift#">";
  5554. }
  5555. }
  5556. // Multiclasses wrapping that one, adding selection patterns for the
  5557. // non-writeback loads and all the stores. (The writeback loads must
  5558. // deliver multiple output values, so they have to be selected by C++
  5559. // code.)
  5560. multiclass MVE_VLDR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
  5561. list<MVEVectorVTInfo> DVTIs> {
  5562. defm "" : MVE_VLDRSTR_qi_m<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
  5563. "u" # memsz.TypeBits>;
  5564. defvar Inst = !cast<Instruction>(NAME);
  5565. foreach DVTI = DVTIs in {
  5566. def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base
  5567. (AVTI.Vec MQPR:$addr), (i32 imm:$offset))),
  5568. (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset)))>;
  5569. def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base_predicated
  5570. (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (AVTI.Pred VCCR:$pred))),
  5571. (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset),
  5572. ARMVCCThen, VCCR:$pred))>;
  5573. }
  5574. }
  5575. multiclass MVE_VSTR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
  5576. list<MVEVectorVTInfo> DVTIs> {
  5577. defm "" : MVE_VLDRSTR_qi_m<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
  5578. !cast<string>(memsz.TypeBits)>;
  5579. defvar Inst = !cast<Instruction>(NAME);
  5580. defvar InstPre = !cast<Instruction>(NAME # "_pre");
  5581. foreach DVTI = DVTIs in {
  5582. def : Pat<(int_arm_mve_vstr_scatter_base
  5583. (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data)),
  5584. (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
  5585. (i32 imm:$offset))>;
  5586. def : Pat<(int_arm_mve_vstr_scatter_base_predicated
  5587. (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred)),
  5588. (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
  5589. (i32 imm:$offset), ARMVCCThen, VCCR:$pred)>;
  5590. def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb
  5591. (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data))),
  5592. (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
  5593. (i32 imm:$offset)))>;
  5594. def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb_predicated
  5595. (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred))),
  5596. (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
  5597. (i32 imm:$offset), ARMVCCThen, VCCR:$pred))>;
  5598. }
  5599. }
  5600. // Actual instruction definitions.
  5601. defm MVE_VLDRWU32_qi: MVE_VLDR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
  5602. defm MVE_VLDRDU64_qi: MVE_VLDR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
  5603. defm MVE_VSTRW32_qi: MVE_VSTR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
  5604. defm MVE_VSTRD64_qi: MVE_VSTR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
  5605. // Define aliases for all the instructions where memory size and
  5606. // vector lane size are the same. These are mnemonic aliases, so they
  5607. // apply consistently across all of the above families - contiguous
  5608. // loads, and both the rq and qi types of gather/scatter.
  5609. //
  5610. // Rationale: As long as you're loading (for example) 16-bit memory
  5611. // values into 16-bit vector lanes, you can think of them as signed or
  5612. // unsigned integers, fp16 or just raw 16-bit blobs and it makes no
  5613. // difference. So we permit all of vldrh.16, vldrh.u16, vldrh.s16,
  5614. // vldrh.f16 and treat them all as equivalent to the canonical
  5615. // spelling (which happens to be .u16 for loads, and just .16 for
  5616. // stores).
  5617. foreach vpt_cond = ["", "t", "e"] in
  5618. foreach memsz = [MVE_memB, MVE_memH, MVE_memW, MVE_memD] in
  5619. foreach suffix = memsz.suffixes in {
  5620. // Define an alias with every suffix in the list, except for the one
  5621. // used by the real Instruction record (i.e. the one that all the
  5622. // rest are aliases *for*).
  5623. if !ne(suffix, memsz.CanonLoadSuffix) then {
  5624. def : MnemonicAlias<
  5625. "vldr" # memsz.MnemonicLetter # vpt_cond # suffix,
  5626. "vldr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonLoadSuffix>;
  5627. }
  5628. if !ne(suffix, memsz.CanonStoreSuffix) then {
  5629. def : MnemonicAlias<
  5630. "vstr" # memsz.MnemonicLetter # vpt_cond # suffix,
  5631. "vstr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonStoreSuffix>;
  5632. }
  5633. }
  5634. // end of MVE predicable load/store
  5635. class MVE_VPT<string suffix, bits<2> size, dag iops, string asm, list<dag> pattern=[]>
  5636. : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm, "", pattern> {
  5637. bits<3> fc;
  5638. bits<4> Mk;
  5639. bits<3> Qn;
  5640. let Inst{31-23} = 0b111111100;
  5641. let Inst{22} = Mk{3};
  5642. let Inst{21-20} = size;
  5643. let Inst{19-17} = Qn{2-0};
  5644. let Inst{16} = 0b1;
  5645. let Inst{15-13} = Mk{2-0};
  5646. let Inst{12} = fc{2};
  5647. let Inst{11-8} = 0b1111;
  5648. let Inst{7} = fc{0};
  5649. let Inst{4} = 0b0;
  5650. let Defs = [VPR];
  5651. let validForTailPredication=1;
  5652. }
  5653. class MVE_VPTt1<string suffix, bits<2> size, dag iops>
  5654. : MVE_VPT<suffix, size, iops, "$fc, $Qn, $Qm"> {
  5655. bits<4> Qm;
  5656. bits<4> Mk;
  5657. let Inst{6} = 0b0;
  5658. let Inst{5} = Qm{3};
  5659. let Inst{3-1} = Qm{2-0};
  5660. let Inst{0} = fc{1};
  5661. }
  5662. class MVE_VPTt1i<string suffix, bits<2> size>
  5663. : MVE_VPTt1<suffix, size,
  5664. (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_i:$fc)> {
  5665. let Inst{12} = 0b0;
  5666. let Inst{0} = 0b0;
  5667. }
  5668. def MVE_VPTv4i32 : MVE_VPTt1i<"i32", 0b10>;
  5669. def MVE_VPTv8i16 : MVE_VPTt1i<"i16", 0b01>;
  5670. def MVE_VPTv16i8 : MVE_VPTt1i<"i8", 0b00>;
  5671. class MVE_VPTt1u<string suffix, bits<2> size>
  5672. : MVE_VPTt1<suffix, size,
  5673. (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_u:$fc)> {
  5674. let Inst{12} = 0b0;
  5675. let Inst{0} = 0b1;
  5676. }
  5677. def MVE_VPTv4u32 : MVE_VPTt1u<"u32", 0b10>;
  5678. def MVE_VPTv8u16 : MVE_VPTt1u<"u16", 0b01>;
  5679. def MVE_VPTv16u8 : MVE_VPTt1u<"u8", 0b00>;
  5680. class MVE_VPTt1s<string suffix, bits<2> size>
  5681. : MVE_VPTt1<suffix, size,
  5682. (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_s:$fc)> {
  5683. let Inst{12} = 0b1;
  5684. }
  5685. def MVE_VPTv4s32 : MVE_VPTt1s<"s32", 0b10>;
  5686. def MVE_VPTv8s16 : MVE_VPTt1s<"s16", 0b01>;
  5687. def MVE_VPTv16s8 : MVE_VPTt1s<"s8", 0b00>;
  5688. class MVE_VPTt2<string suffix, bits<2> size, dag iops>
  5689. : MVE_VPT<suffix, size, iops,
  5690. "$fc, $Qn, $Rm"> {
  5691. bits<4> Rm;
  5692. bits<3> fc;
  5693. bits<4> Mk;
  5694. let Inst{6} = 0b1;
  5695. let Inst{5} = fc{1};
  5696. let Inst{3-0} = Rm{3-0};
  5697. }
  5698. class MVE_VPTt2i<string suffix, bits<2> size>
  5699. : MVE_VPTt2<suffix, size,
  5700. (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_i:$fc)> {
  5701. let Inst{12} = 0b0;
  5702. let Inst{5} = 0b0;
  5703. }
  5704. def MVE_VPTv4i32r : MVE_VPTt2i<"i32", 0b10>;
  5705. def MVE_VPTv8i16r : MVE_VPTt2i<"i16", 0b01>;
  5706. def MVE_VPTv16i8r : MVE_VPTt2i<"i8", 0b00>;
  5707. class MVE_VPTt2u<string suffix, bits<2> size>
  5708. : MVE_VPTt2<suffix, size,
  5709. (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_u:$fc)> {
  5710. let Inst{12} = 0b0;
  5711. let Inst{5} = 0b1;
  5712. }
  5713. def MVE_VPTv4u32r : MVE_VPTt2u<"u32", 0b10>;
  5714. def MVE_VPTv8u16r : MVE_VPTt2u<"u16", 0b01>;
  5715. def MVE_VPTv16u8r : MVE_VPTt2u<"u8", 0b00>;
  5716. class MVE_VPTt2s<string suffix, bits<2> size>
  5717. : MVE_VPTt2<suffix, size,
  5718. (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_s:$fc)> {
  5719. let Inst{12} = 0b1;
  5720. }
  5721. def MVE_VPTv4s32r : MVE_VPTt2s<"s32", 0b10>;
  5722. def MVE_VPTv8s16r : MVE_VPTt2s<"s16", 0b01>;
  5723. def MVE_VPTv16s8r : MVE_VPTt2s<"s8", 0b00>;
  5724. class MVE_VPTf<string suffix, bit size, dag iops, string asm, list<dag> pattern=[]>
  5725. : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm,
  5726. "", pattern> {
  5727. bits<3> fc;
  5728. bits<4> Mk;
  5729. bits<3> Qn;
  5730. let Inst{31-29} = 0b111;
  5731. let Inst{28} = size;
  5732. let Inst{27-23} = 0b11100;
  5733. let Inst{22} = Mk{3};
  5734. let Inst{21-20} = 0b11;
  5735. let Inst{19-17} = Qn{2-0};
  5736. let Inst{16} = 0b1;
  5737. let Inst{15-13} = Mk{2-0};
  5738. let Inst{12} = fc{2};
  5739. let Inst{11-8} = 0b1111;
  5740. let Inst{7} = fc{0};
  5741. let Inst{4} = 0b0;
  5742. let Defs = [VPR];
  5743. let Predicates = [HasMVEFloat];
  5744. let validForTailPredication=1;
  5745. }
  5746. class MVE_VPTft1<string suffix, bit size>
  5747. : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_fp:$fc),
  5748. "$fc, $Qn, $Qm"> {
  5749. bits<3> fc;
  5750. bits<4> Qm;
  5751. let Inst{6} = 0b0;
  5752. let Inst{5} = Qm{3};
  5753. let Inst{3-1} = Qm{2-0};
  5754. let Inst{0} = fc{1};
  5755. }
  5756. def MVE_VPTv4f32 : MVE_VPTft1<"f32", 0b0>;
  5757. def MVE_VPTv8f16 : MVE_VPTft1<"f16", 0b1>;
  5758. class MVE_VPTft2<string suffix, bit size>
  5759. : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_fp:$fc),
  5760. "$fc, $Qn, $Rm"> {
  5761. bits<3> fc;
  5762. bits<4> Rm;
  5763. let Inst{6} = 0b1;
  5764. let Inst{5} = fc{1};
  5765. let Inst{3-0} = Rm{3-0};
  5766. }
  5767. def MVE_VPTv4f32r : MVE_VPTft2<"f32", 0b0>;
  5768. def MVE_VPTv8f16r : MVE_VPTft2<"f16", 0b1>;
  5769. def MVE_VPST : MVE_MI<(outs ), (ins vpt_mask:$Mk), NoItinerary,
  5770. !strconcat("vpst", "${Mk}"), "", "", []> {
  5771. bits<4> Mk;
  5772. let Inst{31-23} = 0b111111100;
  5773. let Inst{22} = Mk{3};
  5774. let Inst{21-16} = 0b110001;
  5775. let Inst{15-13} = Mk{2-0};
  5776. let Inst{12-0} = 0b0111101001101;
  5777. let Unpredictable{12} = 0b1;
  5778. let Unpredictable{7} = 0b1;
  5779. let Unpredictable{5} = 0b1;
  5780. let Uses = [VPR];
  5781. let validForTailPredication = 1;
  5782. }
  5783. def MVE_VPSEL : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
  5784. "vpsel", "", "$Qd, $Qn, $Qm", vpred_n, "", []> {
  5785. bits<4> Qn;
  5786. bits<4> Qd;
  5787. bits<4> Qm;
  5788. let Inst{28} = 0b1;
  5789. let Inst{25-23} = 0b100;
  5790. let Inst{22} = Qd{3};
  5791. let Inst{21-20} = 0b11;
  5792. let Inst{19-17} = Qn{2-0};
  5793. let Inst{16} = 0b1;
  5794. let Inst{15-13} = Qd{2-0};
  5795. let Inst{12-9} = 0b0111;
  5796. let Inst{8} = 0b1;
  5797. let Inst{7} = Qn{3};
  5798. let Inst{6} = 0b0;
  5799. let Inst{5} = Qm{3};
  5800. let Inst{4} = 0b0;
  5801. let Inst{3-1} = Qm{2-0};
  5802. let Inst{0} = 0b1;
  5803. }
  5804. foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32",
  5805. "i8", "i16", "i32", "f16", "f32"] in
  5806. def : MVEInstAlias<"vpsel${vp}." # suffix # "\t$Qd, $Qn, $Qm",
  5807. (MVE_VPSEL MQPR:$Qd, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
  5808. let Predicates = [HasMVEInt] in {
  5809. def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
  5810. (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
  5811. def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
  5812. (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
  5813. def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
  5814. (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
  5815. def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
  5816. (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
  5817. def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
  5818. (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
  5819. def : Pat<(v16i8 (vselect (v16i8 MQPR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
  5820. (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
  5821. (MVE_VCMPi8 (v16i8 MQPR:$pred), (MVE_VMOVimmi8 0), ARMCCne)))>;
  5822. def : Pat<(v8i16 (vselect (v8i16 MQPR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
  5823. (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
  5824. (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne)))>;
  5825. def : Pat<(v4i32 (vselect (v4i32 MQPR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
  5826. (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
  5827. (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne)))>;
  5828. def : Pat<(v8f16 (vselect (v8i16 MQPR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
  5829. (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
  5830. (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne)))>;
  5831. def : Pat<(v4f32 (vselect (v4i32 MQPR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
  5832. (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
  5833. (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne)))>;
  5834. // Pred <-> Int
  5835. def : Pat<(v16i8 (zext (v16i1 VCCR:$pred))),
  5836. (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
  5837. def : Pat<(v8i16 (zext (v8i1 VCCR:$pred))),
  5838. (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
  5839. def : Pat<(v4i32 (zext (v4i1 VCCR:$pred))),
  5840. (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
  5841. def : Pat<(v16i8 (sext (v16i1 VCCR:$pred))),
  5842. (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
  5843. def : Pat<(v8i16 (sext (v8i1 VCCR:$pred))),
  5844. (v8i16 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
  5845. def : Pat<(v4i32 (sext (v4i1 VCCR:$pred))),
  5846. (v4i32 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
  5847. def : Pat<(v16i8 (anyext (v16i1 VCCR:$pred))),
  5848. (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
  5849. def : Pat<(v8i16 (anyext (v8i1 VCCR:$pred))),
  5850. (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
  5851. def : Pat<(v4i32 (anyext (v4i1 VCCR:$pred))),
  5852. (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
  5853. }
  5854. let Predicates = [HasMVEFloat] in {
  5855. // Pred <-> Float
  5856. // 112 is 1.0 in float
  5857. def : Pat<(v4f32 (uint_to_fp (v4i1 VCCR:$pred))),
  5858. (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 112)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred))>;
  5859. // 2620 in 1.0 in half
  5860. def : Pat<(v8f16 (uint_to_fp (v8i1 VCCR:$pred))),
  5861. (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2620)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred))>;
  5862. // 240 is -1.0 in float
  5863. def : Pat<(v4f32 (sint_to_fp (v4i1 VCCR:$pred))),
  5864. (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 240)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred))>;
  5865. // 2748 is -1.0 in half
  5866. def : Pat<(v8f16 (sint_to_fp (v8i1 VCCR:$pred))),
  5867. (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2748)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred))>;
  5868. def : Pat<(v4i1 (fp_to_uint (v4f32 MQPR:$v1))),
  5869. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
  5870. def : Pat<(v8i1 (fp_to_uint (v8f16 MQPR:$v1))),
  5871. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
  5872. def : Pat<(v4i1 (fp_to_sint (v4f32 MQPR:$v1))),
  5873. (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
  5874. def : Pat<(v8i1 (fp_to_sint (v8f16 MQPR:$v1))),
  5875. (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
  5876. }
  5877. def MVE_VPNOT : MVE_p<(outs VCCR:$P0), (ins VCCR:$P0_in), NoItinerary,
  5878. "vpnot", "", "", vpred_n, "", []> {
  5879. let Inst{31-0} = 0b11111110001100010000111101001101;
  5880. let Unpredictable{19-17} = 0b111;
  5881. let Unpredictable{12} = 0b1;
  5882. let Unpredictable{7} = 0b1;
  5883. let Unpredictable{5} = 0b1;
  5884. let Constraints = "";
  5885. let DecoderMethod = "DecodeMVEVPNOT";
  5886. }
  5887. let Predicates = [HasMVEInt] in {
  5888. def : Pat<(v4i1 (xor (v4i1 VCCR:$pred), (v4i1 (predicate_cast (i32 65535))))),
  5889. (v4i1 (MVE_VPNOT (v4i1 VCCR:$pred)))>;
  5890. def : Pat<(v8i1 (xor (v8i1 VCCR:$pred), (v8i1 (predicate_cast (i32 65535))))),
  5891. (v8i1 (MVE_VPNOT (v8i1 VCCR:$pred)))>;
  5892. def : Pat<(v16i1 (xor (v16i1 VCCR:$pred), (v16i1 (predicate_cast (i32 65535))))),
  5893. (v16i1 (MVE_VPNOT (v16i1 VCCR:$pred)))>;
  5894. }
  5895. class MVE_loltp_start<dag iops, string asm, string ops, bits<2> size>
  5896. : t2LOL<(outs GPRlr:$LR), iops, asm, ops> {
  5897. bits<4> Rn;
  5898. let Predicates = [HasMVEInt];
  5899. let Inst{22} = 0b0;
  5900. let Inst{21-20} = size;
  5901. let Inst{19-16} = Rn{3-0};
  5902. let Inst{12} = 0b0;
  5903. }
  5904. class MVE_DLSTP<string asm, bits<2> size>
  5905. : MVE_loltp_start<(ins rGPR:$Rn), asm, "$LR, $Rn", size> {
  5906. let Inst{13} = 0b1;
  5907. let Inst{11-1} = 0b00000000000;
  5908. let Unpredictable{10-1} = 0b1111111111;
  5909. }
  5910. class MVE_WLSTP<string asm, bits<2> size>
  5911. : MVE_loltp_start<(ins rGPR:$Rn, wlslabel_u11:$label),
  5912. asm, "$LR, $Rn, $label", size> {
  5913. bits<11> label;
  5914. let Inst{13} = 0b0;
  5915. let Inst{11} = label{0};
  5916. let Inst{10-1} = label{10-1};
  5917. let isBranch = 1;
  5918. let isTerminator = 1;
  5919. }
  5920. def MVE_DLSTP_8 : MVE_DLSTP<"dlstp.8", 0b00>;
  5921. def MVE_DLSTP_16 : MVE_DLSTP<"dlstp.16", 0b01>;
  5922. def MVE_DLSTP_32 : MVE_DLSTP<"dlstp.32", 0b10>;
  5923. def MVE_DLSTP_64 : MVE_DLSTP<"dlstp.64", 0b11>;
  5924. def MVE_WLSTP_8 : MVE_WLSTP<"wlstp.8", 0b00>;
  5925. def MVE_WLSTP_16 : MVE_WLSTP<"wlstp.16", 0b01>;
  5926. def MVE_WLSTP_32 : MVE_WLSTP<"wlstp.32", 0b10>;
  5927. def MVE_WLSTP_64 : MVE_WLSTP<"wlstp.64", 0b11>;
  5928. class MVE_loltp_end<dag oops, dag iops, string asm, string ops>
  5929. : t2LOL<oops, iops, asm, ops> {
  5930. let Predicates = [HasMVEInt];
  5931. let Inst{22-21} = 0b00;
  5932. let Inst{19-16} = 0b1111;
  5933. let Inst{12} = 0b0;
  5934. }
  5935. def MVE_LETP : MVE_loltp_end<(outs GPRlr:$LRout),
  5936. (ins GPRlr:$LRin, lelabel_u11:$label),
  5937. "letp", "$LRin, $label"> {
  5938. bits<11> label;
  5939. let Inst{20} = 0b1;
  5940. let Inst{13} = 0b0;
  5941. let Inst{11} = label{0};
  5942. let Inst{10-1} = label{10-1};
  5943. let isBranch = 1;
  5944. let isTerminator = 1;
  5945. }
  5946. def MVE_LCTP : MVE_loltp_end<(outs), (ins pred:$p), "lctp${p}", ""> {
  5947. let Inst{20} = 0b0;
  5948. let Inst{13} = 0b1;
  5949. let Inst{11-1} = 0b00000000000;
  5950. let Unpredictable{21-20} = 0b11;
  5951. let Unpredictable{11-1} = 0b11111111111;
  5952. }
  5953. //===----------------------------------------------------------------------===//
  5954. // Patterns
  5955. //===----------------------------------------------------------------------===//
  5956. // PatFrags for loads and stores. Often trying to keep semi-consistent names.
  5957. def aligned32_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
  5958. (pre_store node:$val, node:$ptr, node:$offset), [{
  5959. return cast<StoreSDNode>(N)->getAlignment() >= 4;
  5960. }]>;
  5961. def aligned32_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
  5962. (post_store node:$val, node:$ptr, node:$offset), [{
  5963. return cast<StoreSDNode>(N)->getAlignment() >= 4;
  5964. }]>;
  5965. def aligned16_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
  5966. (pre_store node:$val, node:$ptr, node:$offset), [{
  5967. return cast<StoreSDNode>(N)->getAlignment() >= 2;
  5968. }]>;
  5969. def aligned16_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
  5970. (post_store node:$val, node:$ptr, node:$offset), [{
  5971. return cast<StoreSDNode>(N)->getAlignment() >= 2;
  5972. }]>;
  5973. def aligned_maskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5974. (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
  5975. auto *Ld = cast<MaskedLoadSDNode>(N);
  5976. return Ld->getMemoryVT().getScalarType() == MVT::i8;
  5977. }]>;
  5978. def aligned_sextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5979. (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
  5980. return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
  5981. }]>;
  5982. def aligned_zextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5983. (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
  5984. return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
  5985. }]>;
  5986. def aligned_extmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5987. (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
  5988. auto *Ld = cast<MaskedLoadSDNode>(N);
  5989. EVT ScalarVT = Ld->getMemoryVT().getScalarType();
  5990. return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
  5991. }]>;
  5992. def aligned_maskedloadvi16: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5993. (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
  5994. auto *Ld = cast<MaskedLoadSDNode>(N);
  5995. EVT ScalarVT = Ld->getMemoryVT().getScalarType();
  5996. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && Ld->getAlignment() >= 2;
  5997. }]>;
  5998. def aligned_sextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  5999. (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
  6000. return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
  6001. }]>;
  6002. def aligned_zextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  6003. (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
  6004. return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
  6005. }]>;
  6006. def aligned_extmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  6007. (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
  6008. auto *Ld = cast<MaskedLoadSDNode>(N);
  6009. EVT ScalarVT = Ld->getMemoryVT().getScalarType();
  6010. return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
  6011. }]>;
  6012. def aligned_maskedloadvi32: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
  6013. (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
  6014. auto *Ld = cast<MaskedLoadSDNode>(N);
  6015. EVT ScalarVT = Ld->getMemoryVT().getScalarType();
  6016. return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && Ld->getAlignment() >= 4;
  6017. }]>;
  6018. def aligned_maskedstvi8 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
  6019. (masked_st node:$val, node:$ptr, undef, node:$pred), [{
  6020. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6021. }]>;
  6022. def aligned_maskedstvi16 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
  6023. (masked_st node:$val, node:$ptr, undef, node:$pred), [{
  6024. auto *St = cast<MaskedStoreSDNode>(N);
  6025. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6026. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6027. }]>;
  6028. def aligned_maskedstvi32 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
  6029. (masked_st node:$val, node:$ptr, undef, node:$pred), [{
  6030. auto *St = cast<MaskedStoreSDNode>(N);
  6031. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6032. return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
  6033. }]>;
  6034. def pre_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
  6035. (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
  6036. ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
  6037. return AM == ISD::PRE_INC || AM == ISD::PRE_DEC;
  6038. }]>;
  6039. def post_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
  6040. (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
  6041. ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
  6042. return AM == ISD::POST_INC || AM == ISD::POST_DEC;
  6043. }]>;
  6044. def aligned_pre_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6045. (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6046. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6047. }]>;
  6048. def aligned_post_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6049. (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6050. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6051. }]>;
  6052. def aligned_pre_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6053. (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6054. auto *St = cast<MaskedStoreSDNode>(N);
  6055. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6056. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6057. }]>;
  6058. def aligned_post_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6059. (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6060. auto *St = cast<MaskedStoreSDNode>(N);
  6061. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6062. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6063. }]>;
  6064. def aligned_pre_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6065. (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6066. auto *St = cast<MaskedStoreSDNode>(N);
  6067. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6068. return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
  6069. }]>;
  6070. def aligned_post_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
  6071. (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
  6072. auto *St = cast<MaskedStoreSDNode>(N);
  6073. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6074. return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
  6075. }]>;
  6076. // PatFrags for "Aligned" extending / truncating
  6077. def aligned_extloadvi8 : PatFrag<(ops node:$ptr), (extloadvi8 node:$ptr)>;
  6078. def aligned_sextloadvi8 : PatFrag<(ops node:$ptr), (sextloadvi8 node:$ptr)>;
  6079. def aligned_zextloadvi8 : PatFrag<(ops node:$ptr), (zextloadvi8 node:$ptr)>;
  6080. def aligned_truncstvi8 : PatFrag<(ops node:$val, node:$ptr),
  6081. (truncstorevi8 node:$val, node:$ptr)>;
  6082. def aligned_post_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
  6083. (post_truncstvi8 node:$val, node:$base, node:$offset)>;
  6084. def aligned_pre_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
  6085. (pre_truncstvi8 node:$val, node:$base, node:$offset)>;
  6086. let MinAlignment = 2 in {
  6087. def aligned_extloadvi16 : PatFrag<(ops node:$ptr), (extloadvi16 node:$ptr)>;
  6088. def aligned_sextloadvi16 : PatFrag<(ops node:$ptr), (sextloadvi16 node:$ptr)>;
  6089. def aligned_zextloadvi16 : PatFrag<(ops node:$ptr), (zextloadvi16 node:$ptr)>;
  6090. def aligned_truncstvi16 : PatFrag<(ops node:$val, node:$ptr),
  6091. (truncstorevi16 node:$val, node:$ptr)>;
  6092. def aligned_post_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
  6093. (post_truncstvi16 node:$val, node:$base, node:$offset)>;
  6094. def aligned_pre_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
  6095. (pre_truncstvi16 node:$val, node:$base, node:$offset)>;
  6096. }
  6097. def truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$pred),
  6098. (masked_st node:$val, node:$base, undef, node:$pred), [{
  6099. return cast<MaskedStoreSDNode>(N)->isTruncatingStore();
  6100. }]>;
  6101. def aligned_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$pred),
  6102. (truncmaskedst node:$val, node:$base, node:$pred), [{
  6103. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6104. }]>;
  6105. def aligned_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$pred),
  6106. (truncmaskedst node:$val, node:$base, node:$pred), [{
  6107. auto *St = cast<MaskedStoreSDNode>(N);
  6108. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6109. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6110. }]>;
  6111. def pre_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
  6112. (masked_st node:$val, node:$base, node:$offset, node:$pred), [{
  6113. ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
  6114. return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::PRE_INC || AM == ISD::PRE_DEC);
  6115. }]>;
  6116. def aligned_pre_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
  6117. (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
  6118. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6119. }]>;
  6120. def aligned_pre_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
  6121. (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
  6122. auto *St = cast<MaskedStoreSDNode>(N);
  6123. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6124. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6125. }]>;
  6126. def post_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
  6127. (masked_st node:$val, node:$base, node:$offset, node:$postd), [{
  6128. ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
  6129. return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::POST_INC || AM == ISD::POST_DEC);
  6130. }]>;
  6131. def aligned_post_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
  6132. (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
  6133. return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
  6134. }]>;
  6135. def aligned_post_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
  6136. (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
  6137. auto *St = cast<MaskedStoreSDNode>(N);
  6138. EVT ScalarVT = St->getMemoryVT().getScalarType();
  6139. return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
  6140. }]>;
  6141. // Load/store patterns
  6142. class MVE_vector_store_typed<ValueType Ty, Instruction RegImmInst,
  6143. PatFrag StoreKind, int shift>
  6144. : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr),
  6145. (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr)>;
  6146. class MVE_vector_maskedstore_typed<ValueType Ty, Instruction RegImmInst,
  6147. PatFrag StoreKind, int shift>
  6148. : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, VCCR:$pred),
  6149. (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred)>;
  6150. multiclass MVE_vector_store<Instruction RegImmInst, PatFrag StoreKind,
  6151. int shift> {
  6152. def : MVE_vector_store_typed<v16i8, RegImmInst, StoreKind, shift>;
  6153. def : MVE_vector_store_typed<v8i16, RegImmInst, StoreKind, shift>;
  6154. def : MVE_vector_store_typed<v8f16, RegImmInst, StoreKind, shift>;
  6155. def : MVE_vector_store_typed<v4i32, RegImmInst, StoreKind, shift>;
  6156. def : MVE_vector_store_typed<v4f32, RegImmInst, StoreKind, shift>;
  6157. def : MVE_vector_store_typed<v2i64, RegImmInst, StoreKind, shift>;
  6158. def : MVE_vector_store_typed<v2f64, RegImmInst, StoreKind, shift>;
  6159. }
  6160. class MVE_vector_load_typed<ValueType Ty, Instruction RegImmInst,
  6161. PatFrag LoadKind, int shift>
  6162. : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr)),
  6163. (Ty (RegImmInst t2addrmode_imm7<shift>:$addr))>;
  6164. class MVE_vector_maskedload_typed<ValueType Ty, Instruction RegImmInst,
  6165. PatFrag LoadKind, int shift>
  6166. : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr, VCCR:$pred, (Ty (ARMvmovImm (i32 0))))),
  6167. (Ty (RegImmInst t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred))>;
  6168. multiclass MVE_vector_load<Instruction RegImmInst, PatFrag LoadKind,
  6169. int shift> {
  6170. def : MVE_vector_load_typed<v16i8, RegImmInst, LoadKind, shift>;
  6171. def : MVE_vector_load_typed<v8i16, RegImmInst, LoadKind, shift>;
  6172. def : MVE_vector_load_typed<v8f16, RegImmInst, LoadKind, shift>;
  6173. def : MVE_vector_load_typed<v4i32, RegImmInst, LoadKind, shift>;
  6174. def : MVE_vector_load_typed<v4f32, RegImmInst, LoadKind, shift>;
  6175. def : MVE_vector_load_typed<v2i64, RegImmInst, LoadKind, shift>;
  6176. def : MVE_vector_load_typed<v2f64, RegImmInst, LoadKind, shift>;
  6177. }
  6178. class MVE_vector_offset_store_typed<ValueType Ty, Instruction Opcode,
  6179. PatFrag StoreKind, int shift>
  6180. : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr),
  6181. (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr)>;
  6182. class MVE_vector_offset_maskedstore_typed<ValueType Ty, Instruction Opcode,
  6183. PatFrag StoreKind, int shift>
  6184. : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr, VCCR:$pred),
  6185. (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr, ARMVCCThen, VCCR:$pred)>;
  6186. multiclass MVE_vector_offset_store<Instruction RegImmInst, PatFrag StoreKind,
  6187. int shift> {
  6188. def : MVE_vector_offset_store_typed<v16i8, RegImmInst, StoreKind, shift>;
  6189. def : MVE_vector_offset_store_typed<v8i16, RegImmInst, StoreKind, shift>;
  6190. def : MVE_vector_offset_store_typed<v8f16, RegImmInst, StoreKind, shift>;
  6191. def : MVE_vector_offset_store_typed<v4i32, RegImmInst, StoreKind, shift>;
  6192. def : MVE_vector_offset_store_typed<v4f32, RegImmInst, StoreKind, shift>;
  6193. def : MVE_vector_offset_store_typed<v2i64, RegImmInst, StoreKind, shift>;
  6194. def : MVE_vector_offset_store_typed<v2f64, RegImmInst, StoreKind, shift>;
  6195. }
  6196. let Predicates = [HasMVEInt, IsLE] in {
  6197. // Stores
  6198. defm : MVE_vector_store<MVE_VSTRBU8, byte_alignedstore, 0>;
  6199. defm : MVE_vector_store<MVE_VSTRHU16, hword_alignedstore, 1>;
  6200. defm : MVE_vector_store<MVE_VSTRWU32, alignedstore32, 2>;
  6201. // Loads
  6202. defm : MVE_vector_load<MVE_VLDRBU8, byte_alignedload, 0>;
  6203. defm : MVE_vector_load<MVE_VLDRHU16, hword_alignedload, 1>;
  6204. defm : MVE_vector_load<MVE_VLDRWU32, alignedload32, 2>;
  6205. // Pre/post inc stores
  6206. defm : MVE_vector_offset_store<MVE_VSTRBU8_pre, pre_store, 0>;
  6207. defm : MVE_vector_offset_store<MVE_VSTRBU8_post, post_store, 0>;
  6208. defm : MVE_vector_offset_store<MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
  6209. defm : MVE_vector_offset_store<MVE_VSTRHU16_post, aligned16_post_store, 1>;
  6210. defm : MVE_vector_offset_store<MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
  6211. defm : MVE_vector_offset_store<MVE_VSTRWU32_post, aligned32_post_store, 2>;
  6212. }
  6213. let Predicates = [HasMVEInt, IsBE] in {
  6214. // Aligned Stores
  6215. def : MVE_vector_store_typed<v16i8, MVE_VSTRBU8, store, 0>;
  6216. def : MVE_vector_store_typed<v8i16, MVE_VSTRHU16, alignedstore16, 1>;
  6217. def : MVE_vector_store_typed<v8f16, MVE_VSTRHU16, alignedstore16, 1>;
  6218. def : MVE_vector_store_typed<v4i32, MVE_VSTRWU32, alignedstore32, 2>;
  6219. def : MVE_vector_store_typed<v4f32, MVE_VSTRWU32, alignedstore32, 2>;
  6220. // Aligned Loads
  6221. def : MVE_vector_load_typed<v16i8, MVE_VLDRBU8, load, 0>;
  6222. def : MVE_vector_load_typed<v8i16, MVE_VLDRHU16, alignedload16, 1>;
  6223. def : MVE_vector_load_typed<v8f16, MVE_VLDRHU16, alignedload16, 1>;
  6224. def : MVE_vector_load_typed<v4i32, MVE_VLDRWU32, alignedload32, 2>;
  6225. def : MVE_vector_load_typed<v4f32, MVE_VLDRWU32, alignedload32, 2>;
  6226. // Other unaligned loads/stores need to go though a VREV
  6227. def : Pat<(v2f64 (load t2addrmode_imm7<0>:$addr)),
  6228. (v2f64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6229. def : Pat<(v2i64 (load t2addrmode_imm7<0>:$addr)),
  6230. (v2i64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6231. def : Pat<(v4i32 (load t2addrmode_imm7<0>:$addr)),
  6232. (v4i32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6233. def : Pat<(v4f32 (load t2addrmode_imm7<0>:$addr)),
  6234. (v4f32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6235. def : Pat<(v8i16 (load t2addrmode_imm7<0>:$addr)),
  6236. (v8i16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6237. def : Pat<(v8f16 (load t2addrmode_imm7<0>:$addr)),
  6238. (v8f16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
  6239. def : Pat<(store (v2f64 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6240. (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6241. def : Pat<(store (v2i64 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6242. (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6243. def : Pat<(store (v4i32 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6244. (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6245. def : Pat<(store (v4f32 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6246. (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6247. def : Pat<(store (v8i16 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6248. (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6249. def : Pat<(store (v8f16 MQPR:$val), t2addrmode_imm7<0>:$addr),
  6250. (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
  6251. // Pre/Post inc stores
  6252. def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_pre, pre_store, 0>;
  6253. def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_post, post_store, 0>;
  6254. def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
  6255. def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
  6256. def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
  6257. def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
  6258. def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
  6259. def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
  6260. def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
  6261. def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
  6262. }
  6263. let Predicates = [HasMVEInt] in {
  6264. // Aligned masked store, shared between LE and BE
  6265. def : MVE_vector_maskedstore_typed<v16i8, MVE_VSTRBU8, aligned_maskedstvi8, 0>;
  6266. def : MVE_vector_maskedstore_typed<v8i16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
  6267. def : MVE_vector_maskedstore_typed<v8f16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
  6268. def : MVE_vector_maskedstore_typed<v4i32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
  6269. def : MVE_vector_maskedstore_typed<v4f32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
  6270. // Pre/Post inc masked stores
  6271. def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_pre, aligned_pre_maskedstorevi8, 0>;
  6272. def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_post, aligned_post_maskedstorevi8, 0>;
  6273. def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
  6274. def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
  6275. def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
  6276. def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
  6277. def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
  6278. def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
  6279. def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
  6280. def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
  6281. // Aligned masked loads
  6282. def : MVE_vector_maskedload_typed<v16i8, MVE_VLDRBU8, aligned_maskedloadvi8, 0>;
  6283. def : MVE_vector_maskedload_typed<v8i16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
  6284. def : MVE_vector_maskedload_typed<v8f16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
  6285. def : MVE_vector_maskedload_typed<v4i32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
  6286. def : MVE_vector_maskedload_typed<v4f32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
  6287. }
  6288. // Widening/Narrowing Loads/Stores
  6289. multiclass MVEExtLoadStore<Instruction LoadSInst, Instruction LoadUInst, string StoreInst,
  6290. string Amble, ValueType VT, int Shift> {
  6291. // Trunc stores
  6292. def : Pat<(!cast<PatFrag>("aligned_truncst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr),
  6293. (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr)>;
  6294. def : Pat<(!cast<PatFrag>("aligned_post_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
  6295. (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
  6296. def : Pat<(!cast<PatFrag>("aligned_pre_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
  6297. (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
  6298. // Masked trunc stores
  6299. def : Pat<(!cast<PatFrag>("aligned_truncmaskedst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr, VCCR:$pred),
  6300. (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
  6301. def : Pat<(!cast<PatFrag>("aligned_post_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
  6302. (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
  6303. def : Pat<(!cast<PatFrag>("aligned_pre_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
  6304. (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
  6305. // Ext loads
  6306. def : Pat<(VT (!cast<PatFrag>("aligned_extload"#Amble) taddrmode_imm7<Shift>:$addr)),
  6307. (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
  6308. def : Pat<(VT (!cast<PatFrag>("aligned_sextload"#Amble) taddrmode_imm7<Shift>:$addr)),
  6309. (VT (LoadSInst taddrmode_imm7<Shift>:$addr))>;
  6310. def : Pat<(VT (!cast<PatFrag>("aligned_zextload"#Amble) taddrmode_imm7<Shift>:$addr)),
  6311. (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
  6312. // Masked ext loads
  6313. def : Pat<(VT (!cast<PatFrag>("aligned_extmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
  6314. (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
  6315. def : Pat<(VT (!cast<PatFrag>("aligned_sextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
  6316. (VT (LoadSInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
  6317. def : Pat<(VT (!cast<PatFrag>("aligned_zextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
  6318. (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
  6319. }
  6320. let Predicates = [HasMVEInt] in {
  6321. defm : MVEExtLoadStore<MVE_VLDRBS16, MVE_VLDRBU16, "MVE_VSTRB16", "vi8", v8i16, 0>;
  6322. defm : MVEExtLoadStore<MVE_VLDRBS32, MVE_VLDRBU32, "MVE_VSTRB32", "vi8", v4i32, 0>;
  6323. defm : MVEExtLoadStore<MVE_VLDRHS32, MVE_VLDRHU32, "MVE_VSTRH32", "vi16", v4i32, 1>;
  6324. }
  6325. // Bit convert patterns
  6326. let Predicates = [HasMVEInt] in {
  6327. def : Pat<(v2f64 (bitconvert (v2i64 MQPR:$src))), (v2f64 MQPR:$src)>;
  6328. def : Pat<(v2i64 (bitconvert (v2f64 MQPR:$src))), (v2i64 MQPR:$src)>;
  6329. def : Pat<(v4i32 (bitconvert (v4f32 MQPR:$src))), (v4i32 MQPR:$src)>;
  6330. def : Pat<(v4f32 (bitconvert (v4i32 MQPR:$src))), (v4f32 MQPR:$src)>;
  6331. def : Pat<(v8i16 (bitconvert (v8f16 MQPR:$src))), (v8i16 MQPR:$src)>;
  6332. def : Pat<(v8f16 (bitconvert (v8i16 MQPR:$src))), (v8f16 MQPR:$src)>;
  6333. }
  6334. let Predicates = [IsLE,HasMVEInt] in {
  6335. def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 MQPR:$src)>;
  6336. def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 MQPR:$src)>;
  6337. def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 MQPR:$src)>;
  6338. def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 MQPR:$src)>;
  6339. def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 MQPR:$src)>;
  6340. def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 MQPR:$src)>;
  6341. def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 MQPR:$src)>;
  6342. def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 MQPR:$src)>;
  6343. def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 MQPR:$src)>;
  6344. def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 MQPR:$src)>;
  6345. def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 MQPR:$src)>;
  6346. def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 MQPR:$src)>;
  6347. def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 MQPR:$src)>;
  6348. def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 MQPR:$src)>;
  6349. def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 MQPR:$src)>;
  6350. def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 MQPR:$src)>;
  6351. def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 MQPR:$src)>;
  6352. def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 MQPR:$src)>;
  6353. def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 MQPR:$src)>;
  6354. def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 MQPR:$src)>;
  6355. def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 MQPR:$src)>;
  6356. def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 MQPR:$src)>;
  6357. def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 MQPR:$src)>;
  6358. def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 MQPR:$src)>;
  6359. def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 MQPR:$src)>;
  6360. def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 MQPR:$src)>;
  6361. def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 MQPR:$src)>;
  6362. def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 MQPR:$src)>;
  6363. def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 MQPR:$src)>;
  6364. def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 MQPR:$src)>;
  6365. def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 MQPR:$src)>;
  6366. def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 MQPR:$src)>;
  6367. def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 MQPR:$src)>;
  6368. def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 MQPR:$src)>;
  6369. def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 MQPR:$src)>;
  6370. def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 MQPR:$src)>;
  6371. }
  6372. let Predicates = [IsBE,HasMVEInt] in {
  6373. def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
  6374. def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
  6375. def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
  6376. def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
  6377. def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 (MVE_VREV64_8 MQPR:$src))>;
  6378. def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
  6379. def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
  6380. def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
  6381. def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
  6382. def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 (MVE_VREV64_8 MQPR:$src))>;
  6383. def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
  6384. def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
  6385. def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
  6386. def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
  6387. def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 (MVE_VREV32_8 MQPR:$src))>;
  6388. def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
  6389. def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
  6390. def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
  6391. def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
  6392. def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 (MVE_VREV32_8 MQPR:$src))>;
  6393. def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
  6394. def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
  6395. def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
  6396. def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
  6397. def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 (MVE_VREV16_8 MQPR:$src))>;
  6398. def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
  6399. def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
  6400. def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
  6401. def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
  6402. def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 (MVE_VREV16_8 MQPR:$src))>;
  6403. def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
  6404. def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
  6405. def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
  6406. def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
  6407. def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
  6408. def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
  6409. }