1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900490149024903490449054906490749084909491049114912491349144915491649174918491949204921492249234924492549264927492849294930493149324933493449354936493749384939494049414942494349444945494649474948494949504951495249534954495549564957495849594960496149624963496449654966496749684969497049714972497349744975497649774978497949804981498249834984498549864987498849894990499149924993499449954996499749984999500050015002500350045005500650075008500950105011501250135014501550165017501850195020502150225023502450255026502750285029503050315032503350345035503650375038503950405041504250435044504550465047504850495050505150525053505450555056505750585059506050615062506350645065506650675068506950705071507250735074507550765077507850795080508150825083508450855086508750885089509050915092509350945095509650975098509951005101510251035104510551065107510851095110511151125113511451155116511751185119512051215122512351245125512651275128512951305131513251335134513551365137513851395140514151425143514451455146514751485149515051515152515351545155515651575158515951605161516251635164516551665167516851695170517151725173517451755176517751785179518051815182518351845185518651875188518951905191519251935194519551965197519851995200520152025203520452055206520752085209521052115212521352145215521652175218521952205221522252235224522552265227522852295230523152325233523452355236523752385239524052415242524352445245524652475248524952505251525252535254525552565257525852595260526152625263526452655266526752685269527052715272527352745275527652775278527952805281528252835284528552865287528852895290529152925293529452955296529752985299530053015302530353045305530653075308530953105311531253135314531553165317531853195320532153225323532453255326532753285329533053315332533353345335533653375338533953405341534253435344534553465347534853495350535153525353535453555356535753585359536053615362536353645365536653675368536953705371537253735374537553765377537853795380538153825383538453855386538753885389539053915392539353945395539653975398539954005401540254035404540554065407540854095410541154125413541454155416541754185419542054215422542354245425542654275428542954305431543254335434543554365437543854395440544154425443544454455446544754485449545054515452545354545455545654575458545954605461546254635464546554665467546854695470547154725473547454755476547754785479548054815482548354845485548654875488548954905491549254935494549554965497549854995500550155025503550455055506550755085509551055115512551355145515551655175518551955205521552255235524552555265527552855295530553155325533553455355536553755385539554055415542554355445545554655475548554955505551555255535554555555565557555855595560556155625563556455655566556755685569557055715572557355745575557655775578557955805581558255835584558555865587558855895590559155925593559455955596559755985599560056015602560356045605560656075608560956105611561256135614561556165617561856195620562156225623562456255626562756285629563056315632563356345635563656375638563956405641564256435644564556465647564856495650565156525653565456555656565756585659566056615662566356645665566656675668566956705671567256735674567556765677567856795680568156825683568456855686568756885689569056915692569356945695569656975698569957005701570257035704570557065707570857095710571157125713571457155716571757185719572057215722572357245725572657275728572957305731573257335734573557365737573857395740574157425743574457455746574757485749575057515752575357545755575657575758575957605761576257635764576557665767576857695770577157725773577457755776577757785779578057815782578357845785578657875788578957905791579257935794579557965797579857995800580158025803580458055806580758085809581058115812581358145815581658175818581958205821582258235824582558265827582858295830583158325833583458355836583758385839584058415842584358445845584658475848584958505851585258535854585558565857585858595860586158625863586458655866586758685869587058715872587358745875587658775878587958805881588258835884588558865887588858895890589158925893589458955896589758985899590059015902590359045905590659075908590959105911591259135914591559165917591859195920592159225923592459255926592759285929593059315932593359345935593659375938593959405941594259435944594559465947594859495950595159525953595459555956595759585959596059615962596359645965596659675968596959705971597259735974597559765977597859795980598159825983598459855986598759885989599059915992599359945995599659975998599960006001600260036004600560066007600860096010601160126013601460156016601760186019602060216022602360246025602660276028602960306031603260336034603560366037603860396040604160426043604460456046604760486049605060516052605360546055605660576058605960606061606260636064606560666067606860696070607160726073607460756076607760786079608060816082608360846085608660876088608960906091609260936094609560966097609860996100610161026103610461056106610761086109611061116112611361146115611661176118611961206121612261236124612561266127612861296130613161326133613461356136613761386139614061416142614361446145614661476148614961506151615261536154615561566157615861596160616161626163616461656166616761686169617061716172617361746175617661776178617961806181618261836184618561866187618861896190619161926193619461956196619761986199620062016202620362046205620662076208620962106211621262136214621562166217621862196220622162226223622462256226622762286229623062316232623362346235623662376238623962406241624262436244624562466247624862496250625162526253625462556256625762586259626062616262626362646265626662676268626962706271627262736274627562766277627862796280628162826283628462856286628762886289629062916292629362946295629662976298629963006301630263036304630563066307 |
- /******************************************************************************
- *******************************************************************************
- **
- ** Copyright (C) 2005-2010 Red Hat, Inc. All rights reserved.
- **
- ** This copyrighted material is made available to anyone wishing to use,
- ** modify, copy, or redistribute it subject to the terms and conditions
- ** of the GNU General Public License v.2.
- **
- *******************************************************************************
- ******************************************************************************/
- /* Central locking logic has four stages:
- dlm_lock()
- dlm_unlock()
- request_lock(ls, lkb)
- convert_lock(ls, lkb)
- unlock_lock(ls, lkb)
- cancel_lock(ls, lkb)
- _request_lock(r, lkb)
- _convert_lock(r, lkb)
- _unlock_lock(r, lkb)
- _cancel_lock(r, lkb)
- do_request(r, lkb)
- do_convert(r, lkb)
- do_unlock(r, lkb)
- do_cancel(r, lkb)
- Stage 1 (lock, unlock) is mainly about checking input args and
- splitting into one of the four main operations:
- dlm_lock = request_lock
- dlm_lock+CONVERT = convert_lock
- dlm_unlock = unlock_lock
- dlm_unlock+CANCEL = cancel_lock
- Stage 2, xxxx_lock(), just finds and locks the relevant rsb which is
- provided to the next stage.
- Stage 3, _xxxx_lock(), determines if the operation is local or remote.
- When remote, it calls send_xxxx(), when local it calls do_xxxx().
- Stage 4, do_xxxx(), is the guts of the operation. It manipulates the
- given rsb and lkb and queues callbacks.
- For remote operations, send_xxxx() results in the corresponding do_xxxx()
- function being executed on the remote node. The connecting send/receive
- calls on local (L) and remote (R) nodes:
- L: send_xxxx() -> R: receive_xxxx()
- R: do_xxxx()
- L: receive_xxxx_reply() <- R: send_xxxx_reply()
- */
- #include <linux/types.h>
- #include <linux/rbtree.h>
- #include <linux/slab.h>
- #include "dlm_internal.h"
- #include <linux/dlm_device.h>
- #include "memory.h"
- #include "lowcomms.h"
- #include "requestqueue.h"
- #include "util.h"
- #include "dir.h"
- #include "member.h"
- #include "lockspace.h"
- #include "ast.h"
- #include "lock.h"
- #include "rcom.h"
- #include "recover.h"
- #include "lvb_table.h"
- #include "user.h"
- #include "config.h"
- static int send_request(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_convert(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_unlock(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_cancel(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_grant(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_bast(struct dlm_rsb *r, struct dlm_lkb *lkb, int mode);
- static int send_lookup(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int send_remove(struct dlm_rsb *r);
- static int _request_lock(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static int _cancel_lock(struct dlm_rsb *r, struct dlm_lkb *lkb);
- static void __receive_convert_reply(struct dlm_rsb *r, struct dlm_lkb *lkb,
- struct dlm_message *ms);
- static int receive_extralen(struct dlm_message *ms);
- static void do_purge(struct dlm_ls *ls, int nodeid, int pid);
- static void del_timeout(struct dlm_lkb *lkb);
- static void toss_rsb(struct kref *kref);
- /*
- * Lock compatibilty matrix - thanks Steve
- * UN = Unlocked state. Not really a state, used as a flag
- * PD = Padding. Used to make the matrix a nice power of two in size
- * Other states are the same as the VMS DLM.
- * Usage: matrix[grmode+1][rqmode+1] (although m[rq+1][gr+1] is the same)
- */
- static const int __dlm_compat_matrix[8][8] = {
- /* UN NL CR CW PR PW EX PD */
- {1, 1, 1, 1, 1, 1, 1, 0}, /* UN */
- {1, 1, 1, 1, 1, 1, 1, 0}, /* NL */
- {1, 1, 1, 1, 1, 1, 0, 0}, /* CR */
- {1, 1, 1, 1, 0, 0, 0, 0}, /* CW */
- {1, 1, 1, 0, 1, 0, 0, 0}, /* PR */
- {1, 1, 1, 0, 0, 0, 0, 0}, /* PW */
- {1, 1, 0, 0, 0, 0, 0, 0}, /* EX */
- {0, 0, 0, 0, 0, 0, 0, 0} /* PD */
- };
- /*
- * This defines the direction of transfer of LVB data.
- * Granted mode is the row; requested mode is the column.
- * Usage: matrix[grmode+1][rqmode+1]
- * 1 = LVB is returned to the caller
- * 0 = LVB is written to the resource
- * -1 = nothing happens to the LVB
- */
- const int dlm_lvb_operations[8][8] = {
- /* UN NL CR CW PR PW EX PD*/
- { -1, 1, 1, 1, 1, 1, 1, -1 }, /* UN */
- { -1, 1, 1, 1, 1, 1, 1, 0 }, /* NL */
- { -1, -1, 1, 1, 1, 1, 1, 0 }, /* CR */
- { -1, -1, -1, 1, 1, 1, 1, 0 }, /* CW */
- { -1, -1, -1, -1, 1, 1, 1, 0 }, /* PR */
- { -1, 0, 0, 0, 0, 0, 1, 0 }, /* PW */
- { -1, 0, 0, 0, 0, 0, 0, 0 }, /* EX */
- { -1, 0, 0, 0, 0, 0, 0, 0 } /* PD */
- };
- #define modes_compat(gr, rq) \
- __dlm_compat_matrix[(gr)->lkb_grmode + 1][(rq)->lkb_rqmode + 1]
- int dlm_modes_compat(int mode1, int mode2)
- {
- return __dlm_compat_matrix[mode1 + 1][mode2 + 1];
- }
- /*
- * Compatibility matrix for conversions with QUECVT set.
- * Granted mode is the row; requested mode is the column.
- * Usage: matrix[grmode+1][rqmode+1]
- */
- static const int __quecvt_compat_matrix[8][8] = {
- /* UN NL CR CW PR PW EX PD */
- {0, 0, 0, 0, 0, 0, 0, 0}, /* UN */
- {0, 0, 1, 1, 1, 1, 1, 0}, /* NL */
- {0, 0, 0, 1, 1, 1, 1, 0}, /* CR */
- {0, 0, 0, 0, 1, 1, 1, 0}, /* CW */
- {0, 0, 0, 1, 0, 1, 1, 0}, /* PR */
- {0, 0, 0, 0, 0, 0, 1, 0}, /* PW */
- {0, 0, 0, 0, 0, 0, 0, 0}, /* EX */
- {0, 0, 0, 0, 0, 0, 0, 0} /* PD */
- };
- void dlm_print_lkb(struct dlm_lkb *lkb)
- {
- printk(KERN_ERR "lkb: nodeid %d id %x remid %x exflags %x flags %x "
- "sts %d rq %d gr %d wait_type %d wait_nodeid %d seq %llu\n",
- lkb->lkb_nodeid, lkb->lkb_id, lkb->lkb_remid, lkb->lkb_exflags,
- lkb->lkb_flags, lkb->lkb_status, lkb->lkb_rqmode,
- lkb->lkb_grmode, lkb->lkb_wait_type, lkb->lkb_wait_nodeid,
- (unsigned long long)lkb->lkb_recover_seq);
- }
- static void dlm_print_rsb(struct dlm_rsb *r)
- {
- printk(KERN_ERR "rsb: nodeid %d master %d dir %d flags %lx first %x "
- "rlc %d name %s\n",
- r->res_nodeid, r->res_master_nodeid, r->res_dir_nodeid,
- r->res_flags, r->res_first_lkid, r->res_recover_locks_count,
- r->res_name);
- }
- void dlm_dump_rsb(struct dlm_rsb *r)
- {
- struct dlm_lkb *lkb;
- dlm_print_rsb(r);
- printk(KERN_ERR "rsb: root_list empty %d recover_list empty %d\n",
- list_empty(&r->res_root_list), list_empty(&r->res_recover_list));
- printk(KERN_ERR "rsb lookup list\n");
- list_for_each_entry(lkb, &r->res_lookup, lkb_rsb_lookup)
- dlm_print_lkb(lkb);
- printk(KERN_ERR "rsb grant queue:\n");
- list_for_each_entry(lkb, &r->res_grantqueue, lkb_statequeue)
- dlm_print_lkb(lkb);
- printk(KERN_ERR "rsb convert queue:\n");
- list_for_each_entry(lkb, &r->res_convertqueue, lkb_statequeue)
- dlm_print_lkb(lkb);
- printk(KERN_ERR "rsb wait queue:\n");
- list_for_each_entry(lkb, &r->res_waitqueue, lkb_statequeue)
- dlm_print_lkb(lkb);
- }
- /* Threads cannot use the lockspace while it's being recovered */
- static inline void dlm_lock_recovery(struct dlm_ls *ls)
- {
- down_read(&ls->ls_in_recovery);
- }
- void dlm_unlock_recovery(struct dlm_ls *ls)
- {
- up_read(&ls->ls_in_recovery);
- }
- int dlm_lock_recovery_try(struct dlm_ls *ls)
- {
- return down_read_trylock(&ls->ls_in_recovery);
- }
- static inline int can_be_queued(struct dlm_lkb *lkb)
- {
- return !(lkb->lkb_exflags & DLM_LKF_NOQUEUE);
- }
- static inline int force_blocking_asts(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_exflags & DLM_LKF_NOQUEUEBAST);
- }
- static inline int is_demoted(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_sbflags & DLM_SBF_DEMOTED);
- }
- static inline int is_altmode(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_sbflags & DLM_SBF_ALTMODE);
- }
- static inline int is_granted(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_status == DLM_LKSTS_GRANTED);
- }
- static inline int is_remote(struct dlm_rsb *r)
- {
- DLM_ASSERT(r->res_nodeid >= 0, dlm_print_rsb(r););
- return !!r->res_nodeid;
- }
- static inline int is_process_copy(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_nodeid && !(lkb->lkb_flags & DLM_IFL_MSTCPY));
- }
- static inline int is_master_copy(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_flags & DLM_IFL_MSTCPY) ? 1 : 0;
- }
- static inline int middle_conversion(struct dlm_lkb *lkb)
- {
- if ((lkb->lkb_grmode==DLM_LOCK_PR && lkb->lkb_rqmode==DLM_LOCK_CW) ||
- (lkb->lkb_rqmode==DLM_LOCK_PR && lkb->lkb_grmode==DLM_LOCK_CW))
- return 1;
- return 0;
- }
- static inline int down_conversion(struct dlm_lkb *lkb)
- {
- return (!middle_conversion(lkb) && lkb->lkb_rqmode < lkb->lkb_grmode);
- }
- static inline int is_overlap_unlock(struct dlm_lkb *lkb)
- {
- return lkb->lkb_flags & DLM_IFL_OVERLAP_UNLOCK;
- }
- static inline int is_overlap_cancel(struct dlm_lkb *lkb)
- {
- return lkb->lkb_flags & DLM_IFL_OVERLAP_CANCEL;
- }
- static inline int is_overlap(struct dlm_lkb *lkb)
- {
- return (lkb->lkb_flags & (DLM_IFL_OVERLAP_UNLOCK |
- DLM_IFL_OVERLAP_CANCEL));
- }
- static void queue_cast(struct dlm_rsb *r, struct dlm_lkb *lkb, int rv)
- {
- if (is_master_copy(lkb))
- return;
- del_timeout(lkb);
- DLM_ASSERT(lkb->lkb_lksb, dlm_print_lkb(lkb););
- /* if the operation was a cancel, then return -DLM_ECANCEL, if a
- timeout caused the cancel then return -ETIMEDOUT */
- if (rv == -DLM_ECANCEL && (lkb->lkb_flags & DLM_IFL_TIMEOUT_CANCEL)) {
- lkb->lkb_flags &= ~DLM_IFL_TIMEOUT_CANCEL;
- rv = -ETIMEDOUT;
- }
- if (rv == -DLM_ECANCEL && (lkb->lkb_flags & DLM_IFL_DEADLOCK_CANCEL)) {
- lkb->lkb_flags &= ~DLM_IFL_DEADLOCK_CANCEL;
- rv = -EDEADLK;
- }
- dlm_add_cb(lkb, DLM_CB_CAST, lkb->lkb_grmode, rv, lkb->lkb_sbflags);
- }
- static inline void queue_cast_overlap(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- queue_cast(r, lkb,
- is_overlap_unlock(lkb) ? -DLM_EUNLOCK : -DLM_ECANCEL);
- }
- static void queue_bast(struct dlm_rsb *r, struct dlm_lkb *lkb, int rqmode)
- {
- if (is_master_copy(lkb)) {
- send_bast(r, lkb, rqmode);
- } else {
- dlm_add_cb(lkb, DLM_CB_BAST, rqmode, 0, 0);
- }
- }
- /*
- * Basic operations on rsb's and lkb's
- */
- /* This is only called to add a reference when the code already holds
- a valid reference to the rsb, so there's no need for locking. */
- static inline void hold_rsb(struct dlm_rsb *r)
- {
- kref_get(&r->res_ref);
- }
- void dlm_hold_rsb(struct dlm_rsb *r)
- {
- hold_rsb(r);
- }
- /* When all references to the rsb are gone it's transferred to
- the tossed list for later disposal. */
- static void put_rsb(struct dlm_rsb *r)
- {
- struct dlm_ls *ls = r->res_ls;
- uint32_t bucket = r->res_bucket;
- spin_lock(&ls->ls_rsbtbl[bucket].lock);
- kref_put(&r->res_ref, toss_rsb);
- spin_unlock(&ls->ls_rsbtbl[bucket].lock);
- }
- void dlm_put_rsb(struct dlm_rsb *r)
- {
- put_rsb(r);
- }
- static int pre_rsb_struct(struct dlm_ls *ls)
- {
- struct dlm_rsb *r1, *r2;
- int count = 0;
- spin_lock(&ls->ls_new_rsb_spin);
- if (ls->ls_new_rsb_count > dlm_config.ci_new_rsb_count / 2) {
- spin_unlock(&ls->ls_new_rsb_spin);
- return 0;
- }
- spin_unlock(&ls->ls_new_rsb_spin);
- r1 = dlm_allocate_rsb(ls);
- r2 = dlm_allocate_rsb(ls);
- spin_lock(&ls->ls_new_rsb_spin);
- if (r1) {
- list_add(&r1->res_hashchain, &ls->ls_new_rsb);
- ls->ls_new_rsb_count++;
- }
- if (r2) {
- list_add(&r2->res_hashchain, &ls->ls_new_rsb);
- ls->ls_new_rsb_count++;
- }
- count = ls->ls_new_rsb_count;
- spin_unlock(&ls->ls_new_rsb_spin);
- if (!count)
- return -ENOMEM;
- return 0;
- }
- /* If ls->ls_new_rsb is empty, return -EAGAIN, so the caller can
- unlock any spinlocks, go back and call pre_rsb_struct again.
- Otherwise, take an rsb off the list and return it. */
- static int get_rsb_struct(struct dlm_ls *ls, char *name, int len,
- struct dlm_rsb **r_ret)
- {
- struct dlm_rsb *r;
- int count;
- spin_lock(&ls->ls_new_rsb_spin);
- if (list_empty(&ls->ls_new_rsb)) {
- count = ls->ls_new_rsb_count;
- spin_unlock(&ls->ls_new_rsb_spin);
- log_debug(ls, "find_rsb retry %d %d %s",
- count, dlm_config.ci_new_rsb_count, name);
- return -EAGAIN;
- }
- r = list_first_entry(&ls->ls_new_rsb, struct dlm_rsb, res_hashchain);
- list_del(&r->res_hashchain);
- /* Convert the empty list_head to a NULL rb_node for tree usage: */
- memset(&r->res_hashnode, 0, sizeof(struct rb_node));
- ls->ls_new_rsb_count--;
- spin_unlock(&ls->ls_new_rsb_spin);
- r->res_ls = ls;
- r->res_length = len;
- memcpy(r->res_name, name, len);
- mutex_init(&r->res_mutex);
- INIT_LIST_HEAD(&r->res_lookup);
- INIT_LIST_HEAD(&r->res_grantqueue);
- INIT_LIST_HEAD(&r->res_convertqueue);
- INIT_LIST_HEAD(&r->res_waitqueue);
- INIT_LIST_HEAD(&r->res_root_list);
- INIT_LIST_HEAD(&r->res_recover_list);
- *r_ret = r;
- return 0;
- }
- static int rsb_cmp(struct dlm_rsb *r, const char *name, int nlen)
- {
- char maxname[DLM_RESNAME_MAXLEN];
- memset(maxname, 0, DLM_RESNAME_MAXLEN);
- memcpy(maxname, name, nlen);
- return memcmp(r->res_name, maxname, DLM_RESNAME_MAXLEN);
- }
- int dlm_search_rsb_tree(struct rb_root *tree, char *name, int len,
- struct dlm_rsb **r_ret)
- {
- struct rb_node *node = tree->rb_node;
- struct dlm_rsb *r;
- int rc;
- while (node) {
- r = rb_entry(node, struct dlm_rsb, res_hashnode);
- rc = rsb_cmp(r, name, len);
- if (rc < 0)
- node = node->rb_left;
- else if (rc > 0)
- node = node->rb_right;
- else
- goto found;
- }
- *r_ret = NULL;
- return -EBADR;
- found:
- *r_ret = r;
- return 0;
- }
- static int rsb_insert(struct dlm_rsb *rsb, struct rb_root *tree)
- {
- struct rb_node **newn = &tree->rb_node;
- struct rb_node *parent = NULL;
- int rc;
- while (*newn) {
- struct dlm_rsb *cur = rb_entry(*newn, struct dlm_rsb,
- res_hashnode);
- parent = *newn;
- rc = rsb_cmp(cur, rsb->res_name, rsb->res_length);
- if (rc < 0)
- newn = &parent->rb_left;
- else if (rc > 0)
- newn = &parent->rb_right;
- else {
- log_print("rsb_insert match");
- dlm_dump_rsb(rsb);
- dlm_dump_rsb(cur);
- return -EEXIST;
- }
- }
- rb_link_node(&rsb->res_hashnode, parent, newn);
- rb_insert_color(&rsb->res_hashnode, tree);
- return 0;
- }
- /*
- * Find rsb in rsbtbl and potentially create/add one
- *
- * Delaying the release of rsb's has a similar benefit to applications keeping
- * NL locks on an rsb, but without the guarantee that the cached master value
- * will still be valid when the rsb is reused. Apps aren't always smart enough
- * to keep NL locks on an rsb that they may lock again shortly; this can lead
- * to excessive master lookups and removals if we don't delay the release.
- *
- * Searching for an rsb means looking through both the normal list and toss
- * list. When found on the toss list the rsb is moved to the normal list with
- * ref count of 1; when found on normal list the ref count is incremented.
- *
- * rsb's on the keep list are being used locally and refcounted.
- * rsb's on the toss list are not being used locally, and are not refcounted.
- *
- * The toss list rsb's were either
- * - previously used locally but not any more (were on keep list, then
- * moved to toss list when last refcount dropped)
- * - created and put on toss list as a directory record for a lookup
- * (we are the dir node for the res, but are not using the res right now,
- * but some other node is)
- *
- * The purpose of find_rsb() is to return a refcounted rsb for local use.
- * So, if the given rsb is on the toss list, it is moved to the keep list
- * before being returned.
- *
- * toss_rsb() happens when all local usage of the rsb is done, i.e. no
- * more refcounts exist, so the rsb is moved from the keep list to the
- * toss list.
- *
- * rsb's on both keep and toss lists are used for doing a name to master
- * lookups. rsb's that are in use locally (and being refcounted) are on
- * the keep list, rsb's that are not in use locally (not refcounted) and
- * only exist for name/master lookups are on the toss list.
- *
- * rsb's on the toss list who's dir_nodeid is not local can have stale
- * name/master mappings. So, remote requests on such rsb's can potentially
- * return with an error, which means the mapping is stale and needs to
- * be updated with a new lookup. (The idea behind MASTER UNCERTAIN and
- * first_lkid is to keep only a single outstanding request on an rsb
- * while that rsb has a potentially stale master.)
- */
- static int find_rsb_dir(struct dlm_ls *ls, char *name, int len,
- uint32_t hash, uint32_t b,
- int dir_nodeid, int from_nodeid,
- unsigned int flags, struct dlm_rsb **r_ret)
- {
- struct dlm_rsb *r = NULL;
- int our_nodeid = dlm_our_nodeid();
- int from_local = 0;
- int from_other = 0;
- int from_dir = 0;
- int create = 0;
- int error;
- if (flags & R_RECEIVE_REQUEST) {
- if (from_nodeid == dir_nodeid)
- from_dir = 1;
- else
- from_other = 1;
- } else if (flags & R_REQUEST) {
- from_local = 1;
- }
- /*
- * flags & R_RECEIVE_RECOVER is from dlm_recover_master_copy, so
- * from_nodeid has sent us a lock in dlm_recover_locks, believing
- * we're the new master. Our local recovery may not have set
- * res_master_nodeid to our_nodeid yet, so allow either. Don't
- * create the rsb; dlm_recover_process_copy() will handle EBADR
- * by resending.
- *
- * If someone sends us a request, we are the dir node, and we do
- * not find the rsb anywhere, then recreate it. This happens if
- * someone sends us a request after we have removed/freed an rsb
- * from our toss list. (They sent a request instead of lookup
- * because they are using an rsb from their toss list.)
- */
- if (from_local || from_dir ||
- (from_other && (dir_nodeid == our_nodeid))) {
- create = 1;
- }
- retry:
- if (create) {
- error = pre_rsb_struct(ls);
- if (error < 0)
- goto out;
- }
- spin_lock(&ls->ls_rsbtbl[b].lock);
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (error)
- goto do_toss;
-
- /*
- * rsb is active, so we can't check master_nodeid without lock_rsb.
- */
- kref_get(&r->res_ref);
- error = 0;
- goto out_unlock;
- do_toss:
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (error)
- goto do_new;
- /*
- * rsb found inactive (master_nodeid may be out of date unless
- * we are the dir_nodeid or were the master) No other thread
- * is using this rsb because it's on the toss list, so we can
- * look at or update res_master_nodeid without lock_rsb.
- */
- if ((r->res_master_nodeid != our_nodeid) && from_other) {
- /* our rsb was not master, and another node (not the dir node)
- has sent us a request */
- log_debug(ls, "find_rsb toss from_other %d master %d dir %d %s",
- from_nodeid, r->res_master_nodeid, dir_nodeid,
- r->res_name);
- error = -ENOTBLK;
- goto out_unlock;
- }
- if ((r->res_master_nodeid != our_nodeid) && from_dir) {
- /* don't think this should ever happen */
- log_error(ls, "find_rsb toss from_dir %d master %d",
- from_nodeid, r->res_master_nodeid);
- dlm_print_rsb(r);
- /* fix it and go on */
- r->res_master_nodeid = our_nodeid;
- r->res_nodeid = 0;
- rsb_clear_flag(r, RSB_MASTER_UNCERTAIN);
- r->res_first_lkid = 0;
- }
- if (from_local && (r->res_master_nodeid != our_nodeid)) {
- /* Because we have held no locks on this rsb,
- res_master_nodeid could have become stale. */
- rsb_set_flag(r, RSB_MASTER_UNCERTAIN);
- r->res_first_lkid = 0;
- }
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss);
- error = rsb_insert(r, &ls->ls_rsbtbl[b].keep);
- goto out_unlock;
- do_new:
- /*
- * rsb not found
- */
- if (error == -EBADR && !create)
- goto out_unlock;
- error = get_rsb_struct(ls, name, len, &r);
- if (error == -EAGAIN) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- goto retry;
- }
- if (error)
- goto out_unlock;
- r->res_hash = hash;
- r->res_bucket = b;
- r->res_dir_nodeid = dir_nodeid;
- kref_init(&r->res_ref);
- if (from_dir) {
- /* want to see how often this happens */
- log_debug(ls, "find_rsb new from_dir %d recreate %s",
- from_nodeid, r->res_name);
- r->res_master_nodeid = our_nodeid;
- r->res_nodeid = 0;
- goto out_add;
- }
- if (from_other && (dir_nodeid != our_nodeid)) {
- /* should never happen */
- log_error(ls, "find_rsb new from_other %d dir %d our %d %s",
- from_nodeid, dir_nodeid, our_nodeid, r->res_name);
- dlm_free_rsb(r);
- r = NULL;
- error = -ENOTBLK;
- goto out_unlock;
- }
- if (from_other) {
- log_debug(ls, "find_rsb new from_other %d dir %d %s",
- from_nodeid, dir_nodeid, r->res_name);
- }
- if (dir_nodeid == our_nodeid) {
- /* When we are the dir nodeid, we can set the master
- node immediately */
- r->res_master_nodeid = our_nodeid;
- r->res_nodeid = 0;
- } else {
- /* set_master will send_lookup to dir_nodeid */
- r->res_master_nodeid = 0;
- r->res_nodeid = -1;
- }
- out_add:
- error = rsb_insert(r, &ls->ls_rsbtbl[b].keep);
- out_unlock:
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- out:
- *r_ret = r;
- return error;
- }
- /* During recovery, other nodes can send us new MSTCPY locks (from
- dlm_recover_locks) before we've made ourself master (in
- dlm_recover_masters). */
- static int find_rsb_nodir(struct dlm_ls *ls, char *name, int len,
- uint32_t hash, uint32_t b,
- int dir_nodeid, int from_nodeid,
- unsigned int flags, struct dlm_rsb **r_ret)
- {
- struct dlm_rsb *r = NULL;
- int our_nodeid = dlm_our_nodeid();
- int recover = (flags & R_RECEIVE_RECOVER);
- int error;
- retry:
- error = pre_rsb_struct(ls);
- if (error < 0)
- goto out;
- spin_lock(&ls->ls_rsbtbl[b].lock);
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (error)
- goto do_toss;
- /*
- * rsb is active, so we can't check master_nodeid without lock_rsb.
- */
- kref_get(&r->res_ref);
- goto out_unlock;
- do_toss:
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (error)
- goto do_new;
- /*
- * rsb found inactive. No other thread is using this rsb because
- * it's on the toss list, so we can look at or update
- * res_master_nodeid without lock_rsb.
- */
- if (!recover && (r->res_master_nodeid != our_nodeid) && from_nodeid) {
- /* our rsb is not master, and another node has sent us a
- request; this should never happen */
- log_error(ls, "find_rsb toss from_nodeid %d master %d dir %d",
- from_nodeid, r->res_master_nodeid, dir_nodeid);
- dlm_print_rsb(r);
- error = -ENOTBLK;
- goto out_unlock;
- }
- if (!recover && (r->res_master_nodeid != our_nodeid) &&
- (dir_nodeid == our_nodeid)) {
- /* our rsb is not master, and we are dir; may as well fix it;
- this should never happen */
- log_error(ls, "find_rsb toss our %d master %d dir %d",
- our_nodeid, r->res_master_nodeid, dir_nodeid);
- dlm_print_rsb(r);
- r->res_master_nodeid = our_nodeid;
- r->res_nodeid = 0;
- }
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss);
- error = rsb_insert(r, &ls->ls_rsbtbl[b].keep);
- goto out_unlock;
- do_new:
- /*
- * rsb not found
- */
- error = get_rsb_struct(ls, name, len, &r);
- if (error == -EAGAIN) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- goto retry;
- }
- if (error)
- goto out_unlock;
- r->res_hash = hash;
- r->res_bucket = b;
- r->res_dir_nodeid = dir_nodeid;
- r->res_master_nodeid = dir_nodeid;
- r->res_nodeid = (dir_nodeid == our_nodeid) ? 0 : dir_nodeid;
- kref_init(&r->res_ref);
- error = rsb_insert(r, &ls->ls_rsbtbl[b].keep);
- out_unlock:
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- out:
- *r_ret = r;
- return error;
- }
- static int find_rsb(struct dlm_ls *ls, char *name, int len, int from_nodeid,
- unsigned int flags, struct dlm_rsb **r_ret)
- {
- uint32_t hash, b;
- int dir_nodeid;
- if (len > DLM_RESNAME_MAXLEN)
- return -EINVAL;
- hash = jhash(name, len, 0);
- b = hash & (ls->ls_rsbtbl_size - 1);
- dir_nodeid = dlm_hash2nodeid(ls, hash);
- if (dlm_no_directory(ls))
- return find_rsb_nodir(ls, name, len, hash, b, dir_nodeid,
- from_nodeid, flags, r_ret);
- else
- return find_rsb_dir(ls, name, len, hash, b, dir_nodeid,
- from_nodeid, flags, r_ret);
- }
- /* we have received a request and found that res_master_nodeid != our_nodeid,
- so we need to return an error or make ourself the master */
- static int validate_master_nodeid(struct dlm_ls *ls, struct dlm_rsb *r,
- int from_nodeid)
- {
- if (dlm_no_directory(ls)) {
- log_error(ls, "find_rsb keep from_nodeid %d master %d dir %d",
- from_nodeid, r->res_master_nodeid,
- r->res_dir_nodeid);
- dlm_print_rsb(r);
- return -ENOTBLK;
- }
- if (from_nodeid != r->res_dir_nodeid) {
- /* our rsb is not master, and another node (not the dir node)
- has sent us a request. this is much more common when our
- master_nodeid is zero, so limit debug to non-zero. */
- if (r->res_master_nodeid) {
- log_debug(ls, "validate master from_other %d master %d "
- "dir %d first %x %s", from_nodeid,
- r->res_master_nodeid, r->res_dir_nodeid,
- r->res_first_lkid, r->res_name);
- }
- return -ENOTBLK;
- } else {
- /* our rsb is not master, but the dir nodeid has sent us a
- request; this could happen with master 0 / res_nodeid -1 */
- if (r->res_master_nodeid) {
- log_error(ls, "validate master from_dir %d master %d "
- "first %x %s",
- from_nodeid, r->res_master_nodeid,
- r->res_first_lkid, r->res_name);
- }
- r->res_master_nodeid = dlm_our_nodeid();
- r->res_nodeid = 0;
- return 0;
- }
- }
- /*
- * We're the dir node for this res and another node wants to know the
- * master nodeid. During normal operation (non recovery) this is only
- * called from receive_lookup(); master lookups when the local node is
- * the dir node are done by find_rsb().
- *
- * normal operation, we are the dir node for a resource
- * . _request_lock
- * . set_master
- * . send_lookup
- * . receive_lookup
- * . dlm_master_lookup flags 0
- *
- * recover directory, we are rebuilding dir for all resources
- * . dlm_recover_directory
- * . dlm_rcom_names
- * remote node sends back the rsb names it is master of and we are dir of
- * . dlm_master_lookup RECOVER_DIR (fix_master 0, from_master 1)
- * we either create new rsb setting remote node as master, or find existing
- * rsb and set master to be the remote node.
- *
- * recover masters, we are finding the new master for resources
- * . dlm_recover_masters
- * . recover_master
- * . dlm_send_rcom_lookup
- * . receive_rcom_lookup
- * . dlm_master_lookup RECOVER_MASTER (fix_master 1, from_master 0)
- */
- int dlm_master_lookup(struct dlm_ls *ls, int from_nodeid, char *name, int len,
- unsigned int flags, int *r_nodeid, int *result)
- {
- struct dlm_rsb *r = NULL;
- uint32_t hash, b;
- int from_master = (flags & DLM_LU_RECOVER_DIR);
- int fix_master = (flags & DLM_LU_RECOVER_MASTER);
- int our_nodeid = dlm_our_nodeid();
- int dir_nodeid, error, toss_list = 0;
- if (len > DLM_RESNAME_MAXLEN)
- return -EINVAL;
- if (from_nodeid == our_nodeid) {
- log_error(ls, "dlm_master_lookup from our_nodeid %d flags %x",
- our_nodeid, flags);
- return -EINVAL;
- }
- hash = jhash(name, len, 0);
- b = hash & (ls->ls_rsbtbl_size - 1);
- dir_nodeid = dlm_hash2nodeid(ls, hash);
- if (dir_nodeid != our_nodeid) {
- log_error(ls, "dlm_master_lookup from %d dir %d our %d h %x %d",
- from_nodeid, dir_nodeid, our_nodeid, hash,
- ls->ls_num_nodes);
- *r_nodeid = -1;
- return -EINVAL;
- }
- retry:
- error = pre_rsb_struct(ls);
- if (error < 0)
- return error;
- spin_lock(&ls->ls_rsbtbl[b].lock);
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (!error) {
- /* because the rsb is active, we need to lock_rsb before
- checking/changing re_master_nodeid */
- hold_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- lock_rsb(r);
- goto found;
- }
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (error)
- goto not_found;
- /* because the rsb is inactive (on toss list), it's not refcounted
- and lock_rsb is not used, but is protected by the rsbtbl lock */
- toss_list = 1;
- found:
- if (r->res_dir_nodeid != our_nodeid) {
- /* should not happen, but may as well fix it and carry on */
- log_error(ls, "dlm_master_lookup res_dir %d our %d %s",
- r->res_dir_nodeid, our_nodeid, r->res_name);
- r->res_dir_nodeid = our_nodeid;
- }
- if (fix_master && dlm_is_removed(ls, r->res_master_nodeid)) {
- /* Recovery uses this function to set a new master when
- the previous master failed. Setting NEW_MASTER will
- force dlm_recover_masters to call recover_master on this
- rsb even though the res_nodeid is no longer removed. */
- r->res_master_nodeid = from_nodeid;
- r->res_nodeid = from_nodeid;
- rsb_set_flag(r, RSB_NEW_MASTER);
- if (toss_list) {
- /* I don't think we should ever find it on toss list. */
- log_error(ls, "dlm_master_lookup fix_master on toss");
- dlm_dump_rsb(r);
- }
- }
- if (from_master && (r->res_master_nodeid != from_nodeid)) {
- /* this will happen if from_nodeid became master during
- a previous recovery cycle, and we aborted the previous
- cycle before recovering this master value */
- log_limit(ls, "dlm_master_lookup from_master %d "
- "master_nodeid %d res_nodeid %d first %x %s",
- from_nodeid, r->res_master_nodeid, r->res_nodeid,
- r->res_first_lkid, r->res_name);
- if (r->res_master_nodeid == our_nodeid) {
- log_error(ls, "from_master %d our_master", from_nodeid);
- dlm_dump_rsb(r);
- dlm_send_rcom_lookup_dump(r, from_nodeid);
- goto out_found;
- }
- r->res_master_nodeid = from_nodeid;
- r->res_nodeid = from_nodeid;
- rsb_set_flag(r, RSB_NEW_MASTER);
- }
- if (!r->res_master_nodeid) {
- /* this will happen if recovery happens while we're looking
- up the master for this rsb */
- log_debug(ls, "dlm_master_lookup master 0 to %d first %x %s",
- from_nodeid, r->res_first_lkid, r->res_name);
- r->res_master_nodeid = from_nodeid;
- r->res_nodeid = from_nodeid;
- }
- if (!from_master && !fix_master &&
- (r->res_master_nodeid == from_nodeid)) {
- /* this can happen when the master sends remove, the dir node
- finds the rsb on the keep list and ignores the remove,
- and the former master sends a lookup */
- log_limit(ls, "dlm_master_lookup from master %d flags %x "
- "first %x %s", from_nodeid, flags,
- r->res_first_lkid, r->res_name);
- }
- out_found:
- *r_nodeid = r->res_master_nodeid;
- if (result)
- *result = DLM_LU_MATCH;
- if (toss_list) {
- r->res_toss_time = jiffies;
- /* the rsb was inactive (on toss list) */
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- } else {
- /* the rsb was active */
- unlock_rsb(r);
- put_rsb(r);
- }
- return 0;
- not_found:
- error = get_rsb_struct(ls, name, len, &r);
- if (error == -EAGAIN) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- goto retry;
- }
- if (error)
- goto out_unlock;
- r->res_hash = hash;
- r->res_bucket = b;
- r->res_dir_nodeid = our_nodeid;
- r->res_master_nodeid = from_nodeid;
- r->res_nodeid = from_nodeid;
- kref_init(&r->res_ref);
- r->res_toss_time = jiffies;
- error = rsb_insert(r, &ls->ls_rsbtbl[b].toss);
- if (error) {
- /* should never happen */
- dlm_free_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- goto retry;
- }
- if (result)
- *result = DLM_LU_ADD;
- *r_nodeid = from_nodeid;
- error = 0;
- out_unlock:
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return error;
- }
- static void dlm_dump_rsb_hash(struct dlm_ls *ls, uint32_t hash)
- {
- struct rb_node *n;
- struct dlm_rsb *r;
- int i;
- for (i = 0; i < ls->ls_rsbtbl_size; i++) {
- spin_lock(&ls->ls_rsbtbl[i].lock);
- for (n = rb_first(&ls->ls_rsbtbl[i].keep); n; n = rb_next(n)) {
- r = rb_entry(n, struct dlm_rsb, res_hashnode);
- if (r->res_hash == hash)
- dlm_dump_rsb(r);
- }
- spin_unlock(&ls->ls_rsbtbl[i].lock);
- }
- }
- void dlm_dump_rsb_name(struct dlm_ls *ls, char *name, int len)
- {
- struct dlm_rsb *r = NULL;
- uint32_t hash, b;
- int error;
- hash = jhash(name, len, 0);
- b = hash & (ls->ls_rsbtbl_size - 1);
- spin_lock(&ls->ls_rsbtbl[b].lock);
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (!error)
- goto out_dump;
- error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (error)
- goto out;
- out_dump:
- dlm_dump_rsb(r);
- out:
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- }
- static void toss_rsb(struct kref *kref)
- {
- struct dlm_rsb *r = container_of(kref, struct dlm_rsb, res_ref);
- struct dlm_ls *ls = r->res_ls;
- DLM_ASSERT(list_empty(&r->res_root_list), dlm_print_rsb(r););
- kref_init(&r->res_ref);
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[r->res_bucket].keep);
- rsb_insert(r, &ls->ls_rsbtbl[r->res_bucket].toss);
- r->res_toss_time = jiffies;
- ls->ls_rsbtbl[r->res_bucket].flags |= DLM_RTF_SHRINK;
- if (r->res_lvbptr) {
- dlm_free_lvb(r->res_lvbptr);
- r->res_lvbptr = NULL;
- }
- }
- /* See comment for unhold_lkb */
- static void unhold_rsb(struct dlm_rsb *r)
- {
- int rv;
- rv = kref_put(&r->res_ref, toss_rsb);
- DLM_ASSERT(!rv, dlm_dump_rsb(r););
- }
- static void kill_rsb(struct kref *kref)
- {
- struct dlm_rsb *r = container_of(kref, struct dlm_rsb, res_ref);
- /* All work is done after the return from kref_put() so we
- can release the write_lock before the remove and free. */
- DLM_ASSERT(list_empty(&r->res_lookup), dlm_dump_rsb(r););
- DLM_ASSERT(list_empty(&r->res_grantqueue), dlm_dump_rsb(r););
- DLM_ASSERT(list_empty(&r->res_convertqueue), dlm_dump_rsb(r););
- DLM_ASSERT(list_empty(&r->res_waitqueue), dlm_dump_rsb(r););
- DLM_ASSERT(list_empty(&r->res_root_list), dlm_dump_rsb(r););
- DLM_ASSERT(list_empty(&r->res_recover_list), dlm_dump_rsb(r););
- }
- /* Attaching/detaching lkb's from rsb's is for rsb reference counting.
- The rsb must exist as long as any lkb's for it do. */
- static void attach_lkb(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- hold_rsb(r);
- lkb->lkb_resource = r;
- }
- static void detach_lkb(struct dlm_lkb *lkb)
- {
- if (lkb->lkb_resource) {
- put_rsb(lkb->lkb_resource);
- lkb->lkb_resource = NULL;
- }
- }
- static int create_lkb(struct dlm_ls *ls, struct dlm_lkb **lkb_ret)
- {
- struct dlm_lkb *lkb;
- int rv;
- lkb = dlm_allocate_lkb(ls);
- if (!lkb)
- return -ENOMEM;
- lkb->lkb_nodeid = -1;
- lkb->lkb_grmode = DLM_LOCK_IV;
- kref_init(&lkb->lkb_ref);
- INIT_LIST_HEAD(&lkb->lkb_ownqueue);
- INIT_LIST_HEAD(&lkb->lkb_rsb_lookup);
- INIT_LIST_HEAD(&lkb->lkb_time_list);
- INIT_LIST_HEAD(&lkb->lkb_cb_list);
- mutex_init(&lkb->lkb_cb_mutex);
- INIT_WORK(&lkb->lkb_cb_work, dlm_callback_work);
- idr_preload(GFP_NOFS);
- spin_lock(&ls->ls_lkbidr_spin);
- rv = idr_alloc(&ls->ls_lkbidr, lkb, 1, 0, GFP_NOWAIT);
- if (rv >= 0)
- lkb->lkb_id = rv;
- spin_unlock(&ls->ls_lkbidr_spin);
- idr_preload_end();
- if (rv < 0) {
- log_error(ls, "create_lkb idr error %d", rv);
- dlm_free_lkb(lkb);
- return rv;
- }
- *lkb_ret = lkb;
- return 0;
- }
- static int find_lkb(struct dlm_ls *ls, uint32_t lkid, struct dlm_lkb **lkb_ret)
- {
- struct dlm_lkb *lkb;
- spin_lock(&ls->ls_lkbidr_spin);
- lkb = idr_find(&ls->ls_lkbidr, lkid);
- if (lkb)
- kref_get(&lkb->lkb_ref);
- spin_unlock(&ls->ls_lkbidr_spin);
- *lkb_ret = lkb;
- return lkb ? 0 : -ENOENT;
- }
- static void kill_lkb(struct kref *kref)
- {
- struct dlm_lkb *lkb = container_of(kref, struct dlm_lkb, lkb_ref);
- /* All work is done after the return from kref_put() so we
- can release the write_lock before the detach_lkb */
- DLM_ASSERT(!lkb->lkb_status, dlm_print_lkb(lkb););
- }
- /* __put_lkb() is used when an lkb may not have an rsb attached to
- it so we need to provide the lockspace explicitly */
- static int __put_lkb(struct dlm_ls *ls, struct dlm_lkb *lkb)
- {
- uint32_t lkid = lkb->lkb_id;
- spin_lock(&ls->ls_lkbidr_spin);
- if (kref_put(&lkb->lkb_ref, kill_lkb)) {
- idr_remove(&ls->ls_lkbidr, lkid);
- spin_unlock(&ls->ls_lkbidr_spin);
- detach_lkb(lkb);
- /* for local/process lkbs, lvbptr points to caller's lksb */
- if (lkb->lkb_lvbptr && is_master_copy(lkb))
- dlm_free_lvb(lkb->lkb_lvbptr);
- dlm_free_lkb(lkb);
- return 1;
- } else {
- spin_unlock(&ls->ls_lkbidr_spin);
- return 0;
- }
- }
- int dlm_put_lkb(struct dlm_lkb *lkb)
- {
- struct dlm_ls *ls;
- DLM_ASSERT(lkb->lkb_resource, dlm_print_lkb(lkb););
- DLM_ASSERT(lkb->lkb_resource->res_ls, dlm_print_lkb(lkb););
- ls = lkb->lkb_resource->res_ls;
- return __put_lkb(ls, lkb);
- }
- /* This is only called to add a reference when the code already holds
- a valid reference to the lkb, so there's no need for locking. */
- static inline void hold_lkb(struct dlm_lkb *lkb)
- {
- kref_get(&lkb->lkb_ref);
- }
- /* This is called when we need to remove a reference and are certain
- it's not the last ref. e.g. del_lkb is always called between a
- find_lkb/put_lkb and is always the inverse of a previous add_lkb.
- put_lkb would work fine, but would involve unnecessary locking */
- static inline void unhold_lkb(struct dlm_lkb *lkb)
- {
- int rv;
- rv = kref_put(&lkb->lkb_ref, kill_lkb);
- DLM_ASSERT(!rv, dlm_print_lkb(lkb););
- }
- static void lkb_add_ordered(struct list_head *new, struct list_head *head,
- int mode)
- {
- struct dlm_lkb *lkb = NULL;
- list_for_each_entry(lkb, head, lkb_statequeue)
- if (lkb->lkb_rqmode < mode)
- break;
- __list_add(new, lkb->lkb_statequeue.prev, &lkb->lkb_statequeue);
- }
- /* add/remove lkb to rsb's grant/convert/wait queue */
- static void add_lkb(struct dlm_rsb *r, struct dlm_lkb *lkb, int status)
- {
- kref_get(&lkb->lkb_ref);
- DLM_ASSERT(!lkb->lkb_status, dlm_print_lkb(lkb););
- lkb->lkb_timestamp = ktime_get();
- lkb->lkb_status = status;
- switch (status) {
- case DLM_LKSTS_WAITING:
- if (lkb->lkb_exflags & DLM_LKF_HEADQUE)
- list_add(&lkb->lkb_statequeue, &r->res_waitqueue);
- else
- list_add_tail(&lkb->lkb_statequeue, &r->res_waitqueue);
- break;
- case DLM_LKSTS_GRANTED:
- /* convention says granted locks kept in order of grmode */
- lkb_add_ordered(&lkb->lkb_statequeue, &r->res_grantqueue,
- lkb->lkb_grmode);
- break;
- case DLM_LKSTS_CONVERT:
- if (lkb->lkb_exflags & DLM_LKF_HEADQUE)
- list_add(&lkb->lkb_statequeue, &r->res_convertqueue);
- else
- list_add_tail(&lkb->lkb_statequeue,
- &r->res_convertqueue);
- break;
- default:
- DLM_ASSERT(0, dlm_print_lkb(lkb); printk("sts=%d\n", status););
- }
- }
- static void del_lkb(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- lkb->lkb_status = 0;
- list_del(&lkb->lkb_statequeue);
- unhold_lkb(lkb);
- }
- static void move_lkb(struct dlm_rsb *r, struct dlm_lkb *lkb, int sts)
- {
- hold_lkb(lkb);
- del_lkb(r, lkb);
- add_lkb(r, lkb, sts);
- unhold_lkb(lkb);
- }
- static int msg_reply_type(int mstype)
- {
- switch (mstype) {
- case DLM_MSG_REQUEST:
- return DLM_MSG_REQUEST_REPLY;
- case DLM_MSG_CONVERT:
- return DLM_MSG_CONVERT_REPLY;
- case DLM_MSG_UNLOCK:
- return DLM_MSG_UNLOCK_REPLY;
- case DLM_MSG_CANCEL:
- return DLM_MSG_CANCEL_REPLY;
- case DLM_MSG_LOOKUP:
- return DLM_MSG_LOOKUP_REPLY;
- }
- return -1;
- }
- static int nodeid_warned(int nodeid, int num_nodes, int *warned)
- {
- int i;
- for (i = 0; i < num_nodes; i++) {
- if (!warned[i]) {
- warned[i] = nodeid;
- return 0;
- }
- if (warned[i] == nodeid)
- return 1;
- }
- return 0;
- }
- void dlm_scan_waiters(struct dlm_ls *ls)
- {
- struct dlm_lkb *lkb;
- ktime_t zero = ktime_set(0, 0);
- s64 us;
- s64 debug_maxus = 0;
- u32 debug_scanned = 0;
- u32 debug_expired = 0;
- int num_nodes = 0;
- int *warned = NULL;
- if (!dlm_config.ci_waitwarn_us)
- return;
- mutex_lock(&ls->ls_waiters_mutex);
- list_for_each_entry(lkb, &ls->ls_waiters, lkb_wait_reply) {
- if (ktime_equal(lkb->lkb_wait_time, zero))
- continue;
- debug_scanned++;
- us = ktime_to_us(ktime_sub(ktime_get(), lkb->lkb_wait_time));
- if (us < dlm_config.ci_waitwarn_us)
- continue;
- lkb->lkb_wait_time = zero;
- debug_expired++;
- if (us > debug_maxus)
- debug_maxus = us;
- if (!num_nodes) {
- num_nodes = ls->ls_num_nodes;
- warned = kzalloc(num_nodes * sizeof(int), GFP_KERNEL);
- }
- if (!warned)
- continue;
- if (nodeid_warned(lkb->lkb_wait_nodeid, num_nodes, warned))
- continue;
- log_error(ls, "waitwarn %x %lld %d us check connection to "
- "node %d", lkb->lkb_id, (long long)us,
- dlm_config.ci_waitwarn_us, lkb->lkb_wait_nodeid);
- }
- mutex_unlock(&ls->ls_waiters_mutex);
- kfree(warned);
- if (debug_expired)
- log_debug(ls, "scan_waiters %u warn %u over %d us max %lld us",
- debug_scanned, debug_expired,
- dlm_config.ci_waitwarn_us, (long long)debug_maxus);
- }
- /* add/remove lkb from global waiters list of lkb's waiting for
- a reply from a remote node */
- static int add_to_waiters(struct dlm_lkb *lkb, int mstype, int to_nodeid)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- int error = 0;
- mutex_lock(&ls->ls_waiters_mutex);
- if (is_overlap_unlock(lkb) ||
- (is_overlap_cancel(lkb) && (mstype == DLM_MSG_CANCEL))) {
- error = -EINVAL;
- goto out;
- }
- if (lkb->lkb_wait_type || is_overlap_cancel(lkb)) {
- switch (mstype) {
- case DLM_MSG_UNLOCK:
- lkb->lkb_flags |= DLM_IFL_OVERLAP_UNLOCK;
- break;
- case DLM_MSG_CANCEL:
- lkb->lkb_flags |= DLM_IFL_OVERLAP_CANCEL;
- break;
- default:
- error = -EBUSY;
- goto out;
- }
- lkb->lkb_wait_count++;
- hold_lkb(lkb);
- log_debug(ls, "addwait %x cur %d overlap %d count %d f %x",
- lkb->lkb_id, lkb->lkb_wait_type, mstype,
- lkb->lkb_wait_count, lkb->lkb_flags);
- goto out;
- }
- DLM_ASSERT(!lkb->lkb_wait_count,
- dlm_print_lkb(lkb);
- printk("wait_count %d\n", lkb->lkb_wait_count););
- lkb->lkb_wait_count++;
- lkb->lkb_wait_type = mstype;
- lkb->lkb_wait_time = ktime_get();
- lkb->lkb_wait_nodeid = to_nodeid; /* for debugging */
- hold_lkb(lkb);
- list_add(&lkb->lkb_wait_reply, &ls->ls_waiters);
- out:
- if (error)
- log_error(ls, "addwait error %x %d flags %x %d %d %s",
- lkb->lkb_id, error, lkb->lkb_flags, mstype,
- lkb->lkb_wait_type, lkb->lkb_resource->res_name);
- mutex_unlock(&ls->ls_waiters_mutex);
- return error;
- }
- /* We clear the RESEND flag because we might be taking an lkb off the waiters
- list as part of process_requestqueue (e.g. a lookup that has an optimized
- request reply on the requestqueue) between dlm_recover_waiters_pre() which
- set RESEND and dlm_recover_waiters_post() */
- static int _remove_from_waiters(struct dlm_lkb *lkb, int mstype,
- struct dlm_message *ms)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- int overlap_done = 0;
- if (is_overlap_unlock(lkb) && (mstype == DLM_MSG_UNLOCK_REPLY)) {
- log_debug(ls, "remwait %x unlock_reply overlap", lkb->lkb_id);
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_UNLOCK;
- overlap_done = 1;
- goto out_del;
- }
- if (is_overlap_cancel(lkb) && (mstype == DLM_MSG_CANCEL_REPLY)) {
- log_debug(ls, "remwait %x cancel_reply overlap", lkb->lkb_id);
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- overlap_done = 1;
- goto out_del;
- }
- /* Cancel state was preemptively cleared by a successful convert,
- see next comment, nothing to do. */
- if ((mstype == DLM_MSG_CANCEL_REPLY) &&
- (lkb->lkb_wait_type != DLM_MSG_CANCEL)) {
- log_debug(ls, "remwait %x cancel_reply wait_type %d",
- lkb->lkb_id, lkb->lkb_wait_type);
- return -1;
- }
- /* Remove for the convert reply, and premptively remove for the
- cancel reply. A convert has been granted while there's still
- an outstanding cancel on it (the cancel is moot and the result
- in the cancel reply should be 0). We preempt the cancel reply
- because the app gets the convert result and then can follow up
- with another op, like convert. This subsequent op would see the
- lingering state of the cancel and fail with -EBUSY. */
- if ((mstype == DLM_MSG_CONVERT_REPLY) &&
- (lkb->lkb_wait_type == DLM_MSG_CONVERT) &&
- is_overlap_cancel(lkb) && ms && !ms->m_result) {
- log_debug(ls, "remwait %x convert_reply zap overlap_cancel",
- lkb->lkb_id);
- lkb->lkb_wait_type = 0;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- lkb->lkb_wait_count--;
- goto out_del;
- }
- /* N.B. type of reply may not always correspond to type of original
- msg due to lookup->request optimization, verify others? */
- if (lkb->lkb_wait_type) {
- lkb->lkb_wait_type = 0;
- goto out_del;
- }
- log_error(ls, "remwait error %x remote %d %x msg %d flags %x no wait",
- lkb->lkb_id, ms ? ms->m_header.h_nodeid : 0, lkb->lkb_remid,
- mstype, lkb->lkb_flags);
- return -1;
- out_del:
- /* the force-unlock/cancel has completed and we haven't recvd a reply
- to the op that was in progress prior to the unlock/cancel; we
- give up on any reply to the earlier op. FIXME: not sure when/how
- this would happen */
- if (overlap_done && lkb->lkb_wait_type) {
- log_error(ls, "remwait error %x reply %d wait_type %d overlap",
- lkb->lkb_id, mstype, lkb->lkb_wait_type);
- lkb->lkb_wait_count--;
- lkb->lkb_wait_type = 0;
- }
- DLM_ASSERT(lkb->lkb_wait_count, dlm_print_lkb(lkb););
- lkb->lkb_flags &= ~DLM_IFL_RESEND;
- lkb->lkb_wait_count--;
- if (!lkb->lkb_wait_count)
- list_del_init(&lkb->lkb_wait_reply);
- unhold_lkb(lkb);
- return 0;
- }
- static int remove_from_waiters(struct dlm_lkb *lkb, int mstype)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- int error;
- mutex_lock(&ls->ls_waiters_mutex);
- error = _remove_from_waiters(lkb, mstype, NULL);
- mutex_unlock(&ls->ls_waiters_mutex);
- return error;
- }
- /* Handles situations where we might be processing a "fake" or "stub" reply in
- which we can't try to take waiters_mutex again. */
- static int remove_from_waiters_ms(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- int error;
- if (ms->m_flags != DLM_IFL_STUB_MS)
- mutex_lock(&ls->ls_waiters_mutex);
- error = _remove_from_waiters(lkb, ms->m_type, ms);
- if (ms->m_flags != DLM_IFL_STUB_MS)
- mutex_unlock(&ls->ls_waiters_mutex);
- return error;
- }
- /* If there's an rsb for the same resource being removed, ensure
- that the remove message is sent before the new lookup message.
- It should be rare to need a delay here, but if not, then it may
- be worthwhile to add a proper wait mechanism rather than a delay. */
- static void wait_pending_remove(struct dlm_rsb *r)
- {
- struct dlm_ls *ls = r->res_ls;
- restart:
- spin_lock(&ls->ls_remove_spin);
- if (ls->ls_remove_len &&
- !rsb_cmp(r, ls->ls_remove_name, ls->ls_remove_len)) {
- log_debug(ls, "delay lookup for remove dir %d %s",
- r->res_dir_nodeid, r->res_name);
- spin_unlock(&ls->ls_remove_spin);
- msleep(1);
- goto restart;
- }
- spin_unlock(&ls->ls_remove_spin);
- }
- /*
- * ls_remove_spin protects ls_remove_name and ls_remove_len which are
- * read by other threads in wait_pending_remove. ls_remove_names
- * and ls_remove_lens are only used by the scan thread, so they do
- * not need protection.
- */
- static void shrink_bucket(struct dlm_ls *ls, int b)
- {
- struct rb_node *n, *next;
- struct dlm_rsb *r;
- char *name;
- int our_nodeid = dlm_our_nodeid();
- int remote_count = 0;
- int need_shrink = 0;
- int i, len, rv;
- memset(&ls->ls_remove_lens, 0, sizeof(int) * DLM_REMOVE_NAMES_MAX);
- spin_lock(&ls->ls_rsbtbl[b].lock);
- if (!(ls->ls_rsbtbl[b].flags & DLM_RTF_SHRINK)) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return;
- }
- for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) {
- next = rb_next(n);
- r = rb_entry(n, struct dlm_rsb, res_hashnode);
- /* If we're the directory record for this rsb, and
- we're not the master of it, then we need to wait
- for the master node to send us a dir remove for
- before removing the dir record. */
- if (!dlm_no_directory(ls) &&
- (r->res_master_nodeid != our_nodeid) &&
- (dlm_dir_nodeid(r) == our_nodeid)) {
- continue;
- }
- need_shrink = 1;
- if (!time_after_eq(jiffies, r->res_toss_time +
- dlm_config.ci_toss_secs * HZ)) {
- continue;
- }
- if (!dlm_no_directory(ls) &&
- (r->res_master_nodeid == our_nodeid) &&
- (dlm_dir_nodeid(r) != our_nodeid)) {
- /* We're the master of this rsb but we're not
- the directory record, so we need to tell the
- dir node to remove the dir record. */
- ls->ls_remove_lens[remote_count] = r->res_length;
- memcpy(ls->ls_remove_names[remote_count], r->res_name,
- DLM_RESNAME_MAXLEN);
- remote_count++;
- if (remote_count >= DLM_REMOVE_NAMES_MAX)
- break;
- continue;
- }
- if (!kref_put(&r->res_ref, kill_rsb)) {
- log_error(ls, "tossed rsb in use %s", r->res_name);
- continue;
- }
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss);
- dlm_free_rsb(r);
- }
- if (need_shrink)
- ls->ls_rsbtbl[b].flags |= DLM_RTF_SHRINK;
- else
- ls->ls_rsbtbl[b].flags &= ~DLM_RTF_SHRINK;
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- /*
- * While searching for rsb's to free, we found some that require
- * remote removal. We leave them in place and find them again here
- * so there is a very small gap between removing them from the toss
- * list and sending the removal. Keeping this gap small is
- * important to keep us (the master node) from being out of sync
- * with the remote dir node for very long.
- *
- * From the time the rsb is removed from toss until just after
- * send_remove, the rsb name is saved in ls_remove_name. A new
- * lookup checks this to ensure that a new lookup message for the
- * same resource name is not sent just before the remove message.
- */
- for (i = 0; i < remote_count; i++) {
- name = ls->ls_remove_names[i];
- len = ls->ls_remove_lens[i];
- spin_lock(&ls->ls_rsbtbl[b].lock);
- rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (rv) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_debug(ls, "remove_name not toss %s", name);
- continue;
- }
- if (r->res_master_nodeid != our_nodeid) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_debug(ls, "remove_name master %d dir %d our %d %s",
- r->res_master_nodeid, r->res_dir_nodeid,
- our_nodeid, name);
- continue;
- }
- if (r->res_dir_nodeid == our_nodeid) {
- /* should never happen */
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_error(ls, "remove_name dir %d master %d our %d %s",
- r->res_dir_nodeid, r->res_master_nodeid,
- our_nodeid, name);
- continue;
- }
- if (!time_after_eq(jiffies, r->res_toss_time +
- dlm_config.ci_toss_secs * HZ)) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_debug(ls, "remove_name toss_time %lu now %lu %s",
- r->res_toss_time, jiffies, name);
- continue;
- }
- if (!kref_put(&r->res_ref, kill_rsb)) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_error(ls, "remove_name in use %s", name);
- continue;
- }
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss);
- /* block lookup of same name until we've sent remove */
- spin_lock(&ls->ls_remove_spin);
- ls->ls_remove_len = len;
- memcpy(ls->ls_remove_name, name, DLM_RESNAME_MAXLEN);
- spin_unlock(&ls->ls_remove_spin);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- send_remove(r);
- /* allow lookup of name again */
- spin_lock(&ls->ls_remove_spin);
- ls->ls_remove_len = 0;
- memset(ls->ls_remove_name, 0, DLM_RESNAME_MAXLEN);
- spin_unlock(&ls->ls_remove_spin);
- dlm_free_rsb(r);
- }
- }
- void dlm_scan_rsbs(struct dlm_ls *ls)
- {
- int i;
- for (i = 0; i < ls->ls_rsbtbl_size; i++) {
- shrink_bucket(ls, i);
- if (dlm_locking_stopped(ls))
- break;
- cond_resched();
- }
- }
- static void add_timeout(struct dlm_lkb *lkb)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- if (is_master_copy(lkb))
- return;
- if (test_bit(LSFL_TIMEWARN, &ls->ls_flags) &&
- !(lkb->lkb_exflags & DLM_LKF_NODLCKWT)) {
- lkb->lkb_flags |= DLM_IFL_WATCH_TIMEWARN;
- goto add_it;
- }
- if (lkb->lkb_exflags & DLM_LKF_TIMEOUT)
- goto add_it;
- return;
- add_it:
- DLM_ASSERT(list_empty(&lkb->lkb_time_list), dlm_print_lkb(lkb););
- mutex_lock(&ls->ls_timeout_mutex);
- hold_lkb(lkb);
- list_add_tail(&lkb->lkb_time_list, &ls->ls_timeout);
- mutex_unlock(&ls->ls_timeout_mutex);
- }
- static void del_timeout(struct dlm_lkb *lkb)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- mutex_lock(&ls->ls_timeout_mutex);
- if (!list_empty(&lkb->lkb_time_list)) {
- list_del_init(&lkb->lkb_time_list);
- unhold_lkb(lkb);
- }
- mutex_unlock(&ls->ls_timeout_mutex);
- }
- /* FIXME: is it safe to look at lkb_exflags, lkb_flags, lkb_timestamp, and
- lkb_lksb_timeout without lock_rsb? Note: we can't lock timeout_mutex
- and then lock rsb because of lock ordering in add_timeout. We may need
- to specify some special timeout-related bits in the lkb that are just to
- be accessed under the timeout_mutex. */
- void dlm_scan_timeout(struct dlm_ls *ls)
- {
- struct dlm_rsb *r;
- struct dlm_lkb *lkb;
- int do_cancel, do_warn;
- s64 wait_us;
- for (;;) {
- if (dlm_locking_stopped(ls))
- break;
- do_cancel = 0;
- do_warn = 0;
- mutex_lock(&ls->ls_timeout_mutex);
- list_for_each_entry(lkb, &ls->ls_timeout, lkb_time_list) {
- wait_us = ktime_to_us(ktime_sub(ktime_get(),
- lkb->lkb_timestamp));
- if ((lkb->lkb_exflags & DLM_LKF_TIMEOUT) &&
- wait_us >= (lkb->lkb_timeout_cs * 10000))
- do_cancel = 1;
- if ((lkb->lkb_flags & DLM_IFL_WATCH_TIMEWARN) &&
- wait_us >= dlm_config.ci_timewarn_cs * 10000)
- do_warn = 1;
- if (!do_cancel && !do_warn)
- continue;
- hold_lkb(lkb);
- break;
- }
- mutex_unlock(&ls->ls_timeout_mutex);
- if (!do_cancel && !do_warn)
- break;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- if (do_warn) {
- /* clear flag so we only warn once */
- lkb->lkb_flags &= ~DLM_IFL_WATCH_TIMEWARN;
- if (!(lkb->lkb_exflags & DLM_LKF_TIMEOUT))
- del_timeout(lkb);
- dlm_timeout_warn(lkb);
- }
- if (do_cancel) {
- log_debug(ls, "timeout cancel %x node %d %s",
- lkb->lkb_id, lkb->lkb_nodeid, r->res_name);
- lkb->lkb_flags &= ~DLM_IFL_WATCH_TIMEWARN;
- lkb->lkb_flags |= DLM_IFL_TIMEOUT_CANCEL;
- del_timeout(lkb);
- _cancel_lock(r, lkb);
- }
- unlock_rsb(r);
- unhold_rsb(r);
- dlm_put_lkb(lkb);
- }
- }
- /* This is only called by dlm_recoverd, and we rely on dlm_ls_stop() stopping
- dlm_recoverd before checking/setting ls_recover_begin. */
- void dlm_adjust_timeouts(struct dlm_ls *ls)
- {
- struct dlm_lkb *lkb;
- u64 adj_us = jiffies_to_usecs(jiffies - ls->ls_recover_begin);
- ls->ls_recover_begin = 0;
- mutex_lock(&ls->ls_timeout_mutex);
- list_for_each_entry(lkb, &ls->ls_timeout, lkb_time_list)
- lkb->lkb_timestamp = ktime_add_us(lkb->lkb_timestamp, adj_us);
- mutex_unlock(&ls->ls_timeout_mutex);
- if (!dlm_config.ci_waitwarn_us)
- return;
- mutex_lock(&ls->ls_waiters_mutex);
- list_for_each_entry(lkb, &ls->ls_waiters, lkb_wait_reply) {
- if (ktime_to_us(lkb->lkb_wait_time))
- lkb->lkb_wait_time = ktime_get();
- }
- mutex_unlock(&ls->ls_waiters_mutex);
- }
- /* lkb is master or local copy */
- static void set_lvb_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int b, len = r->res_ls->ls_lvblen;
- /* b=1 lvb returned to caller
- b=0 lvb written to rsb or invalidated
- b=-1 do nothing */
- b = dlm_lvb_operations[lkb->lkb_grmode + 1][lkb->lkb_rqmode + 1];
- if (b == 1) {
- if (!lkb->lkb_lvbptr)
- return;
- if (!(lkb->lkb_exflags & DLM_LKF_VALBLK))
- return;
- if (!r->res_lvbptr)
- return;
- memcpy(lkb->lkb_lvbptr, r->res_lvbptr, len);
- lkb->lkb_lvbseq = r->res_lvbseq;
- } else if (b == 0) {
- if (lkb->lkb_exflags & DLM_LKF_IVVALBLK) {
- rsb_set_flag(r, RSB_VALNOTVALID);
- return;
- }
- if (!lkb->lkb_lvbptr)
- return;
- if (!(lkb->lkb_exflags & DLM_LKF_VALBLK))
- return;
- if (!r->res_lvbptr)
- r->res_lvbptr = dlm_allocate_lvb(r->res_ls);
- if (!r->res_lvbptr)
- return;
- memcpy(r->res_lvbptr, lkb->lkb_lvbptr, len);
- r->res_lvbseq++;
- lkb->lkb_lvbseq = r->res_lvbseq;
- rsb_clear_flag(r, RSB_VALNOTVALID);
- }
- if (rsb_flag(r, RSB_VALNOTVALID))
- lkb->lkb_sbflags |= DLM_SBF_VALNOTVALID;
- }
- static void set_lvb_unlock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- if (lkb->lkb_grmode < DLM_LOCK_PW)
- return;
- if (lkb->lkb_exflags & DLM_LKF_IVVALBLK) {
- rsb_set_flag(r, RSB_VALNOTVALID);
- return;
- }
- if (!lkb->lkb_lvbptr)
- return;
- if (!(lkb->lkb_exflags & DLM_LKF_VALBLK))
- return;
- if (!r->res_lvbptr)
- r->res_lvbptr = dlm_allocate_lvb(r->res_ls);
- if (!r->res_lvbptr)
- return;
- memcpy(r->res_lvbptr, lkb->lkb_lvbptr, r->res_ls->ls_lvblen);
- r->res_lvbseq++;
- rsb_clear_flag(r, RSB_VALNOTVALID);
- }
- /* lkb is process copy (pc) */
- static void set_lvb_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- int b;
- if (!lkb->lkb_lvbptr)
- return;
- if (!(lkb->lkb_exflags & DLM_LKF_VALBLK))
- return;
- b = dlm_lvb_operations[lkb->lkb_grmode + 1][lkb->lkb_rqmode + 1];
- if (b == 1) {
- int len = receive_extralen(ms);
- if (len > r->res_ls->ls_lvblen)
- len = r->res_ls->ls_lvblen;
- memcpy(lkb->lkb_lvbptr, ms->m_extra, len);
- lkb->lkb_lvbseq = ms->m_lvbseq;
- }
- }
- /* Manipulate lkb's on rsb's convert/granted/waiting queues
- remove_lock -- used for unlock, removes lkb from granted
- revert_lock -- used for cancel, moves lkb from convert to granted
- grant_lock -- used for request and convert, adds lkb to granted or
- moves lkb from convert or waiting to granted
- Each of these is used for master or local copy lkb's. There is
- also a _pc() variation used to make the corresponding change on
- a process copy (pc) lkb. */
- static void _remove_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- del_lkb(r, lkb);
- lkb->lkb_grmode = DLM_LOCK_IV;
- /* this unhold undoes the original ref from create_lkb()
- so this leads to the lkb being freed */
- unhold_lkb(lkb);
- }
- static void remove_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- set_lvb_unlock(r, lkb);
- _remove_lock(r, lkb);
- }
- static void remove_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- _remove_lock(r, lkb);
- }
- /* returns: 0 did nothing
- 1 moved lock to granted
- -1 removed lock */
- static int revert_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int rv = 0;
- lkb->lkb_rqmode = DLM_LOCK_IV;
- switch (lkb->lkb_status) {
- case DLM_LKSTS_GRANTED:
- break;
- case DLM_LKSTS_CONVERT:
- move_lkb(r, lkb, DLM_LKSTS_GRANTED);
- rv = 1;
- break;
- case DLM_LKSTS_WAITING:
- del_lkb(r, lkb);
- lkb->lkb_grmode = DLM_LOCK_IV;
- /* this unhold undoes the original ref from create_lkb()
- so this leads to the lkb being freed */
- unhold_lkb(lkb);
- rv = -1;
- break;
- default:
- log_print("invalid status for revert %d", lkb->lkb_status);
- }
- return rv;
- }
- static int revert_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- return revert_lock(r, lkb);
- }
- static void _grant_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- if (lkb->lkb_grmode != lkb->lkb_rqmode) {
- lkb->lkb_grmode = lkb->lkb_rqmode;
- if (lkb->lkb_status)
- move_lkb(r, lkb, DLM_LKSTS_GRANTED);
- else
- add_lkb(r, lkb, DLM_LKSTS_GRANTED);
- }
- lkb->lkb_rqmode = DLM_LOCK_IV;
- lkb->lkb_highbast = 0;
- }
- static void grant_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- set_lvb_lock(r, lkb);
- _grant_lock(r, lkb);
- }
- static void grant_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- set_lvb_lock_pc(r, lkb, ms);
- _grant_lock(r, lkb);
- }
- /* called by grant_pending_locks() which means an async grant message must
- be sent to the requesting node in addition to granting the lock if the
- lkb belongs to a remote node. */
- static void grant_lock_pending(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- grant_lock(r, lkb);
- if (is_master_copy(lkb))
- send_grant(r, lkb);
- else
- queue_cast(r, lkb, 0);
- }
- /* The special CONVDEADLK, ALTPR and ALTCW flags allow the master to
- change the granted/requested modes. We're munging things accordingly in
- the process copy.
- CONVDEADLK: our grmode may have been forced down to NL to resolve a
- conversion deadlock
- ALTPR/ALTCW: our rqmode may have been changed to PR or CW to become
- compatible with other granted locks */
- static void munge_demoted(struct dlm_lkb *lkb)
- {
- if (lkb->lkb_rqmode == DLM_LOCK_IV || lkb->lkb_grmode == DLM_LOCK_IV) {
- log_print("munge_demoted %x invalid modes gr %d rq %d",
- lkb->lkb_id, lkb->lkb_grmode, lkb->lkb_rqmode);
- return;
- }
- lkb->lkb_grmode = DLM_LOCK_NL;
- }
- static void munge_altmode(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- if (ms->m_type != DLM_MSG_REQUEST_REPLY &&
- ms->m_type != DLM_MSG_GRANT) {
- log_print("munge_altmode %x invalid reply type %d",
- lkb->lkb_id, ms->m_type);
- return;
- }
- if (lkb->lkb_exflags & DLM_LKF_ALTPR)
- lkb->lkb_rqmode = DLM_LOCK_PR;
- else if (lkb->lkb_exflags & DLM_LKF_ALTCW)
- lkb->lkb_rqmode = DLM_LOCK_CW;
- else {
- log_print("munge_altmode invalid exflags %x", lkb->lkb_exflags);
- dlm_print_lkb(lkb);
- }
- }
- static inline int first_in_list(struct dlm_lkb *lkb, struct list_head *head)
- {
- struct dlm_lkb *first = list_entry(head->next, struct dlm_lkb,
- lkb_statequeue);
- if (lkb->lkb_id == first->lkb_id)
- return 1;
- return 0;
- }
- /* Check if the given lkb conflicts with another lkb on the queue. */
- static int queue_conflict(struct list_head *head, struct dlm_lkb *lkb)
- {
- struct dlm_lkb *this;
- list_for_each_entry(this, head, lkb_statequeue) {
- if (this == lkb)
- continue;
- if (!modes_compat(this, lkb))
- return 1;
- }
- return 0;
- }
- /*
- * "A conversion deadlock arises with a pair of lock requests in the converting
- * queue for one resource. The granted mode of each lock blocks the requested
- * mode of the other lock."
- *
- * Part 2: if the granted mode of lkb is preventing an earlier lkb in the
- * convert queue from being granted, then deadlk/demote lkb.
- *
- * Example:
- * Granted Queue: empty
- * Convert Queue: NL->EX (first lock)
- * PR->EX (second lock)
- *
- * The first lock can't be granted because of the granted mode of the second
- * lock and the second lock can't be granted because it's not first in the
- * list. We either cancel lkb's conversion (PR->EX) and return EDEADLK, or we
- * demote the granted mode of lkb (from PR to NL) if it has the CONVDEADLK
- * flag set and return DEMOTED in the lksb flags.
- *
- * Originally, this function detected conv-deadlk in a more limited scope:
- * - if !modes_compat(lkb1, lkb2) && !modes_compat(lkb2, lkb1), or
- * - if lkb1 was the first entry in the queue (not just earlier), and was
- * blocked by the granted mode of lkb2, and there was nothing on the
- * granted queue preventing lkb1 from being granted immediately, i.e.
- * lkb2 was the only thing preventing lkb1 from being granted.
- *
- * That second condition meant we'd only say there was conv-deadlk if
- * resolving it (by demotion) would lead to the first lock on the convert
- * queue being granted right away. It allowed conversion deadlocks to exist
- * between locks on the convert queue while they couldn't be granted anyway.
- *
- * Now, we detect and take action on conversion deadlocks immediately when
- * they're created, even if they may not be immediately consequential. If
- * lkb1 exists anywhere in the convert queue and lkb2 comes in with a granted
- * mode that would prevent lkb1's conversion from being granted, we do a
- * deadlk/demote on lkb2 right away and don't let it onto the convert queue.
- * I think this means that the lkb_is_ahead condition below should always
- * be zero, i.e. there will never be conv-deadlk between two locks that are
- * both already on the convert queue.
- */
- static int conversion_deadlock_detect(struct dlm_rsb *r, struct dlm_lkb *lkb2)
- {
- struct dlm_lkb *lkb1;
- int lkb_is_ahead = 0;
- list_for_each_entry(lkb1, &r->res_convertqueue, lkb_statequeue) {
- if (lkb1 == lkb2) {
- lkb_is_ahead = 1;
- continue;
- }
- if (!lkb_is_ahead) {
- if (!modes_compat(lkb2, lkb1))
- return 1;
- } else {
- if (!modes_compat(lkb2, lkb1) &&
- !modes_compat(lkb1, lkb2))
- return 1;
- }
- }
- return 0;
- }
- /*
- * Return 1 if the lock can be granted, 0 otherwise.
- * Also detect and resolve conversion deadlocks.
- *
- * lkb is the lock to be granted
- *
- * now is 1 if the function is being called in the context of the
- * immediate request, it is 0 if called later, after the lock has been
- * queued.
- *
- * recover is 1 if dlm_recover_grant() is trying to grant conversions
- * after recovery.
- *
- * References are from chapter 6 of "VAXcluster Principles" by Roy Davis
- */
- static int _can_be_granted(struct dlm_rsb *r, struct dlm_lkb *lkb, int now,
- int recover)
- {
- int8_t conv = (lkb->lkb_grmode != DLM_LOCK_IV);
- /*
- * 6-10: Version 5.4 introduced an option to address the phenomenon of
- * a new request for a NL mode lock being blocked.
- *
- * 6-11: If the optional EXPEDITE flag is used with the new NL mode
- * request, then it would be granted. In essence, the use of this flag
- * tells the Lock Manager to expedite theis request by not considering
- * what may be in the CONVERTING or WAITING queues... As of this
- * writing, the EXPEDITE flag can be used only with new requests for NL
- * mode locks. This flag is not valid for conversion requests.
- *
- * A shortcut. Earlier checks return an error if EXPEDITE is used in a
- * conversion or used with a non-NL requested mode. We also know an
- * EXPEDITE request is always granted immediately, so now must always
- * be 1. The full condition to grant an expedite request: (now &&
- * !conv && lkb->rqmode == DLM_LOCK_NL && (flags & EXPEDITE)) can
- * therefore be shortened to just checking the flag.
- */
- if (lkb->lkb_exflags & DLM_LKF_EXPEDITE)
- return 1;
- /*
- * A shortcut. Without this, !queue_conflict(grantqueue, lkb) would be
- * added to the remaining conditions.
- */
- if (queue_conflict(&r->res_grantqueue, lkb))
- return 0;
- /*
- * 6-3: By default, a conversion request is immediately granted if the
- * requested mode is compatible with the modes of all other granted
- * locks
- */
- if (queue_conflict(&r->res_convertqueue, lkb))
- return 0;
- /*
- * The RECOVER_GRANT flag means dlm_recover_grant() is granting
- * locks for a recovered rsb, on which lkb's have been rebuilt.
- * The lkb's may have been rebuilt on the queues in a different
- * order than they were in on the previous master. So, granting
- * queued conversions in order after recovery doesn't make sense
- * since the order hasn't been preserved anyway. The new order
- * could also have created a new "in place" conversion deadlock.
- * (e.g. old, failed master held granted EX, with PR->EX, NL->EX.
- * After recovery, there would be no granted locks, and possibly
- * NL->EX, PR->EX, an in-place conversion deadlock.) So, after
- * recovery, grant conversions without considering order.
- */
- if (conv && recover)
- return 1;
- /*
- * 6-5: But the default algorithm for deciding whether to grant or
- * queue conversion requests does not by itself guarantee that such
- * requests are serviced on a "first come first serve" basis. This, in
- * turn, can lead to a phenomenon known as "indefinate postponement".
- *
- * 6-7: This issue is dealt with by using the optional QUECVT flag with
- * the system service employed to request a lock conversion. This flag
- * forces certain conversion requests to be queued, even if they are
- * compatible with the granted modes of other locks on the same
- * resource. Thus, the use of this flag results in conversion requests
- * being ordered on a "first come first servce" basis.
- *
- * DCT: This condition is all about new conversions being able to occur
- * "in place" while the lock remains on the granted queue (assuming
- * nothing else conflicts.) IOW if QUECVT isn't set, a conversion
- * doesn't _have_ to go onto the convert queue where it's processed in
- * order. The "now" variable is necessary to distinguish converts
- * being received and processed for the first time now, because once a
- * convert is moved to the conversion queue the condition below applies
- * requiring fifo granting.
- */
- if (now && conv && !(lkb->lkb_exflags & DLM_LKF_QUECVT))
- return 1;
- /*
- * Even if the convert is compat with all granted locks,
- * QUECVT forces it behind other locks on the convert queue.
- */
- if (now && conv && (lkb->lkb_exflags & DLM_LKF_QUECVT)) {
- if (list_empty(&r->res_convertqueue))
- return 1;
- else
- return 0;
- }
- /*
- * The NOORDER flag is set to avoid the standard vms rules on grant
- * order.
- */
- if (lkb->lkb_exflags & DLM_LKF_NOORDER)
- return 1;
- /*
- * 6-3: Once in that queue [CONVERTING], a conversion request cannot be
- * granted until all other conversion requests ahead of it are granted
- * and/or canceled.
- */
- if (!now && conv && first_in_list(lkb, &r->res_convertqueue))
- return 1;
- /*
- * 6-4: By default, a new request is immediately granted only if all
- * three of the following conditions are satisfied when the request is
- * issued:
- * - The queue of ungranted conversion requests for the resource is
- * empty.
- * - The queue of ungranted new requests for the resource is empty.
- * - The mode of the new request is compatible with the most
- * restrictive mode of all granted locks on the resource.
- */
- if (now && !conv && list_empty(&r->res_convertqueue) &&
- list_empty(&r->res_waitqueue))
- return 1;
- /*
- * 6-4: Once a lock request is in the queue of ungranted new requests,
- * it cannot be granted until the queue of ungranted conversion
- * requests is empty, all ungranted new requests ahead of it are
- * granted and/or canceled, and it is compatible with the granted mode
- * of the most restrictive lock granted on the resource.
- */
- if (!now && !conv && list_empty(&r->res_convertqueue) &&
- first_in_list(lkb, &r->res_waitqueue))
- return 1;
- return 0;
- }
- static int can_be_granted(struct dlm_rsb *r, struct dlm_lkb *lkb, int now,
- int recover, int *err)
- {
- int rv;
- int8_t alt = 0, rqmode = lkb->lkb_rqmode;
- int8_t is_convert = (lkb->lkb_grmode != DLM_LOCK_IV);
- if (err)
- *err = 0;
- rv = _can_be_granted(r, lkb, now, recover);
- if (rv)
- goto out;
- /*
- * The CONVDEADLK flag is non-standard and tells the dlm to resolve
- * conversion deadlocks by demoting grmode to NL, otherwise the dlm
- * cancels one of the locks.
- */
- if (is_convert && can_be_queued(lkb) &&
- conversion_deadlock_detect(r, lkb)) {
- if (lkb->lkb_exflags & DLM_LKF_CONVDEADLK) {
- lkb->lkb_grmode = DLM_LOCK_NL;
- lkb->lkb_sbflags |= DLM_SBF_DEMOTED;
- } else if (!(lkb->lkb_exflags & DLM_LKF_NODLCKWT)) {
- if (err)
- *err = -EDEADLK;
- else {
- log_print("can_be_granted deadlock %x now %d",
- lkb->lkb_id, now);
- dlm_dump_rsb(r);
- }
- }
- goto out;
- }
- /*
- * The ALTPR and ALTCW flags are non-standard and tell the dlm to try
- * to grant a request in a mode other than the normal rqmode. It's a
- * simple way to provide a big optimization to applications that can
- * use them.
- */
- if (rqmode != DLM_LOCK_PR && (lkb->lkb_exflags & DLM_LKF_ALTPR))
- alt = DLM_LOCK_PR;
- else if (rqmode != DLM_LOCK_CW && (lkb->lkb_exflags & DLM_LKF_ALTCW))
- alt = DLM_LOCK_CW;
- if (alt) {
- lkb->lkb_rqmode = alt;
- rv = _can_be_granted(r, lkb, now, 0);
- if (rv)
- lkb->lkb_sbflags |= DLM_SBF_ALTMODE;
- else
- lkb->lkb_rqmode = rqmode;
- }
- out:
- return rv;
- }
- /* FIXME: I don't think that can_be_granted() can/will demote or find deadlock
- for locks pending on the convert list. Once verified (watch for these
- log_prints), we should be able to just call _can_be_granted() and not
- bother with the demote/deadlk cases here (and there's no easy way to deal
- with a deadlk here, we'd have to generate something like grant_lock with
- the deadlk error.) */
- /* Returns the highest requested mode of all blocked conversions; sets
- cw if there's a blocked conversion to DLM_LOCK_CW. */
- static int grant_pending_convert(struct dlm_rsb *r, int high, int *cw,
- unsigned int *count)
- {
- struct dlm_lkb *lkb, *s;
- int recover = rsb_flag(r, RSB_RECOVER_GRANT);
- int hi, demoted, quit, grant_restart, demote_restart;
- int deadlk;
- quit = 0;
- restart:
- grant_restart = 0;
- demote_restart = 0;
- hi = DLM_LOCK_IV;
- list_for_each_entry_safe(lkb, s, &r->res_convertqueue, lkb_statequeue) {
- demoted = is_demoted(lkb);
- deadlk = 0;
- if (can_be_granted(r, lkb, 0, recover, &deadlk)) {
- grant_lock_pending(r, lkb);
- grant_restart = 1;
- if (count)
- (*count)++;
- continue;
- }
- if (!demoted && is_demoted(lkb)) {
- log_print("WARN: pending demoted %x node %d %s",
- lkb->lkb_id, lkb->lkb_nodeid, r->res_name);
- demote_restart = 1;
- continue;
- }
- if (deadlk) {
- log_print("WARN: pending deadlock %x node %d %s",
- lkb->lkb_id, lkb->lkb_nodeid, r->res_name);
- dlm_dump_rsb(r);
- continue;
- }
- hi = max_t(int, lkb->lkb_rqmode, hi);
- if (cw && lkb->lkb_rqmode == DLM_LOCK_CW)
- *cw = 1;
- }
- if (grant_restart)
- goto restart;
- if (demote_restart && !quit) {
- quit = 1;
- goto restart;
- }
- return max_t(int, high, hi);
- }
- static int grant_pending_wait(struct dlm_rsb *r, int high, int *cw,
- unsigned int *count)
- {
- struct dlm_lkb *lkb, *s;
- list_for_each_entry_safe(lkb, s, &r->res_waitqueue, lkb_statequeue) {
- if (can_be_granted(r, lkb, 0, 0, NULL)) {
- grant_lock_pending(r, lkb);
- if (count)
- (*count)++;
- } else {
- high = max_t(int, lkb->lkb_rqmode, high);
- if (lkb->lkb_rqmode == DLM_LOCK_CW)
- *cw = 1;
- }
- }
- return high;
- }
- /* cw of 1 means there's a lock with a rqmode of DLM_LOCK_CW that's blocked
- on either the convert or waiting queue.
- high is the largest rqmode of all locks blocked on the convert or
- waiting queue. */
- static int lock_requires_bast(struct dlm_lkb *gr, int high, int cw)
- {
- if (gr->lkb_grmode == DLM_LOCK_PR && cw) {
- if (gr->lkb_highbast < DLM_LOCK_EX)
- return 1;
- return 0;
- }
- if (gr->lkb_highbast < high &&
- !__dlm_compat_matrix[gr->lkb_grmode+1][high+1])
- return 1;
- return 0;
- }
- static void grant_pending_locks(struct dlm_rsb *r, unsigned int *count)
- {
- struct dlm_lkb *lkb, *s;
- int high = DLM_LOCK_IV;
- int cw = 0;
- if (!is_master(r)) {
- log_print("grant_pending_locks r nodeid %d", r->res_nodeid);
- dlm_dump_rsb(r);
- return;
- }
- high = grant_pending_convert(r, high, &cw, count);
- high = grant_pending_wait(r, high, &cw, count);
- if (high == DLM_LOCK_IV)
- return;
- /*
- * If there are locks left on the wait/convert queue then send blocking
- * ASTs to granted locks based on the largest requested mode (high)
- * found above.
- */
- list_for_each_entry_safe(lkb, s, &r->res_grantqueue, lkb_statequeue) {
- if (lkb->lkb_bastfn && lock_requires_bast(lkb, high, cw)) {
- if (cw && high == DLM_LOCK_PR &&
- lkb->lkb_grmode == DLM_LOCK_PR)
- queue_bast(r, lkb, DLM_LOCK_CW);
- else
- queue_bast(r, lkb, high);
- lkb->lkb_highbast = high;
- }
- }
- }
- static int modes_require_bast(struct dlm_lkb *gr, struct dlm_lkb *rq)
- {
- if ((gr->lkb_grmode == DLM_LOCK_PR && rq->lkb_rqmode == DLM_LOCK_CW) ||
- (gr->lkb_grmode == DLM_LOCK_CW && rq->lkb_rqmode == DLM_LOCK_PR)) {
- if (gr->lkb_highbast < DLM_LOCK_EX)
- return 1;
- return 0;
- }
- if (gr->lkb_highbast < rq->lkb_rqmode && !modes_compat(gr, rq))
- return 1;
- return 0;
- }
- static void send_bast_queue(struct dlm_rsb *r, struct list_head *head,
- struct dlm_lkb *lkb)
- {
- struct dlm_lkb *gr;
- list_for_each_entry(gr, head, lkb_statequeue) {
- /* skip self when sending basts to convertqueue */
- if (gr == lkb)
- continue;
- if (gr->lkb_bastfn && modes_require_bast(gr, lkb)) {
- queue_bast(r, gr, lkb->lkb_rqmode);
- gr->lkb_highbast = lkb->lkb_rqmode;
- }
- }
- }
- static void send_blocking_asts(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- send_bast_queue(r, &r->res_grantqueue, lkb);
- }
- static void send_blocking_asts_all(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- send_bast_queue(r, &r->res_grantqueue, lkb);
- send_bast_queue(r, &r->res_convertqueue, lkb);
- }
- /* set_master(r, lkb) -- set the master nodeid of a resource
- The purpose of this function is to set the nodeid field in the given
- lkb using the nodeid field in the given rsb. If the rsb's nodeid is
- known, it can just be copied to the lkb and the function will return
- 0. If the rsb's nodeid is _not_ known, it needs to be looked up
- before it can be copied to the lkb.
- When the rsb nodeid is being looked up remotely, the initial lkb
- causing the lookup is kept on the ls_waiters list waiting for the
- lookup reply. Other lkb's waiting for the same rsb lookup are kept
- on the rsb's res_lookup list until the master is verified.
- Return values:
- 0: nodeid is set in rsb/lkb and the caller should go ahead and use it
- 1: the rsb master is not available and the lkb has been placed on
- a wait queue
- */
- static int set_master(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int our_nodeid = dlm_our_nodeid();
- if (rsb_flag(r, RSB_MASTER_UNCERTAIN)) {
- rsb_clear_flag(r, RSB_MASTER_UNCERTAIN);
- r->res_first_lkid = lkb->lkb_id;
- lkb->lkb_nodeid = r->res_nodeid;
- return 0;
- }
- if (r->res_first_lkid && r->res_first_lkid != lkb->lkb_id) {
- list_add_tail(&lkb->lkb_rsb_lookup, &r->res_lookup);
- return 1;
- }
- if (r->res_master_nodeid == our_nodeid) {
- lkb->lkb_nodeid = 0;
- return 0;
- }
- if (r->res_master_nodeid) {
- lkb->lkb_nodeid = r->res_master_nodeid;
- return 0;
- }
- if (dlm_dir_nodeid(r) == our_nodeid) {
- /* This is a somewhat unusual case; find_rsb will usually
- have set res_master_nodeid when dir nodeid is local, but
- there are cases where we become the dir node after we've
- past find_rsb and go through _request_lock again.
- confirm_master() or process_lookup_list() needs to be
- called after this. */
- log_debug(r->res_ls, "set_master %x self master %d dir %d %s",
- lkb->lkb_id, r->res_master_nodeid, r->res_dir_nodeid,
- r->res_name);
- r->res_master_nodeid = our_nodeid;
- r->res_nodeid = 0;
- lkb->lkb_nodeid = 0;
- return 0;
- }
- wait_pending_remove(r);
- r->res_first_lkid = lkb->lkb_id;
- send_lookup(r, lkb);
- return 1;
- }
- static void process_lookup_list(struct dlm_rsb *r)
- {
- struct dlm_lkb *lkb, *safe;
- list_for_each_entry_safe(lkb, safe, &r->res_lookup, lkb_rsb_lookup) {
- list_del_init(&lkb->lkb_rsb_lookup);
- _request_lock(r, lkb);
- schedule();
- }
- }
- /* confirm_master -- confirm (or deny) an rsb's master nodeid */
- static void confirm_master(struct dlm_rsb *r, int error)
- {
- struct dlm_lkb *lkb;
- if (!r->res_first_lkid)
- return;
- switch (error) {
- case 0:
- case -EINPROGRESS:
- r->res_first_lkid = 0;
- process_lookup_list(r);
- break;
- case -EAGAIN:
- case -EBADR:
- case -ENOTBLK:
- /* the remote request failed and won't be retried (it was
- a NOQUEUE, or has been canceled/unlocked); make a waiting
- lkb the first_lkid */
- r->res_first_lkid = 0;
- if (!list_empty(&r->res_lookup)) {
- lkb = list_entry(r->res_lookup.next, struct dlm_lkb,
- lkb_rsb_lookup);
- list_del_init(&lkb->lkb_rsb_lookup);
- r->res_first_lkid = lkb->lkb_id;
- _request_lock(r, lkb);
- }
- break;
- default:
- log_error(r->res_ls, "confirm_master unknown error %d", error);
- }
- }
- static int set_lock_args(int mode, struct dlm_lksb *lksb, uint32_t flags,
- int namelen, unsigned long timeout_cs,
- void (*ast) (void *astparam),
- void *astparam,
- void (*bast) (void *astparam, int mode),
- struct dlm_args *args)
- {
- int rv = -EINVAL;
- /* check for invalid arg usage */
- if (mode < 0 || mode > DLM_LOCK_EX)
- goto out;
- if (!(flags & DLM_LKF_CONVERT) && (namelen > DLM_RESNAME_MAXLEN))
- goto out;
- if (flags & DLM_LKF_CANCEL)
- goto out;
- if (flags & DLM_LKF_QUECVT && !(flags & DLM_LKF_CONVERT))
- goto out;
- if (flags & DLM_LKF_CONVDEADLK && !(flags & DLM_LKF_CONVERT))
- goto out;
- if (flags & DLM_LKF_CONVDEADLK && flags & DLM_LKF_NOQUEUE)
- goto out;
- if (flags & DLM_LKF_EXPEDITE && flags & DLM_LKF_CONVERT)
- goto out;
- if (flags & DLM_LKF_EXPEDITE && flags & DLM_LKF_QUECVT)
- goto out;
- if (flags & DLM_LKF_EXPEDITE && flags & DLM_LKF_NOQUEUE)
- goto out;
- if (flags & DLM_LKF_EXPEDITE && mode != DLM_LOCK_NL)
- goto out;
- if (!ast || !lksb)
- goto out;
- if (flags & DLM_LKF_VALBLK && !lksb->sb_lvbptr)
- goto out;
- if (flags & DLM_LKF_CONVERT && !lksb->sb_lkid)
- goto out;
- /* these args will be copied to the lkb in validate_lock_args,
- it cannot be done now because when converting locks, fields in
- an active lkb cannot be modified before locking the rsb */
- args->flags = flags;
- args->astfn = ast;
- args->astparam = astparam;
- args->bastfn = bast;
- args->timeout = timeout_cs;
- args->mode = mode;
- args->lksb = lksb;
- rv = 0;
- out:
- return rv;
- }
- static int set_unlock_args(uint32_t flags, void *astarg, struct dlm_args *args)
- {
- if (flags & ~(DLM_LKF_CANCEL | DLM_LKF_VALBLK | DLM_LKF_IVVALBLK |
- DLM_LKF_FORCEUNLOCK))
- return -EINVAL;
- if (flags & DLM_LKF_CANCEL && flags & DLM_LKF_FORCEUNLOCK)
- return -EINVAL;
- args->flags = flags;
- args->astparam = astarg;
- return 0;
- }
- static int validate_lock_args(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_args *args)
- {
- int rv = -EINVAL;
- if (args->flags & DLM_LKF_CONVERT) {
- if (lkb->lkb_flags & DLM_IFL_MSTCPY)
- goto out;
- if (args->flags & DLM_LKF_QUECVT &&
- !__quecvt_compat_matrix[lkb->lkb_grmode+1][args->mode+1])
- goto out;
- rv = -EBUSY;
- if (lkb->lkb_status != DLM_LKSTS_GRANTED)
- goto out;
- if (lkb->lkb_wait_type)
- goto out;
- if (is_overlap(lkb))
- goto out;
- }
- lkb->lkb_exflags = args->flags;
- lkb->lkb_sbflags = 0;
- lkb->lkb_astfn = args->astfn;
- lkb->lkb_astparam = args->astparam;
- lkb->lkb_bastfn = args->bastfn;
- lkb->lkb_rqmode = args->mode;
- lkb->lkb_lksb = args->lksb;
- lkb->lkb_lvbptr = args->lksb->sb_lvbptr;
- lkb->lkb_ownpid = (int) current->pid;
- lkb->lkb_timeout_cs = args->timeout;
- rv = 0;
- out:
- if (rv)
- log_debug(ls, "validate_lock_args %d %x %x %x %d %d %s",
- rv, lkb->lkb_id, lkb->lkb_flags, args->flags,
- lkb->lkb_status, lkb->lkb_wait_type,
- lkb->lkb_resource->res_name);
- return rv;
- }
- /* when dlm_unlock() sees -EBUSY with CANCEL/FORCEUNLOCK it returns 0
- for success */
- /* note: it's valid for lkb_nodeid/res_nodeid to be -1 when we get here
- because there may be a lookup in progress and it's valid to do
- cancel/unlockf on it */
- static int validate_unlock_args(struct dlm_lkb *lkb, struct dlm_args *args)
- {
- struct dlm_ls *ls = lkb->lkb_resource->res_ls;
- int rv = -EINVAL;
- if (lkb->lkb_flags & DLM_IFL_MSTCPY) {
- log_error(ls, "unlock on MSTCPY %x", lkb->lkb_id);
- dlm_print_lkb(lkb);
- goto out;
- }
- /* an lkb may still exist even though the lock is EOL'ed due to a
- cancel, unlock or failed noqueue request; an app can't use these
- locks; return same error as if the lkid had not been found at all */
- if (lkb->lkb_flags & DLM_IFL_ENDOFLIFE) {
- log_debug(ls, "unlock on ENDOFLIFE %x", lkb->lkb_id);
- rv = -ENOENT;
- goto out;
- }
- /* an lkb may be waiting for an rsb lookup to complete where the
- lookup was initiated by another lock */
- if (!list_empty(&lkb->lkb_rsb_lookup)) {
- if (args->flags & (DLM_LKF_CANCEL | DLM_LKF_FORCEUNLOCK)) {
- log_debug(ls, "unlock on rsb_lookup %x", lkb->lkb_id);
- list_del_init(&lkb->lkb_rsb_lookup);
- queue_cast(lkb->lkb_resource, lkb,
- args->flags & DLM_LKF_CANCEL ?
- -DLM_ECANCEL : -DLM_EUNLOCK);
- unhold_lkb(lkb); /* undoes create_lkb() */
- }
- /* caller changes -EBUSY to 0 for CANCEL and FORCEUNLOCK */
- rv = -EBUSY;
- goto out;
- }
- /* cancel not allowed with another cancel/unlock in progress */
- if (args->flags & DLM_LKF_CANCEL) {
- if (lkb->lkb_exflags & DLM_LKF_CANCEL)
- goto out;
- if (is_overlap(lkb))
- goto out;
- /* don't let scand try to do a cancel */
- del_timeout(lkb);
- if (lkb->lkb_flags & DLM_IFL_RESEND) {
- lkb->lkb_flags |= DLM_IFL_OVERLAP_CANCEL;
- rv = -EBUSY;
- goto out;
- }
- /* there's nothing to cancel */
- if (lkb->lkb_status == DLM_LKSTS_GRANTED &&
- !lkb->lkb_wait_type) {
- rv = -EBUSY;
- goto out;
- }
- switch (lkb->lkb_wait_type) {
- case DLM_MSG_LOOKUP:
- case DLM_MSG_REQUEST:
- lkb->lkb_flags |= DLM_IFL_OVERLAP_CANCEL;
- rv = -EBUSY;
- goto out;
- case DLM_MSG_UNLOCK:
- case DLM_MSG_CANCEL:
- goto out;
- }
- /* add_to_waiters() will set OVERLAP_CANCEL */
- goto out_ok;
- }
- /* do we need to allow a force-unlock if there's a normal unlock
- already in progress? in what conditions could the normal unlock
- fail such that we'd want to send a force-unlock to be sure? */
- if (args->flags & DLM_LKF_FORCEUNLOCK) {
- if (lkb->lkb_exflags & DLM_LKF_FORCEUNLOCK)
- goto out;
- if (is_overlap_unlock(lkb))
- goto out;
- /* don't let scand try to do a cancel */
- del_timeout(lkb);
- if (lkb->lkb_flags & DLM_IFL_RESEND) {
- lkb->lkb_flags |= DLM_IFL_OVERLAP_UNLOCK;
- rv = -EBUSY;
- goto out;
- }
- switch (lkb->lkb_wait_type) {
- case DLM_MSG_LOOKUP:
- case DLM_MSG_REQUEST:
- lkb->lkb_flags |= DLM_IFL_OVERLAP_UNLOCK;
- rv = -EBUSY;
- goto out;
- case DLM_MSG_UNLOCK:
- goto out;
- }
- /* add_to_waiters() will set OVERLAP_UNLOCK */
- goto out_ok;
- }
- /* normal unlock not allowed if there's any op in progress */
- rv = -EBUSY;
- if (lkb->lkb_wait_type || lkb->lkb_wait_count)
- goto out;
- out_ok:
- /* an overlapping op shouldn't blow away exflags from other op */
- lkb->lkb_exflags |= args->flags;
- lkb->lkb_sbflags = 0;
- lkb->lkb_astparam = args->astparam;
- rv = 0;
- out:
- if (rv)
- log_debug(ls, "validate_unlock_args %d %x %x %x %x %d %s", rv,
- lkb->lkb_id, lkb->lkb_flags, lkb->lkb_exflags,
- args->flags, lkb->lkb_wait_type,
- lkb->lkb_resource->res_name);
- return rv;
- }
- /*
- * Four stage 4 varieties:
- * do_request(), do_convert(), do_unlock(), do_cancel()
- * These are called on the master node for the given lock and
- * from the central locking logic.
- */
- static int do_request(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error = 0;
- if (can_be_granted(r, lkb, 1, 0, NULL)) {
- grant_lock(r, lkb);
- queue_cast(r, lkb, 0);
- goto out;
- }
- if (can_be_queued(lkb)) {
- error = -EINPROGRESS;
- add_lkb(r, lkb, DLM_LKSTS_WAITING);
- add_timeout(lkb);
- goto out;
- }
- error = -EAGAIN;
- queue_cast(r, lkb, -EAGAIN);
- out:
- return error;
- }
- static void do_request_effects(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int error)
- {
- switch (error) {
- case -EAGAIN:
- if (force_blocking_asts(lkb))
- send_blocking_asts_all(r, lkb);
- break;
- case -EINPROGRESS:
- send_blocking_asts(r, lkb);
- break;
- }
- }
- static int do_convert(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error = 0;
- int deadlk = 0;
- /* changing an existing lock may allow others to be granted */
- if (can_be_granted(r, lkb, 1, 0, &deadlk)) {
- grant_lock(r, lkb);
- queue_cast(r, lkb, 0);
- goto out;
- }
- /* can_be_granted() detected that this lock would block in a conversion
- deadlock, so we leave it on the granted queue and return EDEADLK in
- the ast for the convert. */
- if (deadlk) {
- /* it's left on the granted queue */
- revert_lock(r, lkb);
- queue_cast(r, lkb, -EDEADLK);
- error = -EDEADLK;
- goto out;
- }
- /* is_demoted() means the can_be_granted() above set the grmode
- to NL, and left us on the granted queue. This auto-demotion
- (due to CONVDEADLK) might mean other locks, and/or this lock, are
- now grantable. We have to try to grant other converting locks
- before we try again to grant this one. */
- if (is_demoted(lkb)) {
- grant_pending_convert(r, DLM_LOCK_IV, NULL, NULL);
- if (_can_be_granted(r, lkb, 1, 0)) {
- grant_lock(r, lkb);
- queue_cast(r, lkb, 0);
- goto out;
- }
- /* else fall through and move to convert queue */
- }
- if (can_be_queued(lkb)) {
- error = -EINPROGRESS;
- del_lkb(r, lkb);
- add_lkb(r, lkb, DLM_LKSTS_CONVERT);
- add_timeout(lkb);
- goto out;
- }
- error = -EAGAIN;
- queue_cast(r, lkb, -EAGAIN);
- out:
- return error;
- }
- static void do_convert_effects(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int error)
- {
- switch (error) {
- case 0:
- grant_pending_locks(r, NULL);
- /* grant_pending_locks also sends basts */
- break;
- case -EAGAIN:
- if (force_blocking_asts(lkb))
- send_blocking_asts_all(r, lkb);
- break;
- case -EINPROGRESS:
- send_blocking_asts(r, lkb);
- break;
- }
- }
- static int do_unlock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- remove_lock(r, lkb);
- queue_cast(r, lkb, -DLM_EUNLOCK);
- return -DLM_EUNLOCK;
- }
- static void do_unlock_effects(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int error)
- {
- grant_pending_locks(r, NULL);
- }
- /* returns: 0 did nothing, -DLM_ECANCEL canceled lock */
- static int do_cancel(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- error = revert_lock(r, lkb);
- if (error) {
- queue_cast(r, lkb, -DLM_ECANCEL);
- return -DLM_ECANCEL;
- }
- return 0;
- }
- static void do_cancel_effects(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int error)
- {
- if (error)
- grant_pending_locks(r, NULL);
- }
- /*
- * Four stage 3 varieties:
- * _request_lock(), _convert_lock(), _unlock_lock(), _cancel_lock()
- */
- /* add a new lkb to a possibly new rsb, called by requesting process */
- static int _request_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- /* set_master: sets lkb nodeid from r */
- error = set_master(r, lkb);
- if (error < 0)
- goto out;
- if (error) {
- error = 0;
- goto out;
- }
- if (is_remote(r)) {
- /* receive_request() calls do_request() on remote node */
- error = send_request(r, lkb);
- } else {
- error = do_request(r, lkb);
- /* for remote locks the request_reply is sent
- between do_request and do_request_effects */
- do_request_effects(r, lkb, error);
- }
- out:
- return error;
- }
- /* change some property of an existing lkb, e.g. mode */
- static int _convert_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- if (is_remote(r)) {
- /* receive_convert() calls do_convert() on remote node */
- error = send_convert(r, lkb);
- } else {
- error = do_convert(r, lkb);
- /* for remote locks the convert_reply is sent
- between do_convert and do_convert_effects */
- do_convert_effects(r, lkb, error);
- }
- return error;
- }
- /* remove an existing lkb from the granted queue */
- static int _unlock_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- if (is_remote(r)) {
- /* receive_unlock() calls do_unlock() on remote node */
- error = send_unlock(r, lkb);
- } else {
- error = do_unlock(r, lkb);
- /* for remote locks the unlock_reply is sent
- between do_unlock and do_unlock_effects */
- do_unlock_effects(r, lkb, error);
- }
- return error;
- }
- /* remove an existing lkb from the convert or wait queue */
- static int _cancel_lock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- if (is_remote(r)) {
- /* receive_cancel() calls do_cancel() on remote node */
- error = send_cancel(r, lkb);
- } else {
- error = do_cancel(r, lkb);
- /* for remote locks the cancel_reply is sent
- between do_cancel and do_cancel_effects */
- do_cancel_effects(r, lkb, error);
- }
- return error;
- }
- /*
- * Four stage 2 varieties:
- * request_lock(), convert_lock(), unlock_lock(), cancel_lock()
- */
- static int request_lock(struct dlm_ls *ls, struct dlm_lkb *lkb, char *name,
- int len, struct dlm_args *args)
- {
- struct dlm_rsb *r;
- int error;
- error = validate_lock_args(ls, lkb, args);
- if (error)
- return error;
- error = find_rsb(ls, name, len, 0, R_REQUEST, &r);
- if (error)
- return error;
- lock_rsb(r);
- attach_lkb(r, lkb);
- lkb->lkb_lksb->sb_lkid = lkb->lkb_id;
- error = _request_lock(r, lkb);
- unlock_rsb(r);
- put_rsb(r);
- return error;
- }
- static int convert_lock(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_args *args)
- {
- struct dlm_rsb *r;
- int error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_lock_args(ls, lkb, args);
- if (error)
- goto out;
- error = _convert_lock(r, lkb);
- out:
- unlock_rsb(r);
- put_rsb(r);
- return error;
- }
- static int unlock_lock(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_args *args)
- {
- struct dlm_rsb *r;
- int error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_unlock_args(lkb, args);
- if (error)
- goto out;
- error = _unlock_lock(r, lkb);
- out:
- unlock_rsb(r);
- put_rsb(r);
- return error;
- }
- static int cancel_lock(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_args *args)
- {
- struct dlm_rsb *r;
- int error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_unlock_args(lkb, args);
- if (error)
- goto out;
- error = _cancel_lock(r, lkb);
- out:
- unlock_rsb(r);
- put_rsb(r);
- return error;
- }
- /*
- * Two stage 1 varieties: dlm_lock() and dlm_unlock()
- */
- int dlm_lock(dlm_lockspace_t *lockspace,
- int mode,
- struct dlm_lksb *lksb,
- uint32_t flags,
- void *name,
- unsigned int namelen,
- uint32_t parent_lkid,
- void (*ast) (void *astarg),
- void *astarg,
- void (*bast) (void *astarg, int mode))
- {
- struct dlm_ls *ls;
- struct dlm_lkb *lkb;
- struct dlm_args args;
- int error, convert = flags & DLM_LKF_CONVERT;
- ls = dlm_find_lockspace_local(lockspace);
- if (!ls)
- return -EINVAL;
- dlm_lock_recovery(ls);
- if (convert)
- error = find_lkb(ls, lksb->sb_lkid, &lkb);
- else
- error = create_lkb(ls, &lkb);
- if (error)
- goto out;
- error = set_lock_args(mode, lksb, flags, namelen, 0, ast,
- astarg, bast, &args);
- if (error)
- goto out_put;
- if (convert)
- error = convert_lock(ls, lkb, &args);
- else
- error = request_lock(ls, lkb, name, namelen, &args);
- if (error == -EINPROGRESS)
- error = 0;
- out_put:
- if (convert || error)
- __put_lkb(ls, lkb);
- if (error == -EAGAIN || error == -EDEADLK)
- error = 0;
- out:
- dlm_unlock_recovery(ls);
- dlm_put_lockspace(ls);
- return error;
- }
- int dlm_unlock(dlm_lockspace_t *lockspace,
- uint32_t lkid,
- uint32_t flags,
- struct dlm_lksb *lksb,
- void *astarg)
- {
- struct dlm_ls *ls;
- struct dlm_lkb *lkb;
- struct dlm_args args;
- int error;
- ls = dlm_find_lockspace_local(lockspace);
- if (!ls)
- return -EINVAL;
- dlm_lock_recovery(ls);
- error = find_lkb(ls, lkid, &lkb);
- if (error)
- goto out;
- error = set_unlock_args(flags, astarg, &args);
- if (error)
- goto out_put;
- if (flags & DLM_LKF_CANCEL)
- error = cancel_lock(ls, lkb, &args);
- else
- error = unlock_lock(ls, lkb, &args);
- if (error == -DLM_EUNLOCK || error == -DLM_ECANCEL)
- error = 0;
- if (error == -EBUSY && (flags & (DLM_LKF_CANCEL | DLM_LKF_FORCEUNLOCK)))
- error = 0;
- out_put:
- dlm_put_lkb(lkb);
- out:
- dlm_unlock_recovery(ls);
- dlm_put_lockspace(ls);
- return error;
- }
- /*
- * send/receive routines for remote operations and replies
- *
- * send_args
- * send_common
- * send_request receive_request
- * send_convert receive_convert
- * send_unlock receive_unlock
- * send_cancel receive_cancel
- * send_grant receive_grant
- * send_bast receive_bast
- * send_lookup receive_lookup
- * send_remove receive_remove
- *
- * send_common_reply
- * receive_request_reply send_request_reply
- * receive_convert_reply send_convert_reply
- * receive_unlock_reply send_unlock_reply
- * receive_cancel_reply send_cancel_reply
- * receive_lookup_reply send_lookup_reply
- */
- static int _create_message(struct dlm_ls *ls, int mb_len,
- int to_nodeid, int mstype,
- struct dlm_message **ms_ret,
- struct dlm_mhandle **mh_ret)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- char *mb;
- /* get_buffer gives us a message handle (mh) that we need to
- pass into lowcomms_commit and a message buffer (mb) that we
- write our data into */
- mh = dlm_lowcomms_get_buffer(to_nodeid, mb_len, GFP_NOFS, &mb);
- if (!mh)
- return -ENOBUFS;
- memset(mb, 0, mb_len);
- ms = (struct dlm_message *) mb;
- ms->m_header.h_version = (DLM_HEADER_MAJOR | DLM_HEADER_MINOR);
- ms->m_header.h_lockspace = ls->ls_global_id;
- ms->m_header.h_nodeid = dlm_our_nodeid();
- ms->m_header.h_length = mb_len;
- ms->m_header.h_cmd = DLM_MSG;
- ms->m_type = mstype;
- *mh_ret = mh;
- *ms_ret = ms;
- return 0;
- }
- static int create_message(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int to_nodeid, int mstype,
- struct dlm_message **ms_ret,
- struct dlm_mhandle **mh_ret)
- {
- int mb_len = sizeof(struct dlm_message);
- switch (mstype) {
- case DLM_MSG_REQUEST:
- case DLM_MSG_LOOKUP:
- case DLM_MSG_REMOVE:
- mb_len += r->res_length;
- break;
- case DLM_MSG_CONVERT:
- case DLM_MSG_UNLOCK:
- case DLM_MSG_REQUEST_REPLY:
- case DLM_MSG_CONVERT_REPLY:
- case DLM_MSG_GRANT:
- if (lkb && lkb->lkb_lvbptr)
- mb_len += r->res_ls->ls_lvblen;
- break;
- }
- return _create_message(r->res_ls, mb_len, to_nodeid, mstype,
- ms_ret, mh_ret);
- }
- /* further lowcomms enhancements or alternate implementations may make
- the return value from this function useful at some point */
- static int send_message(struct dlm_mhandle *mh, struct dlm_message *ms)
- {
- dlm_message_out(ms);
- dlm_lowcomms_commit_buffer(mh);
- return 0;
- }
- static void send_args(struct dlm_rsb *r, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- ms->m_nodeid = lkb->lkb_nodeid;
- ms->m_pid = lkb->lkb_ownpid;
- ms->m_lkid = lkb->lkb_id;
- ms->m_remid = lkb->lkb_remid;
- ms->m_exflags = lkb->lkb_exflags;
- ms->m_sbflags = lkb->lkb_sbflags;
- ms->m_flags = lkb->lkb_flags;
- ms->m_lvbseq = lkb->lkb_lvbseq;
- ms->m_status = lkb->lkb_status;
- ms->m_grmode = lkb->lkb_grmode;
- ms->m_rqmode = lkb->lkb_rqmode;
- ms->m_hash = r->res_hash;
- /* m_result and m_bastmode are set from function args,
- not from lkb fields */
- if (lkb->lkb_bastfn)
- ms->m_asts |= DLM_CB_BAST;
- if (lkb->lkb_astfn)
- ms->m_asts |= DLM_CB_CAST;
- /* compare with switch in create_message; send_remove() doesn't
- use send_args() */
- switch (ms->m_type) {
- case DLM_MSG_REQUEST:
- case DLM_MSG_LOOKUP:
- memcpy(ms->m_extra, r->res_name, r->res_length);
- break;
- case DLM_MSG_CONVERT:
- case DLM_MSG_UNLOCK:
- case DLM_MSG_REQUEST_REPLY:
- case DLM_MSG_CONVERT_REPLY:
- case DLM_MSG_GRANT:
- if (!lkb->lkb_lvbptr)
- break;
- memcpy(ms->m_extra, lkb->lkb_lvbptr, r->res_ls->ls_lvblen);
- break;
- }
- }
- static int send_common(struct dlm_rsb *r, struct dlm_lkb *lkb, int mstype)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = r->res_nodeid;
- error = add_to_waiters(lkb, mstype, to_nodeid);
- if (error)
- return error;
- error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh);
- if (error)
- goto fail;
- send_args(r, lkb, ms);
- error = send_message(mh, ms);
- if (error)
- goto fail;
- return 0;
- fail:
- remove_from_waiters(lkb, msg_reply_type(mstype));
- return error;
- }
- static int send_request(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- return send_common(r, lkb, DLM_MSG_REQUEST);
- }
- static int send_convert(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- int error;
- error = send_common(r, lkb, DLM_MSG_CONVERT);
- /* down conversions go without a reply from the master */
- if (!error && down_conversion(lkb)) {
- remove_from_waiters(lkb, DLM_MSG_CONVERT_REPLY);
- r->res_ls->ls_stub_ms.m_flags = DLM_IFL_STUB_MS;
- r->res_ls->ls_stub_ms.m_type = DLM_MSG_CONVERT_REPLY;
- r->res_ls->ls_stub_ms.m_result = 0;
- __receive_convert_reply(r, lkb, &r->res_ls->ls_stub_ms);
- }
- return error;
- }
- /* FIXME: if this lkb is the only lock we hold on the rsb, then set
- MASTER_UNCERTAIN to force the next request on the rsb to confirm
- that the master is still correct. */
- static int send_unlock(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- return send_common(r, lkb, DLM_MSG_UNLOCK);
- }
- static int send_cancel(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- return send_common(r, lkb, DLM_MSG_CANCEL);
- }
- static int send_grant(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = lkb->lkb_nodeid;
- error = create_message(r, lkb, to_nodeid, DLM_MSG_GRANT, &ms, &mh);
- if (error)
- goto out;
- send_args(r, lkb, ms);
- ms->m_result = 0;
- error = send_message(mh, ms);
- out:
- return error;
- }
- static int send_bast(struct dlm_rsb *r, struct dlm_lkb *lkb, int mode)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = lkb->lkb_nodeid;
- error = create_message(r, NULL, to_nodeid, DLM_MSG_BAST, &ms, &mh);
- if (error)
- goto out;
- send_args(r, lkb, ms);
- ms->m_bastmode = mode;
- error = send_message(mh, ms);
- out:
- return error;
- }
- static int send_lookup(struct dlm_rsb *r, struct dlm_lkb *lkb)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = dlm_dir_nodeid(r);
- error = add_to_waiters(lkb, DLM_MSG_LOOKUP, to_nodeid);
- if (error)
- return error;
- error = create_message(r, NULL, to_nodeid, DLM_MSG_LOOKUP, &ms, &mh);
- if (error)
- goto fail;
- send_args(r, lkb, ms);
- error = send_message(mh, ms);
- if (error)
- goto fail;
- return 0;
- fail:
- remove_from_waiters(lkb, DLM_MSG_LOOKUP_REPLY);
- return error;
- }
- static int send_remove(struct dlm_rsb *r)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = dlm_dir_nodeid(r);
- error = create_message(r, NULL, to_nodeid, DLM_MSG_REMOVE, &ms, &mh);
- if (error)
- goto out;
- memcpy(ms->m_extra, r->res_name, r->res_length);
- ms->m_hash = r->res_hash;
- error = send_message(mh, ms);
- out:
- return error;
- }
- static int send_common_reply(struct dlm_rsb *r, struct dlm_lkb *lkb,
- int mstype, int rv)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int to_nodeid, error;
- to_nodeid = lkb->lkb_nodeid;
- error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh);
- if (error)
- goto out;
- send_args(r, lkb, ms);
- ms->m_result = rv;
- error = send_message(mh, ms);
- out:
- return error;
- }
- static int send_request_reply(struct dlm_rsb *r, struct dlm_lkb *lkb, int rv)
- {
- return send_common_reply(r, lkb, DLM_MSG_REQUEST_REPLY, rv);
- }
- static int send_convert_reply(struct dlm_rsb *r, struct dlm_lkb *lkb, int rv)
- {
- return send_common_reply(r, lkb, DLM_MSG_CONVERT_REPLY, rv);
- }
- static int send_unlock_reply(struct dlm_rsb *r, struct dlm_lkb *lkb, int rv)
- {
- return send_common_reply(r, lkb, DLM_MSG_UNLOCK_REPLY, rv);
- }
- static int send_cancel_reply(struct dlm_rsb *r, struct dlm_lkb *lkb, int rv)
- {
- return send_common_reply(r, lkb, DLM_MSG_CANCEL_REPLY, rv);
- }
- static int send_lookup_reply(struct dlm_ls *ls, struct dlm_message *ms_in,
- int ret_nodeid, int rv)
- {
- struct dlm_rsb *r = &ls->ls_stub_rsb;
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int error, nodeid = ms_in->m_header.h_nodeid;
- error = create_message(r, NULL, nodeid, DLM_MSG_LOOKUP_REPLY, &ms, &mh);
- if (error)
- goto out;
- ms->m_lkid = ms_in->m_lkid;
- ms->m_result = rv;
- ms->m_nodeid = ret_nodeid;
- error = send_message(mh, ms);
- out:
- return error;
- }
- /* which args we save from a received message depends heavily on the type
- of message, unlike the send side where we can safely send everything about
- the lkb for any type of message */
- static void receive_flags(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- lkb->lkb_exflags = ms->m_exflags;
- lkb->lkb_sbflags = ms->m_sbflags;
- lkb->lkb_flags = (lkb->lkb_flags & 0xFFFF0000) |
- (ms->m_flags & 0x0000FFFF);
- }
- static void receive_flags_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- if (ms->m_flags == DLM_IFL_STUB_MS)
- return;
- lkb->lkb_sbflags = ms->m_sbflags;
- lkb->lkb_flags = (lkb->lkb_flags & 0xFFFF0000) |
- (ms->m_flags & 0x0000FFFF);
- }
- static int receive_extralen(struct dlm_message *ms)
- {
- return (ms->m_header.h_length - sizeof(struct dlm_message));
- }
- static int receive_lvb(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- int len;
- if (lkb->lkb_exflags & DLM_LKF_VALBLK) {
- if (!lkb->lkb_lvbptr)
- lkb->lkb_lvbptr = dlm_allocate_lvb(ls);
- if (!lkb->lkb_lvbptr)
- return -ENOMEM;
- len = receive_extralen(ms);
- if (len > ls->ls_lvblen)
- len = ls->ls_lvblen;
- memcpy(lkb->lkb_lvbptr, ms->m_extra, len);
- }
- return 0;
- }
- static void fake_bastfn(void *astparam, int mode)
- {
- log_print("fake_bastfn should not be called");
- }
- static void fake_astfn(void *astparam)
- {
- log_print("fake_astfn should not be called");
- }
- static int receive_request_args(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- lkb->lkb_nodeid = ms->m_header.h_nodeid;
- lkb->lkb_ownpid = ms->m_pid;
- lkb->lkb_remid = ms->m_lkid;
- lkb->lkb_grmode = DLM_LOCK_IV;
- lkb->lkb_rqmode = ms->m_rqmode;
- lkb->lkb_bastfn = (ms->m_asts & DLM_CB_BAST) ? &fake_bastfn : NULL;
- lkb->lkb_astfn = (ms->m_asts & DLM_CB_CAST) ? &fake_astfn : NULL;
- if (lkb->lkb_exflags & DLM_LKF_VALBLK) {
- /* lkb was just created so there won't be an lvb yet */
- lkb->lkb_lvbptr = dlm_allocate_lvb(ls);
- if (!lkb->lkb_lvbptr)
- return -ENOMEM;
- }
- return 0;
- }
- static int receive_convert_args(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- if (lkb->lkb_status != DLM_LKSTS_GRANTED)
- return -EBUSY;
- if (receive_lvb(ls, lkb, ms))
- return -ENOMEM;
- lkb->lkb_rqmode = ms->m_rqmode;
- lkb->lkb_lvbseq = ms->m_lvbseq;
- return 0;
- }
- static int receive_unlock_args(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- if (receive_lvb(ls, lkb, ms))
- return -ENOMEM;
- return 0;
- }
- /* We fill in the stub-lkb fields with the info that send_xxxx_reply()
- uses to send a reply and that the remote end uses to process the reply. */
- static void setup_stub_lkb(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb = &ls->ls_stub_lkb;
- lkb->lkb_nodeid = ms->m_header.h_nodeid;
- lkb->lkb_remid = ms->m_lkid;
- }
- /* This is called after the rsb is locked so that we can safely inspect
- fields in the lkb. */
- static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- int from = ms->m_header.h_nodeid;
- int error = 0;
- switch (ms->m_type) {
- case DLM_MSG_CONVERT:
- case DLM_MSG_UNLOCK:
- case DLM_MSG_CANCEL:
- if (!is_master_copy(lkb) || lkb->lkb_nodeid != from)
- error = -EINVAL;
- break;
- case DLM_MSG_CONVERT_REPLY:
- case DLM_MSG_UNLOCK_REPLY:
- case DLM_MSG_CANCEL_REPLY:
- case DLM_MSG_GRANT:
- case DLM_MSG_BAST:
- if (!is_process_copy(lkb) || lkb->lkb_nodeid != from)
- error = -EINVAL;
- break;
- case DLM_MSG_REQUEST_REPLY:
- if (!is_process_copy(lkb))
- error = -EINVAL;
- else if (lkb->lkb_nodeid != -1 && lkb->lkb_nodeid != from)
- error = -EINVAL;
- break;
- default:
- error = -EINVAL;
- }
- if (error)
- log_error(lkb->lkb_resource->res_ls,
- "ignore invalid message %d from %d %x %x %x %d",
- ms->m_type, from, lkb->lkb_id, lkb->lkb_remid,
- lkb->lkb_flags, lkb->lkb_nodeid);
- return error;
- }
- static void send_repeat_remove(struct dlm_ls *ls, char *ms_name, int len)
- {
- char name[DLM_RESNAME_MAXLEN + 1];
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- struct dlm_rsb *r;
- uint32_t hash, b;
- int rv, dir_nodeid;
- memset(name, 0, sizeof(name));
- memcpy(name, ms_name, len);
- hash = jhash(name, len, 0);
- b = hash & (ls->ls_rsbtbl_size - 1);
- dir_nodeid = dlm_hash2nodeid(ls, hash);
- log_error(ls, "send_repeat_remove dir %d %s", dir_nodeid, name);
- spin_lock(&ls->ls_rsbtbl[b].lock);
- rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (!rv) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_error(ls, "repeat_remove on keep %s", name);
- return;
- }
- rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (!rv) {
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- log_error(ls, "repeat_remove on toss %s", name);
- return;
- }
- /* use ls->remove_name2 to avoid conflict with shrink? */
- spin_lock(&ls->ls_remove_spin);
- ls->ls_remove_len = len;
- memcpy(ls->ls_remove_name, name, DLM_RESNAME_MAXLEN);
- spin_unlock(&ls->ls_remove_spin);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- rv = _create_message(ls, sizeof(struct dlm_message) + len,
- dir_nodeid, DLM_MSG_REMOVE, &ms, &mh);
- if (rv)
- return;
- memcpy(ms->m_extra, name, len);
- ms->m_hash = hash;
- send_message(mh, ms);
- spin_lock(&ls->ls_remove_spin);
- ls->ls_remove_len = 0;
- memset(ls->ls_remove_name, 0, DLM_RESNAME_MAXLEN);
- spin_unlock(&ls->ls_remove_spin);
- }
- static int receive_request(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int from_nodeid;
- int error, namelen = 0;
- from_nodeid = ms->m_header.h_nodeid;
- error = create_lkb(ls, &lkb);
- if (error)
- goto fail;
- receive_flags(lkb, ms);
- lkb->lkb_flags |= DLM_IFL_MSTCPY;
- error = receive_request_args(ls, lkb, ms);
- if (error) {
- __put_lkb(ls, lkb);
- goto fail;
- }
- /* The dir node is the authority on whether we are the master
- for this rsb or not, so if the master sends us a request, we should
- recreate the rsb if we've destroyed it. This race happens when we
- send a remove message to the dir node at the same time that the dir
- node sends us a request for the rsb. */
- namelen = receive_extralen(ms);
- error = find_rsb(ls, ms->m_extra, namelen, from_nodeid,
- R_RECEIVE_REQUEST, &r);
- if (error) {
- __put_lkb(ls, lkb);
- goto fail;
- }
- lock_rsb(r);
- if (r->res_master_nodeid != dlm_our_nodeid()) {
- error = validate_master_nodeid(ls, r, from_nodeid);
- if (error) {
- unlock_rsb(r);
- put_rsb(r);
- __put_lkb(ls, lkb);
- goto fail;
- }
- }
- attach_lkb(r, lkb);
- error = do_request(r, lkb);
- send_request_reply(r, lkb, error);
- do_request_effects(r, lkb, error);
- unlock_rsb(r);
- put_rsb(r);
- if (error == -EINPROGRESS)
- error = 0;
- if (error)
- dlm_put_lkb(lkb);
- return 0;
- fail:
- /* TODO: instead of returning ENOTBLK, add the lkb to res_lookup
- and do this receive_request again from process_lookup_list once
- we get the lookup reply. This would avoid a many repeated
- ENOTBLK request failures when the lookup reply designating us
- as master is delayed. */
- /* We could repeatedly return -EBADR here if our send_remove() is
- delayed in being sent/arriving/being processed on the dir node.
- Another node would repeatedly lookup up the master, and the dir
- node would continue returning our nodeid until our send_remove
- took effect.
- We send another remove message in case our previous send_remove
- was lost/ignored/missed somehow. */
- if (error != -ENOTBLK) {
- log_limit(ls, "receive_request %x from %d %d",
- ms->m_lkid, from_nodeid, error);
- }
- if (namelen && error == -EBADR) {
- send_repeat_remove(ls, ms->m_extra, namelen);
- msleep(1000);
- }
- setup_stub_lkb(ls, ms);
- send_request_reply(&ls->ls_stub_rsb, &ls->ls_stub_lkb, error);
- return error;
- }
- static int receive_convert(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error, reply = 1;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- goto fail;
- if (lkb->lkb_remid != ms->m_lkid) {
- log_error(ls, "receive_convert %x remid %x recover_seq %llu "
- "remote %d %x", lkb->lkb_id, lkb->lkb_remid,
- (unsigned long long)lkb->lkb_recover_seq,
- ms->m_header.h_nodeid, ms->m_lkid);
- error = -ENOENT;
- dlm_put_lkb(lkb);
- goto fail;
- }
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- receive_flags(lkb, ms);
- error = receive_convert_args(ls, lkb, ms);
- if (error) {
- send_convert_reply(r, lkb, error);
- goto out;
- }
- reply = !down_conversion(lkb);
- error = do_convert(r, lkb);
- if (reply)
- send_convert_reply(r, lkb, error);
- do_convert_effects(r, lkb, error);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- fail:
- setup_stub_lkb(ls, ms);
- send_convert_reply(&ls->ls_stub_rsb, &ls->ls_stub_lkb, error);
- return error;
- }
- static int receive_unlock(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- goto fail;
- if (lkb->lkb_remid != ms->m_lkid) {
- log_error(ls, "receive_unlock %x remid %x remote %d %x",
- lkb->lkb_id, lkb->lkb_remid,
- ms->m_header.h_nodeid, ms->m_lkid);
- error = -ENOENT;
- dlm_put_lkb(lkb);
- goto fail;
- }
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- receive_flags(lkb, ms);
- error = receive_unlock_args(ls, lkb, ms);
- if (error) {
- send_unlock_reply(r, lkb, error);
- goto out;
- }
- error = do_unlock(r, lkb);
- send_unlock_reply(r, lkb, error);
- do_unlock_effects(r, lkb, error);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- fail:
- setup_stub_lkb(ls, ms);
- send_unlock_reply(&ls->ls_stub_rsb, &ls->ls_stub_lkb, error);
- return error;
- }
- static int receive_cancel(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- goto fail;
- receive_flags(lkb, ms);
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- error = do_cancel(r, lkb);
- send_cancel_reply(r, lkb, error);
- do_cancel_effects(r, lkb, error);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- fail:
- setup_stub_lkb(ls, ms);
- send_cancel_reply(&ls->ls_stub_rsb, &ls->ls_stub_lkb, error);
- return error;
- }
- static int receive_grant(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- receive_flags_reply(lkb, ms);
- if (is_altmode(lkb))
- munge_altmode(lkb, ms);
- grant_lock_pc(r, lkb, ms);
- queue_cast(r, lkb, 0);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- }
- static int receive_bast(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- queue_bast(r, lkb, ms->m_bastmode);
- lkb->lkb_highbast = ms->m_bastmode;
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- }
- static void receive_lookup(struct dlm_ls *ls, struct dlm_message *ms)
- {
- int len, error, ret_nodeid, from_nodeid, our_nodeid;
- from_nodeid = ms->m_header.h_nodeid;
- our_nodeid = dlm_our_nodeid();
- len = receive_extralen(ms);
- error = dlm_master_lookup(ls, from_nodeid, ms->m_extra, len, 0,
- &ret_nodeid, NULL);
- /* Optimization: we're master so treat lookup as a request */
- if (!error && ret_nodeid == our_nodeid) {
- receive_request(ls, ms);
- return;
- }
- send_lookup_reply(ls, ms, ret_nodeid, error);
- }
- static void receive_remove(struct dlm_ls *ls, struct dlm_message *ms)
- {
- char name[DLM_RESNAME_MAXLEN+1];
- struct dlm_rsb *r;
- uint32_t hash, b;
- int rv, len, dir_nodeid, from_nodeid;
- from_nodeid = ms->m_header.h_nodeid;
- len = receive_extralen(ms);
- if (len > DLM_RESNAME_MAXLEN) {
- log_error(ls, "receive_remove from %d bad len %d",
- from_nodeid, len);
- return;
- }
- dir_nodeid = dlm_hash2nodeid(ls, ms->m_hash);
- if (dir_nodeid != dlm_our_nodeid()) {
- log_error(ls, "receive_remove from %d bad nodeid %d",
- from_nodeid, dir_nodeid);
- return;
- }
- /* Look for name on rsbtbl.toss, if it's there, kill it.
- If it's on rsbtbl.keep, it's being used, and we should ignore this
- message. This is an expected race between the dir node sending a
- request to the master node at the same time as the master node sends
- a remove to the dir node. The resolution to that race is for the
- dir node to ignore the remove message, and the master node to
- recreate the master rsb when it gets a request from the dir node for
- an rsb it doesn't have. */
- memset(name, 0, sizeof(name));
- memcpy(name, ms->m_extra, len);
- hash = jhash(name, len, 0);
- b = hash & (ls->ls_rsbtbl_size - 1);
- spin_lock(&ls->ls_rsbtbl[b].lock);
- rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r);
- if (rv) {
- /* verify the rsb is on keep list per comment above */
- rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r);
- if (rv) {
- /* should not happen */
- log_error(ls, "receive_remove from %d not found %s",
- from_nodeid, name);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return;
- }
- if (r->res_master_nodeid != from_nodeid) {
- /* should not happen */
- log_error(ls, "receive_remove keep from %d master %d",
- from_nodeid, r->res_master_nodeid);
- dlm_print_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return;
- }
- log_debug(ls, "receive_remove from %d master %d first %x %s",
- from_nodeid, r->res_master_nodeid, r->res_first_lkid,
- name);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return;
- }
- if (r->res_master_nodeid != from_nodeid) {
- log_error(ls, "receive_remove toss from %d master %d",
- from_nodeid, r->res_master_nodeid);
- dlm_print_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- return;
- }
- if (kref_put(&r->res_ref, kill_rsb)) {
- rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- dlm_free_rsb(r);
- } else {
- log_error(ls, "receive_remove from %d rsb ref error",
- from_nodeid);
- dlm_print_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[b].lock);
- }
- }
- static void receive_purge(struct dlm_ls *ls, struct dlm_message *ms)
- {
- do_purge(ls, ms->m_nodeid, ms->m_pid);
- }
- static int receive_request_reply(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error, mstype, result;
- int from_nodeid = ms->m_header.h_nodeid;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- mstype = lkb->lkb_wait_type;
- error = remove_from_waiters(lkb, DLM_MSG_REQUEST_REPLY);
- if (error) {
- log_error(ls, "receive_request_reply %x remote %d %x result %d",
- lkb->lkb_id, from_nodeid, ms->m_lkid, ms->m_result);
- dlm_dump_rsb(r);
- goto out;
- }
- /* Optimization: the dir node was also the master, so it took our
- lookup as a request and sent request reply instead of lookup reply */
- if (mstype == DLM_MSG_LOOKUP) {
- r->res_master_nodeid = from_nodeid;
- r->res_nodeid = from_nodeid;
- lkb->lkb_nodeid = from_nodeid;
- }
- /* this is the value returned from do_request() on the master */
- result = ms->m_result;
- switch (result) {
- case -EAGAIN:
- /* request would block (be queued) on remote master */
- queue_cast(r, lkb, -EAGAIN);
- confirm_master(r, -EAGAIN);
- unhold_lkb(lkb); /* undoes create_lkb() */
- break;
- case -EINPROGRESS:
- case 0:
- /* request was queued or granted on remote master */
- receive_flags_reply(lkb, ms);
- lkb->lkb_remid = ms->m_lkid;
- if (is_altmode(lkb))
- munge_altmode(lkb, ms);
- if (result) {
- add_lkb(r, lkb, DLM_LKSTS_WAITING);
- add_timeout(lkb);
- } else {
- grant_lock_pc(r, lkb, ms);
- queue_cast(r, lkb, 0);
- }
- confirm_master(r, result);
- break;
- case -EBADR:
- case -ENOTBLK:
- /* find_rsb failed to find rsb or rsb wasn't master */
- log_limit(ls, "receive_request_reply %x from %d %d "
- "master %d dir %d first %x %s", lkb->lkb_id,
- from_nodeid, result, r->res_master_nodeid,
- r->res_dir_nodeid, r->res_first_lkid, r->res_name);
- if (r->res_dir_nodeid != dlm_our_nodeid() &&
- r->res_master_nodeid != dlm_our_nodeid()) {
- /* cause _request_lock->set_master->send_lookup */
- r->res_master_nodeid = 0;
- r->res_nodeid = -1;
- lkb->lkb_nodeid = -1;
- }
- if (is_overlap(lkb)) {
- /* we'll ignore error in cancel/unlock reply */
- queue_cast_overlap(r, lkb);
- confirm_master(r, result);
- unhold_lkb(lkb); /* undoes create_lkb() */
- } else {
- _request_lock(r, lkb);
- if (r->res_master_nodeid == dlm_our_nodeid())
- confirm_master(r, 0);
- }
- break;
- default:
- log_error(ls, "receive_request_reply %x error %d",
- lkb->lkb_id, result);
- }
- if (is_overlap_unlock(lkb) && (result == 0 || result == -EINPROGRESS)) {
- log_debug(ls, "receive_request_reply %x result %d unlock",
- lkb->lkb_id, result);
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_UNLOCK;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- send_unlock(r, lkb);
- } else if (is_overlap_cancel(lkb) && (result == -EINPROGRESS)) {
- log_debug(ls, "receive_request_reply %x cancel", lkb->lkb_id);
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_UNLOCK;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- send_cancel(r, lkb);
- } else {
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_UNLOCK;
- }
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- }
- static void __receive_convert_reply(struct dlm_rsb *r, struct dlm_lkb *lkb,
- struct dlm_message *ms)
- {
- /* this is the value returned from do_convert() on the master */
- switch (ms->m_result) {
- case -EAGAIN:
- /* convert would block (be queued) on remote master */
- queue_cast(r, lkb, -EAGAIN);
- break;
- case -EDEADLK:
- receive_flags_reply(lkb, ms);
- revert_lock_pc(r, lkb);
- queue_cast(r, lkb, -EDEADLK);
- break;
- case -EINPROGRESS:
- /* convert was queued on remote master */
- receive_flags_reply(lkb, ms);
- if (is_demoted(lkb))
- munge_demoted(lkb);
- del_lkb(r, lkb);
- add_lkb(r, lkb, DLM_LKSTS_CONVERT);
- add_timeout(lkb);
- break;
- case 0:
- /* convert was granted on remote master */
- receive_flags_reply(lkb, ms);
- if (is_demoted(lkb))
- munge_demoted(lkb);
- grant_lock_pc(r, lkb, ms);
- queue_cast(r, lkb, 0);
- break;
- default:
- log_error(r->res_ls, "receive_convert_reply %x remote %d %x %d",
- lkb->lkb_id, ms->m_header.h_nodeid, ms->m_lkid,
- ms->m_result);
- dlm_print_rsb(r);
- dlm_print_lkb(lkb);
- }
- }
- static void _receive_convert_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- struct dlm_rsb *r = lkb->lkb_resource;
- int error;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- /* stub reply can happen with waiters_mutex held */
- error = remove_from_waiters_ms(lkb, ms);
- if (error)
- goto out;
- __receive_convert_reply(r, lkb, ms);
- out:
- unlock_rsb(r);
- put_rsb(r);
- }
- static int receive_convert_reply(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- _receive_convert_reply(lkb, ms);
- dlm_put_lkb(lkb);
- return 0;
- }
- static void _receive_unlock_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- struct dlm_rsb *r = lkb->lkb_resource;
- int error;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- /* stub reply can happen with waiters_mutex held */
- error = remove_from_waiters_ms(lkb, ms);
- if (error)
- goto out;
- /* this is the value returned from do_unlock() on the master */
- switch (ms->m_result) {
- case -DLM_EUNLOCK:
- receive_flags_reply(lkb, ms);
- remove_lock_pc(r, lkb);
- queue_cast(r, lkb, -DLM_EUNLOCK);
- break;
- case -ENOENT:
- break;
- default:
- log_error(r->res_ls, "receive_unlock_reply %x error %d",
- lkb->lkb_id, ms->m_result);
- }
- out:
- unlock_rsb(r);
- put_rsb(r);
- }
- static int receive_unlock_reply(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- _receive_unlock_reply(lkb, ms);
- dlm_put_lkb(lkb);
- return 0;
- }
- static void _receive_cancel_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
- {
- struct dlm_rsb *r = lkb->lkb_resource;
- int error;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_message(lkb, ms);
- if (error)
- goto out;
- /* stub reply can happen with waiters_mutex held */
- error = remove_from_waiters_ms(lkb, ms);
- if (error)
- goto out;
- /* this is the value returned from do_cancel() on the master */
- switch (ms->m_result) {
- case -DLM_ECANCEL:
- receive_flags_reply(lkb, ms);
- revert_lock_pc(r, lkb);
- queue_cast(r, lkb, -DLM_ECANCEL);
- break;
- case 0:
- break;
- default:
- log_error(r->res_ls, "receive_cancel_reply %x error %d",
- lkb->lkb_id, ms->m_result);
- }
- out:
- unlock_rsb(r);
- put_rsb(r);
- }
- static int receive_cancel_reply(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- int error;
- error = find_lkb(ls, ms->m_remid, &lkb);
- if (error)
- return error;
- _receive_cancel_reply(lkb, ms);
- dlm_put_lkb(lkb);
- return 0;
- }
- static void receive_lookup_reply(struct dlm_ls *ls, struct dlm_message *ms)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error, ret_nodeid;
- int do_lookup_list = 0;
- error = find_lkb(ls, ms->m_lkid, &lkb);
- if (error) {
- log_error(ls, "receive_lookup_reply no lkid %x", ms->m_lkid);
- return;
- }
- /* ms->m_result is the value returned by dlm_master_lookup on dir node
- FIXME: will a non-zero error ever be returned? */
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = remove_from_waiters(lkb, DLM_MSG_LOOKUP_REPLY);
- if (error)
- goto out;
- ret_nodeid = ms->m_nodeid;
- /* We sometimes receive a request from the dir node for this
- rsb before we've received the dir node's loookup_reply for it.
- The request from the dir node implies we're the master, so we set
- ourself as master in receive_request_reply, and verify here that
- we are indeed the master. */
- if (r->res_master_nodeid && (r->res_master_nodeid != ret_nodeid)) {
- /* This should never happen */
- log_error(ls, "receive_lookup_reply %x from %d ret %d "
- "master %d dir %d our %d first %x %s",
- lkb->lkb_id, ms->m_header.h_nodeid, ret_nodeid,
- r->res_master_nodeid, r->res_dir_nodeid,
- dlm_our_nodeid(), r->res_first_lkid, r->res_name);
- }
- if (ret_nodeid == dlm_our_nodeid()) {
- r->res_master_nodeid = ret_nodeid;
- r->res_nodeid = 0;
- do_lookup_list = 1;
- r->res_first_lkid = 0;
- } else if (ret_nodeid == -1) {
- /* the remote node doesn't believe it's the dir node */
- log_error(ls, "receive_lookup_reply %x from %d bad ret_nodeid",
- lkb->lkb_id, ms->m_header.h_nodeid);
- r->res_master_nodeid = 0;
- r->res_nodeid = -1;
- lkb->lkb_nodeid = -1;
- } else {
- /* set_master() will set lkb_nodeid from r */
- r->res_master_nodeid = ret_nodeid;
- r->res_nodeid = ret_nodeid;
- }
- if (is_overlap(lkb)) {
- log_debug(ls, "receive_lookup_reply %x unlock %x",
- lkb->lkb_id, lkb->lkb_flags);
- queue_cast_overlap(r, lkb);
- unhold_lkb(lkb); /* undoes create_lkb() */
- goto out_list;
- }
- _request_lock(r, lkb);
- out_list:
- if (do_lookup_list)
- process_lookup_list(r);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- }
- static void _receive_message(struct dlm_ls *ls, struct dlm_message *ms,
- uint32_t saved_seq)
- {
- int error = 0, noent = 0;
- if (!dlm_is_member(ls, ms->m_header.h_nodeid)) {
- log_limit(ls, "receive %d from non-member %d %x %x %d",
- ms->m_type, ms->m_header.h_nodeid, ms->m_lkid,
- ms->m_remid, ms->m_result);
- return;
- }
- switch (ms->m_type) {
- /* messages sent to a master node */
- case DLM_MSG_REQUEST:
- error = receive_request(ls, ms);
- break;
- case DLM_MSG_CONVERT:
- error = receive_convert(ls, ms);
- break;
- case DLM_MSG_UNLOCK:
- error = receive_unlock(ls, ms);
- break;
- case DLM_MSG_CANCEL:
- noent = 1;
- error = receive_cancel(ls, ms);
- break;
- /* messages sent from a master node (replies to above) */
- case DLM_MSG_REQUEST_REPLY:
- error = receive_request_reply(ls, ms);
- break;
- case DLM_MSG_CONVERT_REPLY:
- error = receive_convert_reply(ls, ms);
- break;
- case DLM_MSG_UNLOCK_REPLY:
- error = receive_unlock_reply(ls, ms);
- break;
- case DLM_MSG_CANCEL_REPLY:
- error = receive_cancel_reply(ls, ms);
- break;
- /* messages sent from a master node (only two types of async msg) */
- case DLM_MSG_GRANT:
- noent = 1;
- error = receive_grant(ls, ms);
- break;
- case DLM_MSG_BAST:
- noent = 1;
- error = receive_bast(ls, ms);
- break;
- /* messages sent to a dir node */
- case DLM_MSG_LOOKUP:
- receive_lookup(ls, ms);
- break;
- case DLM_MSG_REMOVE:
- receive_remove(ls, ms);
- break;
- /* messages sent from a dir node (remove has no reply) */
- case DLM_MSG_LOOKUP_REPLY:
- receive_lookup_reply(ls, ms);
- break;
- /* other messages */
- case DLM_MSG_PURGE:
- receive_purge(ls, ms);
- break;
- default:
- log_error(ls, "unknown message type %d", ms->m_type);
- }
- /*
- * When checking for ENOENT, we're checking the result of
- * find_lkb(m_remid):
- *
- * The lock id referenced in the message wasn't found. This may
- * happen in normal usage for the async messages and cancel, so
- * only use log_debug for them.
- *
- * Some errors are expected and normal.
- */
- if (error == -ENOENT && noent) {
- log_debug(ls, "receive %d no %x remote %d %x saved_seq %u",
- ms->m_type, ms->m_remid, ms->m_header.h_nodeid,
- ms->m_lkid, saved_seq);
- } else if (error == -ENOENT) {
- log_error(ls, "receive %d no %x remote %d %x saved_seq %u",
- ms->m_type, ms->m_remid, ms->m_header.h_nodeid,
- ms->m_lkid, saved_seq);
- if (ms->m_type == DLM_MSG_CONVERT)
- dlm_dump_rsb_hash(ls, ms->m_hash);
- }
- if (error == -EINVAL) {
- log_error(ls, "receive %d inval from %d lkid %x remid %x "
- "saved_seq %u",
- ms->m_type, ms->m_header.h_nodeid,
- ms->m_lkid, ms->m_remid, saved_seq);
- }
- }
- /* If the lockspace is in recovery mode (locking stopped), then normal
- messages are saved on the requestqueue for processing after recovery is
- done. When not in recovery mode, we wait for dlm_recoverd to drain saved
- messages off the requestqueue before we process new ones. This occurs right
- after recovery completes when we transition from saving all messages on
- requestqueue, to processing all the saved messages, to processing new
- messages as they arrive. */
- static void dlm_receive_message(struct dlm_ls *ls, struct dlm_message *ms,
- int nodeid)
- {
- if (dlm_locking_stopped(ls)) {
- /* If we were a member of this lockspace, left, and rejoined,
- other nodes may still be sending us messages from the
- lockspace generation before we left. */
- if (!ls->ls_generation) {
- log_limit(ls, "receive %d from %d ignore old gen",
- ms->m_type, nodeid);
- return;
- }
- dlm_add_requestqueue(ls, nodeid, ms);
- } else {
- dlm_wait_requestqueue(ls);
- _receive_message(ls, ms, 0);
- }
- }
- /* This is called by dlm_recoverd to process messages that were saved on
- the requestqueue. */
- void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms,
- uint32_t saved_seq)
- {
- _receive_message(ls, ms, saved_seq);
- }
- /* This is called by the midcomms layer when something is received for
- the lockspace. It could be either a MSG (normal message sent as part of
- standard locking activity) or an RCOM (recovery message sent as part of
- lockspace recovery). */
- void dlm_receive_buffer(union dlm_packet *p, int nodeid)
- {
- struct dlm_header *hd = &p->header;
- struct dlm_ls *ls;
- int type = 0;
- switch (hd->h_cmd) {
- case DLM_MSG:
- dlm_message_in(&p->message);
- type = p->message.m_type;
- break;
- case DLM_RCOM:
- dlm_rcom_in(&p->rcom);
- type = p->rcom.rc_type;
- break;
- default:
- log_print("invalid h_cmd %d from %u", hd->h_cmd, nodeid);
- return;
- }
- if (hd->h_nodeid != nodeid) {
- log_print("invalid h_nodeid %d from %d lockspace %x",
- hd->h_nodeid, nodeid, hd->h_lockspace);
- return;
- }
- ls = dlm_find_lockspace_global(hd->h_lockspace);
- if (!ls) {
- if (dlm_config.ci_log_debug) {
- printk_ratelimited(KERN_DEBUG "dlm: invalid lockspace "
- "%u from %d cmd %d type %d\n",
- hd->h_lockspace, nodeid, hd->h_cmd, type);
- }
- if (hd->h_cmd == DLM_RCOM && type == DLM_RCOM_STATUS)
- dlm_send_ls_not_ready(nodeid, &p->rcom);
- return;
- }
- /* this rwsem allows dlm_ls_stop() to wait for all dlm_recv threads to
- be inactive (in this ls) before transitioning to recovery mode */
- down_read(&ls->ls_recv_active);
- if (hd->h_cmd == DLM_MSG)
- dlm_receive_message(ls, &p->message, nodeid);
- else
- dlm_receive_rcom(ls, &p->rcom, nodeid);
- up_read(&ls->ls_recv_active);
- dlm_put_lockspace(ls);
- }
- static void recover_convert_waiter(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_message *ms_stub)
- {
- if (middle_conversion(lkb)) {
- hold_lkb(lkb);
- memset(ms_stub, 0, sizeof(struct dlm_message));
- ms_stub->m_flags = DLM_IFL_STUB_MS;
- ms_stub->m_type = DLM_MSG_CONVERT_REPLY;
- ms_stub->m_result = -EINPROGRESS;
- ms_stub->m_header.h_nodeid = lkb->lkb_nodeid;
- _receive_convert_reply(lkb, ms_stub);
- /* Same special case as in receive_rcom_lock_args() */
- lkb->lkb_grmode = DLM_LOCK_IV;
- rsb_set_flag(lkb->lkb_resource, RSB_RECOVER_CONVERT);
- unhold_lkb(lkb);
- } else if (lkb->lkb_rqmode >= lkb->lkb_grmode) {
- lkb->lkb_flags |= DLM_IFL_RESEND;
- }
- /* lkb->lkb_rqmode < lkb->lkb_grmode shouldn't happen since down
- conversions are async; there's no reply from the remote master */
- }
- /* A waiting lkb needs recovery if the master node has failed, or
- the master node is changing (only when no directory is used) */
- static int waiter_needs_recovery(struct dlm_ls *ls, struct dlm_lkb *lkb,
- int dir_nodeid)
- {
- if (dlm_no_directory(ls))
- return 1;
- if (dlm_is_removed(ls, lkb->lkb_wait_nodeid))
- return 1;
- return 0;
- }
- /* Recovery for locks that are waiting for replies from nodes that are now
- gone. We can just complete unlocks and cancels by faking a reply from the
- dead node. Requests and up-conversions we flag to be resent after
- recovery. Down-conversions can just be completed with a fake reply like
- unlocks. Conversions between PR and CW need special attention. */
- void dlm_recover_waiters_pre(struct dlm_ls *ls)
- {
- struct dlm_lkb *lkb, *safe;
- struct dlm_message *ms_stub;
- int wait_type, stub_unlock_result, stub_cancel_result;
- int dir_nodeid;
- ms_stub = kmalloc(sizeof(struct dlm_message), GFP_KERNEL);
- if (!ms_stub) {
- log_error(ls, "dlm_recover_waiters_pre no mem");
- return;
- }
- mutex_lock(&ls->ls_waiters_mutex);
- list_for_each_entry_safe(lkb, safe, &ls->ls_waiters, lkb_wait_reply) {
- dir_nodeid = dlm_dir_nodeid(lkb->lkb_resource);
- /* exclude debug messages about unlocks because there can be so
- many and they aren't very interesting */
- if (lkb->lkb_wait_type != DLM_MSG_UNLOCK) {
- log_debug(ls, "waiter %x remote %x msg %d r_nodeid %d "
- "lkb_nodeid %d wait_nodeid %d dir_nodeid %d",
- lkb->lkb_id,
- lkb->lkb_remid,
- lkb->lkb_wait_type,
- lkb->lkb_resource->res_nodeid,
- lkb->lkb_nodeid,
- lkb->lkb_wait_nodeid,
- dir_nodeid);
- }
- /* all outstanding lookups, regardless of destination will be
- resent after recovery is done */
- if (lkb->lkb_wait_type == DLM_MSG_LOOKUP) {
- lkb->lkb_flags |= DLM_IFL_RESEND;
- continue;
- }
- if (!waiter_needs_recovery(ls, lkb, dir_nodeid))
- continue;
- wait_type = lkb->lkb_wait_type;
- stub_unlock_result = -DLM_EUNLOCK;
- stub_cancel_result = -DLM_ECANCEL;
- /* Main reply may have been received leaving a zero wait_type,
- but a reply for the overlapping op may not have been
- received. In that case we need to fake the appropriate
- reply for the overlap op. */
- if (!wait_type) {
- if (is_overlap_cancel(lkb)) {
- wait_type = DLM_MSG_CANCEL;
- if (lkb->lkb_grmode == DLM_LOCK_IV)
- stub_cancel_result = 0;
- }
- if (is_overlap_unlock(lkb)) {
- wait_type = DLM_MSG_UNLOCK;
- if (lkb->lkb_grmode == DLM_LOCK_IV)
- stub_unlock_result = -ENOENT;
- }
- log_debug(ls, "rwpre overlap %x %x %d %d %d",
- lkb->lkb_id, lkb->lkb_flags, wait_type,
- stub_cancel_result, stub_unlock_result);
- }
- switch (wait_type) {
- case DLM_MSG_REQUEST:
- lkb->lkb_flags |= DLM_IFL_RESEND;
- break;
- case DLM_MSG_CONVERT:
- recover_convert_waiter(ls, lkb, ms_stub);
- break;
- case DLM_MSG_UNLOCK:
- hold_lkb(lkb);
- memset(ms_stub, 0, sizeof(struct dlm_message));
- ms_stub->m_flags = DLM_IFL_STUB_MS;
- ms_stub->m_type = DLM_MSG_UNLOCK_REPLY;
- ms_stub->m_result = stub_unlock_result;
- ms_stub->m_header.h_nodeid = lkb->lkb_nodeid;
- _receive_unlock_reply(lkb, ms_stub);
- dlm_put_lkb(lkb);
- break;
- case DLM_MSG_CANCEL:
- hold_lkb(lkb);
- memset(ms_stub, 0, sizeof(struct dlm_message));
- ms_stub->m_flags = DLM_IFL_STUB_MS;
- ms_stub->m_type = DLM_MSG_CANCEL_REPLY;
- ms_stub->m_result = stub_cancel_result;
- ms_stub->m_header.h_nodeid = lkb->lkb_nodeid;
- _receive_cancel_reply(lkb, ms_stub);
- dlm_put_lkb(lkb);
- break;
- default:
- log_error(ls, "invalid lkb wait_type %d %d",
- lkb->lkb_wait_type, wait_type);
- }
- schedule();
- }
- mutex_unlock(&ls->ls_waiters_mutex);
- kfree(ms_stub);
- }
- static struct dlm_lkb *find_resend_waiter(struct dlm_ls *ls)
- {
- struct dlm_lkb *lkb;
- int found = 0;
- mutex_lock(&ls->ls_waiters_mutex);
- list_for_each_entry(lkb, &ls->ls_waiters, lkb_wait_reply) {
- if (lkb->lkb_flags & DLM_IFL_RESEND) {
- hold_lkb(lkb);
- found = 1;
- break;
- }
- }
- mutex_unlock(&ls->ls_waiters_mutex);
- if (!found)
- lkb = NULL;
- return lkb;
- }
- /* Deal with lookups and lkb's marked RESEND from _pre. We may now be the
- master or dir-node for r. Processing the lkb may result in it being placed
- back on waiters. */
- /* We do this after normal locking has been enabled and any saved messages
- (in requestqueue) have been processed. We should be confident that at
- this point we won't get or process a reply to any of these waiting
- operations. But, new ops may be coming in on the rsbs/locks here from
- userspace or remotely. */
- /* there may have been an overlap unlock/cancel prior to recovery or after
- recovery. if before, the lkb may still have a pos wait_count; if after, the
- overlap flag would just have been set and nothing new sent. we can be
- confident here than any replies to either the initial op or overlap ops
- prior to recovery have been received. */
- int dlm_recover_waiters_post(struct dlm_ls *ls)
- {
- struct dlm_lkb *lkb;
- struct dlm_rsb *r;
- int error = 0, mstype, err, oc, ou;
- while (1) {
- if (dlm_locking_stopped(ls)) {
- log_debug(ls, "recover_waiters_post aborted");
- error = -EINTR;
- break;
- }
- lkb = find_resend_waiter(ls);
- if (!lkb)
- break;
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- mstype = lkb->lkb_wait_type;
- oc = is_overlap_cancel(lkb);
- ou = is_overlap_unlock(lkb);
- err = 0;
- log_debug(ls, "waiter %x remote %x msg %d r_nodeid %d "
- "lkb_nodeid %d wait_nodeid %d dir_nodeid %d "
- "overlap %d %d", lkb->lkb_id, lkb->lkb_remid, mstype,
- r->res_nodeid, lkb->lkb_nodeid, lkb->lkb_wait_nodeid,
- dlm_dir_nodeid(r), oc, ou);
- /* At this point we assume that we won't get a reply to any
- previous op or overlap op on this lock. First, do a big
- remove_from_waiters() for all previous ops. */
- lkb->lkb_flags &= ~DLM_IFL_RESEND;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_UNLOCK;
- lkb->lkb_flags &= ~DLM_IFL_OVERLAP_CANCEL;
- lkb->lkb_wait_type = 0;
- lkb->lkb_wait_count = 0;
- mutex_lock(&ls->ls_waiters_mutex);
- list_del_init(&lkb->lkb_wait_reply);
- mutex_unlock(&ls->ls_waiters_mutex);
- unhold_lkb(lkb); /* for waiters list */
- if (oc || ou) {
- /* do an unlock or cancel instead of resending */
- switch (mstype) {
- case DLM_MSG_LOOKUP:
- case DLM_MSG_REQUEST:
- queue_cast(r, lkb, ou ? -DLM_EUNLOCK :
- -DLM_ECANCEL);
- unhold_lkb(lkb); /* undoes create_lkb() */
- break;
- case DLM_MSG_CONVERT:
- if (oc) {
- queue_cast(r, lkb, -DLM_ECANCEL);
- } else {
- lkb->lkb_exflags |= DLM_LKF_FORCEUNLOCK;
- _unlock_lock(r, lkb);
- }
- break;
- default:
- err = 1;
- }
- } else {
- switch (mstype) {
- case DLM_MSG_LOOKUP:
- case DLM_MSG_REQUEST:
- _request_lock(r, lkb);
- if (is_master(r))
- confirm_master(r, 0);
- break;
- case DLM_MSG_CONVERT:
- _convert_lock(r, lkb);
- break;
- default:
- err = 1;
- }
- }
- if (err) {
- log_error(ls, "waiter %x msg %d r_nodeid %d "
- "dir_nodeid %d overlap %d %d",
- lkb->lkb_id, mstype, r->res_nodeid,
- dlm_dir_nodeid(r), oc, ou);
- }
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- }
- return error;
- }
- static void purge_mstcpy_list(struct dlm_ls *ls, struct dlm_rsb *r,
- struct list_head *list)
- {
- struct dlm_lkb *lkb, *safe;
- list_for_each_entry_safe(lkb, safe, list, lkb_statequeue) {
- if (!is_master_copy(lkb))
- continue;
- /* don't purge lkbs we've added in recover_master_copy for
- the current recovery seq */
- if (lkb->lkb_recover_seq == ls->ls_recover_seq)
- continue;
- del_lkb(r, lkb);
- /* this put should free the lkb */
- if (!dlm_put_lkb(lkb))
- log_error(ls, "purged mstcpy lkb not released");
- }
- }
- void dlm_purge_mstcpy_locks(struct dlm_rsb *r)
- {
- struct dlm_ls *ls = r->res_ls;
- purge_mstcpy_list(ls, r, &r->res_grantqueue);
- purge_mstcpy_list(ls, r, &r->res_convertqueue);
- purge_mstcpy_list(ls, r, &r->res_waitqueue);
- }
- static void purge_dead_list(struct dlm_ls *ls, struct dlm_rsb *r,
- struct list_head *list,
- int nodeid_gone, unsigned int *count)
- {
- struct dlm_lkb *lkb, *safe;
- list_for_each_entry_safe(lkb, safe, list, lkb_statequeue) {
- if (!is_master_copy(lkb))
- continue;
- if ((lkb->lkb_nodeid == nodeid_gone) ||
- dlm_is_removed(ls, lkb->lkb_nodeid)) {
- /* tell recover_lvb to invalidate the lvb
- because a node holding EX/PW failed */
- if ((lkb->lkb_exflags & DLM_LKF_VALBLK) &&
- (lkb->lkb_grmode >= DLM_LOCK_PW)) {
- rsb_set_flag(r, RSB_RECOVER_LVB_INVAL);
- }
- del_lkb(r, lkb);
- /* this put should free the lkb */
- if (!dlm_put_lkb(lkb))
- log_error(ls, "purged dead lkb not released");
- rsb_set_flag(r, RSB_RECOVER_GRANT);
- (*count)++;
- }
- }
- }
- /* Get rid of locks held by nodes that are gone. */
- void dlm_recover_purge(struct dlm_ls *ls)
- {
- struct dlm_rsb *r;
- struct dlm_member *memb;
- int nodes_count = 0;
- int nodeid_gone = 0;
- unsigned int lkb_count = 0;
- /* cache one removed nodeid to optimize the common
- case of a single node removed */
- list_for_each_entry(memb, &ls->ls_nodes_gone, list) {
- nodes_count++;
- nodeid_gone = memb->nodeid;
- }
- if (!nodes_count)
- return;
- down_write(&ls->ls_root_sem);
- list_for_each_entry(r, &ls->ls_root_list, res_root_list) {
- hold_rsb(r);
- lock_rsb(r);
- if (is_master(r)) {
- purge_dead_list(ls, r, &r->res_grantqueue,
- nodeid_gone, &lkb_count);
- purge_dead_list(ls, r, &r->res_convertqueue,
- nodeid_gone, &lkb_count);
- purge_dead_list(ls, r, &r->res_waitqueue,
- nodeid_gone, &lkb_count);
- }
- unlock_rsb(r);
- unhold_rsb(r);
- cond_resched();
- }
- up_write(&ls->ls_root_sem);
- if (lkb_count)
- log_rinfo(ls, "dlm_recover_purge %u locks for %u nodes",
- lkb_count, nodes_count);
- }
- static struct dlm_rsb *find_grant_rsb(struct dlm_ls *ls, int bucket)
- {
- struct rb_node *n;
- struct dlm_rsb *r;
- spin_lock(&ls->ls_rsbtbl[bucket].lock);
- for (n = rb_first(&ls->ls_rsbtbl[bucket].keep); n; n = rb_next(n)) {
- r = rb_entry(n, struct dlm_rsb, res_hashnode);
- if (!rsb_flag(r, RSB_RECOVER_GRANT))
- continue;
- if (!is_master(r)) {
- rsb_clear_flag(r, RSB_RECOVER_GRANT);
- continue;
- }
- hold_rsb(r);
- spin_unlock(&ls->ls_rsbtbl[bucket].lock);
- return r;
- }
- spin_unlock(&ls->ls_rsbtbl[bucket].lock);
- return NULL;
- }
- /*
- * Attempt to grant locks on resources that we are the master of.
- * Locks may have become grantable during recovery because locks
- * from departed nodes have been purged (or not rebuilt), allowing
- * previously blocked locks to now be granted. The subset of rsb's
- * we are interested in are those with lkb's on either the convert or
- * waiting queues.
- *
- * Simplest would be to go through each master rsb and check for non-empty
- * convert or waiting queues, and attempt to grant on those rsbs.
- * Checking the queues requires lock_rsb, though, for which we'd need
- * to release the rsbtbl lock. This would make iterating through all
- * rsb's very inefficient. So, we rely on earlier recovery routines
- * to set RECOVER_GRANT on any rsb's that we should attempt to grant
- * locks for.
- */
- void dlm_recover_grant(struct dlm_ls *ls)
- {
- struct dlm_rsb *r;
- int bucket = 0;
- unsigned int count = 0;
- unsigned int rsb_count = 0;
- unsigned int lkb_count = 0;
- while (1) {
- r = find_grant_rsb(ls, bucket);
- if (!r) {
- if (bucket == ls->ls_rsbtbl_size - 1)
- break;
- bucket++;
- continue;
- }
- rsb_count++;
- count = 0;
- lock_rsb(r);
- /* the RECOVER_GRANT flag is checked in the grant path */
- grant_pending_locks(r, &count);
- rsb_clear_flag(r, RSB_RECOVER_GRANT);
- lkb_count += count;
- confirm_master(r, 0);
- unlock_rsb(r);
- put_rsb(r);
- cond_resched();
- }
- if (lkb_count)
- log_rinfo(ls, "dlm_recover_grant %u locks on %u resources",
- lkb_count, rsb_count);
- }
- static struct dlm_lkb *search_remid_list(struct list_head *head, int nodeid,
- uint32_t remid)
- {
- struct dlm_lkb *lkb;
- list_for_each_entry(lkb, head, lkb_statequeue) {
- if (lkb->lkb_nodeid == nodeid && lkb->lkb_remid == remid)
- return lkb;
- }
- return NULL;
- }
- static struct dlm_lkb *search_remid(struct dlm_rsb *r, int nodeid,
- uint32_t remid)
- {
- struct dlm_lkb *lkb;
- lkb = search_remid_list(&r->res_grantqueue, nodeid, remid);
- if (lkb)
- return lkb;
- lkb = search_remid_list(&r->res_convertqueue, nodeid, remid);
- if (lkb)
- return lkb;
- lkb = search_remid_list(&r->res_waitqueue, nodeid, remid);
- if (lkb)
- return lkb;
- return NULL;
- }
- /* needs at least dlm_rcom + rcom_lock */
- static int receive_rcom_lock_args(struct dlm_ls *ls, struct dlm_lkb *lkb,
- struct dlm_rsb *r, struct dlm_rcom *rc)
- {
- struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
- lkb->lkb_nodeid = rc->rc_header.h_nodeid;
- lkb->lkb_ownpid = le32_to_cpu(rl->rl_ownpid);
- lkb->lkb_remid = le32_to_cpu(rl->rl_lkid);
- lkb->lkb_exflags = le32_to_cpu(rl->rl_exflags);
- lkb->lkb_flags = le32_to_cpu(rl->rl_flags) & 0x0000FFFF;
- lkb->lkb_flags |= DLM_IFL_MSTCPY;
- lkb->lkb_lvbseq = le32_to_cpu(rl->rl_lvbseq);
- lkb->lkb_rqmode = rl->rl_rqmode;
- lkb->lkb_grmode = rl->rl_grmode;
- /* don't set lkb_status because add_lkb wants to itself */
- lkb->lkb_bastfn = (rl->rl_asts & DLM_CB_BAST) ? &fake_bastfn : NULL;
- lkb->lkb_astfn = (rl->rl_asts & DLM_CB_CAST) ? &fake_astfn : NULL;
- if (lkb->lkb_exflags & DLM_LKF_VALBLK) {
- int lvblen = rc->rc_header.h_length - sizeof(struct dlm_rcom) -
- sizeof(struct rcom_lock);
- if (lvblen > ls->ls_lvblen)
- return -EINVAL;
- lkb->lkb_lvbptr = dlm_allocate_lvb(ls);
- if (!lkb->lkb_lvbptr)
- return -ENOMEM;
- memcpy(lkb->lkb_lvbptr, rl->rl_lvb, lvblen);
- }
- /* Conversions between PR and CW (middle modes) need special handling.
- The real granted mode of these converting locks cannot be determined
- until all locks have been rebuilt on the rsb (recover_conversion) */
- if (rl->rl_wait_type == cpu_to_le16(DLM_MSG_CONVERT) &&
- middle_conversion(lkb)) {
- rl->rl_status = DLM_LKSTS_CONVERT;
- lkb->lkb_grmode = DLM_LOCK_IV;
- rsb_set_flag(r, RSB_RECOVER_CONVERT);
- }
- return 0;
- }
- /* This lkb may have been recovered in a previous aborted recovery so we need
- to check if the rsb already has an lkb with the given remote nodeid/lkid.
- If so we just send back a standard reply. If not, we create a new lkb with
- the given values and send back our lkid. We send back our lkid by sending
- back the rcom_lock struct we got but with the remid field filled in. */
- /* needs at least dlm_rcom + rcom_lock */
- int dlm_recover_master_copy(struct dlm_ls *ls, struct dlm_rcom *rc)
- {
- struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
- struct dlm_rsb *r;
- struct dlm_lkb *lkb;
- uint32_t remid = 0;
- int from_nodeid = rc->rc_header.h_nodeid;
- int error;
- if (rl->rl_parent_lkid) {
- error = -EOPNOTSUPP;
- goto out;
- }
- remid = le32_to_cpu(rl->rl_lkid);
- /* In general we expect the rsb returned to be R_MASTER, but we don't
- have to require it. Recovery of masters on one node can overlap
- recovery of locks on another node, so one node can send us MSTCPY
- locks before we've made ourselves master of this rsb. We can still
- add new MSTCPY locks that we receive here without any harm; when
- we make ourselves master, dlm_recover_masters() won't touch the
- MSTCPY locks we've received early. */
- error = find_rsb(ls, rl->rl_name, le16_to_cpu(rl->rl_namelen),
- from_nodeid, R_RECEIVE_RECOVER, &r);
- if (error)
- goto out;
- lock_rsb(r);
- if (dlm_no_directory(ls) && (dlm_dir_nodeid(r) != dlm_our_nodeid())) {
- log_error(ls, "dlm_recover_master_copy remote %d %x not dir",
- from_nodeid, remid);
- error = -EBADR;
- goto out_unlock;
- }
- lkb = search_remid(r, from_nodeid, remid);
- if (lkb) {
- error = -EEXIST;
- goto out_remid;
- }
- error = create_lkb(ls, &lkb);
- if (error)
- goto out_unlock;
- error = receive_rcom_lock_args(ls, lkb, r, rc);
- if (error) {
- __put_lkb(ls, lkb);
- goto out_unlock;
- }
- attach_lkb(r, lkb);
- add_lkb(r, lkb, rl->rl_status);
- error = 0;
- ls->ls_recover_locks_in++;
- if (!list_empty(&r->res_waitqueue) || !list_empty(&r->res_convertqueue))
- rsb_set_flag(r, RSB_RECOVER_GRANT);
- out_remid:
- /* this is the new value returned to the lock holder for
- saving in its process-copy lkb */
- rl->rl_remid = cpu_to_le32(lkb->lkb_id);
- lkb->lkb_recover_seq = ls->ls_recover_seq;
- out_unlock:
- unlock_rsb(r);
- put_rsb(r);
- out:
- if (error && error != -EEXIST)
- log_rinfo(ls, "dlm_recover_master_copy remote %d %x error %d",
- from_nodeid, remid, error);
- rl->rl_result = cpu_to_le32(error);
- return error;
- }
- /* needs at least dlm_rcom + rcom_lock */
- int dlm_recover_process_copy(struct dlm_ls *ls, struct dlm_rcom *rc)
- {
- struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
- struct dlm_rsb *r;
- struct dlm_lkb *lkb;
- uint32_t lkid, remid;
- int error, result;
- lkid = le32_to_cpu(rl->rl_lkid);
- remid = le32_to_cpu(rl->rl_remid);
- result = le32_to_cpu(rl->rl_result);
- error = find_lkb(ls, lkid, &lkb);
- if (error) {
- log_error(ls, "dlm_recover_process_copy no %x remote %d %x %d",
- lkid, rc->rc_header.h_nodeid, remid, result);
- return error;
- }
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- if (!is_process_copy(lkb)) {
- log_error(ls, "dlm_recover_process_copy bad %x remote %d %x %d",
- lkid, rc->rc_header.h_nodeid, remid, result);
- dlm_dump_rsb(r);
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return -EINVAL;
- }
- switch (result) {
- case -EBADR:
- /* There's a chance the new master received our lock before
- dlm_recover_master_reply(), this wouldn't happen if we did
- a barrier between recover_masters and recover_locks. */
- log_debug(ls, "dlm_recover_process_copy %x remote %d %x %d",
- lkid, rc->rc_header.h_nodeid, remid, result);
-
- dlm_send_rcom_lock(r, lkb);
- goto out;
- case -EEXIST:
- case 0:
- lkb->lkb_remid = remid;
- break;
- default:
- log_error(ls, "dlm_recover_process_copy %x remote %d %x %d unk",
- lkid, rc->rc_header.h_nodeid, remid, result);
- }
- /* an ack for dlm_recover_locks() which waits for replies from
- all the locks it sends to new masters */
- dlm_recovered_lock(r);
- out:
- unlock_rsb(r);
- put_rsb(r);
- dlm_put_lkb(lkb);
- return 0;
- }
- int dlm_user_request(struct dlm_ls *ls, struct dlm_user_args *ua,
- int mode, uint32_t flags, void *name, unsigned int namelen,
- unsigned long timeout_cs)
- {
- struct dlm_lkb *lkb;
- struct dlm_args args;
- int error;
- dlm_lock_recovery(ls);
- error = create_lkb(ls, &lkb);
- if (error) {
- kfree(ua);
- goto out;
- }
- if (flags & DLM_LKF_VALBLK) {
- ua->lksb.sb_lvbptr = kzalloc(DLM_USER_LVB_LEN, GFP_NOFS);
- if (!ua->lksb.sb_lvbptr) {
- kfree(ua);
- __put_lkb(ls, lkb);
- error = -ENOMEM;
- goto out;
- }
- }
- error = set_lock_args(mode, &ua->lksb, flags, namelen, timeout_cs,
- fake_astfn, ua, fake_bastfn, &args);
- if (error) {
- kfree(ua->lksb.sb_lvbptr);
- ua->lksb.sb_lvbptr = NULL;
- kfree(ua);
- __put_lkb(ls, lkb);
- goto out;
- }
- /* After ua is attached to lkb it will be freed by dlm_free_lkb().
- When DLM_IFL_USER is set, the dlm knows that this is a userspace
- lock and that lkb_astparam is the dlm_user_args structure. */
- lkb->lkb_flags |= DLM_IFL_USER;
- error = request_lock(ls, lkb, name, namelen, &args);
- switch (error) {
- case 0:
- break;
- case -EINPROGRESS:
- error = 0;
- break;
- case -EAGAIN:
- error = 0;
- /* fall through */
- default:
- __put_lkb(ls, lkb);
- goto out;
- }
- /* add this new lkb to the per-process list of locks */
- spin_lock(&ua->proc->locks_spin);
- hold_lkb(lkb);
- list_add_tail(&lkb->lkb_ownqueue, &ua->proc->locks);
- spin_unlock(&ua->proc->locks_spin);
- out:
- dlm_unlock_recovery(ls);
- return error;
- }
- int dlm_user_convert(struct dlm_ls *ls, struct dlm_user_args *ua_tmp,
- int mode, uint32_t flags, uint32_t lkid, char *lvb_in,
- unsigned long timeout_cs)
- {
- struct dlm_lkb *lkb;
- struct dlm_args args;
- struct dlm_user_args *ua;
- int error;
- dlm_lock_recovery(ls);
- error = find_lkb(ls, lkid, &lkb);
- if (error)
- goto out;
- /* user can change the params on its lock when it converts it, or
- add an lvb that didn't exist before */
- ua = lkb->lkb_ua;
- if (flags & DLM_LKF_VALBLK && !ua->lksb.sb_lvbptr) {
- ua->lksb.sb_lvbptr = kzalloc(DLM_USER_LVB_LEN, GFP_NOFS);
- if (!ua->lksb.sb_lvbptr) {
- error = -ENOMEM;
- goto out_put;
- }
- }
- if (lvb_in && ua->lksb.sb_lvbptr)
- memcpy(ua->lksb.sb_lvbptr, lvb_in, DLM_USER_LVB_LEN);
- ua->xid = ua_tmp->xid;
- ua->castparam = ua_tmp->castparam;
- ua->castaddr = ua_tmp->castaddr;
- ua->bastparam = ua_tmp->bastparam;
- ua->bastaddr = ua_tmp->bastaddr;
- ua->user_lksb = ua_tmp->user_lksb;
- error = set_lock_args(mode, &ua->lksb, flags, 0, timeout_cs,
- fake_astfn, ua, fake_bastfn, &args);
- if (error)
- goto out_put;
- error = convert_lock(ls, lkb, &args);
- if (error == -EINPROGRESS || error == -EAGAIN || error == -EDEADLK)
- error = 0;
- out_put:
- dlm_put_lkb(lkb);
- out:
- dlm_unlock_recovery(ls);
- kfree(ua_tmp);
- return error;
- }
- /*
- * The caller asks for an orphan lock on a given resource with a given mode.
- * If a matching lock exists, it's moved to the owner's list of locks and
- * the lkid is returned.
- */
- int dlm_user_adopt_orphan(struct dlm_ls *ls, struct dlm_user_args *ua_tmp,
- int mode, uint32_t flags, void *name, unsigned int namelen,
- unsigned long timeout_cs, uint32_t *lkid)
- {
- struct dlm_lkb *lkb;
- struct dlm_user_args *ua;
- int found_other_mode = 0;
- int found = 0;
- int rv = 0;
- mutex_lock(&ls->ls_orphans_mutex);
- list_for_each_entry(lkb, &ls->ls_orphans, lkb_ownqueue) {
- if (lkb->lkb_resource->res_length != namelen)
- continue;
- if (memcmp(lkb->lkb_resource->res_name, name, namelen))
- continue;
- if (lkb->lkb_grmode != mode) {
- found_other_mode = 1;
- continue;
- }
- found = 1;
- list_del_init(&lkb->lkb_ownqueue);
- lkb->lkb_flags &= ~DLM_IFL_ORPHAN;
- *lkid = lkb->lkb_id;
- break;
- }
- mutex_unlock(&ls->ls_orphans_mutex);
- if (!found && found_other_mode) {
- rv = -EAGAIN;
- goto out;
- }
- if (!found) {
- rv = -ENOENT;
- goto out;
- }
- lkb->lkb_exflags = flags;
- lkb->lkb_ownpid = (int) current->pid;
- ua = lkb->lkb_ua;
- ua->proc = ua_tmp->proc;
- ua->xid = ua_tmp->xid;
- ua->castparam = ua_tmp->castparam;
- ua->castaddr = ua_tmp->castaddr;
- ua->bastparam = ua_tmp->bastparam;
- ua->bastaddr = ua_tmp->bastaddr;
- ua->user_lksb = ua_tmp->user_lksb;
- /*
- * The lkb reference from the ls_orphans list was not
- * removed above, and is now considered the reference
- * for the proc locks list.
- */
- spin_lock(&ua->proc->locks_spin);
- list_add_tail(&lkb->lkb_ownqueue, &ua->proc->locks);
- spin_unlock(&ua->proc->locks_spin);
- out:
- kfree(ua_tmp);
- return rv;
- }
- int dlm_user_unlock(struct dlm_ls *ls, struct dlm_user_args *ua_tmp,
- uint32_t flags, uint32_t lkid, char *lvb_in)
- {
- struct dlm_lkb *lkb;
- struct dlm_args args;
- struct dlm_user_args *ua;
- int error;
- dlm_lock_recovery(ls);
- error = find_lkb(ls, lkid, &lkb);
- if (error)
- goto out;
- ua = lkb->lkb_ua;
- if (lvb_in && ua->lksb.sb_lvbptr)
- memcpy(ua->lksb.sb_lvbptr, lvb_in, DLM_USER_LVB_LEN);
- if (ua_tmp->castparam)
- ua->castparam = ua_tmp->castparam;
- ua->user_lksb = ua_tmp->user_lksb;
- error = set_unlock_args(flags, ua, &args);
- if (error)
- goto out_put;
- error = unlock_lock(ls, lkb, &args);
- if (error == -DLM_EUNLOCK)
- error = 0;
- /* from validate_unlock_args() */
- if (error == -EBUSY && (flags & DLM_LKF_FORCEUNLOCK))
- error = 0;
- if (error)
- goto out_put;
- spin_lock(&ua->proc->locks_spin);
- /* dlm_user_add_cb() may have already taken lkb off the proc list */
- if (!list_empty(&lkb->lkb_ownqueue))
- list_move(&lkb->lkb_ownqueue, &ua->proc->unlocking);
- spin_unlock(&ua->proc->locks_spin);
- out_put:
- dlm_put_lkb(lkb);
- out:
- dlm_unlock_recovery(ls);
- kfree(ua_tmp);
- return error;
- }
- int dlm_user_cancel(struct dlm_ls *ls, struct dlm_user_args *ua_tmp,
- uint32_t flags, uint32_t lkid)
- {
- struct dlm_lkb *lkb;
- struct dlm_args args;
- struct dlm_user_args *ua;
- int error;
- dlm_lock_recovery(ls);
- error = find_lkb(ls, lkid, &lkb);
- if (error)
- goto out;
- ua = lkb->lkb_ua;
- if (ua_tmp->castparam)
- ua->castparam = ua_tmp->castparam;
- ua->user_lksb = ua_tmp->user_lksb;
- error = set_unlock_args(flags, ua, &args);
- if (error)
- goto out_put;
- error = cancel_lock(ls, lkb, &args);
- if (error == -DLM_ECANCEL)
- error = 0;
- /* from validate_unlock_args() */
- if (error == -EBUSY)
- error = 0;
- out_put:
- dlm_put_lkb(lkb);
- out:
- dlm_unlock_recovery(ls);
- kfree(ua_tmp);
- return error;
- }
- int dlm_user_deadlock(struct dlm_ls *ls, uint32_t flags, uint32_t lkid)
- {
- struct dlm_lkb *lkb;
- struct dlm_args args;
- struct dlm_user_args *ua;
- struct dlm_rsb *r;
- int error;
- dlm_lock_recovery(ls);
- error = find_lkb(ls, lkid, &lkb);
- if (error)
- goto out;
- ua = lkb->lkb_ua;
- error = set_unlock_args(flags, ua, &args);
- if (error)
- goto out_put;
- /* same as cancel_lock(), but set DEADLOCK_CANCEL after lock_rsb */
- r = lkb->lkb_resource;
- hold_rsb(r);
- lock_rsb(r);
- error = validate_unlock_args(lkb, &args);
- if (error)
- goto out_r;
- lkb->lkb_flags |= DLM_IFL_DEADLOCK_CANCEL;
- error = _cancel_lock(r, lkb);
- out_r:
- unlock_rsb(r);
- put_rsb(r);
- if (error == -DLM_ECANCEL)
- error = 0;
- /* from validate_unlock_args() */
- if (error == -EBUSY)
- error = 0;
- out_put:
- dlm_put_lkb(lkb);
- out:
- dlm_unlock_recovery(ls);
- return error;
- }
- /* lkb's that are removed from the waiters list by revert are just left on the
- orphans list with the granted orphan locks, to be freed by purge */
- static int orphan_proc_lock(struct dlm_ls *ls, struct dlm_lkb *lkb)
- {
- struct dlm_args args;
- int error;
- hold_lkb(lkb); /* reference for the ls_orphans list */
- mutex_lock(&ls->ls_orphans_mutex);
- list_add_tail(&lkb->lkb_ownqueue, &ls->ls_orphans);
- mutex_unlock(&ls->ls_orphans_mutex);
- set_unlock_args(0, lkb->lkb_ua, &args);
- error = cancel_lock(ls, lkb, &args);
- if (error == -DLM_ECANCEL)
- error = 0;
- return error;
- }
- /* The FORCEUNLOCK flag allows the unlock to go ahead even if the lkb isn't
- granted. Regardless of what rsb queue the lock is on, it's removed and
- freed. The IVVALBLK flag causes the lvb on the resource to be invalidated
- if our lock is PW/EX (it's ignored if our granted mode is smaller.) */
- static int unlock_proc_lock(struct dlm_ls *ls, struct dlm_lkb *lkb)
- {
- struct dlm_args args;
- int error;
- set_unlock_args(DLM_LKF_FORCEUNLOCK | DLM_LKF_IVVALBLK,
- lkb->lkb_ua, &args);
- error = unlock_lock(ls, lkb, &args);
- if (error == -DLM_EUNLOCK)
- error = 0;
- return error;
- }
- /* We have to release clear_proc_locks mutex before calling unlock_proc_lock()
- (which does lock_rsb) due to deadlock with receiving a message that does
- lock_rsb followed by dlm_user_add_cb() */
- static struct dlm_lkb *del_proc_lock(struct dlm_ls *ls,
- struct dlm_user_proc *proc)
- {
- struct dlm_lkb *lkb = NULL;
- mutex_lock(&ls->ls_clear_proc_locks);
- if (list_empty(&proc->locks))
- goto out;
- lkb = list_entry(proc->locks.next, struct dlm_lkb, lkb_ownqueue);
- list_del_init(&lkb->lkb_ownqueue);
- if (lkb->lkb_exflags & DLM_LKF_PERSISTENT)
- lkb->lkb_flags |= DLM_IFL_ORPHAN;
- else
- lkb->lkb_flags |= DLM_IFL_DEAD;
- out:
- mutex_unlock(&ls->ls_clear_proc_locks);
- return lkb;
- }
- /* The ls_clear_proc_locks mutex protects against dlm_user_add_cb() which
- 1) references lkb->ua which we free here and 2) adds lkbs to proc->asts,
- which we clear here. */
- /* proc CLOSING flag is set so no more device_reads should look at proc->asts
- list, and no more device_writes should add lkb's to proc->locks list; so we
- shouldn't need to take asts_spin or locks_spin here. this assumes that
- device reads/writes/closes are serialized -- FIXME: we may need to serialize
- them ourself. */
- void dlm_clear_proc_locks(struct dlm_ls *ls, struct dlm_user_proc *proc)
- {
- struct dlm_lkb *lkb, *safe;
- dlm_lock_recovery(ls);
- while (1) {
- lkb = del_proc_lock(ls, proc);
- if (!lkb)
- break;
- del_timeout(lkb);
- if (lkb->lkb_exflags & DLM_LKF_PERSISTENT)
- orphan_proc_lock(ls, lkb);
- else
- unlock_proc_lock(ls, lkb);
- /* this removes the reference for the proc->locks list
- added by dlm_user_request, it may result in the lkb
- being freed */
- dlm_put_lkb(lkb);
- }
- mutex_lock(&ls->ls_clear_proc_locks);
- /* in-progress unlocks */
- list_for_each_entry_safe(lkb, safe, &proc->unlocking, lkb_ownqueue) {
- list_del_init(&lkb->lkb_ownqueue);
- lkb->lkb_flags |= DLM_IFL_DEAD;
- dlm_put_lkb(lkb);
- }
- list_for_each_entry_safe(lkb, safe, &proc->asts, lkb_cb_list) {
- memset(&lkb->lkb_callbacks, 0,
- sizeof(struct dlm_callback) * DLM_CALLBACKS_SIZE);
- list_del_init(&lkb->lkb_cb_list);
- dlm_put_lkb(lkb);
- }
- mutex_unlock(&ls->ls_clear_proc_locks);
- dlm_unlock_recovery(ls);
- }
- static void purge_proc_locks(struct dlm_ls *ls, struct dlm_user_proc *proc)
- {
- struct dlm_lkb *lkb, *safe;
- while (1) {
- lkb = NULL;
- spin_lock(&proc->locks_spin);
- if (!list_empty(&proc->locks)) {
- lkb = list_entry(proc->locks.next, struct dlm_lkb,
- lkb_ownqueue);
- list_del_init(&lkb->lkb_ownqueue);
- }
- spin_unlock(&proc->locks_spin);
- if (!lkb)
- break;
- lkb->lkb_flags |= DLM_IFL_DEAD;
- unlock_proc_lock(ls, lkb);
- dlm_put_lkb(lkb); /* ref from proc->locks list */
- }
- spin_lock(&proc->locks_spin);
- list_for_each_entry_safe(lkb, safe, &proc->unlocking, lkb_ownqueue) {
- list_del_init(&lkb->lkb_ownqueue);
- lkb->lkb_flags |= DLM_IFL_DEAD;
- dlm_put_lkb(lkb);
- }
- spin_unlock(&proc->locks_spin);
- spin_lock(&proc->asts_spin);
- list_for_each_entry_safe(lkb, safe, &proc->asts, lkb_cb_list) {
- memset(&lkb->lkb_callbacks, 0,
- sizeof(struct dlm_callback) * DLM_CALLBACKS_SIZE);
- list_del_init(&lkb->lkb_cb_list);
- dlm_put_lkb(lkb);
- }
- spin_unlock(&proc->asts_spin);
- }
- /* pid of 0 means purge all orphans */
- static void do_purge(struct dlm_ls *ls, int nodeid, int pid)
- {
- struct dlm_lkb *lkb, *safe;
- mutex_lock(&ls->ls_orphans_mutex);
- list_for_each_entry_safe(lkb, safe, &ls->ls_orphans, lkb_ownqueue) {
- if (pid && lkb->lkb_ownpid != pid)
- continue;
- unlock_proc_lock(ls, lkb);
- list_del_init(&lkb->lkb_ownqueue);
- dlm_put_lkb(lkb);
- }
- mutex_unlock(&ls->ls_orphans_mutex);
- }
- static int send_purge(struct dlm_ls *ls, int nodeid, int pid)
- {
- struct dlm_message *ms;
- struct dlm_mhandle *mh;
- int error;
- error = _create_message(ls, sizeof(struct dlm_message), nodeid,
- DLM_MSG_PURGE, &ms, &mh);
- if (error)
- return error;
- ms->m_nodeid = nodeid;
- ms->m_pid = pid;
- return send_message(mh, ms);
- }
- int dlm_user_purge(struct dlm_ls *ls, struct dlm_user_proc *proc,
- int nodeid, int pid)
- {
- int error = 0;
- if (nodeid && (nodeid != dlm_our_nodeid())) {
- error = send_purge(ls, nodeid, pid);
- } else {
- dlm_lock_recovery(ls);
- if (pid == current->pid)
- purge_proc_locks(ls, proc);
- else
- do_purge(ls, nodeid, pid);
- dlm_unlock_recovery(ls);
- }
- return error;
- }
|